You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
278 lines
10 KiB
278 lines
10 KiB
// indexCachedRI.java
|
|
// -----------------------------
|
|
// (C) 2006 by Michael Peter Christen; mc@anomic.de, Frankfurt a. M., Germany
|
|
// first published 7.11.2006 on http://www.anomic.de
|
|
//
|
|
// This is a part of YaCy, a peer-to-peer based web search engine
|
|
//
|
|
// $LastChangedDate: 2006-04-02 22:40:07 +0200 (So, 02 Apr 2006) $
|
|
// $LastChangedRevision: 1986 $
|
|
// $LastChangedBy: orbiter $
|
|
//
|
|
// LICENSE
|
|
//
|
|
// This program is free software; you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation; either version 2 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program; if not, write to the Free Software
|
|
// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
package de.anomic.index;
|
|
|
|
import java.util.ArrayList;
|
|
import java.util.HashMap;
|
|
import java.util.Iterator;
|
|
import java.util.Map;
|
|
import java.util.Set;
|
|
import java.util.TreeSet;
|
|
|
|
import de.anomic.kelondro.kelondroBase64Order;
|
|
import de.anomic.kelondro.kelondroCloneableIterator;
|
|
import de.anomic.kelondro.kelondroMergeIterator;
|
|
import de.anomic.kelondro.kelondroOrder;
|
|
import de.anomic.kelondro.kelondroRotateIterator;
|
|
import de.anomic.kelondro.kelondroRow;
|
|
import de.anomic.server.logging.serverLog;
|
|
|
|
public class indexCachedRI implements indexRI {
|
|
|
|
private kelondroRow payloadrow;
|
|
private kelondroOrder indexOrder;
|
|
private indexRAMRI riExtern, riIntern;
|
|
private indexCollectionRI backend;
|
|
public boolean busyCacheFlush; // shows if a cache flush is currently performed
|
|
private int flushsize;
|
|
|
|
public indexCachedRI(indexRAMRI riExtern, indexRAMRI riIntern, indexCollectionRI backend, kelondroOrder payloadorder, kelondroRow payloadrow, serverLog log) {
|
|
this.riExtern = riExtern;
|
|
this.riIntern = riIntern;
|
|
this.backend = backend;
|
|
this.indexOrder = payloadorder;
|
|
this.payloadrow = payloadrow;
|
|
this.busyCacheFlush = false;
|
|
this.flushsize = 2000;
|
|
}
|
|
|
|
public kelondroRow payloadrow() {
|
|
return payloadrow;
|
|
}
|
|
|
|
public int minMem() {
|
|
return 1024 * 1024;
|
|
}
|
|
|
|
public void setWordFlushSize(int flushsize) {
|
|
this.flushsize = flushsize;
|
|
}
|
|
|
|
public void flushControl() {
|
|
// check for forced flush
|
|
synchronized (this) {
|
|
if (riExtern.size() > riExtern.getMaxWordCount()) {
|
|
flushCache(riExtern, riExtern.size() + flushsize - riExtern.getMaxWordCount());
|
|
}
|
|
if (riIntern.size() > riIntern.getMaxWordCount()) {
|
|
flushCache(riIntern, riIntern.size() + flushsize - riIntern.getMaxWordCount());
|
|
}
|
|
}
|
|
}
|
|
|
|
public long getUpdateTime(String wordHash) {
|
|
indexContainer entries = getContainer(wordHash, null);
|
|
if (entries == null) return 0;
|
|
return entries.updated();
|
|
}
|
|
|
|
public void addEntries(indexContainer entries, long updateTime, boolean intern) {
|
|
// add the entry
|
|
if (intern) {
|
|
riIntern.addEntries(entries, updateTime, true);
|
|
} else {
|
|
riExtern.addEntries(entries, updateTime, false);
|
|
flushControl();
|
|
}
|
|
}
|
|
|
|
public void flushCacheSome() {
|
|
flushCacheSome(riExtern);
|
|
flushCacheSome(riIntern);
|
|
}
|
|
|
|
private void flushCacheSome(indexRAMRI ram) {
|
|
flushCache(ram, flushsize);
|
|
while (ram.maxURLinCache() > 2048) flushCache(ram, 1);
|
|
}
|
|
|
|
private void flushCache(indexRAMRI ram, int count) {
|
|
if (count <= 0) return;
|
|
if (count > 5000) count = 5000;
|
|
busyCacheFlush = true;
|
|
String wordHash;
|
|
ArrayList containerList = new ArrayList();
|
|
synchronized (this) {
|
|
for (int i = 0; i < count; i++) { // possible position of outOfMemoryError ?
|
|
if (ram.size() == 0) break;
|
|
// select one word to flush
|
|
wordHash = ram.bestFlushWordHash();
|
|
|
|
// move one container from ram to flush list
|
|
indexContainer c = ram.deleteContainer(wordHash);
|
|
if (c != null) containerList.add(c);
|
|
}
|
|
// flush the containers
|
|
backend.addMultipleEntries(containerList);
|
|
}
|
|
busyCacheFlush = false;
|
|
}
|
|
|
|
public boolean hasContainer(String wordHash) {
|
|
if (riExtern.hasContainer(wordHash)) return true;
|
|
if (riIntern.hasContainer(wordHash)) return true;
|
|
if (backend.hasContainer(wordHash)) return true;
|
|
return false;
|
|
}
|
|
|
|
public indexContainer getContainer(String wordHash, Set urlselection) {
|
|
// get from cache
|
|
indexContainer container = riExtern.getContainer(wordHash, urlselection);
|
|
if (container == null) {
|
|
container = riIntern.getContainer(wordHash, urlselection);
|
|
} else {
|
|
container.addAllUnique(riIntern.getContainer(wordHash, urlselection));
|
|
}
|
|
|
|
// get from collection index
|
|
if (container == null) {
|
|
container = backend.getContainer(wordHash, urlselection);
|
|
} else {
|
|
container.addAllUnique(backend.getContainer(wordHash, urlselection));
|
|
}
|
|
return container;
|
|
}
|
|
|
|
public Map getContainers(Set wordHashes, Set urlselection, boolean interruptIfEmpty) {
|
|
// return map of wordhash:indexContainer
|
|
|
|
// retrieve entities that belong to the hashes
|
|
HashMap containers = new HashMap();
|
|
String singleHash;
|
|
indexContainer singleContainer;
|
|
Iterator i = wordHashes.iterator();
|
|
while (i.hasNext()) {
|
|
// get next word hash:
|
|
singleHash = (String) i.next();
|
|
|
|
// retrieve index
|
|
singleContainer = getContainer(singleHash, urlselection);
|
|
|
|
// check result
|
|
if (((singleContainer == null) || (singleContainer.size() == 0)) && (interruptIfEmpty)) return new HashMap();
|
|
|
|
containers.put(singleHash, singleContainer);
|
|
}
|
|
return containers;
|
|
}
|
|
|
|
public int size() {
|
|
return java.lang.Math.max(backend.size(), java.lang.Math.max(riIntern.size(), riExtern.size()));
|
|
}
|
|
|
|
public int indexSize(String wordHash) {
|
|
int size = backend.indexSize(wordHash);
|
|
size += riIntern.indexSize(wordHash);
|
|
size += riExtern.indexSize(wordHash);
|
|
return size;
|
|
}
|
|
|
|
public void close() {
|
|
synchronized (this) {
|
|
riIntern.close();
|
|
riExtern.close();
|
|
backend.close();
|
|
}
|
|
}
|
|
|
|
public indexContainer deleteContainer(String wordHash) {
|
|
indexContainer c = riIntern.deleteContainer(wordHash);
|
|
if (c == null) c = riExtern.deleteContainer(wordHash); else c.addAllUnique(riExtern.deleteContainer(wordHash));
|
|
if (c == null) c = backend.deleteContainer(wordHash); else c.addAllUnique(backend.deleteContainer(wordHash));
|
|
return c;
|
|
}
|
|
|
|
public boolean removeEntry(String wordHash, String urlHash) {
|
|
boolean removed = false;
|
|
removed = removed | (riIntern.removeEntry(wordHash, urlHash));
|
|
removed = removed | (riExtern.removeEntry(wordHash, urlHash));
|
|
removed = removed | (backend.removeEntry(wordHash, urlHash));
|
|
return removed;
|
|
}
|
|
|
|
public int removeEntries(String wordHash, Set urlHashes) {
|
|
int removed = 0;
|
|
removed += riIntern.removeEntries(wordHash, urlHashes);
|
|
removed += riExtern.removeEntries(wordHash, urlHashes);
|
|
removed += backend.removeEntries(wordHash, urlHashes);
|
|
return removed;
|
|
}
|
|
|
|
public String removeEntriesExpl(String wordHash, Set urlHashes) {
|
|
String removed = "";
|
|
removed += riIntern.removeEntries(wordHash, urlHashes) + ", ";
|
|
removed += riExtern.removeEntries(wordHash, urlHashes) + ", ";
|
|
removed += backend.removeEntries(wordHash, urlHashes) + ", ";
|
|
return removed;
|
|
}
|
|
|
|
public TreeSet indexContainerSet(String startHash, boolean ramOnly, boolean rot, int count) {
|
|
// creates a set of indexContainers
|
|
// this does not use the dhtInCache
|
|
kelondroOrder containerOrder = new indexContainerOrder((kelondroOrder) indexOrder.clone());
|
|
containerOrder.rotate(startHash.getBytes());
|
|
TreeSet containers = new TreeSet(containerOrder);
|
|
Iterator i = wordContainers(startHash, ramOnly, rot);
|
|
if (ramOnly) count = Math.min(riExtern.size(), count);
|
|
indexContainer container;
|
|
while ((count > 0) && (i.hasNext())) {
|
|
container = (indexContainer) i.next();
|
|
if ((container != null) && (container.size() > 0)) {
|
|
containers.add(container);
|
|
count--;
|
|
}
|
|
}
|
|
return containers;
|
|
}
|
|
|
|
public kelondroCloneableIterator wordContainers(String startHash, boolean rot) {
|
|
// returns an iteration of indexContainers
|
|
return wordContainers(startHash, false, rot);
|
|
}
|
|
|
|
public kelondroCloneableIterator wordContainers(String startHash, boolean ramOnly, boolean rot) {
|
|
kelondroCloneableIterator i;
|
|
if (ramOnly) {
|
|
i = riExtern.wordContainers(startHash, false);
|
|
} else {
|
|
i = new kelondroMergeIterator(
|
|
riExtern.wordContainers(startHash, false),
|
|
backend.wordContainers(startHash, false),
|
|
new indexContainerOrder(this.indexOrder),
|
|
indexContainer.containerMergeMethod,
|
|
true);
|
|
}
|
|
if (rot) {
|
|
return new kelondroRotateIterator(i, new String(kelondroBase64Order.zero(startHash.length())));
|
|
} else {
|
|
return i;
|
|
}
|
|
}
|
|
|
|
}
|