git-svn-id: https://svn.berlios.de/svnroot/repos/yacy/trunk@4343 6c8d7289-2bf4-0310-a012-ef5d649a1542pull/1/head
parent
cbefc651ac
commit
f4e9ff6ce9
@ -1,277 +0,0 @@
|
||||
// indexCachedRI.java
|
||||
// -----------------------------
|
||||
// (C) 2006 by Michael Peter Christen; mc@anomic.de, Frankfurt a. M., Germany
|
||||
// first published 7.11.2006 on http://www.anomic.de
|
||||
//
|
||||
// This is a part of YaCy, a peer-to-peer based web search engine
|
||||
//
|
||||
// $LastChangedDate: 2006-04-02 22:40:07 +0200 (So, 02 Apr 2006) $
|
||||
// $LastChangedRevision: 1986 $
|
||||
// $LastChangedBy: orbiter $
|
||||
//
|
||||
// LICENSE
|
||||
//
|
||||
// This program is free software; you can redistribute it and/or modify
|
||||
// it under the terms of the GNU General Public License as published by
|
||||
// the Free Software Foundation; either version 2 of the License, or
|
||||
// (at your option) any later version.
|
||||
//
|
||||
// This program is distributed in the hope that it will be useful,
|
||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
// GNU General Public License for more details.
|
||||
//
|
||||
// You should have received a copy of the GNU General Public License
|
||||
// along with this program; if not, write to the Free Software
|
||||
// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
||||
|
||||
package de.anomic.index;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.HashMap;
|
||||
import java.util.Iterator;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.TreeSet;
|
||||
|
||||
import de.anomic.kelondro.kelondroBase64Order;
|
||||
import de.anomic.kelondro.kelondroCloneableIterator;
|
||||
import de.anomic.kelondro.kelondroMergeIterator;
|
||||
import de.anomic.kelondro.kelondroOrder;
|
||||
import de.anomic.kelondro.kelondroRotateIterator;
|
||||
import de.anomic.kelondro.kelondroRow;
|
||||
import de.anomic.server.logging.serverLog;
|
||||
|
||||
public class indexCachedRI implements indexRI {
|
||||
|
||||
private kelondroRow payloadrow;
|
||||
private kelondroOrder indexOrder;
|
||||
private indexRAMRI riExtern, riIntern;
|
||||
private indexCollectionRI backend;
|
||||
public boolean busyCacheFlush; // shows if a cache flush is currently performed
|
||||
private int flushsize;
|
||||
|
||||
public indexCachedRI(indexRAMRI riExtern, indexRAMRI riIntern, indexCollectionRI backend, kelondroOrder payloadorder, kelondroRow payloadrow, serverLog log) {
|
||||
this.riExtern = riExtern;
|
||||
this.riIntern = riIntern;
|
||||
this.backend = backend;
|
||||
this.indexOrder = payloadorder;
|
||||
this.payloadrow = payloadrow;
|
||||
this.busyCacheFlush = false;
|
||||
this.flushsize = 2000;
|
||||
}
|
||||
|
||||
public kelondroRow payloadrow() {
|
||||
return payloadrow;
|
||||
}
|
||||
|
||||
public int minMem() {
|
||||
return 1024 * 1024;
|
||||
}
|
||||
|
||||
public void setWordFlushSize(int flushsize) {
|
||||
this.flushsize = flushsize;
|
||||
}
|
||||
|
||||
public void flushControl() {
|
||||
// check for forced flush
|
||||
synchronized (this) {
|
||||
if (riExtern.size() > riExtern.getMaxWordCount()) {
|
||||
flushCache(riExtern, riExtern.size() + flushsize - riExtern.getMaxWordCount());
|
||||
}
|
||||
if (riIntern.size() > riIntern.getMaxWordCount()) {
|
||||
flushCache(riIntern, riIntern.size() + flushsize - riIntern.getMaxWordCount());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
public long getUpdateTime(String wordHash) {
|
||||
indexContainer entries = getContainer(wordHash, null);
|
||||
if (entries == null) return 0;
|
||||
return entries.updated();
|
||||
}
|
||||
|
||||
public void addEntries(indexContainer entries, long updateTime, boolean intern) {
|
||||
// add the entry
|
||||
if (intern) {
|
||||
riIntern.addEntries(entries, updateTime, true);
|
||||
} else {
|
||||
riExtern.addEntries(entries, updateTime, false);
|
||||
flushControl();
|
||||
}
|
||||
}
|
||||
|
||||
public void flushCacheSome() {
|
||||
flushCacheSome(riExtern);
|
||||
flushCacheSome(riIntern);
|
||||
}
|
||||
|
||||
private void flushCacheSome(indexRAMRI ram) {
|
||||
flushCache(ram, flushsize);
|
||||
while (ram.maxURLinCache() > 2048) flushCache(ram, 1);
|
||||
}
|
||||
|
||||
private void flushCache(indexRAMRI ram, int count) {
|
||||
if (count <= 0) return;
|
||||
if (count > 5000) count = 5000;
|
||||
busyCacheFlush = true;
|
||||
String wordHash;
|
||||
ArrayList<indexContainer> containerList = new ArrayList<indexContainer>();
|
||||
synchronized (this) {
|
||||
for (int i = 0; i < count; i++) { // possible position of outOfMemoryError ?
|
||||
if (ram.size() == 0) break;
|
||||
// select one word to flush
|
||||
wordHash = ram.bestFlushWordHash();
|
||||
|
||||
// move one container from ram to flush list
|
||||
indexContainer c = ram.deleteContainer(wordHash);
|
||||
if (c != null) containerList.add(c);
|
||||
}
|
||||
// flush the containers
|
||||
backend.addMultipleEntries(containerList);
|
||||
}
|
||||
busyCacheFlush = false;
|
||||
}
|
||||
|
||||
public boolean hasContainer(String wordHash) {
|
||||
if (riExtern.hasContainer(wordHash)) return true;
|
||||
if (riIntern.hasContainer(wordHash)) return true;
|
||||
if (backend.hasContainer(wordHash)) return true;
|
||||
return false;
|
||||
}
|
||||
|
||||
public indexContainer getContainer(String wordHash, Set<String> urlselection) {
|
||||
// get from cache
|
||||
indexContainer container = riExtern.getContainer(wordHash, urlselection);
|
||||
if (container == null) {
|
||||
container = riIntern.getContainer(wordHash, urlselection);
|
||||
} else {
|
||||
container.addAllUnique(riIntern.getContainer(wordHash, urlselection));
|
||||
}
|
||||
|
||||
// get from collection index
|
||||
if (container == null) {
|
||||
container = backend.getContainer(wordHash, urlselection);
|
||||
} else {
|
||||
container.addAllUnique(backend.getContainer(wordHash, urlselection));
|
||||
}
|
||||
return container;
|
||||
}
|
||||
|
||||
public Map<String, indexContainer> getContainers(Set<String> wordHashes, Set<String> urlselection, boolean interruptIfEmpty) {
|
||||
// return map of wordhash:indexContainer
|
||||
|
||||
// retrieve entities that belong to the hashes
|
||||
HashMap<String, indexContainer> containers = new HashMap<String, indexContainer>();
|
||||
String singleHash;
|
||||
indexContainer singleContainer;
|
||||
Iterator<String> i = wordHashes.iterator();
|
||||
while (i.hasNext()) {
|
||||
// get next word hash:
|
||||
singleHash = i.next();
|
||||
|
||||
// retrieve index
|
||||
singleContainer = getContainer(singleHash, urlselection);
|
||||
|
||||
// check result
|
||||
if (((singleContainer == null) || (singleContainer.size() == 0)) && (interruptIfEmpty)) return new HashMap<String, indexContainer>();
|
||||
|
||||
containers.put(singleHash, singleContainer);
|
||||
}
|
||||
return containers;
|
||||
}
|
||||
|
||||
public int size() {
|
||||
return java.lang.Math.max(backend.size(), java.lang.Math.max(riIntern.size(), riExtern.size()));
|
||||
}
|
||||
|
||||
public int indexSize(String wordHash) {
|
||||
int size = backend.indexSize(wordHash);
|
||||
size += riIntern.indexSize(wordHash);
|
||||
size += riExtern.indexSize(wordHash);
|
||||
return size;
|
||||
}
|
||||
|
||||
public void close() {
|
||||
synchronized (this) {
|
||||
riIntern.close();
|
||||
riExtern.close();
|
||||
backend.close();
|
||||
}
|
||||
}
|
||||
|
||||
public indexContainer deleteContainer(String wordHash) {
|
||||
indexContainer c = riIntern.deleteContainer(wordHash);
|
||||
if (c == null) c = riExtern.deleteContainer(wordHash); else c.addAllUnique(riExtern.deleteContainer(wordHash));
|
||||
if (c == null) c = backend.deleteContainer(wordHash); else c.addAllUnique(backend.deleteContainer(wordHash));
|
||||
return c;
|
||||
}
|
||||
|
||||
public boolean removeEntry(String wordHash, String urlHash) {
|
||||
boolean removed = false;
|
||||
removed = removed | (riIntern.removeEntry(wordHash, urlHash));
|
||||
removed = removed | (riExtern.removeEntry(wordHash, urlHash));
|
||||
removed = removed | (backend.removeEntry(wordHash, urlHash));
|
||||
return removed;
|
||||
}
|
||||
|
||||
public int removeEntries(String wordHash, Set<String> urlHashes) {
|
||||
int removed = 0;
|
||||
removed += riIntern.removeEntries(wordHash, urlHashes);
|
||||
removed += riExtern.removeEntries(wordHash, urlHashes);
|
||||
removed += backend.removeEntries(wordHash, urlHashes);
|
||||
return removed;
|
||||
}
|
||||
|
||||
public String removeEntriesExpl(String wordHash, Set<String> urlHashes) {
|
||||
String removed = "";
|
||||
removed += riIntern.removeEntries(wordHash, urlHashes) + ", ";
|
||||
removed += riExtern.removeEntries(wordHash, urlHashes) + ", ";
|
||||
removed += backend.removeEntries(wordHash, urlHashes) + ", ";
|
||||
return removed;
|
||||
}
|
||||
|
||||
public TreeSet<indexContainer> indexContainerSet(String startHash, boolean ramOnly, boolean rot, int count) {
|
||||
// creates a set of indexContainers
|
||||
// this does not use the dhtInCache
|
||||
kelondroOrder containerOrder = new indexContainerOrder((kelondroOrder) indexOrder.clone());
|
||||
containerOrder.rotate(startHash.getBytes());
|
||||
TreeSet<indexContainer> containers = new TreeSet<indexContainer>(containerOrder);
|
||||
Iterator<indexContainer> i = wordContainers(startHash, ramOnly, rot);
|
||||
if (ramOnly) count = Math.min(riExtern.size(), count);
|
||||
indexContainer container;
|
||||
while ((count > 0) && (i.hasNext())) {
|
||||
container = i.next();
|
||||
if ((container != null) && (container.size() > 0)) {
|
||||
containers.add(container);
|
||||
count--;
|
||||
}
|
||||
}
|
||||
return containers;
|
||||
}
|
||||
|
||||
public kelondroCloneableIterator<indexContainer> wordContainers(String startHash, boolean rot) {
|
||||
// returns an iteration of indexContainers
|
||||
return wordContainers(startHash, false, rot);
|
||||
}
|
||||
|
||||
public kelondroCloneableIterator<indexContainer> wordContainers(String startHash, boolean ramOnly, boolean rot) {
|
||||
kelondroCloneableIterator<indexContainer> i;
|
||||
if (ramOnly) {
|
||||
i = riExtern.wordContainers(startHash, false);
|
||||
} else {
|
||||
i = new kelondroMergeIterator(
|
||||
riExtern.wordContainers(startHash, false),
|
||||
backend.wordContainers(startHash, false),
|
||||
new indexContainerOrder(this.indexOrder),
|
||||
indexContainer.containerMergeMethod,
|
||||
true);
|
||||
}
|
||||
if (rot) {
|
||||
return new kelondroRotateIterator<indexContainer>(i, new String(kelondroBase64Order.zero(startHash.length())));
|
||||
} else {
|
||||
return i;
|
||||
}
|
||||
}
|
||||
|
||||
}
|
Loading…
Reference in new issue