You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
649 lines
25 KiB
649 lines
25 KiB
// kelondroCache.java
|
|
// (C) 2006 by Michael Peter Christen; mc@anomic.de, Frankfurt a. M., Germany
|
|
// first published 26.10.2006 on http://www.anomic.de
|
|
//
|
|
// This is a part of YaCy, a peer-to-peer based web search engine
|
|
//
|
|
// $LastChangedDate: 2006-04-02 22:40:07 +0200 (So, 02 Apr 2006) $
|
|
// $LastChangedRevision: 1986 $
|
|
// $LastChangedBy: orbiter $
|
|
//
|
|
// LICENSE
|
|
//
|
|
// This program is free software; you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation; either version 2 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program; if not, write to the Free Software
|
|
// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
|
|
package de.anomic.kelondro;
|
|
|
|
import java.io.IOException;
|
|
import java.util.Date;
|
|
import java.util.HashMap;
|
|
import java.util.Iterator;
|
|
import java.util.List;
|
|
import java.util.Map;
|
|
import java.util.TreeMap;
|
|
|
|
import de.anomic.kelondro.kelondroRow.Entry;
|
|
|
|
public class kelondroCache implements kelondroIndex {
|
|
|
|
// this is a combined read cache and write buffer
|
|
// we maintain four tables:
|
|
// - a read-cache
|
|
// - a miss-cache
|
|
// - a write buffer for rows that are not contained in the target index
|
|
// - a write buffer for rows that are known to be contained in target
|
|
// furthermore, if we access a kelondroFlexTable, we can use the ram index of the underlying index
|
|
|
|
// static object tracker; stores information about object cache usage
|
|
private static final TreeMap objectTracker = new TreeMap();
|
|
private static long memStopGrow = 4000000; // a limit for the node cache to stop growing if less than this memory amount is available
|
|
private static long memStartShrink = 2000000; // a limit for the node cache to start with shrinking if less than this memory amount is available
|
|
|
|
// class objects
|
|
private kelondroRowSet readHitCache;
|
|
private kelondroRowSet readMissCache;
|
|
private kelondroRowSet writeBufferUnique; // entries of that buffer are not contained in index
|
|
private kelondroRowSet writeBufferDoubles; // entries of that buffer shall overwrite entries in index
|
|
private kelondroIndex index;
|
|
private kelondroRow keyrow;
|
|
private int readHit, readMiss, writeUnique, writeDouble, cacheDelete, cacheFlush;
|
|
private int hasnotHit, hasnotMiss, hasnotUnique, hasnotDouble, hasnotDelete, hasnotFlush;
|
|
|
|
public kelondroCache(kelondroIndex backupIndex, boolean read, boolean write) throws IOException {
|
|
assert write == false;
|
|
this.index = backupIndex;
|
|
this.keyrow = new kelondroRow(new kelondroColumn[]{index.row().column(index.row().primaryKey)}, index.row().objectOrder, index.row().primaryKey);
|
|
this.readHitCache = (read) ? new kelondroRowSet(index.row(), 0) : null;
|
|
this.readMissCache = (read) ? new kelondroRowSet(this.keyrow, 0) : null;
|
|
this.writeBufferUnique = (write) ? new kelondroRowSet(index.row(), 0) : null;
|
|
this.writeBufferDoubles = (write) ? new kelondroRowSet(index.row(), 0) : null;
|
|
this.readHit = 0;
|
|
this.readMiss = 0;
|
|
this.writeUnique = 0;
|
|
this.writeDouble = 0;
|
|
this.cacheDelete = 0;
|
|
this.cacheFlush = 0;
|
|
this.hasnotHit = 0;
|
|
this.hasnotMiss = 0;
|
|
this.hasnotUnique = 0;
|
|
this.hasnotDouble = 0;
|
|
this.hasnotDelete = 0;
|
|
this.hasnotFlush = 0;
|
|
objectTracker.put(backupIndex.filename(), this);
|
|
}
|
|
|
|
public final int cacheObjectChunkSize() {
|
|
try {
|
|
return index.row().objectsize();
|
|
} catch (IOException e) {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
public int writeBufferSize() {
|
|
return
|
|
((writeBufferUnique == null) ? 0 : writeBufferUnique.size()) +
|
|
((writeBufferDoubles == null) ? 0 : writeBufferDoubles.size());
|
|
}
|
|
|
|
public kelondroProfile profile() {
|
|
return index.profile(); // TODO: implement own profile and merge with global
|
|
}
|
|
|
|
public static void setCacheGrowStati(long memStopGrowNew, long memStartShrinkNew) {
|
|
memStopGrow = memStopGrowNew;
|
|
memStartShrink = memStartShrinkNew;
|
|
}
|
|
|
|
public static long getMemStopGrow() {
|
|
return memStopGrow ;
|
|
}
|
|
|
|
public static long getMemStartShrink() {
|
|
return memStartShrink ;
|
|
}
|
|
|
|
public static final Iterator filenames() {
|
|
// iterates string objects; all file names from record tracker
|
|
return objectTracker.keySet().iterator();
|
|
}
|
|
|
|
public static final Map memoryStats(String filename) {
|
|
// returns a map for each file in the tracker;
|
|
// the map represents properties for each record oobjects,
|
|
// i.e. for cache memory allocation
|
|
kelondroCache theObjectsCache = (kelondroCache) objectTracker.get(filename);
|
|
return theObjectsCache.memoryStats();
|
|
}
|
|
|
|
private final Map memoryStats() {
|
|
// returns statistical data about this object
|
|
HashMap map = new HashMap();
|
|
map.put("objectHitChunkSize", (readHitCache == null) ? "0" : Integer.toString(readHitCache.rowdef.objectsize));
|
|
map.put("objectHitCacheCount", (readHitCache == null) ? "0" : Integer.toString(readHitCache.size()));
|
|
map.put("objectHitMem", (readHitCache == null) ? "0" : Integer.toString((int) (readHitCache.rowdef.objectsize * readHitCache.size() * kelondroRowCollection.growfactor)));
|
|
map.put("objectHitCacheReadHit", Integer.toString(readHit));
|
|
map.put("objectHitCacheReadMiss", Integer.toString(readMiss));
|
|
map.put("objectHitCacheWriteUnique", Integer.toString(writeUnique));
|
|
map.put("objectHitCacheWriteDouble", Integer.toString(writeDouble));
|
|
map.put("objectHitCacheDeletes", Integer.toString(cacheDelete));
|
|
map.put("objectHitCacheFlushes", Integer.toString(cacheFlush));
|
|
|
|
map.put("objectMissChunkSize", (readMissCache == null) ? "0" : Integer.toString(readMissCache.rowdef.objectsize));
|
|
map.put("objectMissCacheCount", (readMissCache == null) ? "0" : Integer.toString(readMissCache.size()));
|
|
map.put("objectMissMem", (readMissCache == null) ? "0" : Integer.toString((int) (readMissCache.rowdef.objectsize * readMissCache.size() * kelondroRowCollection.growfactor)));
|
|
map.put("objectMissCacheReadHit", Integer.toString(hasnotHit));
|
|
map.put("objectMissCacheReadMiss", Integer.toString(hasnotMiss));
|
|
map.put("objectMissCacheWriteUnique", Integer.toString(hasnotUnique));
|
|
map.put("objectMissCacheWriteDouble", Integer.toString(hasnotDouble));
|
|
map.put("objectMissCacheDeletes", Integer.toString(hasnotDelete));
|
|
map.put("objectMissCacheFlushes", Integer.toString(hasnotFlush));
|
|
|
|
// future feature .. map.put("objectElderTimeRead", index.profile().)
|
|
return map;
|
|
}
|
|
|
|
private int cacheGrowStatus() {
|
|
return kelondroRecords.cacheGrowStatus(memStopGrow, memStartShrink);
|
|
}
|
|
|
|
private void flushUnique() throws IOException {
|
|
if (writeBufferUnique == null) return;
|
|
synchronized (writeBufferUnique) {
|
|
Iterator i = writeBufferUnique.rows();
|
|
while (i.hasNext()) {
|
|
this.index.addUnique((kelondroRow.Entry) i.next());
|
|
this.cacheFlush++;
|
|
}
|
|
writeBufferUnique.clear();
|
|
}
|
|
}
|
|
|
|
private void flushUnique(int maxcount) throws IOException {
|
|
if (writeBufferUnique == null) return;
|
|
if (maxcount == 0) return;
|
|
synchronized (writeBufferUnique) {
|
|
kelondroRowCollection delete = new kelondroRowCollection(this.keyrow, maxcount);
|
|
Iterator i = writeBufferUnique.rows();
|
|
kelondroRow.Entry row;
|
|
while ((i.hasNext()) && (maxcount-- > 0)) {
|
|
row = (kelondroRow.Entry) i.next();
|
|
delete.add(row.getColBytes(index.row().primaryKey));
|
|
this.index.addUnique(row);
|
|
this.cacheFlush++;
|
|
}
|
|
i = delete.rows();
|
|
while (i.hasNext()) {
|
|
writeBufferUnique.remove(((kelondroRow.Entry) i.next()).getColBytes(0));
|
|
}
|
|
delete = null;
|
|
writeBufferUnique.trim(true);
|
|
}
|
|
}
|
|
|
|
private void flushDoubles() throws IOException {
|
|
if (writeBufferDoubles == null) return;
|
|
synchronized (writeBufferDoubles) {
|
|
Iterator i = writeBufferDoubles.rows();
|
|
while (i.hasNext()) {
|
|
this.index.put((kelondroRow.Entry) i.next());
|
|
this.cacheFlush++;
|
|
}
|
|
writeBufferDoubles.clear();
|
|
}
|
|
}
|
|
|
|
private void flushDoubles(int maxcount) throws IOException {
|
|
if (writeBufferDoubles == null) return;
|
|
if (maxcount == 0) return;
|
|
synchronized (writeBufferDoubles) {
|
|
kelondroRowCollection delete = new kelondroRowCollection(this.keyrow, maxcount);
|
|
Iterator i = writeBufferDoubles.rows();
|
|
kelondroRow.Entry row;
|
|
while ((i.hasNext()) && (maxcount-- > 0)) {
|
|
row = (kelondroRow.Entry) i.next();
|
|
delete.add(row.getColBytes(index.row().primaryKey));
|
|
this.index.addUnique(row);
|
|
this.cacheFlush++;
|
|
}
|
|
i = delete.rows();
|
|
while (i.hasNext()) writeBufferDoubles.remove(((kelondroRow.Entry) i.next()).getColBytes(0));
|
|
delete = null;
|
|
writeBufferDoubles.trim(true);
|
|
}
|
|
}
|
|
|
|
public void flushSome() throws IOException {
|
|
if (writeBufferUnique != null) flushUnique(writeBufferUnique.size() / 10);
|
|
if (writeBufferDoubles != null) flushDoubles(writeBufferDoubles.size() / 10);
|
|
}
|
|
|
|
private int sumRecords() {
|
|
return
|
|
((readHitCache == null) ? 0 : readHitCache.size()) +
|
|
((writeBufferUnique == null) ? 0 : writeBufferUnique.size()) +
|
|
((writeBufferDoubles == null) ? 0 : writeBufferDoubles.size());
|
|
}
|
|
|
|
private void checkMissSpace() {
|
|
if ((readMissCache != null) && (cacheGrowStatus() < 1)) {
|
|
readMissCache.clear();
|
|
}
|
|
}
|
|
|
|
private void checkHitSpace() throws IOException {
|
|
int s = sumRecords();
|
|
if (cacheGrowStatus() < 2) {flushDoubles(s / 4); s = sumRecords();}
|
|
if (cacheGrowStatus() < 2) {flushUnique(s / 4); s = sumRecords();}
|
|
if ((cacheGrowStatus() < 2) && (readHitCache != null)) {
|
|
readHitCache.clear();
|
|
}
|
|
if (cacheGrowStatus() < 1) {
|
|
flushUnique();
|
|
flushDoubles();
|
|
if (readHitCache != null) {
|
|
readHitCache.clear();
|
|
}
|
|
}
|
|
}
|
|
|
|
public synchronized void close() {
|
|
try {
|
|
flushUnique();
|
|
} catch (IOException e) {
|
|
e.printStackTrace();
|
|
}
|
|
try {
|
|
flushDoubles();
|
|
} catch (IOException e) {
|
|
e.printStackTrace();
|
|
}
|
|
index.close();
|
|
readHitCache = null;
|
|
readMissCache = null;
|
|
writeBufferUnique = null;
|
|
writeBufferDoubles = null;
|
|
}
|
|
|
|
public boolean has(byte[] key) throws IOException {
|
|
return (get(key) != null);
|
|
}
|
|
|
|
public synchronized Entry get(byte[] key) throws IOException {
|
|
// first look into the miss cache
|
|
if (readMissCache != null) {
|
|
if (readMissCache.get(key) != null) {
|
|
this.hasnotHit++;
|
|
return null;
|
|
} else {
|
|
this.hasnotMiss++;
|
|
}
|
|
}
|
|
|
|
Entry entry = null;
|
|
|
|
// then try the hit cache and the buffers
|
|
if (readHitCache != null) {
|
|
entry = readHitCache.get(key);
|
|
if (entry != null) {
|
|
this.readHit++;
|
|
return entry;
|
|
}
|
|
}
|
|
if (writeBufferUnique != null) {
|
|
entry = writeBufferUnique.get(key);
|
|
if (entry != null) {
|
|
this.readHit++;
|
|
return entry;
|
|
}
|
|
}
|
|
if (writeBufferDoubles != null) {
|
|
entry = writeBufferDoubles.get(key);
|
|
if (entry != null) {
|
|
this.readHit++;
|
|
return entry;
|
|
}
|
|
}
|
|
|
|
// finally ask the backend index
|
|
this.readMiss++;
|
|
entry = index.get(key);
|
|
// learn from result
|
|
if (entry == null) {
|
|
checkMissSpace();
|
|
if (readMissCache != null) {
|
|
kelondroRow.Entry dummy = readMissCache.put(readMissCache.row().newEntry(key));
|
|
if (dummy == null) this.hasnotUnique++; else this.hasnotDouble++;
|
|
}
|
|
return null;
|
|
} else {
|
|
checkHitSpace();
|
|
if (readHitCache != null) {
|
|
kelondroRow.Entry dummy = readHitCache.put(entry);
|
|
if (dummy == null) this.writeUnique++; else this.writeDouble++;
|
|
}
|
|
return entry;
|
|
}
|
|
}
|
|
|
|
public synchronized void putMultiple(List rows, Date entryDate) throws IOException {
|
|
Iterator i = rows.iterator();
|
|
while (i.hasNext()) put ((Entry) i.next(), entryDate);
|
|
}
|
|
|
|
public synchronized Entry put(Entry row) throws IOException {
|
|
assert (row != null);
|
|
assert (row.columns() == row().columns());
|
|
//assert (!(serverLog.allZero(row.getColBytes(index.primarykey()))));
|
|
|
|
byte[] key = row.getColBytes(index.row().primaryKey);
|
|
checkHitSpace();
|
|
|
|
// remove entry from miss- and hit-cache
|
|
if (readMissCache != null) {
|
|
if (readMissCache.remove(key) != null) {
|
|
this.hasnotHit++;
|
|
// the entry does not exist before
|
|
if (writeBufferUnique != null) {
|
|
// since we know that the entry does not exist, we know that new
|
|
// entry belongs to the unique buffer
|
|
writeBufferUnique.put(row);
|
|
return null;
|
|
}
|
|
assert (writeBufferDoubles == null);
|
|
index.put(row); // write to backend
|
|
if (readHitCache != null) {
|
|
kelondroRow.Entry dummy = readHitCache.put(row); // learn that entry
|
|
if (dummy == null) this.writeUnique++; else this.writeDouble++;
|
|
}
|
|
return null;
|
|
}
|
|
}
|
|
|
|
Entry entry;
|
|
|
|
if (readHitCache != null) {
|
|
entry = readHitCache.get(key);
|
|
if (entry != null) {
|
|
// since we know that the entry was in the read cache, it cannot be in any write cache
|
|
if (writeBufferDoubles != null) {
|
|
// because the entry exists, it must be written in the doubles buffer
|
|
readHitCache.remove(key);
|
|
this.cacheDelete++;
|
|
writeBufferDoubles.put(row);
|
|
return entry;
|
|
} else {
|
|
// write directly to backend index
|
|
index.put(row);
|
|
// learn from situation
|
|
kelondroRow.Entry dummy = readHitCache.put(row); // overwrite old entry
|
|
if (dummy == null) this.writeUnique++; else this.writeDouble++;
|
|
return entry;
|
|
}
|
|
}
|
|
}
|
|
|
|
// we still don't know if the key exists. Look into the buffers
|
|
if (writeBufferUnique != null) {
|
|
entry = writeBufferUnique.get(key);
|
|
if (entry != null) {
|
|
writeBufferUnique.put(row);
|
|
return entry;
|
|
}
|
|
}
|
|
if (writeBufferDoubles != null) {
|
|
entry = writeBufferDoubles.get(key);
|
|
if (entry != null) {
|
|
writeBufferDoubles.put(row);
|
|
return entry;
|
|
}
|
|
}
|
|
|
|
// finally, we still don't know if this is a double-entry or unique-entry
|
|
// there is a chance to get that information 'cheap':
|
|
// look into the node ram cache of the back-end index.
|
|
// that does only work, if the node cache is complete
|
|
// that is the case for kelondroFlexTables with ram index
|
|
if ((writeBufferUnique != null) &&
|
|
(index instanceof kelondroFlexTable) &&
|
|
(((kelondroFlexTable) index).hasRAMIndex()) &&
|
|
(!(((kelondroFlexTable) index).has(key)))) {
|
|
// this an unique entry
|
|
writeBufferUnique.put(row);
|
|
return null; // since that was unique, there was no entry before
|
|
}
|
|
|
|
// the worst case: we must write to the back-end directly
|
|
entry = index.put(row);
|
|
if (readHitCache != null) {
|
|
kelondroRow.Entry dummy = readHitCache.put(row); // learn that entry
|
|
if (dummy == null) this.writeUnique++; else this.writeDouble++;
|
|
}
|
|
return entry;
|
|
}
|
|
|
|
public synchronized Entry put(Entry row, Date entryDate) throws IOException {
|
|
// a put with a date is bad for the cache: the date cannot be handled
|
|
// The write buffer does not work here, because it does not store dates.
|
|
|
|
if (entryDate == null) return put(row);
|
|
|
|
assert (row != null);
|
|
assert (row.columns() == row().columns());
|
|
//assert (!(serverLog.allZero(row.getColBytes(index.primarykey()))));
|
|
assert (writeBufferUnique == null);
|
|
assert (writeBufferDoubles == null);
|
|
|
|
byte[] key = row.getColBytes(index.row().primaryKey);
|
|
checkHitSpace();
|
|
|
|
// remove entry from miss- and hit-cache
|
|
if (readMissCache != null) {
|
|
this.readMissCache.remove(key);
|
|
this.hasnotDelete++;
|
|
}
|
|
|
|
// the worst case: we must write to the backend directly
|
|
Entry entry = index.put(row);
|
|
if (readHitCache != null) {
|
|
kelondroRow.Entry dummy = readHitCache.put(row); // learn that entry
|
|
if (dummy == null) this.writeUnique++; else this.writeDouble++;
|
|
}
|
|
return entry;
|
|
}
|
|
|
|
public synchronized void addUnique(Entry row) throws IOException {
|
|
assert (row != null);
|
|
assert (row.columns() == row().columns());
|
|
//assert (!(serverLog.allZero(row.getColBytes(index.primarykey()))));
|
|
|
|
byte[] key = row.getColBytes(index.row().primaryKey);
|
|
checkHitSpace();
|
|
|
|
// remove entry from miss- and hit-cache
|
|
if (readMissCache != null) {
|
|
this.readMissCache.remove(key);
|
|
this.hasnotDelete++;
|
|
// the entry does not exist before
|
|
if (writeBufferUnique != null) {
|
|
// since we know that the entry does not exist, we know that new
|
|
// entry belongs to the unique buffer
|
|
writeBufferUnique.put(row);
|
|
return;
|
|
}
|
|
assert (writeBufferDoubles == null);
|
|
index.addUnique(row); // write to backend
|
|
if (readHitCache != null) {
|
|
kelondroRow.Entry dummy = readHitCache.put(row); // learn that entry
|
|
if (dummy == null) this.writeUnique++; else this.writeDouble++;
|
|
}
|
|
return;
|
|
}
|
|
|
|
if ((writeBufferUnique != null) &&
|
|
(index instanceof kelondroFlexTable) &&
|
|
(((kelondroFlexTable) index).hasRAMIndex()) &&
|
|
(!(((kelondroFlexTable) index).has(key)))) {
|
|
// this an unique entry
|
|
writeBufferUnique.addUnique(row);
|
|
return;
|
|
}
|
|
|
|
// the worst case: we must write to the back-end directly
|
|
index.addUnique(row);
|
|
if (readHitCache != null) {
|
|
kelondroRow.Entry dummy = readHitCache.put(row); // learn that entry
|
|
if (dummy == null) this.writeUnique++; else this.writeDouble++;
|
|
}
|
|
}
|
|
|
|
public synchronized void addUnique(Entry row, Date entryDate) throws IOException {
|
|
if (entryDate == null) {
|
|
addUnique(row);
|
|
return;
|
|
}
|
|
|
|
assert (row != null);
|
|
assert (row.columns() == row().columns());
|
|
//assert (!(serverLog.allZero(row.getColBytes(index.primarykey()))));
|
|
assert (writeBufferUnique == null);
|
|
assert (writeBufferDoubles == null);
|
|
|
|
byte[] key = row.getColBytes(index.row().primaryKey);
|
|
checkHitSpace();
|
|
|
|
// remove entry from miss- and hit-cache
|
|
if (readMissCache != null) {
|
|
this.readMissCache.remove(key);
|
|
this.hasnotDelete++;
|
|
}
|
|
|
|
// the worst case: we must write to the backend directly
|
|
index.addUnique(row);
|
|
if (readHitCache != null) {
|
|
kelondroRow.Entry dummy = readHitCache.put(row); // learn that entry
|
|
if (dummy == null) this.writeUnique++; else this.writeDouble++;
|
|
}
|
|
}
|
|
|
|
public synchronized void addUniqueMultiple(List rows, Date entryDate) throws IOException {
|
|
Iterator i = rows.iterator();
|
|
while (i.hasNext()) addUnique((Entry) i.next(), entryDate);
|
|
}
|
|
|
|
public synchronized Entry remove(byte[] key) throws IOException {
|
|
|
|
checkMissSpace();
|
|
|
|
// add entry to miss-cache
|
|
if (readMissCache != null) {
|
|
// set the miss cache; if there was already an entry we know that the return value must be null
|
|
kelondroRow.Entry dummy = readMissCache.put(readMissCache.row().newEntry(key));
|
|
if (dummy == null) {
|
|
this.hasnotUnique++;
|
|
} else {
|
|
this.hasnotHit++;
|
|
this.hasnotDouble++;
|
|
return null;
|
|
}
|
|
}
|
|
|
|
// remove entry from hit-cache
|
|
if (readHitCache != null) {
|
|
Entry entry = readHitCache.remove(key);
|
|
if (entry == null) {
|
|
this.readMiss++;
|
|
} else {
|
|
this.readHit++;
|
|
this.cacheDelete++;
|
|
index.remove(key);
|
|
return entry;
|
|
}
|
|
}
|
|
|
|
// if the key already exists in one buffer, remove that buffer
|
|
if (writeBufferUnique != null) {
|
|
Entry entry = writeBufferUnique.remove(key);
|
|
if (entry != null) return entry;
|
|
}
|
|
if (writeBufferDoubles != null) {
|
|
Entry entry = writeBufferDoubles.remove(key);
|
|
if (entry != null) {
|
|
index.remove(key);
|
|
return entry;
|
|
}
|
|
}
|
|
|
|
return index.remove(key);
|
|
}
|
|
|
|
public synchronized Entry removeOne() throws IOException {
|
|
|
|
checkMissSpace();
|
|
|
|
if ((writeBufferUnique != null) && (writeBufferUnique.size() > 0)) {
|
|
Entry entry = writeBufferUnique.removeOne();
|
|
if (readMissCache != null) {
|
|
kelondroRow.Entry dummy = readMissCache.put(readMissCache.row().newEntry(entry.getColBytes(index.row().primaryKey)));
|
|
if (dummy == null) this.hasnotUnique++; else this.hasnotDouble++;
|
|
}
|
|
return entry;
|
|
}
|
|
|
|
if ((writeBufferDoubles != null) && (writeBufferDoubles.size() > 0)) {
|
|
Entry entry = writeBufferDoubles.removeOne();
|
|
byte[] key = entry.getColBytes(index.row().primaryKey);
|
|
if (readMissCache != null) {
|
|
kelondroRow.Entry dummy = readMissCache.put(readMissCache.row().newEntry(key));
|
|
if (dummy == null) this.hasnotUnique++; else this.hasnotDouble++;
|
|
}
|
|
index.remove(key);
|
|
return entry;
|
|
}
|
|
|
|
Entry entry = index.removeOne();
|
|
if (entry == null) return null;
|
|
byte[] key = entry.getColBytes(index.row().primaryKey);
|
|
if (readMissCache != null) {
|
|
kelondroRow.Entry dummy = readMissCache.put(readMissCache.row().newEntry(key));
|
|
if (dummy == null) this.hasnotUnique++; else this.hasnotDouble++;
|
|
}
|
|
if (readHitCache != null) {
|
|
kelondroRow.Entry dummy = readHitCache.remove(key);
|
|
if (dummy != null) this.cacheDelete++;
|
|
}
|
|
return entry;
|
|
}
|
|
|
|
public synchronized kelondroRow row() throws IOException {
|
|
return index.row();
|
|
}
|
|
|
|
public synchronized kelondroCloneableIterator rows(boolean up, byte[] firstKey) throws IOException {
|
|
flushUnique();
|
|
return index.rows(up, firstKey);
|
|
}
|
|
|
|
public int size() throws IOException {
|
|
return index.size() + ((writeBufferUnique == null) ? 0 : writeBufferUnique.size());
|
|
}
|
|
|
|
public String filename() {
|
|
return index.filename();
|
|
}
|
|
|
|
}
|