diff --git a/htroot/Bookmarks.java b/htroot/Bookmarks.java index 43ed0ec1a..b1b962571 100644 --- a/htroot/Bookmarks.java +++ b/htroot/Bookmarks.java @@ -50,7 +50,6 @@ import net.yacy.kelondro.data.meta.URIMetadataRow; import net.yacy.kelondro.logging.Log; import net.yacy.peers.NewsPool; import net.yacy.search.Switchboard; -import net.yacy.search.index.Segments; import de.anomic.data.BookmarkHelper; import de.anomic.data.BookmarksDB; import de.anomic.data.BookmarksDB.Bookmark; @@ -195,7 +194,7 @@ public class Bookmarks { final BookmarksDB.Bookmark bookmark = sb.bookmarksDB.getBookmark(urlHash); if (bookmark == null) { // try to get the bookmark from the LURL database - final URIMetadataRow urlentry = sb.indexSegments.urlMetadata(Segments.Process.PUBLIC).load(ASCII.getBytes(urlHash)); + final URIMetadataRow urlentry = sb.index.urlMetadata().load(ASCII.getBytes(urlHash)); if (urlentry != null) try { final Document document = Document.mergeDocuments(urlentry.url(), null, sb.loader.loadDocuments(sb.loader.request(urlentry.url(), true, false), CacheStrategy.IFEXIST, 5000, Integer.MAX_VALUE)); prop.put("mode_edit", "0"); // create mode diff --git a/htroot/CrawlResults.java b/htroot/CrawlResults.java index 39c901113..de66ff41e 100644 --- a/htroot/CrawlResults.java +++ b/htroot/CrawlResults.java @@ -39,7 +39,6 @@ import net.yacy.kelondro.data.meta.URIMetadataRow; import net.yacy.kelondro.logging.Log; import net.yacy.peers.Seed; import net.yacy.search.Switchboard; -import net.yacy.search.index.Segments; import de.anomic.crawler.ResultURLs; import de.anomic.crawler.ResultURLs.EventOrigin; import de.anomic.crawler.ResultURLs.InitExecEntry; @@ -117,7 +116,7 @@ public class CrawlResults { final String hash = post.get("hash", null); if (hash != null) { // delete from database - sb.indexSegments.urlMetadata(Segments.Process.LOCALCRAWLING).remove(hash.getBytes()); + sb.index.urlMetadata().remove(hash.getBytes()); } } @@ -127,7 +126,7 @@ public class CrawlResults { if (hashpart != null) { // delete all urls for this domain from database try { - sb.indexSegments.urlMetadata(Segments.Process.LOCALCRAWLING).deleteDomain(hashpart); + sb.index.urlMetadata().deleteDomain(hashpart); ResultURLs.deleteDomain(tabletype, domain, hashpart); } catch (final IOException e) { Log.logException(e); @@ -187,7 +186,7 @@ public class CrawlResults { while (i.hasNext()) { entry = i.next(); try { - urle = sb.indexSegments.urlMetadata(Segments.Process.LOCALCRAWLING).load(UTF8.getBytes(entry.getKey())); + urle = sb.index.urlMetadata().load(UTF8.getBytes(entry.getKey())); if (urle == null) { Log.logWarning("PLASMA", "CrawlResults: URL not in index with url hash " + entry.getKey()); urlstr = null; diff --git a/htroot/Crawler_p.java b/htroot/Crawler_p.java index ceacd2286..6b43fd250 100644 --- a/htroot/Crawler_p.java +++ b/htroot/Crawler_p.java @@ -53,7 +53,6 @@ import net.yacy.peers.NewsPool; import net.yacy.search.Switchboard; import net.yacy.search.SwitchboardConstants; import net.yacy.search.index.Segment; -import net.yacy.search.index.Segments; import de.anomic.crawler.CrawlProfile; import de.anomic.crawler.SitemapImporter; import de.anomic.crawler.ZURL.FailCategory; @@ -95,16 +94,7 @@ public class Crawler_p { prop.put("forwardToCrawlStart", "0"); // get segment - Segment indexSegment = null; - if (post != null && post.containsKey("segment")) { - final String segmentName = post.get("segment"); - if (sb.indexSegments.segmentExist(segmentName)) { - indexSegment = sb.indexSegments.segment(segmentName); - } - } else { - // take default segment - indexSegment = sb.indexSegments.segment(Segments.Process.PUBLIC); - } + Segment indexSegment = sb.index; prop.put("info", "0"); diff --git a/htroot/IndexCleaner_p.java b/htroot/IndexCleaner_p.java index a7e0e990a..c8f7db167 100644 --- a/htroot/IndexCleaner_p.java +++ b/htroot/IndexCleaner_p.java @@ -29,7 +29,6 @@ import net.yacy.cora.protocol.RequestHeader; import net.yacy.search.Switchboard; import net.yacy.search.index.MetadataRepository; import net.yacy.search.index.Segment; -import net.yacy.search.index.Segments; import de.anomic.server.serverObjects; import de.anomic.server.serverSwitch; @@ -41,19 +40,10 @@ public class IndexCleaner_p { final serverObjects prop = new serverObjects(); final Switchboard sb = (Switchboard) env; prop.put("title", "DbCleanup_p"); - + // get segment - Segment indexSegment = null; - if (post != null && post.containsKey("segment")) { - String segmentName = post.get("segment"); - if (sb.indexSegments.segmentExist(segmentName)) { - indexSegment = sb.indexSegments.segment(segmentName); - } - } else { - // take default segment - indexSegment = sb.indexSegments.segment(Segments.Process.PUBLIC); - } - + Segment indexSegment = sb.index; + if (post!=null) { if (post.get("action").equals("ustart")) { if (urldbCleanerThread==null || !urldbCleanerThread.isAlive()) { diff --git a/htroot/IndexControlRWIs_p.html b/htroot/IndexControlRWIs_p.html index 1dc3760e3..448b3d2ee 100644 --- a/htroot/IndexControlRWIs_p.html +++ b/htroot/IndexControlRWIs_p.html @@ -11,16 +11,7 @@
The local index currently contains #[wcount]# reverse word indexes