From 222850414e19f7e24f403ca202c7ec776bdfd48e Mon Sep 17 00:00:00 2001 From: orbiter Date: Tue, 30 Jun 2009 09:27:46 +0000 Subject: [PATCH] simplification of the code: removed unused classes, methods and variables git-svn-id: https://svn.berlios.de/svnroot/repos/yacy/trunk@6154 6c8d7289-2bf4-0310-a012-ef5d649a1542 --- htroot/BlacklistCleaner_p.java | 4 +- htroot/QuickCrawlLink_p.java | 1 - htroot/ViewFile.java | 2 +- htroot/WatchCrawler_p.java | 5 +- htroot/yacy/search.java | 4 +- htroot/yacysearch.java | 3 +- source/dbtest.java | 2 +- source/de/anomic/crawler/CrawlProfile.java | 15 -- source/de/anomic/crawler/CrawlQueues.java | 4 +- .../de/anomic/crawler/CrawlSwitchboard.java | 14 +- source/de/anomic/data/SitemapParser.java | 2 +- source/de/anomic/data/bookmarksDB.java | 2 +- source/de/anomic/data/userDB.java | 8 +- source/de/anomic/data/wiki/knwikiParser.java | 25 +-- source/de/anomic/http/httpSSI.java | 4 +- source/de/anomic/http/httpd.java | 14 -- source/de/anomic/http/httpdFileHandler.java | 2 +- source/de/anomic/http/httpdProxyHandler.java | 16 +- .../anomic/kelondro/index/RowCollection.java | 4 +- .../anomic/kelondro/io/BufferedIOChunks.java | 145 ------------------ .../kelondro/io/BufferedRandomAccess.java | 88 ----------- .../anomic/kelondro/order/NaturalOrder.java | 2 +- source/de/anomic/kelondro/table/Records.java | 33 +--- .../de/anomic/kelondro/table/SplitTable.java | 14 +- source/de/anomic/kelondro/text/IndexCell.java | 4 +- .../kelondro/text/MetadataRepository.java | 2 +- .../text/ReferenceContainerArray.java | 2 +- .../referencePrototype/WordReferenceVars.java | 2 +- source/de/anomic/plasma/parser/Condenser.java | 14 +- .../de/anomic/plasma/plasmaParserConfig.java | 8 +- .../anomic/plasma/plasmaRankingCRProcess.java | 3 +- .../de/anomic/plasma/plasmaSearchEvent.java | 4 +- .../de/anomic/plasma/plasmaSnippetCache.java | 8 +- .../de/anomic/plasma/plasmaSwitchboard.java | 2 +- source/de/anomic/search/Query.java | 2 +- source/de/anomic/tools/crypt.java | 21 +-- .../anomic/urlRedirector/urlRedirectord.java | 2 - source/de/anomic/yacy/dht/PeerSelection.java | 3 +- source/de/anomic/yacy/yacyPeerActions.java | 6 +- source/de/anomic/yacy/yacyRelease.java | 2 +- source/de/anomic/yacy/yacySeedDB.java | 6 - source/de/anomic/ymage/ymageOSM.java | 2 +- 42 files changed, 74 insertions(+), 432 deletions(-) delete mode 100644 source/de/anomic/kelondro/io/BufferedIOChunks.java delete mode 100644 source/de/anomic/kelondro/io/BufferedRandomAccess.java diff --git a/htroot/BlacklistCleaner_p.java b/htroot/BlacklistCleaner_p.java index 0dbea480a..369ea0307 100644 --- a/htroot/BlacklistCleaner_p.java +++ b/htroot/BlacklistCleaner_p.java @@ -113,7 +113,7 @@ public class BlacklistCleaner_p { } // list illegal entries - final HashMap ies = getIllegalEntries(blacklistToUse, supportedBlacklistTypes, plasmaSwitchboard.urlBlacklist); + final HashMap ies = getIllegalEntries(blacklistToUse, plasmaSwitchboard.urlBlacklist); prop.put(RESULTS + "blList", blacklistToUse); prop.put(RESULTS + "entries", ies.size()); prop.putHTML(RESULTS + "blEngine", plasmaSwitchboard.urlBlacklist.getEngineInfo()); @@ -192,7 +192,7 @@ public class BlacklistCleaner_p { return r.toArray(new String[r.size()]); } - private static HashMap/* entry, error-code */ getIllegalEntries(final String blacklistToUse, final String[] supportedBlacklistTypes, final Blacklist blEngine) { + private static HashMap/* entry, error-code */ getIllegalEntries(final String blacklistToUse, final Blacklist blEngine) { final HashMap r = new HashMap(); final HashSet ok = new HashSet(); diff --git a/htroot/QuickCrawlLink_p.java b/htroot/QuickCrawlLink_p.java index b4e021907..187db6d29 100644 --- a/htroot/QuickCrawlLink_p.java +++ b/htroot/QuickCrawlLink_p.java @@ -132,7 +132,6 @@ public class QuickCrawlLink_p { pe = sb.crawler.profilesActiveCrawls.newEntry( crawlingStartURL.getHost(), crawlingStartURL, - CrawlProfile.KEYWORDS_USER, crawlingMustMatch, crawlingMustNotMatch, CrawlingDepth, diff --git a/htroot/ViewFile.java b/htroot/ViewFile.java index 4da078631..037b4347e 100644 --- a/htroot/ViewFile.java +++ b/htroot/ViewFile.java @@ -159,7 +159,7 @@ public class ViewFile { if (resource == null) { Document entry = null; try { - entry = sb.crawlQueues.loadResourceFromWeb(url, 10000, false, true, false); + entry = sb.crawlQueues.loadResourceFromWeb(url, true, false); } catch (final Exception e) { prop.put("error", "4"); prop.putHTML("error_errorText", e.getMessage()); diff --git a/htroot/WatchCrawler_p.java b/htroot/WatchCrawler_p.java index 39d100dce..792034c92 100644 --- a/htroot/WatchCrawler_p.java +++ b/htroot/WatchCrawler_p.java @@ -218,7 +218,6 @@ public class WatchCrawler_p { final CrawlProfile.entry pe = sb.crawler.profilesActiveCrawls.newEntry( crawlingStartURL.getHost(), crawlingStartURL, - CrawlProfile.KEYWORDS_USER, newcrawlingMustMatch, newcrawlingMustNotMatch, newcrawlingdepth, @@ -338,7 +337,7 @@ public class WatchCrawler_p { // creating a crawler profile final yacyURL crawlURL = new yacyURL("file://" + file.toString(), null); final CrawlProfile.entry profile = sb.crawler.profilesActiveCrawls.newEntry( - fileName, crawlURL, CrawlProfile.KEYWORDS_USER, + fileName, crawlURL, newcrawlingMustMatch, CrawlProfile.MATCH_NEVER, newcrawlingdepth, @@ -402,7 +401,7 @@ public class WatchCrawler_p { // create a new profile final CrawlProfile.entry pe = sb.crawler.profilesActiveCrawls.newEntry( - sitemapURLStr, sitemapURL, CrawlProfile.KEYWORDS_USER, + sitemapURLStr, sitemapURL, newcrawlingMustMatch, CrawlProfile.MATCH_NEVER, newcrawlingdepth, diff --git a/htroot/yacy/search.java b/htroot/yacy/search.java index 6a424f982..9f53bd49b 100644 --- a/htroot/yacy/search.java +++ b/htroot/yacy/search.java @@ -201,7 +201,6 @@ public final class search { 0, filter, Query.SEARCHDOM_LOCAL, - null, -1, null, false, @@ -255,7 +254,6 @@ public final class search { 0, filter, Query.SEARCHDOM_LOCAL, - null, -1, constraint, false, @@ -269,7 +267,7 @@ public final class search { RSSFeed.channels(RSSFeed.REMOTESEARCH).addMessage(new RSSMessage("Remote Search Request from " + ((remoteSeed == null) ? "unknown" : remoteSeed.getName()), Query.anonymizedQueryHashes(theQuery.queryHashes), "")); // make event - theSearch = plasmaSearchEvent.getEvent(theQuery, rankingProfile, sb.indexSegment, sb.peers, sb.crawlResults, null, true); + theSearch = plasmaSearchEvent.getEvent(theQuery, sb.indexSegment, sb.peers, sb.crawlResults, null, true); // set statistic details of search result and find best result index set if (theSearch.getRankingResult().getLocalResourceSize() == 0) { diff --git a/htroot/yacysearch.java b/htroot/yacysearch.java index 6f97471ae..2323d51ae 100644 --- a/htroot/yacysearch.java +++ b/htroot/yacysearch.java @@ -427,7 +427,6 @@ public class yacysearch { urlmask, (clustersearch && globalsearch) ? Query.SEARCHDOM_CLUSTERALL : ((globalsearch) ? Query.SEARCHDOM_GLOBALDHT : Query.SEARCHDOM_LOCAL), - "", 20, constraint, true, @@ -454,7 +453,7 @@ public class yacysearch { theQuery.setOffset(0); // in case that this is a new search, always start without a offset offset = 0; } - final plasmaSearchEvent theSearch = plasmaSearchEvent.getEvent(theQuery, ranking, sb.indexSegment, sb.peers, sb.crawlResults, (sb.isRobinsonMode()) ? sb.clusterhashes : null, false); + final plasmaSearchEvent theSearch = plasmaSearchEvent.getEvent(theQuery, sb.indexSegment, sb.peers, sb.crawlResults, (sb.isRobinsonMode()) ? sb.clusterhashes : null, false); // generate result object //serverLog.logFine("LOCAL_SEARCH", "SEARCH TIME AFTER ORDERING OF SEARCH RESULTS: " + (System.currentTimeMillis() - timestamp) + " ms"); diff --git a/source/dbtest.java b/source/dbtest.java index 1ded63bad..bcd5a668e 100644 --- a/source/dbtest.java +++ b/source/dbtest.java @@ -196,7 +196,7 @@ public class dbtest { } if (dbe.equals("kelondroSplitTable")) { final File tablepath = new File(tablename).getParentFile(); - return new SplitTable(tablepath, new File(tablename).getName(), testRow, true); + return new SplitTable(tablepath, new File(tablename).getName(), testRow); } if (dbe.equals("kelondroEcoTable")) { return new Table(new File(tablename), testRow, Table.tailCacheForceUsage, 1000, 0); diff --git a/source/de/anomic/crawler/CrawlProfile.java b/source/de/anomic/crawler/CrawlProfile.java index 7b9528816..7c16c76b0 100644 --- a/source/de/anomic/crawler/CrawlProfile.java +++ b/source/de/anomic/crawler/CrawlProfile.java @@ -28,7 +28,6 @@ import java.util.HashMap; import java.util.HashSet; import java.util.Iterator; import java.util.Map; -import java.util.Set; import java.util.concurrent.ConcurrentHashMap; import java.util.regex.Pattern; @@ -47,17 +46,6 @@ public class CrawlProfile { public static final String MATCH_ALL = ".*"; public static final String MATCH_NEVER = ""; - public static final HashSet NO_KEYWORDS = new HashSet(0); - public static final HashSet KEYWORDS_PROXY = word2set("xproxy"); - public static final HashSet KEYWORDS_REMOTE = word2set("xremote"); - public static final HashSet KEYWORDS_USER = word2set("xuser"); - public static final HashSet KEYWORDS_SNIPPET = word2set("xsnippet"); - - private static final HashSet word2set(String word) { - HashSet s = new HashSet(1); - s.add(word); - return s; - } static HashMap> domsCache = new HashMap>(); @@ -168,7 +156,6 @@ public class CrawlProfile { public entry newEntry( final String name, final yacyURL startURL, - final Set keywords, final String mustmatch, final String mustnotmatch, final int generalDepth, final long recrawlIfOlder /*date*/, final int domFilterDepth, final int domMaxPages, @@ -180,7 +167,6 @@ public class CrawlProfile { final entry ne = new entry( name, startURL, - keywords, mustmatch, mustnotmatch, generalDepth, recrawlIfOlder, domFilterDepth, domMaxPages, @@ -284,7 +270,6 @@ public class CrawlProfile { public entry(final String name, final yacyURL startURL, - final Set keywords, final String mustmatch, final String mustnotmatch, final int depth, diff --git a/source/de/anomic/crawler/CrawlQueues.java b/source/de/anomic/crawler/CrawlQueues.java index 8b226714d..b48e2dd46 100644 --- a/source/de/anomic/crawler/CrawlQueues.java +++ b/source/de/anomic/crawler/CrawlQueues.java @@ -505,9 +505,7 @@ public class CrawlQueues { } public Document loadResourceFromWeb( - final yacyURL url, - final int socketTimeout, - final boolean keepInMemory, + final yacyURL url, final boolean forText, final boolean global ) throws IOException { diff --git a/source/de/anomic/crawler/CrawlSwitchboard.java b/source/de/anomic/crawler/CrawlSwitchboard.java index cca281cf9..794e7069a 100644 --- a/source/de/anomic/crawler/CrawlSwitchboard.java +++ b/source/de/anomic/crawler/CrawlSwitchboard.java @@ -183,7 +183,7 @@ public final class CrawlSwitchboard { if (this.defaultProxyProfile == null) { // generate new default entry for proxy crawling - this.defaultProxyProfile = this.profilesActiveCrawls.newEntry("proxy", null, CrawlProfile.KEYWORDS_PROXY, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, + this.defaultProxyProfile = this.profilesActiveCrawls.newEntry("proxy", null, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0 /*Integer.parseInt(getConfig(PROXY_PREFETCH_DEPTH, "0"))*/, this.profilesActiveCrawls.getRecrawlDate(CRAWL_PROFILE_PROXY_RECRAWL_CYCLE), -1, -1, false, true /*getConfigBool(PROXY_INDEXING_LOCAL_TEXT, true)*/, @@ -193,32 +193,32 @@ public final class CrawlSwitchboard { } if (this.defaultRemoteProfile == null) { // generate new default entry for remote crawling - defaultRemoteProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_REMOTE, null, CrawlProfile.KEYWORDS_REMOTE, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, + defaultRemoteProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_REMOTE, null, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, -1, -1, -1, true, true, true, false, true, false, true, true, false); } if (this.defaultTextSnippetLocalProfile == null) { // generate new default entry for snippet fetch and optional crawling - defaultTextSnippetLocalProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SNIPPET_LOCAL_TEXT, null, CrawlProfile.KEYWORDS_SNIPPET, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, + defaultTextSnippetLocalProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SNIPPET_LOCAL_TEXT, null, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, this.profilesActiveCrawls.getRecrawlDate(CRAWL_PROFILE_SNIPPET_LOCAL_TEXT_RECRAWL_CYCLE), -1, -1, true, false, false, false, false, false, true, true, false); } if (this.defaultTextSnippetGlobalProfile == null) { // generate new default entry for snippet fetch and optional crawling - defaultTextSnippetGlobalProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SNIPPET_GLOBAL_TEXT, null, CrawlProfile.KEYWORDS_SNIPPET, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, + defaultTextSnippetGlobalProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SNIPPET_GLOBAL_TEXT, null, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, this.profilesActiveCrawls.getRecrawlDate(CRAWL_PROFILE_SNIPPET_GLOBAL_TEXT_RECRAWL_CYCLE), -1, -1, true, true, true, true, true, false, true, true, false); } if (this.defaultMediaSnippetLocalProfile == null) { // generate new default entry for snippet fetch and optional crawling - defaultMediaSnippetLocalProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SNIPPET_LOCAL_MEDIA, null, CrawlProfile.KEYWORDS_SNIPPET, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, + defaultMediaSnippetLocalProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SNIPPET_LOCAL_MEDIA, null, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, this.profilesActiveCrawls.getRecrawlDate(CRAWL_PROFILE_SNIPPET_LOCAL_MEDIA_RECRAWL_CYCLE), -1, -1, true, false, false, false, false, false, true, true, false); } if (this.defaultMediaSnippetGlobalProfile == null) { // generate new default entry for snippet fetch and optional crawling - defaultMediaSnippetGlobalProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SNIPPET_GLOBAL_MEDIA, null, CrawlProfile.KEYWORDS_SNIPPET, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, + defaultMediaSnippetGlobalProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SNIPPET_GLOBAL_MEDIA, null, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, this.profilesActiveCrawls.getRecrawlDate(CRAWL_PROFILE_SNIPPET_GLOBAL_MEDIA_RECRAWL_CYCLE), -1, -1, true, false, true, true, true, false, true, true, false); } if (this.defaultSurrogateProfile == null) { // generate new default entry for surrogate parsing - defaultSurrogateProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SURROGATE, null, CrawlProfile.KEYWORDS_SNIPPET, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, + defaultSurrogateProfile = this.profilesActiveCrawls.newEntry(CRAWL_PROFILE_SURROGATE, null, CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, 0, this.profilesActiveCrawls.getRecrawlDate(CRAWL_PROFILE_SURROGATE_RECRAWL_CYCLE), -1, -1, true, true, false, false, false, false, true, true, false); } } diff --git a/source/de/anomic/data/SitemapParser.java b/source/de/anomic/data/SitemapParser.java index f8b8cd87f..f552d7974 100644 --- a/source/de/anomic/data/SitemapParser.java +++ b/source/de/anomic/data/SitemapParser.java @@ -308,7 +308,7 @@ public class SitemapParser extends DefaultHandler { private CrawlProfile.entry createProfile(final String domainName, final yacyURL sitemapURL) { return this.sb.crawler.profilesActiveCrawls.newEntry( - domainName, sitemapURL, CrawlProfile.KEYWORDS_USER, + domainName, sitemapURL, // crawling Filter CrawlProfile.MATCH_ALL, CrawlProfile.MATCH_NEVER, // Depth diff --git a/source/de/anomic/data/bookmarksDB.java b/source/de/anomic/data/bookmarksDB.java index a5705a4f5..8f60e7b50 100644 --- a/source/de/anomic/data/bookmarksDB.java +++ b/source/de/anomic/data/bookmarksDB.java @@ -254,7 +254,7 @@ public class bookmarksDB { // stack url sb.crawler.profilesPassiveCrawls.removeEntry(crawlingStartURL.hash()); // if there is an old entry, delete it CrawlProfile.entry pe = sb.crawler.profilesActiveCrawls.newEntry( - folder+"/"+crawlingStartURL, crawlingStartURL, CrawlProfile.KEYWORDS_USER, + folder+"/"+crawlingStartURL, crawlingStartURL, newcrawlingMustMatch, CrawlProfile.MATCH_NEVER, newcrawlingdepth, diff --git a/source/de/anomic/data/userDB.java b/source/de/anomic/data/userDB.java index abf8c4d63..52c18c982 100644 --- a/source/de/anomic/data/userDB.java +++ b/source/de/anomic/data/userDB.java @@ -145,9 +145,9 @@ public final class userDB { return null; } public Entry getUser(final httpRequestHeader header){ - return getUser(header.get(httpRequestHeader.AUTHORIZATION), header.get(httpRequestHeader.CONNECTION_PROP_CLIENTIP), header.getHeaderCookies()); + return getUser(header.get(httpRequestHeader.AUTHORIZATION), header.getHeaderCookies()); } - public Entry getUser(final String auth, final String ip, final String cookies){ + public Entry getUser(final String auth, final String cookies){ Entry entry=null; if(auth != null) entry=proxyAuth(auth); @@ -161,8 +161,8 @@ public final class userDB { * @param auth * the http-headerline for authorisation */ - public boolean hasAdminRight(final String auth, final String ip, final String cookies) { - final Entry entry = getUser(auth, ip, cookies); + public boolean hasAdminRight(final String auth, final String cookies) { + final Entry entry = getUser(auth, cookies); if (entry != null) return entry.hasAdminRight(); // else if(entry != null && cookieAdminAuth(cookies)) diff --git a/source/de/anomic/data/wiki/knwikiParser.java b/source/de/anomic/data/wiki/knwikiParser.java index 4219d42ec..c2cfc5240 100644 --- a/source/de/anomic/data/wiki/knwikiParser.java +++ b/source/de/anomic/data/wiki/knwikiParser.java @@ -46,7 +46,7 @@ public class knwikiParser implements wikiParser { private String[] BEs; private final plasmaSwitchboard sb; - public knwikiParser(final plasmaSwitchboard sb) { + private knwikiParser(final plasmaSwitchboard sb) { this.sb = sb; } @@ -96,33 +96,16 @@ public class knwikiParser implements wikiParser { return parse(content, null); } - public String transform(final String content, final plasmaSwitchboard sb) { - return parse(content, null); - } - public String transform(final byte[] content) throws UnsupportedEncodingException { return parse(new String(content, "UTF-8"), null); } public String transform( - final byte[] content, final String encoding, - final plasmaSwitchboard switchboard) throws UnsupportedEncodingException { - return parse(new String(content, encoding), null); - } - - public String transform(final byte[] content, final String encoding) throws UnsupportedEncodingException { + final byte[] content, final String encoding) throws UnsupportedEncodingException { return parse(new String(content, encoding), null); } - - public String transform(final byte[] text, final String encoding, final String publicAddress) throws UnsupportedEncodingException { - return parse(new String(text, encoding), publicAddress); - } - - public String transform(final String text, final String publicAddress) { - return parse(text, publicAddress); - } - public String parse(String text, final String publicAddress) { + private String parse(String text, final String publicAddress) { tokens = new Token[] { new SimpleToken('=', '=', new String[][] { null, { "h2" }, { "h3" }, { "h4" } }, true), new SimpleToken('\'', '\'', new String[][] { null, { "i" }, { "b" }, null, { "b", "i" } }, false), @@ -159,7 +142,7 @@ public class knwikiParser implements wikiParser { return Text.mergeTexts(tt); } - public String parseUnescaped(String text) { + private String parseUnescaped(String text) { Token st; Matcher m; StringBuffer sb; diff --git a/source/de/anomic/http/httpSSI.java b/source/de/anomic/http/httpSSI.java index 5c7ce6660..d43517dbb 100644 --- a/source/de/anomic/http/httpSSI.java +++ b/source/de/anomic/http/httpSSI.java @@ -49,7 +49,7 @@ public class httpSSI { } else { out.write(in.getBytes(off, p - off)); } - parseSSI(in, p, q + 3 - p, out, authorization, requesthost); + parseSSI(in, p, out, authorization, requesthost); off = q + 3; p = in.indexOf("