From 5f8a5ca32d50a196c4df1ffd9e9b9e8f66d1b075 Mon Sep 17 00:00:00 2001 From: sixcooler Date: Wed, 24 Aug 2011 12:07:53 +0000 Subject: [PATCH] - not doing merge-jobs while short on Memory - using configuration-values of crawling-max-filesize also for snippetfetching and loading files into Index git-svn-id: https://svn.berlios.de/svnroot/repos/yacy/trunk@7893 6c8d7289-2bf4-0310-a012-ef5d649a1542 --- source/de/anomic/crawler/CrawlQueues.java | 4 +--- source/de/anomic/search/Switchboard.java | 4 ++-- source/de/anomic/search/TextSnippet.java | 2 +- .../net/yacy/kelondro/rwi/IODispatcher.java | 2 +- .../net/yacy/repository/LoaderDispatcher.java | 20 +++++++++++++++---- 5 files changed, 21 insertions(+), 11 deletions(-) diff --git a/source/de/anomic/crawler/CrawlQueues.java b/source/de/anomic/crawler/CrawlQueues.java index 1ddd132e3..4d6e9099f 100644 --- a/source/de/anomic/crawler/CrawlQueues.java +++ b/source/de/anomic/crawler/CrawlQueues.java @@ -48,7 +48,6 @@ import net.yacy.kelondro.util.FileUtils; import net.yacy.kelondro.workflow.WorkflowJob; import de.anomic.crawler.NoticedURL.StackType; import de.anomic.crawler.ZURL.FailCategory; -import de.anomic.crawler.retrieval.HTTPLoader; import de.anomic.crawler.retrieval.Request; import de.anomic.crawler.retrieval.Response; import de.anomic.search.Segments; @@ -595,9 +594,8 @@ public class CrawlQueues { // returns null if everything went fine, a fail reason string if a problem occurred try { this.request.setStatus("loading", WorkflowJob.STATUS_RUNNING); - final int maxFileSize = CrawlQueues.this.sb.getConfigInt("crawler.http.maxFileSize", HTTPLoader.DEFAULT_MAXFILESIZE); final CrawlProfile e = CrawlQueues.this.sb.crawler.getActive(UTF8.getBytes(this.request.profileHandle())); - final Response response = CrawlQueues.this.sb.loader.load(this.request, e == null ? CacheStrategy.IFEXIST : e.cacheStrategy(), maxFileSize, true); + final Response response = CrawlQueues.this.sb.loader.load(this.request, e == null ? CacheStrategy.IFEXIST : e.cacheStrategy(), true); if (response == null) { this.request.setStatus("error", WorkflowJob.STATUS_FINISHED); if (CrawlQueues.this.log.isFine()) CrawlQueues.this.log.logFine("problem loading " + this.request.url().toString() + ": no content (possibly caused by cache policy)"); diff --git a/source/de/anomic/search/Switchboard.java b/source/de/anomic/search/Switchboard.java index 529062e3c..d41f1a9af 100644 --- a/source/de/anomic/search/Switchboard.java +++ b/source/de/anomic/search/Switchboard.java @@ -2216,7 +2216,7 @@ public final class Switchboard extends serverSwitch { @Override public void run() { try { - final Response response = Switchboard.this.loader.load(request, CacheStrategy.IFFRESH, Integer.MAX_VALUE, true); + final Response response = Switchboard.this.loader.load(request, CacheStrategy.IFFRESH, true); if (response == null) { throw new IOException("response == null"); } @@ -2610,7 +2610,7 @@ public final class Switchboard extends serverSwitch { // if we have an url then try to load the rss RSSReader rss = null; try { - final Response response = sb.loader.load(sb.loader.request(url, true, false), CacheStrategy.NOCACHE, Integer.MAX_VALUE, true); + final Response response = sb.loader.load(sb.loader.request(url, true, false), CacheStrategy.NOCACHE, true); final byte[] resource = (response == null) ? null : response.getContent(); //System.out.println("BLEKKO: " + UTF8.String(resource)); rss = resource == null ? null : RSSReader.parse(RSSFeed.DEFAULT_MAXSIZE, resource); diff --git a/source/de/anomic/search/TextSnippet.java b/source/de/anomic/search/TextSnippet.java index 8ee4164c1..6fe34556b 100644 --- a/source/de/anomic/search/TextSnippet.java +++ b/source/de/anomic/search/TextSnippet.java @@ -266,7 +266,7 @@ public class TextSnippet implements Comparable, Comparator