*) Changing robots parser cxclusion policy

- crawling is now allowed if server returned a 403 statuscode 
     when trying to download the robots.txt
   See: http://www.yacy-forum.de/viewtopic.php?t=1612

git-svn-id: https://svn.berlios.de/svnroot/repos/yacy/trunk@1421 6c8d7289-2bf4-0310-a012-ef5d649a1542
pull/1/head
theli 19 years ago
parent c69f7a39a3
commit 754a35877f

@ -293,7 +293,7 @@ public final class robotsParser{
if (modDate != null) reqHeaders.put(httpHeader.IF_MODIFIED_SINCE,httpc.dateString(entry.getModDate())); if (modDate != null) reqHeaders.put(httpHeader.IF_MODIFIED_SINCE,httpc.dateString(entry.getModDate()));
} }
httpc.response res = con.GET(robotsURL.getPath(), reqHeaders); httpc.response res = con.GET(robotsURL.getFile(), reqHeaders);
if (res.status.startsWith("2")) { if (res.status.startsWith("2")) {
if (!res.responseHeader.mime().startsWith("text/plain")) { if (!res.responseHeader.mime().startsWith("text/plain")) {
robotsTxt = null; robotsTxt = null;
@ -336,7 +336,7 @@ public final class robotsParser{
"\nRedirecting request to: " + redirectionUrl); "\nRedirecting request to: " + redirectionUrl);
return downloadRobotsTxt(redirectionUrl,redirectionCount,entry); return downloadRobotsTxt(redirectionUrl,redirectionCount,entry);
} else if (res.status.startsWith("401") || res.status.startsWith("403")) { } else if (res.status.startsWith("401")/* || res.status.startsWith("403") */) {
accessCompletelyRestricted = true; accessCompletelyRestricted = true;
serverLog.logFinest("ROBOTS","Access to Robots.txt not allowed on URL '" + robotsURL + "'."); serverLog.logFinest("ROBOTS","Access to Robots.txt not allowed on URL '" + robotsURL + "'.");
} else { } else {

Loading…
Cancel
Save