|
|
|
// plasmaCrawlerLoader.java
|
|
|
|
// ------------------------
|
|
|
|
// part of YaCy
|
|
|
|
// (C) by Michael Peter Christen; mc@anomic.de
|
|
|
|
// first published on http://www.anomic.de
|
|
|
|
// Frankfurt, Germany, 2004
|
|
|
|
// last major change: 25.02.2004
|
|
|
|
//
|
|
|
|
// This program is free software; you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU General Public License as published by
|
|
|
|
// the Free Software Foundation; either version 2 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
|
|
// along with this program; if not, write to the Free Software
|
|
|
|
// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
//
|
|
|
|
// Using this software in any meaning (reading, learning, copying, compiling,
|
|
|
|
// running) means that you agree that the Author(s) is (are) not responsible
|
|
|
|
// for cost, loss of data or any harm that may be caused directly or indirectly
|
|
|
|
// by usage of this softare or this documentation. The usage of this software
|
|
|
|
// is on your own risk. The installation and usage (starting/running) of this
|
|
|
|
// software may allow other people or application to access your computer and
|
|
|
|
// any attached devices and is highly dependent on the configuration of the
|
|
|
|
// software which must be done by the user of the software; the author(s) is
|
|
|
|
// (are) also not responsible for proper configuration and usage of the
|
|
|
|
// software, even if provoked by documentation provided together with
|
|
|
|
// the software.
|
|
|
|
//
|
|
|
|
// Any changes to this file according to the GPL as documented in the file
|
|
|
|
// gpl.txt aside this file in the shipment you received can be done to the
|
|
|
|
// lines that follows this copyright notice here, but changes must not be
|
|
|
|
// done inside the copyright notive above. A re-distribution must contain
|
|
|
|
// the intact and unchanged copyright notice.
|
|
|
|
// Contributions and changes to the program code must be marked as such.
|
|
|
|
|
|
|
|
package de.anomic.plasma;
|
|
|
|
|
|
|
|
import java.io.*;
|
|
|
|
import java.util.*;
|
|
|
|
import java.net.*;
|
|
|
|
import de.anomic.net.*;
|
|
|
|
import de.anomic.http.*;
|
|
|
|
import de.anomic.server.*;
|
|
|
|
import de.anomic.tools.*;
|
|
|
|
import de.anomic.htmlFilter.*;
|
|
|
|
|
|
|
|
public final class plasmaCrawlLoader {
|
|
|
|
|
|
|
|
private plasmaHTCache cacheManager;
|
|
|
|
private int socketTimeout;
|
|
|
|
private int loadTimeout;
|
|
|
|
private boolean remoteProxyUse;
|
|
|
|
private String remoteProxyHost;
|
|
|
|
private int remoteProxyPort;
|
|
|
|
private int maxSlots;
|
|
|
|
private List slots;
|
|
|
|
private serverLog log;
|
|
|
|
private HashSet acceptMimeTypes;
|
|
|
|
|
|
|
|
public plasmaCrawlLoader(plasmaHTCache cacheManager, serverLog log, int socketTimeout, int loadTimeout, int mslots, boolean proxyUse, String proxyHost, int proxyPort,
|
|
|
|
HashSet acceptMimeTypes) {
|
|
|
|
this.cacheManager = cacheManager;
|
|
|
|
this.log = log;
|
|
|
|
this.socketTimeout = socketTimeout;
|
|
|
|
this.loadTimeout = loadTimeout;
|
|
|
|
this.remoteProxyUse = proxyUse;
|
|
|
|
this.remoteProxyHost = proxyHost;
|
|
|
|
this.remoteProxyPort = proxyPort;
|
|
|
|
this.maxSlots = mslots;
|
|
|
|
this.slots = new LinkedList();
|
|
|
|
this.acceptMimeTypes = acceptMimeTypes;
|
|
|
|
}
|
|
|
|
|
|
|
|
private void killTimeouts() {
|
|
|
|
Exec thread;
|
|
|
|
for (int i = slots.size() - 1; i >= 0; i--) {
|
|
|
|
// check if thread is alive
|
|
|
|
thread = (Exec) slots.get(i);
|
|
|
|
if (thread.isAlive()) {
|
|
|
|
// check the age of the thread
|
|
|
|
if (System.currentTimeMillis() - thread.startdate > loadTimeout) {
|
|
|
|
// we kill that thread
|
|
|
|
thread.interrupt(); // hopefully this wakes him up.
|
|
|
|
slots.remove(i);
|
|
|
|
log.logDebug("IGNORING SLEEPING DOWNLOAD SLOT " + thread.url.toString());
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// thread i is dead, remove it
|
|
|
|
slots.remove(i);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
public synchronized void loadParallel(URL url, String referer, String initiator, int depth, plasmaCrawlProfile.entry profile) {
|
|
|
|
|
|
|
|
// wait until there is space in the download slots
|
|
|
|
Exec thread;
|
|
|
|
while (slots.size() >= maxSlots) {
|
|
|
|
killTimeouts();
|
|
|
|
|
|
|
|
// wait a while
|
|
|
|
try {
|
|
|
|
Thread.currentThread().sleep(1000);
|
|
|
|
} catch (InterruptedException e) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// we found space in the download slots
|
|
|
|
thread = new Exec(url, referer, initiator, depth, profile);
|
|
|
|
thread.start();
|
|
|
|
slots.add(thread);
|
|
|
|
}
|
|
|
|
|
|
|
|
public int size() {
|
|
|
|
killTimeouts();
|
|
|
|
return slots.size();
|
|
|
|
}
|
|
|
|
|
|
|
|
public Exec[] threadStatus() {
|
|
|
|
killTimeouts();
|
|
|
|
Exec[] result = new Exec[slots.size()];
|
|
|
|
for (int i = 0; i < slots.size(); i++) result[i] = (Exec) slots.get(i);
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
public final class Exec extends Thread {
|
|
|
|
|
|
|
|
public URL url;
|
|
|
|
public String referer;
|
|
|
|
public String initiator;
|
|
|
|
public int depth;
|
|
|
|
public long startdate;
|
|
|
|
public plasmaCrawlProfile.entry profile;
|
|
|
|
public String error;
|
|
|
|
|
|
|
|
public Exec(URL url, String referer, String initiator, int depth, plasmaCrawlProfile.entry profile) {
|
|
|
|
this.url = url; // the url to crawl
|
|
|
|
this.referer = referer; // the url that contained this url as link
|
|
|
|
this.initiator = initiator;
|
|
|
|
this.depth = depth; // distance from start-url
|
|
|
|
this.startdate = System.currentTimeMillis();
|
|
|
|
this.profile = profile;
|
|
|
|
this.error = null;
|
|
|
|
}
|
|
|
|
|
|
|
|
public void run() {
|
|
|
|
try {
|
|
|
|
load(url, referer, initiator, depth, profile);
|
|
|
|
} catch (IOException e) {
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private httpc newhttpc(String server, int port, boolean ssl) throws IOException {
|
|
|
|
// a new httpc connection, combined with possible remote proxy
|
|
|
|
if (remoteProxyUse)
|
|
|
|
return httpc.getInstance(server, port, socketTimeout, ssl, remoteProxyHost, remoteProxyPort);
|
|
|
|
else return httpc.getInstance(server, port, socketTimeout, ssl);
|
|
|
|
}
|
|
|
|
|
|
|
|
private void load(URL url, String referer, String initiator, int depth, plasmaCrawlProfile.entry profile) throws IOException {
|
|
|
|
if (url == null) return;
|
|
|
|
Date requestDate = new Date(); // remember the time...
|
|
|
|
String host = url.getHost();
|
|
|
|
String path = url.getPath();
|
|
|
|
int port = url.getPort();
|
|
|
|
boolean ssl = url.getProtocol().equals("https");
|
|
|
|
if (port < 0) port = (ssl) ? 443 : 80;
|
|
|
|
|
|
|
|
// set referrer; in some case advertise a little bit:
|
|
|
|
referer = referer.trim();
|
|
|
|
if (referer.length() == 0) referer = "http://www.yacy.net/yacy/";
|
|
|
|
|
|
|
|
// take a file from the net
|
|
|
|
httpc remote = null;
|
|
|
|
try {
|
|
|
|
// create a request header
|
|
|
|
httpHeader requestHeader = new httpHeader();
|
|
|
|
requestHeader.put("User-Agent", httpdProxyHandler.userAgent);
|
|
|
|
requestHeader.put("Referer", referer);
|
|
|
|
requestHeader.put("Accept-Encoding", "gzip,deflate");
|
|
|
|
|
|
|
|
//System.out.println("CRAWLER_REQUEST_HEADER=" + requestHeader.toString()); // DEBUG
|
|
|
|
|
|
|
|
// open the connection
|
|
|
|
remote = newhttpc(host, port, ssl);
|
|
|
|
|
|
|
|
// send request
|
|
|
|
httpc.response res = remote.GET(path, requestHeader);
|
|
|
|
|
|
|
|
if (res.status.startsWith("200")) {
|
|
|
|
// the transfer is ok
|
|
|
|
long contentLength = res.responseHeader.contentLength();
|
|
|
|
|
|
|
|
// reserve cache entry
|
|
|
|
plasmaHTCache.Entry htCache = cacheManager.newEntry(requestDate, depth, url, requestHeader, res.status, res.responseHeader, initiator, profile);
|
|
|
|
|
|
|
|
// request has been placed and result has been returned. work off response
|
|
|
|
File cacheFile = cacheManager.getCachePath(url);
|
|
|
|
try {
|
|
|
|
if (!(httpd.isTextMime(res.responseHeader.mime().toLowerCase(), acceptMimeTypes))) {
|
|
|
|
// if the response has not the right file type then reject file
|
|
|
|
remote.close();
|
|
|
|
log.logInfo("REJECTED WRONG MIME TYPE " + res.responseHeader.mime() + " for url " + url.toString());
|
|
|
|
htCache.status = plasmaHTCache.CACHE_UNFILLED;
|
|
|
|
} else if ((profile.storeHTCache()) && ((error = htCache.shallStoreCache()) == null)) {
|
|
|
|
// we write the new cache entry to file system directly
|
|
|
|
cacheFile.getParentFile().mkdirs();
|
|
|
|
FileOutputStream fos = new FileOutputStream(cacheFile);
|
|
|
|
htCache.cacheArray = res.writeContent(fos); // writes in cacheArray and cache file
|
|
|
|
fos.close();
|
|
|
|
htCache.status = plasmaHTCache.CACHE_FILL;
|
|
|
|
} else {
|
|
|
|
if (error != null) log.logDebug("CRAWLER NOT STORED RESOURCE " + url.toString() + ": " + error);
|
|
|
|
// anyway, the content still lives in the content scraper
|
|
|
|
htCache.cacheArray = res.writeContent(null); // writes only into cacheArray
|
|
|
|
htCache.status = plasmaHTCache.CACHE_PASSING;
|
|
|
|
}
|
|
|
|
// enQueue new entry with response header
|
|
|
|
if ((initiator == null) || (initiator.length() == 0)) {
|
|
|
|
// enqueued for proxy writings
|
|
|
|
cacheManager.stackProcess(htCache);
|
|
|
|
} else {
|
|
|
|
// direct processing for crawling
|
|
|
|
cacheManager.process(htCache);
|
|
|
|
}
|
|
|
|
} catch (SocketException e) {
|
|
|
|
// this may happen if the client suddenly closes its connection
|
|
|
|
// maybe the user has stopped loading
|
|
|
|
// in that case, we are not responsible and just forget it
|
|
|
|
// but we clean the cache also, since it may be only partial
|
|
|
|
// and most possible corrupted
|
|
|
|
if (cacheFile.exists()) cacheFile.delete();
|
|
|
|
log.logError("CRAWLER LOADER ERROR1: with url=" + url.toString() + ": " + e.toString());
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// if the response has not the right response type then reject file
|
|
|
|
log.logInfo("REJECTED WRONG STATUS TYPE '" + res.status + "' for url " + url.toString());
|
|
|
|
// not processed any further
|
|
|
|
}
|
|
|
|
remote.close();
|
|
|
|
} catch (Exception e) {
|
|
|
|
// this may happen if the targeted host does not exist or anything with the
|
|
|
|
// remote server was wrong.
|
|
|
|
log.logError("CRAWLER LOADER ERROR2 with url=" + url.toString() + ": " + e.toString());
|
|
|
|
e.printStackTrace();
|
|
|
|
} finally {
|
|
|
|
if (remote != null) httpc.returnInstance(remote);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|