You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
204 lines
6.5 KiB
204 lines
6.5 KiB
// YMarkCrawlStart.java
|
|
// (C) 2012 by Stefan Förster, sof@gmx.de, Norderstedt, Germany
|
|
// first published 2011 on http://yacy.net
|
|
//
|
|
// This is a part of YaCy, a peer-to-peer based web search engine
|
|
//
|
|
// $LastChangedDate: 2011-03-09 13:50:39 +0100 (Mi, 09 Mrz 2011) $
|
|
// $LastChangedRevision: 7574 $
|
|
// $LastChangedBy: apfelmaennchen $
|
|
//
|
|
// LICENSE
|
|
//
|
|
// This program is free software; you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation; either version 2 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program; if not, write to the Free Software
|
|
// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
package net.yacy.data.ymark;
|
|
|
|
import java.io.IOException;
|
|
import java.util.Date;
|
|
import java.util.HashMap;
|
|
import java.util.Iterator;
|
|
import java.util.regex.Pattern;
|
|
|
|
import net.yacy.cora.document.UTF8;
|
|
import net.yacy.cora.federate.yacy.CacheStrategy;
|
|
import net.yacy.crawler.CrawlSwitchboard;
|
|
import net.yacy.crawler.data.CrawlProfile;
|
|
import net.yacy.crawler.retrieval.Request;
|
|
import net.yacy.data.WorkTables;
|
|
import net.yacy.kelondro.blob.Tables;
|
|
import net.yacy.kelondro.data.meta.DigestURI;
|
|
import net.yacy.search.Switchboard;
|
|
|
|
public class YMarkCrawlStart extends HashMap<String,String>{
|
|
|
|
private static final long serialVersionUID = 1L;
|
|
private final WorkTables worktables;
|
|
private Date date_last_exec;
|
|
private Date date_next_exec;
|
|
private Date date_recording;
|
|
private String apicall_pk;
|
|
private String url;
|
|
|
|
public static enum CRAWLSTART {
|
|
SINGLE, ONE_LINK, FULL_DOMAIN
|
|
}
|
|
|
|
public YMarkCrawlStart(final WorkTables worktables) {
|
|
super();
|
|
this.date_recording = new Date(0);
|
|
this.worktables = worktables;
|
|
}
|
|
|
|
public YMarkCrawlStart(final WorkTables worktables, final String url) {
|
|
super();
|
|
this.worktables = worktables;
|
|
this.url = url;
|
|
this.date_recording = new Date(0);
|
|
this.clear();
|
|
this.load();
|
|
}
|
|
|
|
public String getPK() {
|
|
if(this.isEmpty())
|
|
return "";
|
|
return this.apicall_pk;
|
|
}
|
|
|
|
public Date date_last_exec() {
|
|
if(this.isEmpty())
|
|
return new Date(0);
|
|
return this.date_last_exec;
|
|
}
|
|
|
|
public Date date_next_exec() {
|
|
if(this.isEmpty())
|
|
return new Date(0);
|
|
return this.date_next_exec;
|
|
}
|
|
|
|
public boolean hasSchedule() {
|
|
return (!this.isEmpty() && this.date_next_exec.after(new Date()));
|
|
}
|
|
|
|
public boolean isRunning(final CrawlSwitchboard crawler) {
|
|
final Iterator<byte[]> iter = crawler.getActive().iterator();
|
|
while(iter.hasNext()) {
|
|
final byte[] key = iter.next();
|
|
final CrawlProfile crawl = crawler.getActive(key);
|
|
if (crawl != null) return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
public Date date_recording() {
|
|
return this.date_recording;
|
|
}
|
|
|
|
public void set_url(final String url) {
|
|
if(!this.url.equals(url)) {
|
|
this.url = url;
|
|
this.clear();
|
|
this.load();
|
|
}
|
|
}
|
|
|
|
public int exec(final String host, final int port, final String realm) {
|
|
return this.worktables.execAPICall(this.apicall_pk, host, port, realm);
|
|
}
|
|
|
|
private void load() {
|
|
try {
|
|
final StringBuilder buffer = new StringBuilder(500);
|
|
buffer.append("^crawl start for ");
|
|
buffer.append(Pattern.quote(this.url));
|
|
buffer.append("?.*");
|
|
final Pattern pattern = Pattern.compile(buffer.toString());
|
|
//final Iterator<Tables.Row> APIcalls = this.worktables.iterator(WorkTables.TABLE_API_NAME, WorkTables.TABLE_API_COL_URL, pattern);
|
|
final Iterator<Tables.Row> APIcalls = this.worktables.iterator(WorkTables.TABLE_API_NAME, WorkTables.TABLE_API_COL_COMMENT, pattern);
|
|
Tables.Row row = null;
|
|
while(APIcalls.hasNext()) {
|
|
row = APIcalls.next();
|
|
if(row.get(WorkTables.TABLE_API_COL_TYPE, "").equals("crawler")) {
|
|
Date date = row.get(WorkTables.TABLE_API_COL_DATE_RECORDING, row.get(WorkTables.TABLE_API_COL_DATE, new Date()));
|
|
if(date.after(this.date_recording)) {
|
|
this.clear();
|
|
this.apicall_pk = UTF8.String(row.getPK());
|
|
this.date_recording = date;
|
|
this.date_next_exec = row.get(WorkTables.TABLE_API_COL_DATE_NEXT_EXEC, new Date(0));
|
|
this.date_last_exec = row.get(WorkTables.TABLE_API_COL_DATE_LAST_EXEC, new Date(0));
|
|
buffer.setLength(0);
|
|
buffer.append(row.get(WorkTables.TABLE_API_COL_URL, ""));
|
|
buffer.delete(0, buffer.indexOf("?")+1);
|
|
int start = 0;
|
|
int end = 0;
|
|
String key;
|
|
String value;
|
|
while(start < buffer.length()) {
|
|
end = buffer.indexOf("=", start);
|
|
key = buffer.substring(start, end);
|
|
start = end+1;
|
|
end = buffer.indexOf("&", start);
|
|
if(end < 0 || end > buffer.length())
|
|
end = buffer.length()-1;
|
|
value = buffer.substring(start, end);
|
|
start = end+1;
|
|
put(key, value);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
} catch (final IOException e) {
|
|
// TODO Auto-generated catch block
|
|
}
|
|
}
|
|
|
|
public static String crawlStart(
|
|
final Switchboard sb,
|
|
final DigestURI startURL,
|
|
final String urlMustMatch,
|
|
final String urlMustNotMatch,
|
|
final int depth,
|
|
final boolean crawlingQ, final boolean medialink) {
|
|
final CrawlProfile pe = new CrawlProfile(
|
|
(startURL.getHost() == null) ? startURL.toNormalform(true) : startURL.getHost(),
|
|
urlMustMatch,
|
|
urlMustNotMatch,
|
|
CrawlProfile.MATCH_ALL_STRING,
|
|
CrawlProfile.MATCH_NEVER_STRING,
|
|
"",
|
|
CrawlProfile.MATCH_NEVER_STRING,
|
|
CrawlProfile.MATCH_ALL_STRING,
|
|
CrawlProfile.MATCH_NEVER_STRING,
|
|
depth,
|
|
medialink,
|
|
CrawlProfile.getRecrawlDate(CrawlSwitchboard.CRAWL_PROFILE_PROXY_RECRAWL_CYCLE),
|
|
-1,
|
|
crawlingQ,
|
|
true, true, true, false, true, true, true,
|
|
CacheStrategy.IFFRESH,
|
|
"robot_" + CrawlSwitchboard.CRAWL_PROFILE_SNIPPET_GLOBAL_MEDIA); // TODO: make this a default profile in CrawlSwitchboard
|
|
sb.crawler.putActive(pe.handle().getBytes(), pe);
|
|
return sb.crawlStacker.stackCrawl(new Request(
|
|
sb.peers.mySeed().hash.getBytes(),
|
|
startURL,
|
|
null,
|
|
"CRAWLING-ROOT",
|
|
new Date(),
|
|
pe.handle(), 0, 0, 0, 0
|
|
));
|
|
}
|
|
}
|