You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
373 lines
16 KiB
373 lines
16 KiB
// IndexControlRWIs_p.java
|
|
// -----------------------
|
|
// (C) 2004-2007 by Michael Peter Christen; mc@yacy.net, Frankfurt a. M., Germany
|
|
// first published 2004 on http://yacy.net
|
|
//
|
|
// This is a part of YaCy, a peer-to-peer based web search engine
|
|
//
|
|
// $LastChangedDate$
|
|
// $LastChangedRevision$
|
|
// $LastChangedBy$
|
|
//
|
|
// LICENSE
|
|
//
|
|
// This program is free software; you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation; either version 2 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program; if not, write to the Free Software
|
|
// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
import java.io.File;
|
|
import java.io.IOException;
|
|
import java.net.MalformedURLException;
|
|
import java.util.Iterator;
|
|
import java.util.List;
|
|
|
|
import net.yacy.cora.date.GenericFormatter;
|
|
import net.yacy.cora.document.ASCII;
|
|
import net.yacy.cora.federate.yacy.CacheStrategy;
|
|
import net.yacy.cora.lod.JenaTripleStore;
|
|
import net.yacy.cora.order.Base64Order;
|
|
import net.yacy.cora.protocol.RequestHeader;
|
|
import net.yacy.crawler.data.Cache;
|
|
import net.yacy.crawler.data.ResultURLs;
|
|
import net.yacy.data.WorkTables;
|
|
import net.yacy.kelondro.data.meta.DigestURI;
|
|
import net.yacy.kelondro.data.meta.URIMetadataNode;
|
|
import net.yacy.kelondro.data.word.Word;
|
|
import net.yacy.kelondro.logging.Log;
|
|
import net.yacy.kelondro.util.RotateIterator;
|
|
import net.yacy.search.Switchboard;
|
|
import net.yacy.search.index.Fulltext;
|
|
import net.yacy.search.index.Segment;
|
|
import net.yacy.server.serverObjects;
|
|
import net.yacy.server.serverSwitch;
|
|
|
|
public class IndexControlURLs_p {
|
|
|
|
public static serverObjects respond(@SuppressWarnings("unused") final RequestHeader header, final serverObjects post, final serverSwitch env) {
|
|
// return variable that accumulates replacements
|
|
final Switchboard sb = (Switchboard) env;
|
|
|
|
final serverObjects prop = new serverObjects();
|
|
|
|
Segment segment = sb.index;
|
|
|
|
// set default values
|
|
prop.put("urlstring", "");
|
|
prop.put("urlhash", "");
|
|
prop.put("result", "");
|
|
prop.putNum("ucount", segment.fulltext().size());
|
|
prop.put("otherHosts", "");
|
|
prop.put("genUrlProfile", 0);
|
|
prop.put("statistics", 1);
|
|
prop.put("statistics_lines", 100);
|
|
prop.put("statisticslines", 0);
|
|
prop.put("reload", 0);
|
|
prop.put("indexdump", 0);
|
|
prop.put("lurlexport", 0);
|
|
prop.put("reload", 0);
|
|
prop.put("dumprestore", 1);
|
|
List<File> dumpFiles = segment.fulltext().dumpFiles();
|
|
prop.put("dumprestore_dumpfile", dumpFiles.size() == 0 ? "" : dumpFiles.get(dumpFiles.size() - 1).getAbsolutePath());
|
|
prop.put("cleanup", post == null ? 1 : 0);
|
|
prop.put("cleanup_solr", sb.index.fulltext().connectedRemoteSolr() ? 1 : 0);
|
|
|
|
// show export messages
|
|
final Fulltext.Export export = segment.fulltext().export();
|
|
if ((export != null) && (export.isAlive())) {
|
|
// there is currently a running export
|
|
prop.put("lurlexport", 2);
|
|
prop.put("lurlexportfinished", 0);
|
|
prop.put("lurlexporterror", 0);
|
|
prop.put("lurlexport_exportfile", export.file().toString());
|
|
prop.put("lurlexport_urlcount", export.count());
|
|
prop.put("reload", 1);
|
|
} else {
|
|
prop.put("lurlexport", 1);
|
|
prop.put("lurlexport_exportfile", sb.getDataPath() + "/DATA/EXPORT/" + GenericFormatter.SHORT_SECOND_FORMATTER.format());
|
|
if (export == null) {
|
|
// there has never been an export
|
|
prop.put("lurlexportfinished", 0);
|
|
prop.put("lurlexporterror", 0);
|
|
} else {
|
|
// an export was running but has finished
|
|
prop.put("lurlexportfinished", 1);
|
|
prop.put("lurlexportfinished_exportfile", export.file().toString());
|
|
prop.put("lurlexportfinished_urlcount", export.count());
|
|
if (export.failed() == null) {
|
|
prop.put("lurlexporterror", 0);
|
|
} else {
|
|
prop.put("lurlexporterror", 1);
|
|
prop.put("lurlexporterror_exportfile", export.file().toString());
|
|
prop.put("lurlexporterror_exportfailmsg", export.failed());
|
|
}
|
|
}
|
|
}
|
|
|
|
if (post == null || env == null) {
|
|
return prop; // nothing to do
|
|
}
|
|
|
|
// post values that are set on numerous input fields with same name
|
|
String urlstring = post.get("urlstring", "").trim();
|
|
String urlhash = post.get("urlhash", "").trim();
|
|
if (urlhash.isEmpty() && urlstring.length() > 0) {
|
|
try {
|
|
urlhash = ASCII.String(new DigestURI(urlstring).hash());
|
|
} catch (final MalformedURLException e) {
|
|
}
|
|
}
|
|
|
|
if (!urlstring.startsWith("http://") &&
|
|
!urlstring.startsWith("https://") &&
|
|
!urlstring.startsWith("ftp://") &&
|
|
!urlstring.startsWith("smb://") &&
|
|
!urlstring.startsWith("file://")) { urlstring = "http://" + urlstring; }
|
|
|
|
prop.putHTML("urlstring", urlstring);
|
|
prop.putHTML("urlhash", urlhash);
|
|
prop.put("result", " ");
|
|
|
|
// delete everything
|
|
if ( post.containsKey("deletecomplete") ) {
|
|
if ( post.get("deleteIndex", "").equals("on") ) {
|
|
segment.clear();
|
|
}
|
|
if ( post.get("deleteRemoteSolr", "").equals("on")) {
|
|
try {
|
|
sb.index.fulltext().getSolr().clear();
|
|
} catch ( final Exception e ) {
|
|
Log.logException(e);
|
|
}
|
|
}
|
|
if ( post.get("deleteCrawlQueues", "").equals("on") ) {
|
|
sb.crawlQueues.clear();
|
|
sb.crawlStacker.clear();
|
|
ResultURLs.clearStacks();
|
|
}
|
|
if ( post.get("deleteTriplestore", "").equals("on") ) {
|
|
JenaTripleStore.clear();
|
|
}
|
|
if ( post.get("deleteCache", "").equals("on") ) {
|
|
Cache.clear();
|
|
}
|
|
if ( post.get("deleteRobots", "").equals("on") ) {
|
|
try {sb.robots.clear();} catch (IOException e) {}
|
|
}
|
|
if ( post.get("deleteSearchFl", "").equals("on") ) {
|
|
sb.tables.clear(WorkTables.TABLE_SEARCH_FAILURE_NAME);
|
|
}
|
|
post.remove("deletecomplete");
|
|
}
|
|
|
|
if (post.containsKey("urlhashdeleteall")) {
|
|
int i = segment.removeAllUrlReferences(urlhash.getBytes(), sb.loader, CacheStrategy.IFEXIST);
|
|
prop.put("result", "Deleted URL and " + i + " references from " + i + " word indexes.");
|
|
}
|
|
|
|
if (post.containsKey("urlhashdelete")) {
|
|
final URIMetadataNode entry = segment.fulltext().getMetadata(ASCII.getBytes(urlhash));
|
|
if (entry == null) {
|
|
prop.putHTML("result", "No Entry for URL hash " + urlhash + "; nothing deleted.");
|
|
} else {
|
|
urlstring = entry.url().toNormalform(true);
|
|
prop.put("urlstring", "");
|
|
sb.urlRemove(segment, urlhash.getBytes());
|
|
prop.putHTML("result", "Removed URL " + urlstring);
|
|
}
|
|
}
|
|
|
|
if (post.containsKey("urldelete")) {
|
|
try {
|
|
urlhash = ASCII.String((new DigestURI(urlstring)).hash());
|
|
} catch (final MalformedURLException e) {
|
|
urlhash = null;
|
|
}
|
|
if ((urlhash == null) || (urlstring == null)) {
|
|
prop.put("result", "No input given; nothing deleted.");
|
|
} else {
|
|
sb.urlRemove(segment, urlhash.getBytes());
|
|
prop.putHTML("result", "Removed URL " + urlstring);
|
|
}
|
|
}
|
|
|
|
if (post.containsKey("urlstringsearch")) {
|
|
try {
|
|
final DigestURI url = new DigestURI(urlstring);
|
|
urlhash = ASCII.String(url.hash());
|
|
prop.put("urlhash", urlhash);
|
|
final URIMetadataNode entry = segment.fulltext().getMetadata(ASCII.getBytes(urlhash));
|
|
if (entry == null) {
|
|
prop.putHTML("result", "No Entry for URL " + url.toNormalform(true));
|
|
prop.putHTML("urlstring", urlstring);
|
|
prop.put("urlhash", "");
|
|
} else {
|
|
prop.putAll(genUrlProfile(segment, entry, urlhash));
|
|
prop.put("statistics", 0);
|
|
}
|
|
} catch (final MalformedURLException e) {
|
|
prop.putHTML("result", "bad url: " + urlstring);
|
|
prop.put("urlhash", "");
|
|
}
|
|
}
|
|
|
|
if (post.containsKey("urlhashsearch")) {
|
|
final URIMetadataNode entry = segment.fulltext().getMetadata(ASCII.getBytes(urlhash));
|
|
if (entry == null) {
|
|
prop.putHTML("result", "No Entry for URL hash " + urlhash);
|
|
} else {
|
|
prop.putHTML("urlstring", entry.url().toNormalform(true));
|
|
prop.putAll(genUrlProfile(segment, entry, urlhash));
|
|
prop.put("statistics", 0);
|
|
}
|
|
}
|
|
|
|
// generate list
|
|
if (post.containsKey("urlhashsimilar")) {
|
|
final Iterator<URIMetadataNode> entryIt = new RotateIterator<URIMetadataNode>(segment.fulltext().entries(), ASCII.String(Base64Order.zero((urlhash == null ? 0 : urlhash.length()))), (int) segment.RWICount());
|
|
final StringBuilder result = new StringBuilder("Sequential List of URL-Hashes:<br />");
|
|
URIMetadataNode entry;
|
|
int i = 0, rows = 0, cols = 0;
|
|
prop.put("urlhashsimilar", "1");
|
|
while (entryIt.hasNext() && i < 256) {
|
|
entry = entryIt.next();
|
|
if (entry == null) break;
|
|
prop.put("urlhashsimilar_rows_"+rows+"_cols_"+cols+"_urlHash", ASCII.String(entry.hash()));
|
|
cols++;
|
|
if (cols==8) {
|
|
prop.put("urlhashsimilar_rows_"+rows+"_cols", cols);
|
|
cols = 0;
|
|
rows++;
|
|
}
|
|
i++;
|
|
}
|
|
prop.put("statistics", 0);
|
|
prop.put("urlhashsimilar_rows", rows);
|
|
prop.put("result", result.toString());
|
|
}
|
|
|
|
if (post.containsKey("lurlexport")) {
|
|
// parse format
|
|
int format = 0;
|
|
final String fname = post.get("format", "url-text");
|
|
final boolean dom = fname.startsWith("dom"); // if dom== false complete urls are exported, otherwise only the domain
|
|
if (fname.endsWith("text")) format = 0;
|
|
if (fname.endsWith("html")) format = 1;
|
|
if (fname.endsWith("rss")) format = 2;
|
|
|
|
// extend export file name
|
|
String s = post.get("exportfile", "");
|
|
if (s.indexOf('.',0) < 0) {
|
|
if (format == 0) s = s + ".txt";
|
|
if (format == 1) s = s + ".html";
|
|
if (format == 2) s = s + ".xml";
|
|
}
|
|
final File f = new File(s);
|
|
f.getParentFile().mkdirs();
|
|
final String filter = post.get("exportfilter", ".*");
|
|
final Fulltext.Export running = segment.fulltext().export(f, filter, null, format, dom);
|
|
|
|
prop.put("lurlexport_exportfile", s);
|
|
prop.put("lurlexport_urlcount", running.count());
|
|
if ((running != null) && (running.failed() == null)) {
|
|
prop.put("lurlexport", 2);
|
|
}
|
|
prop.put("reload", 1);
|
|
}
|
|
|
|
if (post.containsKey("indexdump")) {
|
|
final File dump = segment.fulltext().dumpSolr();
|
|
prop.put("indexdump", 1);
|
|
prop.put("indexdump_dumpfile", dump.getAbsolutePath());
|
|
}
|
|
|
|
if (post.containsKey("indexrestore")) {
|
|
final File dump = new File(post.get("dumpfile", ""));
|
|
segment.fulltext().restoreSolr(dump);
|
|
}
|
|
|
|
if (post.containsKey("deletedomain")) {
|
|
final String hp = post.get("hashpart");
|
|
try {
|
|
segment.fulltext().deleteDomain(hp);
|
|
} catch (final IOException e) {
|
|
// TODO Auto-generated catch block
|
|
Log.logException(e);
|
|
}
|
|
// trigger the loading of the table
|
|
post.put("statistics", "");
|
|
}
|
|
|
|
if (post.containsKey("statistics")) {
|
|
final int count = post.getInt("lines", 100);
|
|
Iterator<Fulltext.HostStat> statsiter;
|
|
prop.put("statistics_lines", count);
|
|
int cnt = 0;
|
|
try {
|
|
final Fulltext metadata = segment.fulltext();
|
|
statsiter = metadata.statistics(count, metadata.urlSampleScores(metadata.domainSampleCollector()));
|
|
boolean dark = true;
|
|
Fulltext.HostStat hs;
|
|
while (statsiter.hasNext() && cnt < count) {
|
|
hs = statsiter.next();
|
|
prop.put("statisticslines_domains_" + cnt + "_dark", (dark) ? "1" : "0");
|
|
prop.put("statisticslines_domains_" + cnt + "_domain", hs.hostname + ((hs.port == 80) ? "" : ":" + hs.port));
|
|
prop.put("statisticslines_domains_" + cnt + "lines", count);
|
|
prop.put("statisticslines_domains_" + cnt + "_hashpart", hs.hosthash);
|
|
prop.put("statisticslines_domains_" + cnt + "_count", hs.count);
|
|
dark = !dark;
|
|
cnt++;
|
|
}
|
|
} catch (final IOException e) {
|
|
Log.logException(e);
|
|
}
|
|
prop.put("statisticslines_domains", cnt);
|
|
prop.put("statisticslines", 1);
|
|
}
|
|
|
|
// insert constants
|
|
prop.putNum("ucount", segment.fulltext().size());
|
|
// return rewrite properties
|
|
return prop;
|
|
}
|
|
|
|
private static serverObjects genUrlProfile(final Segment segment, final URIMetadataNode entry, final String urlhash) {
|
|
final serverObjects prop = new serverObjects();
|
|
if (entry == null) {
|
|
prop.put("genUrlProfile", "1");
|
|
prop.put("genUrlProfile_urlhash", urlhash);
|
|
return prop;
|
|
}
|
|
final URIMetadataNode le = (entry.referrerHash() == null || entry.referrerHash().length != Word.commonHashLength) ? null : segment.fulltext().getMetadata(entry.referrerHash());
|
|
if (entry.url() == null) {
|
|
prop.put("genUrlProfile", "1");
|
|
prop.put("genUrlProfile_urlhash", urlhash);
|
|
return prop;
|
|
}
|
|
prop.put("genUrlProfile", "2");
|
|
prop.putHTML("genUrlProfile_urlNormalform", entry.url().toNormalform(true));
|
|
prop.put("genUrlProfile_urlhash", urlhash);
|
|
prop.put("genUrlProfile_urlDescr", entry.dc_title());
|
|
prop.put("genUrlProfile_moddate", entry.moddate().toString());
|
|
prop.put("genUrlProfile_loaddate", entry.loaddate().toString());
|
|
prop.put("genUrlProfile_referrer", (le == null) ? 0 : 1);
|
|
prop.putHTML("genUrlProfile_referrer_url", (le == null) ? "<unknown>" : le.url().toNormalform(true));
|
|
prop.put("genUrlProfile_referrer_hash", (le == null) ? "" : ASCII.String(le.hash()));
|
|
prop.put("genUrlProfile_doctype", String.valueOf(entry.doctype()));
|
|
prop.put("genUrlProfile_language", entry.language());
|
|
prop.put("genUrlProfile_size", entry.size());
|
|
prop.put("genUrlProfile_wordCount", entry.wordCount());
|
|
return prop;
|
|
}
|
|
|
|
}
|