/** * HostBrowser * Copyright 2012 by Michael Peter Christen, mc@yacy.net, Frankfurt am Main, Germany * First released 27.09.2012 at http://yacy.net * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public License * along with this program in the file lgpl21.txt * If not, see . */ import java.io.IOException; import java.net.MalformedURLException; import java.util.ArrayList; import java.util.Date; import java.util.HashMap; import java.util.HashSet; import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.Set; import java.util.TreeMap; import java.util.concurrent.BlockingQueue; import org.apache.solr.common.SolrDocument; import net.yacy.cora.document.ASCII; import net.yacy.cora.document.MultiProtocolURI; import net.yacy.cora.document.UTF8; import net.yacy.cora.federate.solr.YaCySchema; import net.yacy.cora.federate.solr.connector.AbstractSolrConnector; import net.yacy.cora.protocol.RequestHeader; import net.yacy.cora.sorting.ClusteredScoreMap; import net.yacy.cora.sorting.ReversibleScoreMap; import net.yacy.crawler.retrieval.Request; import net.yacy.kelondro.data.meta.DigestURI; import net.yacy.kelondro.data.meta.URIMetadataNode; import net.yacy.kelondro.logging.Log; import net.yacy.peers.graphics.WebStructureGraph.StructureEntry; import net.yacy.search.Switchboard; import net.yacy.search.index.Fulltext; import net.yacy.server.serverObjects; import net.yacy.server.serverSwitch; public class HostBrowser { public static serverObjects respond(final RequestHeader header, final serverObjects post, final serverSwitch env) { // return variable that accumulates replacements final Switchboard sb = (Switchboard) env; Fulltext fulltext = sb.index.fulltext(); final boolean admin = sb.verifyAuthentication(header); final boolean autoload = admin && sb.getConfigBool("browser.autoload", true); final boolean load4everyone = sb.getConfigBool("browser.load4everyone", false); final boolean loadRight = autoload || load4everyone; // add config later final boolean searchAllowed = sb.getConfigBool("publicSearchpage", true) || admin; final serverObjects prop = new serverObjects(); // set default values prop.put("path", ""); prop.put("result", ""); prop.putNum("ucount", fulltext.size()); prop.put("hosts", 0); prop.put("files", 0); prop.put("admin", 0); if (!searchAllowed) { prop.put("result", "You are not allowed to use this page. Please ask an administrator for permission."); return prop; } String path = post == null ? "" : post.get("path", "").trim(); if (admin && path.length() == 0 && sb.crawler.getActiveSize() > 0) sb.index.fulltext().commit(); if (post == null || env == null) { return prop; } int p = path.lastIndexOf('/'); if (p < 0 && path.length() > 0) path = path + "/"; else if (p > 7) path = path.substring(0, p + 1); // the search path shall always end with "/" if (path.length() > 0 && ( !path.startsWith("http://") && !path.startsWith("https://") && !path.startsWith("ftp://") && !path.startsWith("smb://") && !path.startsWith("file://"))) { path = "http://" + path; } prop.putHTML("path", path); prop.put("delete", admin && path.length() > 0 ? 1 : 0); DigestURI pathURI = null; try {pathURI = new DigestURI(path);} catch (MalformedURLException e) {} String load = post.get("load", ""); boolean wait = false; if (loadRight && autoload && path.length() != 0 && pathURI != null && load.length() == 0 && !sb.index.exists(pathURI.hash())) { // in case that the url does not exist and loading is wanted turn this request into a loading request load = path; wait = true; } if (load.length() > 0 && loadRight) { // stack URL DigestURI url; if (sb.crawlStacker.size() > 2) wait = false; try { url = new DigestURI(load); String reasonString = sb.crawlStacker.stackCrawl(new Request( sb.peers.mySeed().hash.getBytes(), url, null, load, new Date(), sb.crawler.defaultProxyProfile.handle(), 0, 0, 0, 0 )); prop.put("result", reasonString == null ? ("added url to indexer: " + load) : ("not indexed url '" + load + "': " + reasonString)); if (wait) for (int i = 0; i < 30; i++) { if (sb.index.exists(url.hash())) break; try {Thread.sleep(100);} catch (InterruptedException e) {} } } catch (MalformedURLException e) { prop.put("result", "bad url '" + load + "'"); } } if (post.containsKey("hosts")) { // generate host list try { int maxcount = 200; ReversibleScoreMap score = fulltext.getSolr().getFacet(YaCySchema.host_s.name(), maxcount); int c = 0; Iterator i = score.keys(false); String host; while (i.hasNext() && c < maxcount) { host = i.next(); prop.put("hosts_list_" + c + "_host", host); prop.put("hosts_list_" + c + "_count", score.get(host)); c++; } prop.put("hosts_list", c); prop.put("hosts", 1); } catch (IOException e) { Log.logException(e); } } if (path.length() > 0) { boolean delete = false; if (admin && post.containsKey("delete")) { // delete the complete path!! That includes everything that matches with this prefix. delete = true; } boolean complete = post.getBoolean("complete"); if (complete) { // we want only root paths for complete lists p = path.indexOf('/', 10); if (p > 0) path = path.substring(0, p + 1); } prop.put("files_complete", complete ? 1 : 0); prop.put("files_complete_path", path); p = path.substring(0, path.length() - 1).lastIndexOf('/'); if (p < 8) { prop.put("files_root", 1); } else { path = path.substring(0, p + 1); prop.put("files_root", 0); prop.put("files_root_path", path); } try { // generate file list from path DigestURI uri = new DigestURI(path); String host = uri.getHost(); prop.putHTML("outbound_host", host); prop.putHTML("inbound_host", host); String hosthash = ASCII.String(uri.hash(), 6, 6); // get all files for a specific host from the index BlockingQueue docs = fulltext.getSolr().concurrentQuery(YaCySchema.host_s.name() + ":" + host, 0, 100000, 60000); SolrDocument doc; Set storedDocs = new HashSet(); Set inboundLinks = new HashSet(); Map> outboundHosts = new HashMap>(); int hostsize = 0; final List deleteIDs = new ArrayList(); while ((doc = docs.take()) != AbstractSolrConnector.POISON_DOCUMENT) { String u = (String) doc.getFieldValue(YaCySchema.sku.getSolrFieldName()); hostsize++; if (u.startsWith(path)) { if (delete) { deleteIDs.add(ASCII.getBytes((String) doc.getFieldValue(YaCySchema.id.name()))); } else { storedDocs.add(u); } } else if (complete) { storedDocs.add(u); } // collect inboundlinks to browse the host Iterator links = URIMetadataNode.getLinks(doc, true); while (links.hasNext()) { u = links.next(); if ((complete || u.startsWith(path)) && !storedDocs.contains(u)) inboundLinks.add(u); } // collect outboundlinks to browse to the outbound links = URIMetadataNode.getLinks(doc, false); while (links.hasNext()) { u = links.next(); try { MultiProtocolURI mu = new MultiProtocolURI(u); if (mu.getHost() != null) { ReversibleScoreMap lks = outboundHosts.get(mu.getHost()); if (lks == null) { lks = new ClusteredScoreMap(UTF8.insensitiveUTF8Comparator); outboundHosts.put(mu.getHost(), lks); } lks.set(u, u.length()); } } catch (MalformedURLException e) {} } } if (deleteIDs.size() > 0) sb.index.fulltext().remove(deleteIDs, true); // now combine both lists into one Map files = new HashMap(); for (String u: storedDocs) files.put(u, true); for (String u: inboundLinks) if (!storedDocs.contains(u)) files.put(u, false); Log.logInfo("HostBrowser", "collected " + files.size() + " urls for path " + path); // distinguish files and folders Map list = new TreeMap(); int pl = path.length(); String file; for (Map.Entry entry: files.entrySet()) { file = entry.getKey().substring(pl); p = file.indexOf('/'); if (p < 0) { // this is a file list.put(entry.getKey(), entry.getValue()); // Boolean value: this is a file } else { // this is a directory path or a file in a subdirectory String remainingPath = file.substring(0, p + 1); if (complete && remainingPath.indexOf('.') > 0) { list.put(entry.getKey(), entry.getValue()); // Boolean value: this is a file } else { String dir = path + remainingPath; Object c = list.get(dir); if (c == null) { int[] linkedStored = new int[]{0,0}; linkedStored[entry.getValue().booleanValue() ? 1 : 0]++; list.put(dir, linkedStored); } else if (c instanceof int[]) { ((int[]) c)[entry.getValue().booleanValue() ? 1 : 0]++; } } } } int maxcount = 1000; int c = 0; for (Map.Entry entry: list.entrySet()) { if (entry.getValue() instanceof Boolean) { // this is a file prop.put("files_list_" + c + "_type", 0); prop.put("files_list_" + c + "_type_url", entry.getKey()); boolean indexed = ((Boolean) entry.getValue()).booleanValue(); try {uri = new DigestURI(entry.getKey());} catch (MalformedURLException e) {uri = null;} boolean loading = load.equals(entry.getKey()) || (uri != null && sb.crawlQueues.urlExists(uri.hash()) != null); //String failr = fulltext.failReason(ASCII.String(uri.hash())); prop.put("files_list_" + c + "_type_stored", indexed ? 1 : loading ? 2 : 0); prop.put("files_list_" + c + "_type_stored_load", loadRight ? 1 : 0); if (loadRight) { prop.put("files_list_" + c + "_type_stored_load_url", entry.getKey()); prop.put("files_list_" + c + "_type_stored_load_path", path); } } else { // this is a folder prop.put("files_list_" + c + "_type", 1); prop.put("files_list_" + c + "_type_url", entry.getKey()); prop.put("files_list_" + c + "_type_count", ((int[]) entry.getValue())[1] + " stored / " + ((int[]) entry.getValue())[0] + " linked"); } if (++c >= maxcount) break; } prop.put("files_list", c); prop.putHTML("files_path", path); prop.put("files_hostsize", hostsize); prop.put("files_subpathsize", storedDocs.size()); prop.put("files", 1); // generate inbound-links table StructureEntry struct = sb.webStructure.incomingReferences(hosthash); if (struct != null && struct.references.size() > 0) { maxcount = 200; ReversibleScoreMap score = new ClusteredScoreMap(UTF8.insensitiveUTF8Comparator); for (Map.Entry entry: struct.references.entrySet()) score.set(entry.getKey(), entry.getValue()); c = 0; Iterator i = score.keys(false); while (i.hasNext() && c < maxcount) { host = i.next(); prop.put("inbound_list_" + c + "_host", sb.webStructure.hostHash2hostName(host)); prop.put("inbound_list_" + c + "_count", score.get(host)); c++; } prop.put("inbound_list", c); prop.put("inbound", 1); } else { prop.put("inbound", 0); } // generate outbound-links table if (outboundHosts.size() > 0) { maxcount = 200; ReversibleScoreMap score = new ClusteredScoreMap(UTF8.insensitiveUTF8Comparator); for (Map.Entry> entry: outboundHosts.entrySet()) score.set(entry.getKey(), entry.getValue().size()); c = 0; Iterator i = score.keys(false); while (i.hasNext() && c < maxcount) { host = i.next(); prop.put("outbound_list_" + c + "_host", host); prop.put("outbound_list_" + c + "_count", score.get(host)); prop.put("outbound_list_" + c + "_link", outboundHosts.get(host).getMinKey()); c++; } prop.put("outbound_list", c); prop.put("outbound", 1); } else { prop.put("outbound", 0); } } catch (Throwable e) { Log.logException(e); } } // insert constants prop.putNum("ucount", fulltext.size()); // return rewrite properties return prop; } }