You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
347 lines
14 KiB
347 lines
14 KiB
/**
|
|
* htmlParser.java
|
|
* Copyright 2009 by Michael Peter Christen, mc@yacy.net, Frankfurt am Main, Germany
|
|
* First released 09.07.2009 at http://yacy.net
|
|
*
|
|
* $LastChangedDate$
|
|
* $LastChangedRevision$
|
|
* $LastChangedBy$
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program in the file lgpl21.txt
|
|
* If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
package net.yacy.document.parser;
|
|
|
|
import java.io.BufferedInputStream;
|
|
import java.io.ByteArrayInputStream;
|
|
import java.io.IOException;
|
|
import java.io.InputStream;
|
|
import java.io.UnsupportedEncodingException;
|
|
import java.net.MalformedURLException;
|
|
import java.nio.charset.Charset;
|
|
import java.nio.charset.IllegalCharsetNameException;
|
|
import java.nio.charset.UnsupportedCharsetException;
|
|
import java.util.HashSet;
|
|
import java.util.LinkedHashMap;
|
|
import java.util.Set;
|
|
|
|
import net.yacy.cora.document.encoding.UTF8;
|
|
import net.yacy.cora.document.id.AnchorURL;
|
|
import net.yacy.cora.document.id.DigestURL;
|
|
import net.yacy.cora.protocol.ClientIdentification;
|
|
import net.yacy.cora.util.CommonPattern;
|
|
import net.yacy.document.AbstractParser;
|
|
import net.yacy.document.Document;
|
|
import net.yacy.document.Parser;
|
|
import net.yacy.document.VocabularyScraper;
|
|
import net.yacy.document.parser.html.ContentScraper;
|
|
import net.yacy.document.parser.html.ImageEntry;
|
|
import net.yacy.document.parser.html.ScraperInputStream;
|
|
import net.yacy.document.parser.html.TransformerWriter;
|
|
import net.yacy.kelondro.util.FileUtils;
|
|
|
|
import com.ibm.icu.text.CharsetDetector;
|
|
|
|
|
|
public class htmlParser extends AbstractParser implements Parser {
|
|
|
|
private static final int maxLinks = 10000;
|
|
|
|
public final static String[] htmlExtensions = new String[]{
|
|
"htm","html","phtml","shtml","shtm","stm","xhtml","phtml","phtm",
|
|
"tpl","php","php2","php3","php4","php5","cfm","asp","aspx","tex","txt","msg"
|
|
};
|
|
|
|
public final static Set<String> htmlExtensionsSet;
|
|
|
|
static {
|
|
htmlExtensionsSet = new HashSet<>(htmlExtensions.length);
|
|
for (String ext: htmlExtensions) htmlExtensionsSet.add(ext);
|
|
}
|
|
|
|
public htmlParser() {
|
|
super("Streaming HTML Parser");
|
|
this.SUPPORTED_EXTENSIONS.addAll(htmlExtensionsSet);
|
|
this.SUPPORTED_MIME_TYPES.add("text/html");
|
|
this.SUPPORTED_MIME_TYPES.add("text/xhtml+xml");
|
|
this.SUPPORTED_MIME_TYPES.add("application/xhtml+xml");
|
|
this.SUPPORTED_MIME_TYPES.add("application/x-httpd-php");
|
|
this.SUPPORTED_MIME_TYPES.add("application/x-tex");
|
|
this.SUPPORTED_MIME_TYPES.add("application/vnd.ms-outlook");
|
|
this.SUPPORTED_MIME_TYPES.add("text/plain");
|
|
this.SUPPORTED_MIME_TYPES.add("text/csv");
|
|
}
|
|
|
|
@Override
|
|
public Document[] parse(
|
|
final AnchorURL location,
|
|
final String mimeType,
|
|
final String documentCharset,
|
|
final VocabularyScraper vocscraper,
|
|
final int timezoneOffset,
|
|
final InputStream sourceStream) throws Parser.Failure, InterruptedException {
|
|
|
|
try {
|
|
// first get a document from the parsed html
|
|
Charset[] detectedcharsetcontainer = new Charset[]{null};
|
|
final ContentScraper scraper = parseToScraper(location, documentCharset, vocscraper, detectedcharsetcontainer, timezoneOffset, sourceStream, maxLinks);
|
|
// parseToScraper also detects/corrects/sets charset from html content tag
|
|
final Document document = transformScraper(location, mimeType, detectedcharsetcontainer[0].name(), scraper);
|
|
|
|
return new Document[]{document};
|
|
} catch (final IOException e) {
|
|
throw new Parser.Failure("IOException in htmlParser: " + e.getMessage(), location);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* the transformScraper method transforms a scraper object into a document object
|
|
* @param location
|
|
* @param mimeType
|
|
* @param charSet
|
|
* @param scraper
|
|
* @return
|
|
*/
|
|
private static Document transformScraper(final DigestURL location, final String mimeType, final String charSet, final ContentScraper scraper) {
|
|
final String[] sections = new String[
|
|
scraper.getHeadlines(1).length +
|
|
scraper.getHeadlines(2).length +
|
|
scraper.getHeadlines(3).length +
|
|
scraper.getHeadlines(4).length +
|
|
scraper.getHeadlines(5).length +
|
|
scraper.getHeadlines(6).length];
|
|
int p = 0;
|
|
for (int i = 1; i <= 6; i++) {
|
|
for (final String headline : scraper.getHeadlines(i)) {
|
|
sections[p++] = headline;
|
|
}
|
|
}
|
|
LinkedHashMap<DigestURL, ImageEntry> noDoubleImages = new LinkedHashMap<>();
|
|
for (ImageEntry ie: scraper.getImages()) noDoubleImages.put(ie.url(), ie);
|
|
final Document ppd = new Document(
|
|
location,
|
|
mimeType,
|
|
charSet,
|
|
scraper,
|
|
scraper.getContentLanguages(),
|
|
scraper.getKeywords(),
|
|
scraper.getTitles(),
|
|
scraper.getAuthor(),
|
|
scraper.getPublisher(),
|
|
sections,
|
|
scraper.getDescriptions(),
|
|
scraper.getLon(), scraper.getLat(),
|
|
scraper.getText(),
|
|
scraper.getAnchors(),
|
|
scraper.getRSS(),
|
|
noDoubleImages,
|
|
scraper.indexingDenied(),
|
|
scraper.getDate());
|
|
ppd.setFavicon(scraper.getFavicon());
|
|
|
|
return ppd;
|
|
}
|
|
|
|
public static ContentScraper parseToScraper(final DigestURL location, final String documentCharset, final VocabularyScraper vocabularyScraper, final int timezoneOffset, final String input, final int maxLinks) throws IOException {
|
|
Charset[] detectedcharsetcontainer = new Charset[]{null};
|
|
InputStream sourceStream;
|
|
try {
|
|
sourceStream = new ByteArrayInputStream(documentCharset == null ? UTF8.getBytes(input) : input.getBytes(documentCharset));
|
|
} catch (UnsupportedEncodingException e) {
|
|
sourceStream = new ByteArrayInputStream(UTF8.getBytes(input));
|
|
}
|
|
ContentScraper scraper;
|
|
try {
|
|
scraper = parseToScraper(location, documentCharset, vocabularyScraper, detectedcharsetcontainer, timezoneOffset, sourceStream, maxLinks);
|
|
} catch (Failure e) {
|
|
throw new IOException(e.getMessage());
|
|
}
|
|
return scraper;
|
|
}
|
|
|
|
public static ContentScraper parseToScraper(
|
|
final DigestURL location,
|
|
final String documentCharset,
|
|
final VocabularyScraper vocabularyScraper,
|
|
Charset[] detectedcharsetcontainer,
|
|
final int timezoneOffset,
|
|
InputStream sourceStream,
|
|
final int maxLinks) throws Parser.Failure, IOException {
|
|
|
|
// make a scraper
|
|
String charset = null;
|
|
|
|
// ah, we are lucky, we got a character-encoding via HTTP-header
|
|
if (documentCharset != null) {
|
|
charset = patchCharsetEncoding(documentCharset);
|
|
}
|
|
|
|
// nothing found: try to find a meta-tag
|
|
if (charset == null) {
|
|
ScraperInputStream htmlFilter = null;
|
|
try {
|
|
htmlFilter = new ScraperInputStream(sourceStream, documentCharset, vocabularyScraper, location, null, false, maxLinks, timezoneOffset);
|
|
sourceStream = htmlFilter;
|
|
charset = htmlFilter.detectCharset();
|
|
} catch (final IOException e1) {
|
|
throw new Parser.Failure("Charset error:" + e1.getMessage(), location);
|
|
} finally {
|
|
if (htmlFilter != null) htmlFilter.close();
|
|
}
|
|
}
|
|
|
|
// the author didn't tell us the encoding, try the mozilla-heuristic
|
|
if (charset == null) {
|
|
final CharsetDetector det = new CharsetDetector();
|
|
det.enableInputFilter(true);
|
|
final InputStream detStream = new BufferedInputStream(sourceStream);
|
|
det.setText(detStream);
|
|
charset = det.detect().getName();
|
|
sourceStream = detStream;
|
|
}
|
|
|
|
// wtf? still nothing, just take system-standard
|
|
if (charset == null) {
|
|
detectedcharsetcontainer[0] = Charset.defaultCharset();
|
|
} else {
|
|
try {
|
|
detectedcharsetcontainer[0] = Charset.forName(charset);
|
|
} catch (final IllegalCharsetNameException e) {
|
|
detectedcharsetcontainer[0] = Charset.defaultCharset();
|
|
} catch (final UnsupportedCharsetException e) {
|
|
detectedcharsetcontainer[0] = Charset.defaultCharset();
|
|
}
|
|
}
|
|
|
|
// parsing the content
|
|
final ContentScraper scraper = new ContentScraper(location, maxLinks, vocabularyScraper, timezoneOffset);
|
|
final TransformerWriter writer = new TransformerWriter(null,null,scraper,null,false, Math.max(64, Math.min(4096, sourceStream.available())));
|
|
try {
|
|
FileUtils.copy(sourceStream, writer, detectedcharsetcontainer[0]);
|
|
} catch (final IOException e) {
|
|
throw new Parser.Failure("IO error:" + e.getMessage(), location);
|
|
} finally {
|
|
writer.flush();
|
|
//sourceStream.close(); keep open for multipe parsing (close done by caller)
|
|
writer.close();
|
|
}
|
|
//OutputStream hfos = new htmlFilterOutputStream(null, scraper, null, false);
|
|
//serverFileUtils.copy(sourceFile, hfos);
|
|
//hfos.close();
|
|
if (writer.binarySuspect()) {
|
|
final String errorMsg = "Binary data found in resource";
|
|
throw new Parser.Failure(errorMsg, location);
|
|
}
|
|
return scraper;
|
|
}
|
|
|
|
/**
|
|
* some html authors use wrong encoding names, either because they don't know exactly what they
|
|
* are doing or they produce a type. Many times, the upper/downcase scheme of the name is fuzzy
|
|
* This method patches wrong encoding names. The correct names are taken from
|
|
* http://www.iana.org/assignments/character-sets
|
|
* @param encoding
|
|
* @return patched encoding name
|
|
*/
|
|
public static String patchCharsetEncoding(String encoding) {
|
|
|
|
// do nothing with null
|
|
if ((encoding == null) || (encoding.length() < 3)) return null;
|
|
|
|
// trim encoding string
|
|
encoding = encoding.trim();
|
|
|
|
// fix upper/lowercase
|
|
encoding = encoding.toUpperCase();
|
|
if (encoding.startsWith("SHIFT")) return "Shift_JIS";
|
|
if (encoding.startsWith("BIG")) return "Big5";
|
|
// all other names but such with "windows" use uppercase
|
|
if (encoding.startsWith("WINDOWS")) encoding = "windows" + encoding.substring(7);
|
|
if (encoding.startsWith("MACINTOSH")) encoding = "MacRoman";
|
|
|
|
// fix wrong fill characters
|
|
encoding = CommonPattern.UNDERSCORE.matcher(encoding).replaceAll("-");
|
|
|
|
if (encoding.matches("GB[_-]?2312([-_]80)?")) return "GB2312";
|
|
if (encoding.matches(".*UTF[-_]?8.*")) return "UTF-8";
|
|
if (encoding.startsWith("US")) return "US-ASCII";
|
|
if (encoding.startsWith("KOI")) return "KOI8-R";
|
|
|
|
// patch missing '-'
|
|
if (encoding.startsWith("windows") && encoding.length() > 7) {
|
|
final char c = encoding.charAt(7);
|
|
if ((c >= '0') && (c <= '9')) {
|
|
encoding = "windows-" + encoding.substring(7);
|
|
}
|
|
}
|
|
|
|
if (encoding.startsWith("ISO")) {
|
|
// patch typos
|
|
if (encoding.length() > 3) {
|
|
final char c = encoding.charAt(3);
|
|
if ((c >= '0') && (c <= '9')) {
|
|
encoding = "ISO-" + encoding.substring(3);
|
|
}
|
|
}
|
|
if (encoding.length() > 8) {
|
|
final char c = encoding.charAt(8);
|
|
if ((c >= '0') && (c <= '9')) {
|
|
encoding = encoding.substring(0, 8) + "-" + encoding.substring(8);
|
|
}
|
|
}
|
|
}
|
|
|
|
// patch wrong name
|
|
if (encoding.startsWith("ISO-8559")) {
|
|
// popular typo
|
|
encoding = "ISO-8859" + encoding.substring(8);
|
|
}
|
|
|
|
// converting cp\d{4} -> windows-\d{4}
|
|
if (encoding.matches("CP([_-])?125[0-8]")) {
|
|
final char c = encoding.charAt(2);
|
|
if ((c >= '0') && (c <= '9')) {
|
|
encoding = "windows-" + encoding.substring(2);
|
|
} else {
|
|
encoding = "windows" + encoding.substring(2);
|
|
}
|
|
}
|
|
|
|
return encoding;
|
|
}
|
|
|
|
public static void main(final String[] args) {
|
|
// test parsing of a url
|
|
AnchorURL url;
|
|
try {
|
|
url = new AnchorURL(args[0]);
|
|
final byte[] content = url.get(ClientIdentification.yacyInternetCrawlerAgent, null, null);
|
|
final Document[] document = new htmlParser().parse(url, "text/html", "utf-8", new VocabularyScraper(), 0, new ByteArrayInputStream(content));
|
|
final String title = document[0].dc_title();
|
|
System.out.println(title);
|
|
} catch (final MalformedURLException e) {
|
|
e.printStackTrace();
|
|
} catch (final IOException e) {
|
|
e.printStackTrace();
|
|
} catch (final Parser.Failure e) {
|
|
e.printStackTrace();
|
|
} catch (final InterruptedException e) {
|
|
e.printStackTrace();
|
|
}
|
|
System.exit(0);
|
|
}
|
|
|
|
}
|