mirror of
https://github.com/yacy/yacy_search_server.git
synced 2024-09-21 00:00:13 +02:00
67beef657f
stack on html tag objects, not using a recursive parse-again method which may cause bad performance and huge memory allocation. The new method also produced better parsed image objects with exact anchor text references.
347 lines
14 KiB
Java
347 lines
14 KiB
Java
/**
|
|
* htmlParser.java
|
|
* Copyright 2009 by Michael Peter Christen, mc@yacy.net, Frankfurt am Main, Germany
|
|
* First released 09.07.2009 at http://yacy.net
|
|
*
|
|
* $LastChangedDate$
|
|
* $LastChangedRevision$
|
|
* $LastChangedBy$
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program in the file lgpl21.txt
|
|
* If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
package net.yacy.document.parser;
|
|
|
|
import java.io.BufferedInputStream;
|
|
import java.io.ByteArrayInputStream;
|
|
import java.io.IOException;
|
|
import java.io.InputStream;
|
|
import java.io.UnsupportedEncodingException;
|
|
import java.net.MalformedURLException;
|
|
import java.nio.charset.Charset;
|
|
import java.nio.charset.IllegalCharsetNameException;
|
|
import java.nio.charset.UnsupportedCharsetException;
|
|
import java.util.LinkedHashMap;
|
|
|
|
import net.yacy.cora.document.encoding.UTF8;
|
|
import net.yacy.cora.document.id.AnchorURL;
|
|
import net.yacy.cora.document.id.DigestURL;
|
|
import net.yacy.cora.protocol.ClientIdentification;
|
|
import net.yacy.cora.util.CommonPattern;
|
|
import net.yacy.document.AbstractParser;
|
|
import net.yacy.document.Document;
|
|
import net.yacy.document.Parser;
|
|
import net.yacy.document.parser.html.CharacterCoding;
|
|
import net.yacy.document.parser.html.ContentScraper;
|
|
import net.yacy.document.parser.html.ImageEntry;
|
|
import net.yacy.document.parser.html.ScraperInputStream;
|
|
import net.yacy.document.parser.html.TransformerWriter;
|
|
import net.yacy.kelondro.util.FileUtils;
|
|
|
|
import com.ibm.icu.text.CharsetDetector;
|
|
|
|
|
|
public class htmlParser extends AbstractParser implements Parser {
|
|
|
|
private static final int maxLinks = 10000;
|
|
|
|
public htmlParser() {
|
|
super("Streaming HTML Parser");
|
|
this.SUPPORTED_EXTENSIONS.add("htm");
|
|
this.SUPPORTED_EXTENSIONS.add("html");
|
|
this.SUPPORTED_EXTENSIONS.add("phtml");
|
|
this.SUPPORTED_EXTENSIONS.add("shtml");
|
|
this.SUPPORTED_EXTENSIONS.add("xhtml");
|
|
this.SUPPORTED_EXTENSIONS.add("php");
|
|
this.SUPPORTED_EXTENSIONS.add("php3");
|
|
this.SUPPORTED_EXTENSIONS.add("php4");
|
|
this.SUPPORTED_EXTENSIONS.add("php5");
|
|
this.SUPPORTED_EXTENSIONS.add("cfm");
|
|
this.SUPPORTED_EXTENSIONS.add("asp");
|
|
this.SUPPORTED_EXTENSIONS.add("aspx");
|
|
this.SUPPORTED_EXTENSIONS.add("tex");
|
|
this.SUPPORTED_EXTENSIONS.add("txt");
|
|
//SUPPORTED_EXTENSIONS.add("js");
|
|
this.SUPPORTED_EXTENSIONS.add("jsp");
|
|
this.SUPPORTED_EXTENSIONS.add("mf");
|
|
this.SUPPORTED_EXTENSIONS.add("pl");
|
|
this.SUPPORTED_EXTENSIONS.add("py");
|
|
this.SUPPORTED_MIME_TYPES.add("text/html");
|
|
this.SUPPORTED_MIME_TYPES.add("text/xhtml+xml");
|
|
this.SUPPORTED_MIME_TYPES.add("application/xhtml+xml");
|
|
this.SUPPORTED_MIME_TYPES.add("application/x-httpd-php");
|
|
this.SUPPORTED_MIME_TYPES.add("application/x-tex");
|
|
this.SUPPORTED_MIME_TYPES.add("text/plain");
|
|
this.SUPPORTED_MIME_TYPES.add("text/sgml");
|
|
this.SUPPORTED_MIME_TYPES.add("text/csv");
|
|
}
|
|
|
|
@Override
|
|
public Document[] parse(
|
|
final AnchorURL location,
|
|
final String mimeType,
|
|
final String documentCharset,
|
|
final InputStream sourceStream) throws Parser.Failure, InterruptedException {
|
|
|
|
try {
|
|
// first get a document from the parsed html
|
|
Charset[] detectedcharsetcontainer = new Charset[]{null};
|
|
final ContentScraper scraper = parseToScraper(location, documentCharset, detectedcharsetcontainer, sourceStream, maxLinks);
|
|
// parseToScraper also detects/corrects/sets charset from html content tag
|
|
final Document document = transformScraper(location, mimeType, detectedcharsetcontainer[0].name(), scraper);
|
|
|
|
return new Document[]{document};
|
|
} catch (final IOException e) {
|
|
throw new Parser.Failure("IOException in htmlParser: " + e.getMessage(), location);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* the transformScraper method transforms a scraper object into a document object
|
|
* @param location
|
|
* @param mimeType
|
|
* @param charSet
|
|
* @param scraper
|
|
* @return
|
|
*/
|
|
private static Document transformScraper(final DigestURL location, final String mimeType, final String charSet, final ContentScraper scraper) {
|
|
final String[] sections = new String[
|
|
scraper.getHeadlines(1).length +
|
|
scraper.getHeadlines(2).length +
|
|
scraper.getHeadlines(3).length +
|
|
scraper.getHeadlines(4).length +
|
|
scraper.getHeadlines(5).length +
|
|
scraper.getHeadlines(6).length];
|
|
int p = 0;
|
|
for (int i = 1; i <= 6; i++) {
|
|
for (final String headline : scraper.getHeadlines(i)) {
|
|
sections[p++] = headline;
|
|
}
|
|
}
|
|
LinkedHashMap<AnchorURL, ImageEntry> noDoubleImages = new LinkedHashMap<AnchorURL, ImageEntry>();
|
|
for (ImageEntry ie: scraper.getImages()) noDoubleImages.put(ie.url(), ie);
|
|
final Document ppd = new Document(
|
|
location,
|
|
mimeType,
|
|
charSet,
|
|
scraper,
|
|
scraper.getContentLanguages(),
|
|
scraper.getKeywords(),
|
|
scraper.getTitles(),
|
|
scraper.getAuthor(),
|
|
scraper.getPublisher(),
|
|
sections,
|
|
scraper.getDescriptions(),
|
|
scraper.getLon(), scraper.getLat(),
|
|
scraper.getText(),
|
|
scraper.getAnchors(),
|
|
scraper.getRSS(),
|
|
noDoubleImages,
|
|
scraper.indexingDenied(),
|
|
scraper.getDate());
|
|
ppd.setFavicon(scraper.getFavicon());
|
|
|
|
return ppd;
|
|
}
|
|
|
|
public static ContentScraper parseToScraper(final DigestURL location, final String documentCharset, String input, int maxLinks) throws IOException {
|
|
Charset[] detectedcharsetcontainer = new Charset[]{null};
|
|
InputStream sourceStream;
|
|
try {
|
|
sourceStream = new ByteArrayInputStream(documentCharset == null ? UTF8.getBytes(input) : input.getBytes(documentCharset));
|
|
} catch (UnsupportedEncodingException e) {
|
|
sourceStream = new ByteArrayInputStream(UTF8.getBytes(input));
|
|
}
|
|
ContentScraper scraper;
|
|
try {
|
|
scraper = parseToScraper(location, documentCharset, detectedcharsetcontainer, sourceStream, maxLinks);
|
|
} catch (Failure e) {
|
|
throw new IOException(e.getMessage());
|
|
}
|
|
return scraper;
|
|
}
|
|
|
|
public static ContentScraper parseToScraper(
|
|
final DigestURL location,
|
|
final String documentCharset,
|
|
Charset[] detectedcharsetcontainer,
|
|
InputStream sourceStream,
|
|
final int maxLinks) throws Parser.Failure, IOException {
|
|
|
|
// make a scraper
|
|
String charset = null;
|
|
|
|
// ah, we are lucky, we got a character-encoding via HTTP-header
|
|
if (documentCharset != null) {
|
|
charset = patchCharsetEncoding(documentCharset);
|
|
}
|
|
|
|
// nothing found: try to find a meta-tag
|
|
if (charset == null) {
|
|
ScraperInputStream htmlFilter = null;
|
|
try {
|
|
htmlFilter = new ScraperInputStream(sourceStream, documentCharset, location, null, false, maxLinks);
|
|
sourceStream = htmlFilter;
|
|
charset = htmlFilter.detectCharset();
|
|
} catch (final IOException e1) {
|
|
throw new Parser.Failure("Charset error:" + e1.getMessage(), location);
|
|
} finally {
|
|
if (htmlFilter != null) htmlFilter.close();
|
|
}
|
|
}
|
|
|
|
// the author didn't tell us the encoding, try the mozilla-heuristic
|
|
if (charset == null) {
|
|
final CharsetDetector det = new CharsetDetector();
|
|
det.enableInputFilter(true);
|
|
final InputStream detStream = new BufferedInputStream(sourceStream);
|
|
det.setText(detStream);
|
|
charset = det.detect().getName();
|
|
sourceStream = detStream;
|
|
}
|
|
|
|
// wtf? still nothing, just take system-standard
|
|
if (charset == null) {
|
|
detectedcharsetcontainer[0] = Charset.defaultCharset();
|
|
} else {
|
|
try {
|
|
detectedcharsetcontainer[0] = Charset.forName(charset);
|
|
} catch (final IllegalCharsetNameException e) {
|
|
detectedcharsetcontainer[0] = Charset.defaultCharset();
|
|
} catch (final UnsupportedCharsetException e) {
|
|
detectedcharsetcontainer[0] = Charset.defaultCharset();
|
|
}
|
|
}
|
|
|
|
// parsing the content
|
|
final ContentScraper scraper = new ContentScraper(location, maxLinks);
|
|
final TransformerWriter writer = new TransformerWriter(null,null,scraper,null,false, Math.max(64, Math.min(4096, sourceStream.available())));
|
|
try {
|
|
FileUtils.copy(sourceStream, writer, detectedcharsetcontainer[0]);
|
|
} catch (final IOException e) {
|
|
throw new Parser.Failure("IO error:" + e.getMessage(), location);
|
|
} finally {
|
|
writer.flush();
|
|
//sourceStream.close(); keep open for multipe parsing (close done by caller)
|
|
writer.close();
|
|
}
|
|
//OutputStream hfos = new htmlFilterOutputStream(null, scraper, null, false);
|
|
//serverFileUtils.copy(sourceFile, hfos);
|
|
//hfos.close();
|
|
if (writer.binarySuspect()) {
|
|
final String errorMsg = "Binary data found in resource";
|
|
throw new Parser.Failure(errorMsg, location);
|
|
}
|
|
return scraper;
|
|
}
|
|
|
|
/**
|
|
* some html authors use wrong encoding names, either because they don't know exactly what they
|
|
* are doing or they produce a type. Many times, the upper/downcase scheme of the name is fuzzy
|
|
* This method patches wrong encoding names. The correct names are taken from
|
|
* http://www.iana.org/assignments/character-sets
|
|
* @param encoding
|
|
* @return patched encoding name
|
|
*/
|
|
public static String patchCharsetEncoding(String encoding) {
|
|
|
|
// do nothing with null
|
|
if ((encoding == null) || (encoding.length() < 3)) return null;
|
|
|
|
// trim encoding string
|
|
encoding = encoding.trim();
|
|
|
|
// fix upper/lowercase
|
|
encoding = encoding.toUpperCase();
|
|
if (encoding.startsWith("SHIFT")) return "Shift_JIS";
|
|
if (encoding.startsWith("BIG")) return "Big5";
|
|
// all other names but such with "windows" use uppercase
|
|
if (encoding.startsWith("WINDOWS")) encoding = "windows" + encoding.substring(7);
|
|
if (encoding.startsWith("MACINTOSH")) encoding = "MacRoman";
|
|
|
|
// fix wrong fill characters
|
|
encoding = CommonPattern.UNDERSCORE.matcher(encoding).replaceAll("-");
|
|
|
|
if (encoding.matches("GB[_-]?2312([-_]80)?")) return "GB2312";
|
|
if (encoding.matches(".*UTF[-_]?8.*")) return "UTF-8";
|
|
if (encoding.startsWith("US")) return "US-ASCII";
|
|
if (encoding.startsWith("KOI")) return "KOI8-R";
|
|
|
|
// patch missing '-'
|
|
if (encoding.startsWith("windows") && encoding.length() > 7) {
|
|
final char c = encoding.charAt(7);
|
|
if ((c >= '0') && (c <= '9')) {
|
|
encoding = "windows-" + encoding.substring(7);
|
|
}
|
|
}
|
|
|
|
if (encoding.startsWith("ISO")) {
|
|
// patch typos
|
|
if (encoding.length() > 3) {
|
|
final char c = encoding.charAt(3);
|
|
if ((c >= '0') && (c <= '9')) {
|
|
encoding = "ISO-" + encoding.substring(3);
|
|
}
|
|
}
|
|
if (encoding.length() > 8) {
|
|
final char c = encoding.charAt(8);
|
|
if ((c >= '0') && (c <= '9')) {
|
|
encoding = encoding.substring(0, 8) + "-" + encoding.substring(8);
|
|
}
|
|
}
|
|
}
|
|
|
|
// patch wrong name
|
|
if (encoding.startsWith("ISO-8559")) {
|
|
// popular typo
|
|
encoding = "ISO-8859" + encoding.substring(8);
|
|
}
|
|
|
|
// converting cp\d{4} -> windows-\d{4}
|
|
if (encoding.matches("CP([_-])?125[0-8]")) {
|
|
final char c = encoding.charAt(2);
|
|
if ((c >= '0') && (c <= '9')) {
|
|
encoding = "windows-" + encoding.substring(2);
|
|
} else {
|
|
encoding = "windows" + encoding.substring(2);
|
|
}
|
|
}
|
|
|
|
return encoding;
|
|
}
|
|
|
|
public static void main(final String[] args) {
|
|
// test parsing of a url
|
|
AnchorURL url;
|
|
try {
|
|
url = new AnchorURL(args[0]);
|
|
final byte[] content = url.get(ClientIdentification.yacyInternetCrawlerAgent, null, null);
|
|
final Document[] document = new htmlParser().parse(url, "text/html", "utf-8", new ByteArrayInputStream(content));
|
|
final String title = document[0].dc_title();
|
|
System.out.println(title);
|
|
} catch (final MalformedURLException e) {
|
|
e.printStackTrace();
|
|
} catch (final IOException e) {
|
|
e.printStackTrace();
|
|
} catch (final Parser.Failure e) {
|
|
e.printStackTrace();
|
|
} catch (final InterruptedException e) {
|
|
e.printStackTrace();
|
|
}
|
|
System.exit(0);
|
|
}
|
|
|
|
}
|