//ViewFile.java //----------------------- //part of YaCy //(C) by Michael Peter Christen; mc@yacy.net //first published on http://www.anomic.de //Frankfurt, Germany, 2004 //last major change: 12.07.2004 //This program is free software; you can redistribute it and/or modify //it under the terms of the GNU General Public License as published by //the Free Software Foundation; either version 2 of the License, or //(at your option) any later version. //This program is distributed in the hope that it will be useful, //but WITHOUT ANY WARRANTY; without even the implied warranty of //MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the //GNU General Public License for more details. //You should have received a copy of the GNU General Public License //along with this program; if not, write to the Free Software //Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA //you must compile this file with //javac -classpath .:../Classes Status.java //if the shell's current path is HTROOT import java.io.ByteArrayInputStream; import java.io.IOException; import java.io.UnsupportedEncodingException; import java.net.MalformedURLException; import java.net.URLDecoder; import java.util.Enumeration; import java.util.HashMap; import java.util.Iterator; import java.util.Map; import net.yacy.document.Condenser; import net.yacy.document.Document; import net.yacy.document.ParserException; import net.yacy.document.parser.html.CharacterCoding; import net.yacy.document.parser.html.ImageEntry; import net.yacy.kelondro.data.meta.DigestURI; import net.yacy.kelondro.data.meta.URIMetadataRow; import net.yacy.kelondro.logging.Log; import net.yacy.repository.LoaderDispatcher; import de.anomic.crawler.retrieval.Response; import de.anomic.http.client.Client; import de.anomic.http.client.Cache; import de.anomic.http.server.RequestHeader; import de.anomic.http.server.ResponseHeader; import de.anomic.search.Segment; import de.anomic.search.Segments; import de.anomic.search.Switchboard; import de.anomic.server.serverObjects; import de.anomic.server.serverSwitch; public class ViewFile { public static final int VIEW_MODE_NO_TEXT = 0; public static final int VIEW_MODE_AS_PLAIN_TEXT = 1; public static final int VIEW_MODE_AS_PARSED_TEXT = 2; public static final int VIEW_MODE_AS_PARSED_SENTENCES = 3; public static final int VIEW_MODE_AS_IFRAME_FROM_WEB = 4; public static final int VIEW_MODE_AS_IFRAME_FROM_CACHE = 5; public static final int VIEW_MODE_AS_LINKLIST = 6; public static final int VIEW_MODE_AS_PARSED_WORDS = 7; private static final String HIGHLIGHT_CSS = "searchHighlight"; private static final int MAX_HIGHLIGHTS = 6; public static serverObjects respond(final RequestHeader header, final serverObjects post, final serverSwitch env) { final serverObjects prop = new serverObjects(); final Switchboard sb = (Switchboard)env; if (post == null) { prop.put("display", 1); prop.put("error_display", 0); prop.putHTML("error_words", ""); prop.put("error_vMode-sentences", "1"); prop.put("error", "1"); prop.put("url", ""); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } final int display = post.getInt("display", 1); // get segment Segment indexSegment = null; if (post.containsKey("segment")) { String segmentName = post.get("segment"); if (sb.indexSegments.segmentExist(segmentName)) { indexSegment = sb.indexSegments.segment(segmentName); } } else { // take default segment indexSegment = sb.indexSegments.segment(Segments.Process.PUBLIC); } prop.put("display", display); prop.put("error_display", display); if (post.containsKey("words")) prop.putHTML("error_words", post.get("words")); else { prop.putHTML("error_words", ""); } final String viewMode = post.get("viewMode","parsed"); prop.put("error_vMode-" + viewMode, "1"); DigestURI url = null; String descr = ""; final int wordCount = 0; int size = 0; boolean pre = false; // get the url hash from which the content should be loaded String urlHash = post.get("urlHash", ""); URIMetadataRow urlEntry = null; // get the urlEntry that belongs to the url hash if (urlHash.length() > 0 && (urlEntry = indexSegment.urlMetadata().load(urlHash.getBytes(), null, 0)) != null) { // get the url that belongs to the entry final URIMetadataRow.Components metadata = urlEntry.metadata(); if ((metadata == null) || (metadata.url() == null)) { prop.put("error", "3"); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } url = metadata.url(); descr = metadata.dc_title(); //urlEntry.wordCount(); size = urlEntry.size(); pre = urlEntry.flags().get(Condenser.flag_cat_indexof); } prop.put("error_inurldb", urlEntry == null ? 0 : 1); // alternatively, get the url simply from a url String // this can be used as a simple tool to test the text parser final String urlString = post.get("url", ""); if (urlString.length() > 0) try { // this call forces the peer to download web pages // it is therefore protected by the admin password if (!sb.verifyAuthentication(header, false)) { prop.put("AUTHENTICATE", "admin log-in"); // force log-in return prop; } // define an url by post parameter url = new DigestURI(urlString, null); urlHash = new String(url.hash()); pre = post.get("pre", "false").equals("true"); } catch (final MalformedURLException e) {} if (url == null) { prop.put("error", "1"); prop.put("viewMode", VIEW_MODE_NO_TEXT); prop.put("url", ""); return prop; } else { prop.put("url", url.toNormalform(false, true)); } // loading the resource content as byte array prop.put("error_incache", Cache.has(url) ? 1 : 0); byte[] resource = null; ResponseHeader responseHeader = null; String resMime = null; // trying to load the resource body try { resource = Cache.getContent(url); } catch (IOException e) { Log.logException(e); resource = null; } responseHeader = Cache.getResponseHeader(url); // if the resource body was not cached we try to load it from web if (resource == null) { Response entry = null; try { entry = sb.loader.load(url, true, false); } catch (final Exception e) { prop.put("error", "4"); prop.putHTML("error_errorText", e.getMessage()); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } if (entry != null) { resource = entry.getContent(); } if (resource == null) { prop.put("error", "4"); prop.put("error_errorText", "No resource available"); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } } // try to load resource metadata if (responseHeader == null) { // try to load the metadata from cache try { responseHeader = Cache.getResponseHeader(url); } catch (final Exception e) { /* ignore this */ } // if the metadata was not cached try to load it from web if (responseHeader == null) { final String protocol = url.getProtocol(); if (!((protocol.equals("http") || protocol.equals("https")))) { prop.put("error", "6"); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } responseHeader = Client.whead(url.toString()); if (responseHeader == null) { prop.put("error", "4"); prop.put("error_errorText", "Unable to load resource metadata."); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } resMime = responseHeader.mime(); } } else { resMime = responseHeader.mime(); } final String[] wordArray = wordArray(post.get("words", null)); if (viewMode.equals("plain")) { // TODO: how to handle very large files here ? String content; try { content = new String(resource, "UTF-8"); } catch (final Exception e) { prop.put("error", "4"); prop.putHTML("error_errorText", e.getMessage()); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } finally { resource = null; } prop.put("error", "0"); prop.put("viewMode", VIEW_MODE_AS_PLAIN_TEXT); prop.put("viewMode_plainText", markup(wordArray, content).replaceAll("\n", "
").replaceAll("\t", "    ")); } else if (viewMode.equals("iframeWeb")) { prop.put("viewMode", VIEW_MODE_AS_IFRAME_FROM_WEB); prop.put("viewMode_url", url.toNormalform(false, true)); } else if (viewMode.equals("iframeCache")) { prop.put("viewMode", VIEW_MODE_AS_IFRAME_FROM_CACHE); prop.put("viewMode_url", url.toNormalform(false, true)); } else if (viewMode.equals("parsed") || viewMode.equals("sentences") || viewMode.equals("words") || viewMode.equals("links")) { // parsing the resource content Document document = null; try { document = LoaderDispatcher.parseDocument(url, resource.length, new ByteArrayInputStream(resource), null); if (document == null) { prop.put("error", "5"); prop.put("error_errorText", "Unknown error"); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } } catch (final ParserException e) { prop.put("error", "5"); prop.putHTML("error_errorText", e.getMessage()); prop.put("viewMode", VIEW_MODE_NO_TEXT); return prop; } finally { resource = null; } resMime = document.dc_format(); if (viewMode.equals("parsed")) { final String content = new String(document.getTextBytes()); // content = wikiCode.replaceHTML(content); // added by Marc Nause prop.put("viewMode", VIEW_MODE_AS_PARSED_TEXT); prop.put("viewMode_title", document.dc_title()); prop.put("viewMode_creator", document.dc_creator()); prop.put("viewMode_subject", document.dc_subject(',')); prop.put("viewMode_description", document.dc_description()); prop.put("viewMode_publisher", document.dc_publisher()); prop.put("viewMode_format", document.dc_format()); prop.put("viewMode_identifier", document.dc_identifier()); prop.put("viewMode_source", document.dc_source().toString()); prop.put("viewMode_parsedText", markup(wordArray, content).replaceAll("\n", "
").replaceAll("\t", "    ")); } else if (viewMode.equals("sentences")) { prop.put("viewMode", VIEW_MODE_AS_PARSED_SENTENCES); final Iterator sentences = document.getSentences(pre); boolean dark = true; int i = 0; String sentence; if (sentences != null) { // Search word highlighting while (sentences.hasNext()) { sentence = sentences.next().toString(); if (sentence.trim().length() > 0) { prop.put("viewMode_sentences_" + i + "_nr", i + 1); prop.put("viewMode_sentences_" + i + "_text", markup(wordArray, sentence)); prop.put("viewMode_sentences_" + i + "_dark", dark ? "1" : "0"); dark = !dark; i++; } } } prop.put("viewMode_sentences", i); } else if (viewMode.equals("words")) { prop.put("viewMode", VIEW_MODE_AS_PARSED_WORDS); final Iterator sentences = document.getSentences(pre); boolean dark = true; int i = 0; String sentence, token; if (sentences != null) { // Search word highlighting while (sentences.hasNext()) { sentence = sentences.next().toString(); Enumeration tokens = Condenser.wordTokenizer(sentence, "UTF-8"); while (tokens.hasMoreElements()) { token = tokens.nextElement().toString(); if (token.length() > 0) { prop.put("viewMode_words_" + i + "_nr", i + 1); prop.put("viewMode_words_" + i + "_word", token); prop.put("viewMode_words_" + i + "_dark", dark ? "1" : "0"); dark = !dark; i++; } } } } prop.put("viewMode_words", i); } else if (viewMode.equals("links")) { prop.put("viewMode", VIEW_MODE_AS_LINKLIST); boolean dark = true; int i = 0; i += putMediaInfo(prop, wordArray, i, document.getVideolinks(), "video", (i % 2 == 0)); i += putMediaInfo(prop, wordArray, i, document.getAudiolinks(), "audio", (i % 2 == 0)); dark = (i % 2 == 0); final HashMap ts = document.getImages(); final Iterator tsi = ts.values().iterator(); ImageEntry entry; while (tsi.hasNext()) { entry = tsi.next(); prop.put("viewMode_links_" + i + "_nr", i); prop.put("viewMode_links_" + i + "_dark", dark ? "1" : "0"); prop.put("viewMode_links_" + i + "_type", "image"); prop.put("viewMode_links_" + i + "_text", markup(wordArray, entry.alt())); prop.put("viewMode_links_" + i + "_url", entry.url().toNormalform(false, true)); prop.put("viewMode_links_" + i + "_link", markup(wordArray, entry.url().toNormalform(false, true))); if (entry.width() > 0 && entry.height() > 0) prop.put("viewMode_links_" + i + "_attr", entry.width() + "x" + entry.height() + " Pixel"); else prop.put("viewMode_links_" + i + "_attr", "unknown"); dark = !dark; i++; } i += putMediaInfo(prop, wordArray, i, document.getApplinks(), "app", (i % 2 == 0)); i += putMediaInfo(prop, wordArray, i, document.getHyperlinks(), "link", (i % 2 == 0)); prop.put("viewMode_links", i); } if (document != null) document.close(); } prop.put("error", "0"); prop.put("error_url", url.toNormalform(false, true)); prop.put("error_hash", urlHash); prop.put("error_wordCount", wordCount); prop.putHTML("error_desc", descr); prop.putNum("error_size", size); prop.put("error_mimeTypeAvailable", (resMime == null) ? "0" : "1"); prop.put("error_mimeTypeAvailable_mimeType", resMime); return prop; } private static final String[] wordArray(String words) { String[] w = new String[0]; if (words == null || words.length() == 0) return w; try { words = URLDecoder.decode(words, "UTF-8"); w = words.substring(1, words.length() - 1).split(","); } catch (final UnsupportedEncodingException e) {} return w; } private static final String markup(final String[] wordArray, String message) { message = CharacterCoding.unicode2html(message, true); if (wordArray != null) for (int j = 0; j < wordArray.length; j++) { final String currentWord = wordArray[j].trim(); // TODO: replace upper-/lowercase words as well message = message.replaceAll(currentWord, "" + currentWord + ""); } return message; } private static int putMediaInfo(final serverObjects prop, final String[] wordArray, int c, final Map media, final String name, boolean dark) { final Iterator> mi = media.entrySet().iterator(); Map.Entry entry; int i = 0; while (mi.hasNext()) { entry = mi.next(); prop.put("viewMode_links_" + c + "_nr", c); prop.put("viewMode_links_" + c + "_dark", ((dark) ? 1 : 0)); prop.putHTML("viewMode_links_" + c + "_type", name); prop.put("viewMode_links_" + c + "_text", markup(wordArray, entry.getValue())); prop.put("viewMode_links_" + c + "_link", markup(wordArray, entry.getKey().toNormalform(true, false))); prop.put("viewMode_links_" + c + "_url", entry.getKey().toNormalform(true, false)); prop.putHTML("viewMode_links_" + c + "_attr", ""); dark = !dark; c++; i++; } return i; } }