yacy_search_server/htroot/CrawlStartScanner_p.java
2011-12-06 11:37:05 +01:00

278 lines
12 KiB
Java

/**
* CrawlStartScanner_p
* Copyright 2010 by Michael Peter Christen, mc@yacy.net, Frankfurt am Main, Germany
* First released 12.12.2010 at http://yacy.net
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public License
* along with this program in the file lgpl21.txt
* If not, see <http://www.gnu.org/licenses/>.
*/
import java.net.InetAddress;
import java.net.MalformedURLException;
import java.util.ConcurrentModificationException;
import java.util.HashSet;
import java.util.Iterator;
import java.util.Map;
import java.util.Set;
import java.util.TreeMap;
import net.yacy.cora.protocol.Domains;
import net.yacy.cora.protocol.RequestHeader;
import net.yacy.cora.protocol.Scanner;
import net.yacy.cora.protocol.Scanner.Access;
import net.yacy.kelondro.data.meta.DigestURI;
import net.yacy.kelondro.logging.Log;
import net.yacy.kelondro.order.Base64Order;
import net.yacy.search.Switchboard;
import net.yacy.search.SwitchboardConstants;
import net.yacy.search.query.SearchEventCache;
import de.anomic.data.WorkTables;
import de.anomic.server.serverObjects;
import de.anomic.server.serverSwitch;
public class CrawlStartScanner_p
{
private final static int CONCURRENT_RUNNER = 100;
public static serverObjects respond(
final RequestHeader header,
final serverObjects post,
final serverSwitch env) {
final serverObjects prop = new serverObjects();
final Switchboard sb = (Switchboard) env;
// clean up all search events
SearchEventCache.cleanupEvents(true);
prop.put("noserverdetected", 0);
prop.put("hosts", "");
prop.put("intranet.checked", sb.isIntranetMode() ? 1 : 0);
int timeout = sb.isIntranetMode() ? 200 : 3000;
timeout = post == null ? timeout : post.getInt("timeout", timeout);
// make a scanhosts entry
String hosts = post == null ? "" : post.get("scanhosts", "");
final Set<InetAddress> ips = Domains.myIntranetIPs();
prop.put("intranethosts", ips.toString());
prop.put("intranetHint", sb.isIntranetMode() ? 0 : 1);
if ( hosts.length() == 0 ) {
InetAddress ip;
if ( sb.isIntranetMode() ) {
if ( ips.size() > 0 ) {
ip = ips.iterator().next();
} else {
ip = Domains.dnsResolve("192.168.0.1");
}
} else {
ip = Domains.myPublicLocalIP();
if ( Domains.isThisHostIP(ip) ) {
ip = sb.peers.mySeed().getInetAddress();
}
}
if ( ip != null ) {
hosts = ip.getHostAddress();
}
}
prop.put("scanhosts", hosts);
// parse post requests
if ( post != null ) {
int repeat_time = 0;
String repeat_unit = "seldays";
long validTime = 0;
// check scheduler
if ( post.get("rescan", "").equals("scheduler") ) {
repeat_time = post.getInt("repeat_time", -1);
repeat_unit = post.get("repeat_unit", "selminutes"); // selminutes, selhours, seldays
if ( repeat_unit.equals("selminutes") ) {
validTime = repeat_time * 60 * 1000;
}
if ( repeat_unit.equals("selhours") ) {
validTime = repeat_time * 60 * 60 * 1000;
}
if ( repeat_unit.equals("seldays") ) {
validTime = repeat_time * 24 * 60 * 60 * 1000;
}
}
final boolean bigrange = post.get("subnet", "24").equals("16");
// case: an IP range was given; scan the range for services and display result
if ( post.containsKey("scan") && "hosts".equals(post.get("source", "")) ) {
final Set<InetAddress> ia = new HashSet<InetAddress>();
for ( String host : hosts.split(",") ) {
if ( host.startsWith("http://") ) {
host = host.substring(7);
}
if ( host.startsWith("https://") ) {
host = host.substring(8);
}
if ( host.startsWith("ftp://") ) {
host = host.substring(6);
}
if ( host.startsWith("smb://") ) {
host = host.substring(6);
}
final int p = host.indexOf('/', 0);
if ( p >= 0 ) {
host = host.substring(0, p);
}
ia.add(Domains.dnsResolve(host));
}
final Scanner scanner = new Scanner(ia, CONCURRENT_RUNNER, timeout);
if ( post.get("scanftp", "").equals("on") ) {
scanner.addFTP(bigrange);
}
if ( post.get("scanhttp", "").equals("on") ) {
scanner.addHTTP(bigrange);
}
if ( post.get("scanhttps", "").equals("on") ) {
scanner.addHTTPS(bigrange);
}
if ( post.get("scansmb", "").equals("on") ) {
scanner.addSMB(bigrange);
}
scanner.start();
scanner.terminate();
if ( "on".equals(post.get("accumulatescancache", ""))
&& !"scheduler".equals(post.get("rescan", "")) ) {
Scanner.scancacheExtend(scanner, validTime);
} else {
Scanner.scancacheReplace(scanner, validTime);
}
}
if ( post.containsKey("scan") && "intranet".equals(post.get("source", "")) ) {
final Scanner scanner = new Scanner(Domains.myIntranetIPs(), CONCURRENT_RUNNER, timeout);
if ( "on".equals(post.get("scanftp", "")) ) {
scanner.addFTP(bigrange);
}
if ( "on".equals(post.get("scanhttp", "")) ) {
scanner.addHTTP(bigrange);
}
if ( "on".equals(post.get("scanhttps", "")) ) {
scanner.addHTTPS(bigrange);
}
if ( "on".equals(post.get("scansmb", "")) ) {
scanner.addSMB(bigrange);
}
scanner.start();
scanner.terminate();
if ( "on".equals(post.get("accumulatescancache", ""))
&& !"scheduler".equals(post.get("rescan", "")) ) {
Scanner.scancacheExtend(scanner, validTime);
} else {
Scanner.scancacheReplace(scanner, validTime);
}
}
// check crawl request
if ( post.containsKey("crawl") ) {
// make a pk/url mapping
final Iterator<Map.Entry<Scanner.Service, Scanner.Access>> se = Scanner.scancacheEntries();
final Map<byte[], DigestURI> pkmap =
new TreeMap<byte[], DigestURI>(Base64Order.enhancedCoder);
while ( se.hasNext() ) {
final Scanner.Service u = se.next().getKey();
DigestURI uu;
try {
uu = new DigestURI(u.url());
pkmap.put(uu.hash(), uu);
} catch ( final MalformedURLException e ) {
Log.logException(e);
}
}
// search for crawl start requests in this mapping
for ( final Map.Entry<String, String> entry : post.entrySet() ) {
if ( entry.getValue().startsWith("mark_") ) {
final byte[] pk = entry.getValue().substring(5).getBytes();
final DigestURI url = pkmap.get(pk);
if ( url != null ) {
String path =
"/Crawler_p.html?createBookmark=off&xsstopw=off&crawlingDomMaxPages=10000&intention=&range=domain&indexMedia=on&recrawl=nodoubles&xdstopw=off&storeHTCache=on&sitemapURL=&repeat_time=7&crawlingQ=on&cachePolicy=iffresh&indexText=on&crawlingMode=url&mustnotmatch=&crawlingDomFilterDepth=1&crawlingDomFilterCheck=off&crawlingstart=Start%20New%20Crawl&xpstopw=off&repeat_unit=seldays&crawlingDepth=99";
path += "&crawlingURL=" + url.toNormalform(true, false);
WorkTables.execAPICall(
"localhost",
(int) sb.getConfigLong("port", 8090),
sb.getConfig(SwitchboardConstants.ADMIN_ACCOUNT_B64MD5, ""),
path,
pk);
}
}
}
}
// check scheduler
if ( "scheduler".equals(post.get("rescan", "")) ) {
// store this call as api call
if ( repeat_time > 0 ) {
// store as scheduled api call
sb.tables.recordAPICall(
post,
"CrawlStartScanner_p.html",
WorkTables.TABLE_API_TYPE_CRAWLER,
"network scanner for hosts: " + hosts,
repeat_time,
repeat_unit.substring(3));
}
// execute the scan results
if ( Scanner.scancacheSize() > 0 ) {
// make a comment cache
final Map<byte[], String> apiCommentCache = WorkTables.commentCache(sb);
String urlString;
DigestURI u;
try {
final Iterator<Map.Entry<Scanner.Service, Scanner.Access>> se =
Scanner.scancacheEntries();
Map.Entry<Scanner.Service, Scanner.Access> host;
while ( se.hasNext() ) {
host = se.next();
try {
u = new DigestURI(host.getKey().url());
urlString = u.toNormalform(true, false);
if ( host.getValue() == Access.granted
&& Scanner.inIndex(apiCommentCache, urlString) == null ) {
String path =
"/Crawler_p.html?createBookmark=off&xsstopw=off&crawlingDomMaxPages=10000&intention=&range=domain&indexMedia=on&recrawl=nodoubles&xdstopw=off&storeHTCache=on&sitemapURL=&repeat_time=7&crawlingQ=on&cachePolicy=iffresh&indexText=on&crawlingMode=url&mustnotmatch=&crawlingDomFilterDepth=1&crawlingDomFilterCheck=off&crawlingstart=Start%20New%20Crawl&xpstopw=off&repeat_unit=seldays&crawlingDepth=99";
path += "&crawlingURL=" + urlString;
WorkTables.execAPICall(
"localhost",
(int) sb.getConfigLong("port", 8090),
sb.getConfig(SwitchboardConstants.ADMIN_ACCOUNT_B64MD5, ""),
path,
u.hash());
}
} catch ( final MalformedURLException e ) {
Log.logException(e);
}
}
} catch ( final ConcurrentModificationException e ) {
}
}
}
}
return prop;
}
}