mirror of
https://github.com/yacy/yacy_search_server.git
synced 2024-09-19 00:01:41 +02:00
6491270b3a
removed preferred IPv4 in start options and added a new field IP6 in peer seeds which will contain one or more IPv6 addresses. Now every peer has one or more IP addresses assigned, even several IPv6 addresses are possible. The peer-ping process must check all given and possible IP addresses for a backping and return the one IP which was successful when pinging the peer. The ping-ing peer must be able to recognize which of the given IPs are available for outside access of the peer and store this accordingly. If only one IPv6 address is available and no IPv4, then the IPv6 is stored in the old IP field of the seed DNA. Many methods in Seed.java are now marked as @deprecated because they had been used for a single IP only. There is still a large construction site left in YaCy now where all these deprecated methods must be replaced with new method calls. The 'extra'-IPs, used by cluster assignment had been removed since that can be replaced with IPv6 usage in p2p clusters. All clusters must now use IPv6 if they want an intranet-routing.
287 lines
14 KiB
Java
287 lines
14 KiB
Java
/**
|
|
* CrawlStartScanner_p
|
|
* Copyright 2010 by Michael Peter Christen, mc@yacy.net, Frankfurt am Main, Germany
|
|
* First released 12.12.2010 at http://yacy.net
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this program in the file lgpl21.txt
|
|
* If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
import java.io.IOException;
|
|
import java.net.InetAddress;
|
|
import java.net.MalformedURLException;
|
|
import java.util.ArrayList;
|
|
import java.util.Collection;
|
|
import java.util.ConcurrentModificationException;
|
|
import java.util.Iterator;
|
|
import java.util.LinkedHashSet;
|
|
import java.util.Map;
|
|
import java.util.Set;
|
|
import java.util.TreeMap;
|
|
import java.util.regex.Pattern;
|
|
|
|
import net.yacy.cora.document.id.DigestURL;
|
|
import net.yacy.cora.federate.solr.connector.AbstractSolrConnector;
|
|
import net.yacy.cora.order.Base64Order;
|
|
import net.yacy.cora.protocol.Domains;
|
|
import net.yacy.cora.protocol.RequestHeader;
|
|
import net.yacy.cora.protocol.Scanner;
|
|
import net.yacy.cora.protocol.Scanner.Access;
|
|
import net.yacy.cora.sorting.ReversibleScoreMap;
|
|
import net.yacy.cora.util.ConcurrentLog;
|
|
import net.yacy.data.WorkTables;
|
|
import net.yacy.search.Switchboard;
|
|
import net.yacy.search.SwitchboardConstants;
|
|
import net.yacy.search.query.SearchEventCache;
|
|
import net.yacy.search.schema.CollectionSchema;
|
|
import net.yacy.server.serverObjects;
|
|
import net.yacy.server.serverSwitch;
|
|
|
|
public class CrawlStartScanner_p
|
|
{
|
|
|
|
private final static int CONCURRENT_RUNNER = 200;
|
|
|
|
public static serverObjects respond(
|
|
@SuppressWarnings("unused") final RequestHeader header,
|
|
final serverObjects post,
|
|
final serverSwitch env) {
|
|
|
|
final serverObjects prop = new serverObjects();
|
|
final Switchboard sb = (Switchboard) env;
|
|
|
|
// clean up all search events
|
|
SearchEventCache.cleanupEvents(true);
|
|
|
|
prop.put("noserverdetected", 0);
|
|
prop.put("hosts", "");
|
|
prop.put("intranet.checked", sb.isIntranetMode() ? 1 : 0);
|
|
|
|
int timeout = sb.isIntranetMode() ? 200 : 3000;
|
|
timeout = post == null ? timeout : post.getInt("timeout", timeout);
|
|
|
|
// make a scanhosts entry
|
|
String hostt = post == null ? "" : post.get("scanhosts", "").trim();
|
|
boolean listall = false;
|
|
if (hostt.equals("*") || (post != null && "all".equals(post.get("source", "")))) {
|
|
hostt = "";
|
|
listall = true;
|
|
post.put("source", "hosts");
|
|
}
|
|
String[] hosts0 = hostt.indexOf('\n') > 0 || hostt.indexOf('\r') > 0 ? hostt.split("[\\r\\n]+") : hostt.split(Pattern.quote(","));
|
|
Set<String> hostSet = new LinkedHashSet<String>();
|
|
for (String s: hosts0) if (s != null && s.length() > 0) hostSet.add(s);
|
|
|
|
final Set<InetAddress> ips = Domains.myIntranetIPs();
|
|
prop.put("intranethosts", ips.toString());
|
|
prop.put("intranetHint", sb.isIntranetMode() ? 0 : 1);
|
|
if ( hostSet.isEmpty() ) {
|
|
InetAddress ip;
|
|
if ( sb.isIntranetMode() ) {
|
|
if ( !ips.isEmpty() ) {
|
|
ip = ips.iterator().next();
|
|
} else {
|
|
ip = Domains.dnsResolve("192.168.0.1");
|
|
}
|
|
} else {
|
|
ip = Domains.myPublicLocalIP();
|
|
if ( Domains.isThisHostIP(ip) ) {
|
|
ip = Domains.dnsResolve(sb.peers.mySeed().getIP());
|
|
}
|
|
}
|
|
if ( ip != null ) {
|
|
hostSet.add(ip.getHostAddress());
|
|
}
|
|
}
|
|
String hos = ""; for (String s: hostSet) hos += s + "\n";
|
|
prop.put("scanhosts", hos.trim());
|
|
|
|
if (listall) {
|
|
// get a list of all hosts in the index
|
|
ReversibleScoreMap<String> hostscore = null;
|
|
try {
|
|
hostscore = sb.index.fulltext().getDefaultConnector().getFacets(AbstractSolrConnector.CATCHALL_QUERY, 1000, CollectionSchema.host_s.getSolrFieldName()).get(CollectionSchema.host_s.getSolrFieldName());
|
|
} catch (final IOException e) {}
|
|
if (hostscore != null) {
|
|
for (String s: hostscore) hostSet.add(s);
|
|
}
|
|
}
|
|
|
|
// parse post requests
|
|
if ( post != null ) {
|
|
int repeat_time = 0;
|
|
String repeat_unit = "seldays";
|
|
|
|
// check scheduler
|
|
if ( post.get("rescan", "").equals("scheduler") ) {
|
|
repeat_time = post.getInt("repeat_time", -1);
|
|
repeat_unit = post.get("repeat_unit", "selminutes"); // selminutes, selhours, seldays
|
|
}
|
|
|
|
final int subnet = post.getInt("subnet", 24);
|
|
|
|
// scan a range of ips
|
|
if (post.containsKey("scan")) {
|
|
|
|
// start a scanner
|
|
final Scanner scanner = new Scanner(CONCURRENT_RUNNER, timeout);
|
|
|
|
boolean scanhttp = "on".equals(post.get("scanhttp", ""));
|
|
boolean scanhttps = "on".equals(post.get("scanhttps", ""));
|
|
boolean scanftp = "on".equals(post.get("scanftp", ""));
|
|
boolean scansmb = "on".equals(post.get("scansmb", ""));
|
|
|
|
// select host base to scan
|
|
if ("hosts".equals(post.get("source", ""))) {
|
|
for (String host: hostSet) {
|
|
if (host.startsWith("http://")) host = host.substring(7);
|
|
if (host.startsWith("https://")) host = host.substring(8);
|
|
if (host.startsWith("ftp://")) host = host.substring(6);
|
|
if (host.startsWith("smb://")) host = host.substring(6);
|
|
final int p = host.indexOf('/', 0);
|
|
if (p >= 0) host = host.substring(0, p);
|
|
InetAddress ip;
|
|
Collection<InetAddress> scanbase = new ArrayList<InetAddress>();
|
|
if (host.length() > 0) {
|
|
ip = Domains.dnsResolve(host); if (ip != null) scanbase.add(ip);
|
|
if (host.startsWith("ftp.") || host.startsWith("www.") ||
|
|
host.startsWith("blog.") || host.startsWith("wiki.") ||
|
|
host.startsWith("forum.") || host.startsWith("forums.") ||
|
|
host.startsWith("events.") || host.startsWith("bugs.")) {
|
|
host = host.substring(host.indexOf('.') + 1);
|
|
if (!hostSet.contains(host)) {
|
|
ip = Domains.dnsResolve(host);
|
|
if (ip != null) scanbase.add(ip);
|
|
}
|
|
}
|
|
if (scanftp && !hostSet.contains("ftp." + host)) {
|
|
ip = Domains.dnsResolve("ftp." + host);
|
|
if (ip != null) scanbase.add(ip);
|
|
}
|
|
if ((scanhttp || scanhttps) && !hostSet.contains("www." + host)) {
|
|
ip = Domains.dnsResolve("www." + host);
|
|
if (ip != null) scanbase.add(ip);
|
|
}
|
|
}
|
|
scanner.addProtocols(Scanner.genlist(scanbase, subnet), scanhttp, scanhttps, scanftp, scansmb);
|
|
}
|
|
}
|
|
if ("intranet".equals(post.get("source", ""))) {
|
|
scanner.addProtocols(Scanner.genlist(Domains.myIntranetIPs(), subnet), scanhttp, scanhttps, scanftp, scansmb);
|
|
}
|
|
|
|
scanner.terminate();
|
|
if ("on".equals(post.get("accumulatescancache", "")) && !"scheduler".equals(post.get("rescan", ""))) {
|
|
Scanner.scancacheExtend(scanner);
|
|
} else {
|
|
Scanner.scancacheReplace(scanner);
|
|
}
|
|
}
|
|
|
|
// check crawl request
|
|
if ( post.containsKey("crawl") ) {
|
|
// make a pk/url mapping
|
|
final Iterator<Map.Entry<Scanner.Service, Scanner.Access>> se = Scanner.scancacheEntries();
|
|
final Map<byte[], DigestURL> pkmap = new TreeMap<byte[], DigestURL>(Base64Order.enhancedCoder);
|
|
while (se.hasNext()) {
|
|
final Scanner.Service u = se.next().getKey();
|
|
DigestURL uu;
|
|
try {
|
|
uu = u.url();
|
|
pkmap.put(uu.hash(), uu);
|
|
} catch (final MalformedURLException e ) {
|
|
ConcurrentLog.logException(e);
|
|
}
|
|
}
|
|
// search for crawl start requests in this mapping
|
|
for ( final Map.Entry<String, String> entry : post.entrySet() ) {
|
|
if ( entry.getValue().startsWith("mark_") ) {
|
|
final byte[] pk = entry.getValue().substring(5).getBytes();
|
|
final DigestURL url = pkmap.get(pk);
|
|
if ( url != null ) {
|
|
String path = "/Crawler_p.html?createBookmark=off&xsstopw=off&crawlingDomMaxPages=10000&intention=&range=domain&indexMedia=on&recrawl=nodoubles&xdstopw=off&storeHTCache=on&sitemapURL=&repeat_time=7&crawlingQ=on&cachePolicy=iffresh&indexText=on&crawlingMode=url&mustnotmatch=&crawlingDomFilterDepth=1&crawlingDomFilterCheck=off&crawlingstart=Start%20New%20Crawl&xpstopw=off&repeat_unit=seldays&crawlingDepth=99&directDocByURL=off";
|
|
path += "&crawlingURL=" + url.toNormalform(true);
|
|
WorkTables.execAPICall(
|
|
Domains.LOCALHOST,
|
|
(int) sb.getConfigLong("port", 8090),
|
|
path,
|
|
pk,
|
|
sb.getConfig(SwitchboardConstants.ADMIN_ACCOUNT_USER_NAME, "admin"),
|
|
sb.getConfig(SwitchboardConstants.ADMIN_ACCOUNT_B64MD5, ""));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// check scheduler
|
|
if ( "scheduler".equals(post.get("rescan", "")) ) {
|
|
|
|
// store this call as api call
|
|
if ( repeat_time > 0 ) {
|
|
// store as scheduled api call
|
|
sb.tables.recordAPICall(
|
|
post,
|
|
"CrawlStartScanner_p.html",
|
|
WorkTables.TABLE_API_TYPE_CRAWLER,
|
|
"network scanner for hosts: " + hostSet.toString(),
|
|
repeat_time,
|
|
repeat_unit.substring(3));
|
|
}
|
|
|
|
// execute the scan results
|
|
if ( Scanner.scancacheSize() > 0 ) {
|
|
// make a comment cache
|
|
final Map<byte[], String> apiCommentCache = WorkTables.commentCache(sb);
|
|
|
|
String urlString;
|
|
DigestURL u;
|
|
try {
|
|
final Iterator<Map.Entry<Scanner.Service, Scanner.Access>> se = Scanner.scancacheEntries();
|
|
Map.Entry<Scanner.Service, Scanner.Access> host;
|
|
while ( se.hasNext() ) {
|
|
host = se.next();
|
|
try {
|
|
u = host.getKey().url();
|
|
urlString = u.toNormalform(true);
|
|
if (host.getValue() == Access.granted
|
|
&& Scanner.inIndex(apiCommentCache, urlString) == null ) {
|
|
String path =
|
|
"/Crawler_p.html?createBookmark=off&xsstopw=off&crawlingDomMaxPages=10000&intention=&range=domain&indexMedia=on&recrawl=nodoubles&xdstopw=off&storeHTCache=on&sitemapURL=&repeat_time=7&crawlingQ=on&cachePolicy=iffresh&indexText=on&crawlingMode=url&mustnotmatch=&crawlingDomFilterDepth=1&crawlingDomFilterCheck=off&crawlingstart=Start%20New%20Crawl&xpstopw=off&repeat_unit=seldays&crawlingDepth=99";
|
|
path += "&crawlingURL=" + urlString;
|
|
WorkTables.execAPICall(
|
|
Domains.LOCALHOST,
|
|
(int) sb.getConfigLong("port", 8090),
|
|
path,
|
|
u.hash(),
|
|
sb.getConfig(SwitchboardConstants.ADMIN_ACCOUNT_USER_NAME, "admin"),
|
|
sb.getConfig(SwitchboardConstants.ADMIN_ACCOUNT_B64MD5, ""));
|
|
}
|
|
} catch (final MalformedURLException e ) {
|
|
ConcurrentLog.logException(e);
|
|
}
|
|
}
|
|
} catch (final ConcurrentModificationException e) {
|
|
// we don't want to synchronize this
|
|
ConcurrentLog.logException(e);
|
|
}
|
|
}
|
|
|
|
}
|
|
}
|
|
|
|
return prop;
|
|
}
|
|
|
|
}
|