This images shows incoming connections to your YaCy peer and outgoing connections from your peer to other peers and web servers==这幅图显示了到你节点的传入连接,以及从你节点到其他节点或网站服务器的传出连接
The right '*', after the '/', can be replaced by a <a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">regular expression</a>.== 在 '/' 后边的 '*' ,可用<a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">正则表达式</a>表示.
#domain.net/fullpath<==domain.net/绝对路径<
#>domain.net/*<==>domain.net/*<
#*.domain.net/*<==*.domain.net/*<
#*.sub.domain.net/*<==*.sub.domain.net/*<
#sub.domain.*/*<==sub.domain.*/*<
#domain.*/*<==domain.*/*<
#was removed from blacklist==wurde aus Blacklist entfernt
#was added to the blacklist==wurde zur Blacklist hinzugefügt
To see a list of all APIs, please visit the <a href="https://wiki.yacy.net/index.php/Dev:API" target="_blank">API wiki page</a>.==要查看所有API的列表,请访问<a href="https://wiki.yacy.net/index.php/Dev:API" target="_blank">API wiki page</a>。
To see a list of all APIs==要查看所有API的列表,请访问<a href="https://wiki.yacy.net/index.php/Dev:API" target="_blank">API wiki page</a>。
To see a list of all APIs, please visit the <a href="https://wiki.yacy.net/index.php/Dev:API">API wiki page</a>.==获取所有API, 请访问<a href="https://wiki.yacy.net/index.php/Dev:API">API Wiki</a>.
Access to your peer from your own computer (localhost access) is granted with administrator rights. No need to configure an administration account.==通过管理员权限授予从你自己的计算机访问你的节点(localhost访问权限).无需配置管理帐户.
This setting is convenient but less secure than using a qualified admin account.==此设置很方便,但比使用合格的管理员帐户安全性低.
Please use with care, notably when you browse untrusted and potentially malicious websites while running your YaCy peer on the same computer.==请谨慎使用,尤其是在计算机上运行YaCy节点并浏览不受信任和可能有恶意的网站时.
This is required if you want a remote access to your peer, but it also hardens access controls on administration operations of your peer.==如果你希望远端访问你的节点,则这是必需的,但它也会加强节点管理操作的访问控制.
Select one of the default skins. <b>After selection it might be required to reload the web page while holding the shift key to refresh cached style files.</b>==选择一个默认皮肤。<b>选择后,重新加载网页,可能需要在按住shift键的同时刷新缓存的样式文件。</b>
Your port has changed. Please wait 10 seconds.==你的端口已更改。 请等待10秒。
Your browser will be redirected to the new <a href="http://#[host]#:#[port]#/ConfigBasic.html">location</a> in 5 seconds.==你的浏览器将在5秒内重定向到新的<a href="http://#[host]#:#[port]#/ConfigBasic.html">位置</a>。
The peer port was changed successfully.==节点端口已经成功修改。
Your YaCy installation behaves independently from other peers and you define your own web index by starting your own web crawl. This can be used to search your own web pages or to define a topic-oriented search portal.==你的YaCy安装独立于其他节点,你可以通过开始自己的网络爬虫来创建自己的网络索引。这可用于搜索你的个人网站或创建专题搜索门户。
Your peer can be reached by other peers==外部能访问你的节点
Your peer cannot be reached from outside==外部不能访问你的节点
which is not fatal, but would be good for the YaCy network==此举不是强制的,但有利于YaCy网络
please open your firewall for this port and/or set a virtual server option in your router to allow connections on this port.==请改变你的防火墙或者虚拟机路由设置, 从而让外网能访问这个端口。
Opening a router port is <i>not</i> a YaCy-specific task;==打开一个路由器端口不是一个YaCy特定的任务;
you can see instruction videos everywhere in the internet, just search for <a href="http://www.youtube.com/results?search_query=Open+Ports+on+a+Router">Open Ports on a <our-router-type> Router</a> and add your router type as search term.==你可以在互联网上的任何地方查看说明视频,只需搜索<a href="http://www.youtube.com/results?search_query=Open Ports on a Router">在<我的路由器类型>路由器打开一个端口</a>并添加你的路由器类型作为搜索词。
However: if you fail to open a router port, you can nevertheless use YaCy with full functionality, the only function that is missing is on the side of the other YaCy users because they cannot see your peer.==但是:如果你无法打开路由器端口,你仍然可以使用YaCy的全部功能,唯一缺失的功能是对其他YaCy用户而言的,因为他们无法看到你的YaCy节点。
Your Peer name is a default name; please set an individual peer name.==你的节点名称为系统默认,请另外设置一个名称。
You did not set a user name and/or a password.==你未设置用户名和/或密码。
Some pages are protected by passwords.==一些页面受密码保护。
You should set a password at the <a href="ConfigAccounts_p.html">Accounts Menu</a> to secure your YaCy peer.</p>::==你可以在 <a href="ConfigAccounts_p.html">账户菜单</a> 设置密码, 从而加强你的YaCy节点安全性。</p>::
You did not open a port in your firewall or your router does not forward the server port to your peer.==你未在防火墙中打开端口,或者你的路由器不能与服务器端口建立有效链接。
This is needed if you want to fully participate in the YaCy network.==如果你想完全加入YaCy网络, 此项是必须的。
You can also use your peer without opening it, but this is not recomended.==不开放端口你也能使用你的节点, 但是不推荐。
A <a href="http://en.wikipedia.org/wiki/Heuristic" target="_blank">heuristic</a> is an 'experience-based technique that help in problem solving, learning and discovery' (wikipedia).==<a href="http://en.wikipedia.org/wiki/Heuristic" target="_blank">启发式</a>是一种“基于经验的技术,有助于解决问题,学习和发现”
search-result: shallow crawl on all displayed search results==搜索结果:浅度爬取所有显示的搜索结果
When a search is made then all displayed result links are crawled with a depth-1 crawl.==当进行搜索时,所有显示的结果网址的爬网深度-1。
When using this heuristic, then every new search request line is used for a call to listed opensearch systems.==使用这种启发式时,每个新的搜索请求行都用于调用列出的opensearch系统。
This means: right after the search request every page is loaded and every page that is linked on this page.==这意味着:在搜索请求之后,就开始加载结果的每个页面及每个页面上的链接。
If you check 'add as global crawl job' the pages to be crawled are added to the global crawl queue (remote peers can pickup pages to be crawled).==如果选中'添加为全球爬取作业',则要爬取的页面将被添加到全球爬取队列中(其他远端YaCy节点可能会帮助爬取这些页面)。
With the button "discover from index" you can search within the metadata of your local index (Web Structure Index) to find systems which support the Opensearch specification.==使用“从索引发现”按钮,你可以在本地索引(Web结构索引)的元数据中搜索,以查找支持Opensearch规范的系统。
The task is started in the background. It may take some minutes before new entries appear (after refreshing the page).==任务在后台启动。 出现新词条可能需要几分钟时间(在刷新页面之后)。
located in <i>defaults/heuristicopensearch.conf</i> to the DATA/SETTINGS directory.==位于DATA / SETTINGS目录的<i> defaults / heuristicopensearch.conf </i>中。
For the discover function the <i>web graph</i> option of the web structure index and the fields <i>target_rel_s, target_protocol_s, target_urlstub_s</i> have to be switched on in the <a href="IndexSchema_p.html?core=webgraph">webgraph Solr schema</a>.==对于发现功能,Web结构索引的<i> web图表</i>选项和字段<i> target_rel_s,target_protocol_s,target_urlstub_s </i>必须在<a href="IndexSchema_p.html?core=webgraph">webgraph Solr模式</a>。
When using this heuristic==使用这种启发式时,每个新的搜索请求行都用于调用列出的opensearch系统。
For the discover function the <i>web graph</i> option of the web structure index and the fields <i>target_rel_s==对于发现功能,Web结构索引的<i> web图表</ i>选项和字段<i> target_rel_s,target_protocol_s,target_urlstub_s </ i>必须在<a href =“IndexSchema_p.html ?core = webgraph“> webgraph Solr模式</a>。
The search heuristics that can be switched on here are techniques that help the discovery of possible search results based on link guessing, in-search crawling and requests to other search engines.==你可以在这里开启启发式搜索, 通过猜测链接, 嵌套搜索和访问其他搜索引擎, 从而找到更多符合你期望的结果.
When a search heuristic is used, the resulting links are not used directly as search result but the loaded pages are indexed and stored like other content.==开启启发式搜索时, 搜索结果给出的链接并不是直接搜索的链接, 而是已经缓存在其他服务器上的结果.
This ensures that blacklists can be used and that the searched word actually appears on the page that was discovered by the heuristic.==这保证了黑名单的有效性, 并且搜索关键字是通过启发式搜索找到的.
The success of heuristics are marked with an image==启发式搜索找到的结果会被特定图标标记
heuristic:<name>==启发式:<名称>
#(redundant)==(redundant)
(new link)==(新链接)
below the favicon left from the search result entry:==搜索结果中使用的图标:
The search result was discovered by a heuristic, but the link was already known by YaCy==搜索结果通过启发式搜索, 且链接已知
The search result was discovered by a heuristic, not previously known by YaCy==搜索结果通过启发式搜索, 且链接未知
When a search is made using a 'site'-operator (like: 'download site:yacy.net') then the host of the site-operator is instantly crawled with a host-restricted depth-1 crawl.==当使用'站点'-操作符搜索时(比如: 'download site:yacy.net') ,服务器就会立即爬取层数为 最大限制深度-1 的内容.
That means: right after the search request the portal page of the host is loaded and every page that is linked on this page that points to a page on the same host.==意即: 在链接请求发出后, 搜索引擎就会载入在同一服务器中每一个与此页面相连的网页.
Because this 'instant crawl' must obey the robots.txt and a minimum access time for two consecutive pages, this heuristic is rather slow, but may discover all wanted search results using a second search (after a small pause of some seconds).==因为'立即爬取'依赖于爬虫协议和两个相连页面的最小访问时间, 所以这个启发式选项会相当慢, 但是在第二次搜索时会搜索到更多词条(需要间隔几秒钟).
The HTCache stores content retrieved by the HTTP and FTP protocol. Documents from smb:// and file:// locations are not cached.==超文本缓存存储着从HTTP和FTP协议获得的内容. 其中从smb:// 和 file:// 取得的内容不会被缓存.
The cache is a rotating cache: if it is full, then the oldest entries are deleted and new one can fill the space.==此缓存是队列式的: 队列满时, 会删除旧内容, 从而加入新内容.
#HTCache Configuration
HTCache Configuration==超文本缓存配置
Cache hits==缓存命中率
The path where the cache is stored==缓存存储路径
The current size of the cache==当前缓存容量
>#[actualCacheSize]# MB for #[actualCacheDocCount]# files, #[docSizeAverage]# KB / file in average==>#[actualCacheSize]#MB为#[actualCacheDocCount]#文件, #[docSizeAverage]#平均KB /文件
For P2P operation, at least DHT distribution or DHT receive (or both) must be set. You have thus defined a Robinson configuration==关于P2P操作,必须至少勾选DHT分发或DHT接收(或两者)。 因此,你已被确定为漂流配置
Global Search in P2P configuration is only allowed, if index receive is switched on. You have a P2P configuration, but are not allowed to search other peers.==P2P配置中的全局搜索仅在打开接受索引时才被允许。你已有P2P配置,但不被允许搜索其他节点。
If your peer runs in 'Robinson Mode' you run YaCy as a search engine for your own search portal without data exchange to other peers==如果你的节点运行在'漂流模式', 你能在不与其他节点交换数据的情况下进行搜索
There is no index receive and no index distribution between your peer and any other peer==你不会与其他节点进行索引传递
When you allow access from the YaCy network, your data is recognized using keywords==当你允许YaCy网络的访问时, 你的数据会以关键字形式表示
Please describe your search portal with some keywords (comma-separated)==请用关键字描述你的搜索门户 (以逗号隔开)
If you leave the field empty, no peer asks your peer. If you fill in a '*', your peer is always asked.==如果此部分留空, 那么你的节点不会被其他节点访问. 如果内容是 '*' 则标示你的节点永远被允许访问.
it should be used to encrypt outgoing communications with them (for operations such as network presence, index transfer, remote crawl==它应该被用来加密与它们的传出通信(操作:网络存在、索引传输、远端爬行
Note also that encryption of remote search queries is configured with a dedicated setting in the <a href="ConfigPortal_p.html">Config Portal</a> page.==另请注意,请在<a href="ConfigPortal_p.html">门户配置</a>页面中设置远端搜索加密功能。
Every page will become a single index hit and the url is artifically extended with a post/get attribute value containing the page number as value==每个页面都将成为单个索引匹配,并且使用包含页码作为值的post/get属性值人为扩展url
If you like to integrate YaCy as portal for your web pages, you may want to change icons and messages on the search page.==如果你想将YaCy作为你的网站搜索门户, 你可能需要在这改变搜索页面的图标和信息。
and a link to a home page that is reached when the 'corporate identity'-images are clicked.==和一个点击'企业标志'图像后转到主页的超链接。
To change also colours and styles use the <a href="ConfigAppearance_p.html">Appearance Servlet</a> for different skins and languages.==若要改变颜色和风格,请到<a href="ConfigAppearance_p.html">外观选项</a>选择你喜欢的皮肤和语言。
Control whether media search results are as default strictly limited to indexed documents matching exactly the desired content domain==控制媒体搜索结果是否默认严格限制为与所需内容域完全匹配的索引文档
Automated results resorting with JavaScript makes the browser load the full result set of each search request.==基于JavaScript的自动结果重新排序,使浏览器加载每个搜索请求的完整结果集。
This may lead to high system loads on the server.==这可能会导致服务器上的系统负载过高。
Please check the 'Peer-to-peer search with JavaScript results resorting' section in the <a href="SearchAccessRate_p.html">Local Search access rate</a> configuration page to set up proper limitations on this mode by unauthenticated users.==请查看<a href="SearchAccessRate_p.html">本地搜索访问率</a> 配置页面中的“使用JavaScript对P2P搜索结果重排”部分,对未经身份验证的用户使用该模式加以适当限制。
Prefer https for search queries on remote peers.==首选https用于远端节点上的搜索查询。
When SSL/TLS is enabled on remote peers, https should be used to encrypt data exchanged with them when performing peer-to-peer searches.==在远端节点上启用SSL/TLS时,应使用https来加密在执行P2P搜索时与它们交换的数据。
Please note that contrary to strict TLS, certificates are not validated against trusted certificate authorities (CA), thus allowing YaCy peers to use self-signed certificates.==请注意,与严格TLS相反,证书不会针对受信任的证书颁发机构(CA)进行验证,因此允许YaCy节点使用自签名证书。
Statistics on text snippets generation can be enabled in the <a href="Settings_p.html?page=debug">Debug/Analysis Settings</a> page.==可以在<a href="Settings_p.html?page=debug">调试/分析设置</a>页面中启用文本摘录生成的统计信息。
List of hosts that shall be excluded from search results by default but can be included using the site:<host> operator:==默认情况下将被排除在搜索结果之外的服务器列表,但可以使用site:<host>操作符包括进来
or <a href="ViewProfile.html?hash=localhash">in the public</a> using a <a href="ViewProfile.rdf?hash=localhash">FOAF RDF file</a>.==或者<a href="ViewProfile.html?hash=localhash">在公共场所时</a>使用<a href="ViewProfile.rdf?hash=localhash">FOAF RDF 文件</a>.
Homepage (appears on every <a href="Supporter.html">Supporter Page</a> as long as your peer is online)==首页(显示在每个<a href="Supporter.html">支持者</a> 页面中, 前提是你的节点在线).
You can change anything, but some options need a restart, and some options can crash YaCy, if wrong values are used.==你可以改变任何这里的设置, 当然, 有的需要重启才能生效, 有的甚至能引起YaCy崩溃.
We give information how to integrate a search box on any web page that==如何将一个搜索框集成到任意
calls the normal YaCy search window.==调用YaCy搜索的页面.
Simply use the following code:==使用以下代码:
MySearch== 我的搜索
"Search"=="搜索"
This would look like:==示例:
This does not use a style sheet file to make the integration into another web page with a different style sheet easier.==在这里并没有使用样式文件, 因为这样会比较容易将其嵌入到不同样式的页面里.
Below is a generic template of the search result page. Mark the check boxes for features you would like to be displayed.==以下是搜索结果页面的通用模板.选中你希望显示的功能复选框.
To change colors and styles use the <a href="ConfigAppearance_p.html">Appearance</a> menu for different skins.==要改变颜色和样式,使用<a href="ConfigAppearance_p.html">外观</a>菜单以改变皮肤。
Other portal settings can be adjusted in <a href="ConfigPortal_p.html">Generic Search Portal</a> menu.==其他门户网站设置可以在<a href="ConfigPortal_p.html">通用搜索门户</a>菜单中调整.
Maximum days number in the histogram. Beware that a large value may trigger high CPU loads both on the server and on the browser with large result sets.==直方图中的最大天数. 请注意, 较大的值可能会在服务器和具有大结果集的浏览器上触发高CPU负载.
Showing #[numActiveRunning]# active connections from a max. of #[numMax]# allowed incoming connections==正在显示 #[numActiveRunning]# 活动连接,最大允许传入连接 #[numMax]#
Double-Content detection is done using a ranking on a 'unique'-Field, named 'fuzzy_signature_unique_b'.==重复内容检测是使用名为'fuzzy_signature_unique_b'的'unique'字段上的排名完成的。
This field is set during parsing and is influenced by two attributes for the <a href="https://lucene.apache.org/solr/5_5_2/solr-core/org/apache/solr/update/processor/TextProfileSignature.html" target="_blank">TextProfileSignature</a> class.==此字段在解析期间设置,并受<a href="https://lucene.apache.org/solr/5_5_2/solr-core/org/apache/solr/update/processor/TextProfileSignature.html" target="_blank">TextProfileSignature</a>类的两个属性影响。
>minTokenLen<==>最小令牌长度<
This is the minimum length of a word which shall be considered as element of the signature. Should be either 2 or 3.==这是一个应被视为签名的元素单词的最小长度。应该是2或3。
>quantRate<==>量化率<
The quantRate is a measurement for the number of words that take part in a signature computation. The higher the number, the less==量化率是参与签名计算的单词数量的度量。
words are used for the signature==数字越大,用于签名的单词就越少。
For minTokenLen = 2 the quantRate value should not be below 0.24; for minTokenLen = 3 the quantRate value must be not below 0.5.==对于最小令牌长度=2,量化率值不应低于0.24; 对于最小令牌长度=3,量化率值必须不低于0.5。
The quantRate is a measurement for the number of words that take part in a signature computation. The higher the number==quantRate是参与签名计算的单词数量的度量。 数字越高,越少
#-----------------------------
#File: ContentIntegrationPHPBB3_p.html
#---------------------------
Content Integration: Retrieval from phpBB3 Databases==内容集成: 从phpBB3数据库中导入
It is possible to extract texts directly from mySQL and postgreSQL databases.==能直接从mysql或者postgresql中解压出内容.
If you read from an imported database, here are some hints to get around problems when importing dumps in phpMyAdmin:==如果从使用phpMyAdmin读取数据库内容, 你可能会用到以下建议:
in phpmyadmin/config.inc.php and place your dump file in /tmp (Otherwise it is not possible to upload files larger than 2MB)==设置phpmyadmin/config.inc.php的内容, 并将你的数据库文件放到 /tmp 目录下(否则不能上传大于2MB的文件)
When an export is started, surrogate files are generated into DATA/SURROGATE/in which are automatically fetched by an indexer thread.==导出过程开始时, 在 DATA/SURROGATE/in 目录下自动生成备份文件, 并且会被索引器自动爬取.
All indexed surrogate files are then moved to DATA/SURROGATE/out and can be re-cycled when an index is deleted.==所有被索引的备份文件都在 DATA/SURROGATE/out 目录下, 并被索引器循环利用.
it may take some seconds until the first result appears there.==在出现第一个搜索结果前需要几秒钟时间.
If you crawl any un-wanted pages, you can delete them <a href="IndexCreateQueues_p.html?stack=LOCAL">here</a>.==如果你爬取了不需要的页面, 你可以 <a href="IndexCreateQueues_p.html?stack=LOCAL">点这</a> 删除它们.
#Crawl profiles hold information about a specific URL which is internally used to perform the crawl it belongs to.==Crawl Profile enthalten Informationen über eine spezifische URL, welche intern genutzt wird, um nachzuvollziehen, wozu der Crawl gehört.
#The profiles for remote crawls, <a href="ProxyIndexingMonitor_p.html">indexing via proxy</a> and snippet fetches==Die Profile für Remote Crawl, <a href="ProxyIndexingMonitor_p.html">Indexierung per Proxy</a> und Snippet Abrufe
#cannot be altered here as they are hard-coded.==können nicht verändert werden, weil sie "hard-coded" sind.
YaCy knows 5 different ways to acquire web indexes. The details of these processes (1-5) are described within the submenu's listed==YaCy使用5种不同的方式来获取网络索引. 详细描述显示在子菜单的进程(1-5)中,
above which also will show you a table with indexing results so far. The information in these tables is considered as private,==以上列表也会显示目前的索引结果. 表中的信息是私有的,
so you need to log-in with your administration password.==所以你需要以管理员账户来查看.
Case (6) is a monitor of the local receipt-generator, the opposed case of (1). It contains also an indexing result monitor but is not considered private==事件(6)是本地回执生成器的监控器, (1)的相反事件. 它也包含一个索引结果监控器, 但不是私有的.
<em>Use Case:</em> You get entries here, if you start a local crawl on the '<a href="CrawlStartExpert.html">Advanced Crawler</a>' page and check the==<em>用法:</em> 你可在此获得词条, 当你在 '<a href="CrawlStartExpert.html">高级爬虫页面</a> 上启动本地爬取并勾选
'Do Remote Indexing'-flag, and if you checked the 'Accept Remote Crawl Requests'-flag on the '<a href="RemoteCrawl_p.html">Remote Crawling</a>' page.=='执行远端索引'-标志时, 这需要你确保在 '<a href="RemoteCrawl_p.html">远端爬取</a>' 页面中勾选了'接受远端爬取请求'-标志.
the logic of the Global Distributed Hash Table.==你的节点是最适合存储它们的.
<em>Use Case:</em> This list may fill if you check the 'Index Receive'-flag on the 'Index Control' page==<em>用法:</em> 如果你在'索引控制'页面上选中'索引接收'-标志, 则此列表会填写
This is the 'mirror'-case of process (1).==这是进程(1)的'镜像'事件.
<em>Use Case:</em> This list may fill if you check the 'Accept Remote Crawl Requests'-flag on the '<a href="RemoteCrawl_p.html">Remote Crawling</a>' page==<em>用法:</em> 如果你在 '<a href="RemoteCrawl_p.html">远端爬取</a>' 页面勾选'接受远端爬取请求'-标记,此列表会填写
<em>Use Case:</em> place files with dublin core metadata content into DATA/SURROGATES/in or use an index import method==将包含Dublin核心元数据的文件放在 DATA/SURROGATES/in 中, 或者使用索引导入方式
Click on this API button to see a documentation of the POST request parameter for crawl starts.==单击此API按钮查看爬取启动的POST请求参数的文档。
Expert Crawl Start==高级爬取开启
Start Crawling Job:==开启爬取任务:
You can define URLs as start points for Web page crawling and start crawling here.==你可以在此指定网页爬取起始点的网址和开启爬取。
"Crawling" means that YaCy will download the given website, extract all links in it and then download the content behind these links.== "爬取中"意即YaCy会下载指定的网站, 并提取出其中的链接,接着下载链接中的全部内容。
This is repeated as long as specified under "Crawling Depth".==它将一直重复上述步骤,直到满足指定的"爬取深度"。
A crawl can also be started using wget and the <a href="http://www.yacy-websearch.net/wiki/index.php/Dev:APICrawler" target="_blank">post arguments</a> for this web page.==也可以使用此网页的wget和<a href="http://www.yacy-websearch.net/wiki/index.php/Dev:APICrawler" target="_blank">post参数</a>开启爬取。
>Crawl Job<==>爬取任务<
A Crawl Job consist of one or more start point, crawl limitations and document freshness rules.==爬取任务由一个或多个起始点、爬取限制和文档更新规则构成。
One Start URL or a list of URLs:<br/>(must start with http:// https:// ftp:// smb:// file://)==起始网址或网址列表:<br/>(必须以http:// https:// ftp:// smb:// file://开头)
Define the start-url(s) here. You can submit more than one URL, each line one URL please.==在此给定起始网址。你可以提交多个网址,请一个网址一行。
Each of these URLs are the root for a crawl start, existing start URLs are always re-loaded.==这些网址中每个都是爬取开始的起点,已存在的起始网址总是会被重新加载。
Other already visited URLs are sorted out as "double", if they are not allowed using the re-crawl option.==对其他已访问过的网址,如果基于重爬选项它们不被允许,则被标记为'重复'。
>From Link-List of URL<==>来自网址的链接列表<
From Sitemap==来自网站地图
From File (enter a path<br/>within your local file system)==来自文件<br/>(输入一个本地文件系统路径)
This defines how often the Crawler will follow links (of links..) embedded in websites.==此选项决定了爬虫将跟随嵌入网址中链接的深度。
0 means that only the page you enter under "Starting Point" will be added==0代表仅将"起始点"网址添加到索引。
to the index. 2-4 is good for normal indexing. Values over 8 are not useful, since a depth-8 crawl will==2-4是常规索引用的值。超过8的值没有用,因为深度为8的爬取将
index approximately 25.600.000.000 pages, maybe this is the whole WWW.==索引接近256亿个网页,这可能是整个互联网的内容。
also all linked non-parsable documents==包括全部链接中不可解析的文档
>Unlimited crawl depth for URLs matching with<==>对这些匹配的网址不不限制爬取深度<
>Maximum Pages per Domain<==>每个域名下最大网页数<
You can limit the maximum number of pages that are fetched and indexed from a single domain with this option.==使用此选项,你可以限制单个域名下爬取和索引的页面数。
You can combine this limitation with the 'Auto-Dom-Filter', so that the limit is applied to all the domains within==你可以将此设置与'Auto-Dom-Filter'结合起来, 以限制给定深度中所有域名。
the given depth. Domains outside the given depth are then sorted-out anyway.==超出深度范围的域名会被自动忽略。
>Use<==>使用<
Page-Count<==页面数<
>misc. Constraints<==>其它限制<
A questionmark is usually a hint for a dynamic page. URLs pointing to dynamic content should usually not be crawled.==问号标记常用作动态网页的提示。指向动态内容的地址通常不应该被爬取。
However, there are sometimes web pages with static content that==然而,也有些含有静态网页地址也包含问号标记。
is accessed with URLs containing question marks. If you are unsure, do not check this to avoid crawl loops.==如果你不确定,不要勾选此项以防爬取陷入循环。
Following frames is NOT done by Gxxg1e, but we do by default to have a richer content. 'nofollow' in robots metadata can be overridden; this does not affect obeying of the robots.txt which is never ignored.==以下框架不是Gxxg1e制作的,但我们默认会制作更丰富的内容。robots元数据中的nofollow可被否决;这并不影响对无法忽视的robots.txt的遵守。
Accept URLs with query-part ('?'): ==接受包含问号标记('?')的地址:
Not loading URLs with unsupported file extension is faster but less accurate.==不加载包含不受支持文件扩展名的网址速度更快,但准确性更低。
Indeed, for some web resources the actual Media Type is not consistent with the URL file extension. Here are some examples:==实际上,对于某些网络资源,实际的媒体类型与网址中文件扩展名不一致。以下是一些例子:
: the .de extension is unknown, but the actual Media Type of this page is text/html==: 这个.de扩展名未知,但此页面的实际媒体类型为text/html
: the .com extension is not supported (executable file format), but the actual Media Type of this page is text/html==: 这个.com扩展名不受支持(可执行文件格式),但此页面的实际媒体类型为text/html
: the .png extension is a supported image format, but the actual Media Type of this page is text/html==: 这个.png扩展名是一种受支持的图像格式,但该页面的实际媒体类型是text/html
The filter is a <b><a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">regular expression</a></b>.==这个过滤器是一个<b><a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">正则表达式</a></b>。
Example: to allow only urls that contain the word 'science', set the must-match filter to '.*science.*'. ==示例:要仅允许包含单词“science”的网址,请将“必须匹配”筛选器设置为'.*science.*'。
You can also use an automatic domain-restriction to fully crawl a single domain.==你还可以使用自动域名限制来完全爬取单个域名。
Attention: you can test the functionality of your regular expressions using the <a href="RegexTest.html">Regular Expression Tester</a> within YaCy.==注意:你可以使用YaCy中的<a href="RegexTest.html">正则表达式测试仪</a>测试正则表达式的功能。
>Load Filter on URL origin of links<==>在链接的地址上加载筛选器<
The filter is a <b><a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">regular expression</a></b>==这个过滤器是一个<b><a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">正则表达式</a></b>
Example: to allow loading only links from pages on example.org domain, set the must-match filter to '.*example.org.*'.==示例:为只允许加载域名example.org网页中链接,将“必须匹配”筛选器设置为'.*example.org.*'。
Crawls can be restricted to specific countries. This uses the country code that can be computed from==爬取可以限制在特定的国家。它使用的国家代码可以从存放网页的服务器的IP计算得出。
the IP of the server that hosts the page. The filter is not a regular expressions but a list of country codes, separated by comma.==过滤器不是正则表达式,而是国家代码列表,用逗号分隔。
The filter is a <b><a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">regular expression</a></b>==这个过滤器是一个<b><a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">正则表达式</a></b>
that <b>must not match</b> with the URLs to allow that the content of the url is indexed.==匹配那些<b>必须排除</b>的网址,以允许对剩下网址的内容进行索引。
Filter on Content of Document<br/>(all visible text, including camel-case-tokenized url and title)==文档内容过滤器<br/>(所有可见文本,包括驼峰大小写标记的网址和标题)
Filter on Document Media Type (aka MIME type)==文档媒体类型过滤器(又名MIME类型)
that <b>must match</b> with the document Media Type (also known as MIME Type) to allow the URL to be indexed. ==对那些有<b>必须匹配</b>文档媒体类型(也称为MIME类型)的网址进行索引。
Standard Media Types are described at the <a href="https://www.iana.org/assignments/media-types/media-types.xhtml" target="_blank">IANA registry</a>.==<a href="https://www.iana.org/assignments/media-types/media-types.xhtml" target="_blank">IANA注册表</a>中描述了标准媒体类型。
Solr query filter on any active <a href="IndexSchema_p.html" target="_blank">indexed</a> field(s)==任何<a href="IndexSchema_p.html" target="_blank">激活索引</a>字段上的Solr查询过滤器
Each parsed document is checked against the given Solr query before being added to the index.==在添加到索引之前,将根据给定的Solr查询检查每个已解析的文档。
The query must be written in respect to the <a href="https://lucene.apache.org/solr/guide/6_6/the-standard-query-parser.html#the-standard-query-parser" target="_blank">standard</a> Solr query syntax.==必须按照<a href="https://lucene.apache.org/solr/guide/6_6/the-standard-query-parser.html#the-standard-query-parser" target="_blank">标准</a>Solr查询语法编写查询。
The embedded local Solr index must be connected to use this kind of filter.==要使用这种过滤器,必须连接嵌入式本地Solr索引。
You can configure this with the <a href="IndexFederated_p.html">Index Sources & targets</a> page.==你可以使用<a href="IndexFederated_p.html">索引源目标</a>页面对此进行配置。
Check this option to be sure to get fresh search results including newly crawled documents. Beware that it will also interrupt any refreshing/resorting of search results currently requested from browser-side.==选中此选项以确保获得新包括新爬取文档的搜索结果.请注意,它也会中断当前从浏览器端请求的搜索结果的刷新/排序.
After a crawl was done in the past, document may become stale and eventually they are also deleted on the target host.==在过去完成爬取后,文档可能会过时,最终它们也会在目标服务器上被删除。
To remove old files from the search index it is not sufficient to just consider them for re-load but it may be necessary==若要从搜索索引中删除旧文件,仅考虑重新加载它们是不够的。
to delete them because they simply do not exist any more. Use this in combination with re-crawl while this time should be longer.==但可能有必要删除它们,因为它们已经不存在了。与重新爬取组合使用,而这一时间应该更长。
A web crawl performs a double-check on all links found in the internet against the internal database. If the same url is found again,==网页爬取参照自身数据库,对所有找到的链接进行重复性检查.如果链接重复,
then the url is treated as double when you check the 'no doubles' option. A url may be loaded again when it has reached a specific age,==并且'无重复'选项打开, 则被以重复链接对待.如果地址存在时间超过一定时间,
never go online, use all content from cache. If no cache exist, treat content as unavailable==从不检查线上内容, 全部使用缓存内容. 如果缓存存在, 将其视为无效
>Robot Behaviour<==>机器人行为<
Use Special User Agent and robot identification==使用特殊的用户代理和机器人识别
Because YaCy can be used as replacement for commercial search appliances==因为YaCy可以替代商业搜索设备
(like the Google Search Appliance aka GSA) the user must be able to crawl all web pages that are granted to such commercial platforms.==(像谷歌搜索设备,又名GSA)用户必须能够抓取所有授予此类商业平台的网页。
Not having this option would be a strong handicap for professional usage of this software. Therefore you are able to select==没有这个选项将是专业使用该软件的一大障碍。
alternative user agents here which have different crawl timings and also identify itself with another user agent and obey the corresponding robots rule.==因此,你可以在此处选择替代用户代理,它具有不同爬取时间,还可以伪装成另一个用户代理标识,并遵守相应的机器人规则。
>Enrich Vocabulary<==>丰富词汇<
>Scraping Fields<==>刮领域<
You can use class names to enrich the terms of a vocabulary based on the text content that appears on web pages. Please write the names of classes into the matrix.==你可以根据网页上显示的文本内容,使用类名丰富词汇表中的术语。请把类名写进表格。
Snapshots are xml metadata and pictures of web pages that can be created during crawling time.==快照是可以在爬取期间创建的xml元数据和网页图片。
The xml data is stored in the same way as a Solr search result with one hit and the pictures will be stored as pdf into subdirectories==xml数据以与Solr搜索结果相同的方式存储,只需点击一次,图片将以pdf格式存储到HTCACHE/snapshots/的子目录中。
of HTCACHE/snapshots/. From the pdfs the jpg thumbnails are computed. Snapshot generation can be controlled using a depth parameter; that==根据PDF计算jpg缩略图。可以使用深度参数控制快照生成;
means a snapshot is only be generated if the crawl depth of a document is smaller or equal to the given number here. If the number is set to -1,==这意味着只有当文档的爬网深度小于或等于此处给定的数字时,才会生成快照。
Only XML snapshots can be generated. as the <a href="https://wkhtmltopdf.org/" target="_blank">wkhtmltopdf</a> util is not found by YaCy on your system.==只能生成XML快照。因为YaCy在你的系统上找不到<a href="https://wkhtmltopdf.org/" target="_blank">wkhtmltopdf</a>工具。
It is required to generate PDF snapshots from crawled pages that can then be converted to images.==需要从爬取的页面中生成PDF快照,然后将其转换为图像。
This enables indexing of the webpages the crawler will download. This should be switched on by default, unless you want to crawl only to fill the==这样就可以对爬虫将下载的网页进行索引。
Document Cache without indexing.==默认情况下,应该打开该选项,除非你只想爬取以填充文档缓存而不建立索引。
A crawl result can be tagged with names which are candidates for a collection request.==爬取结果可以标记为收集请求的候选名称。
These tags can be selected with the <a href="gsa/search?q=www&site=#[collection]#">GSA interface</a> using the 'site' operator.==这些标签可以通过<a href="gsa/search?q=www&site=#[collection]#">GSA界面</a>使用“网站”运算进行选择。
To use this option, the 'collection_sxt'-field must be switched on in the <a href="IndexFederated_p.html">Solr Schema</a>==要使用此选项,必须在<a href="IndexFederated_p.html">Solr模式</a>中打开“collection_sxt”字段
The time zone is required when the parser detects a date in the crawled web page. Content can be searched with the on: - modifier which==当解析器在已爬取的网页中检测到日期时,需要时区。
requires also a time zone when a query is made. To normalize all given dates, the date is stored in UTC time zone. To get the right offset==可以使用on:-修饰符搜索内容,在进行查询时,该修饰符还需要一个时区。为了规范化所有给定的日期,该日期存储在UTC时区中。
from dates without time zones to UTC, this offset must be given here. The offset is given in minutes;==要获得从没有时区的日期到UTC的正确偏移量,必须在此处给出该偏移量。偏移量以分钟为单位;
Time zone offsets for locations east of UTC must be negative; offsets for zones west of UTC must be positve.==UTC以东位置的时区偏移必须为负值;UTC以西区域的偏移量必须为正值。
YaCy can scan a network segment for available http, ftp and smb server.==YaCy可以扫描一个网段以查找可用的http、ftp和smb服务器。
You must first select a IP range and then, after this range is scanned,==你须先指定IP范围,此后该范围将被扫描,
it is possible to select servers that had been found for a full-site crawl.==也可以选择已找到的服务器作全站点爬取。
No servers had been detected in the given IP range #[iprange]#. Please enter a different IP range for another scan.==在给定IP范围内#[iprange]#,未检测到可用服务器,请重新指定IP范围。
No more that four pages are loaded from the same host in one second (not more that 120 document per minute) to limit the load on the target server.==每秒最多从同一服务器中载入4个页面(每分钟不超过120个文件)以减少对目标服务器影响。
A second crawl for a different host increases the throughput to a maximum of 240 documents per minute since the crawler balances the load over all hosts.==因爬虫会平衡全部服务器的负载,对于不同服务器的二次爬取, 生产量会上升到每分钟最多240个文件。
A 'shallow crawl' which is not limited to a single host (or site)==当目标服务器数量很多时, 不局限于单个服务器(或站点)的'浅爬取'模式
can extend the pages per minute (ppm) rate to unlimited documents per minute when the number of target hosts is high.==会将生产量上升到每分钟无限页面数(ppm)。
This can be done using the <a href="CrawlStartExpert.html">Expert Crawl Start</a> servlet.==可在<a href="CrawlStartExpert.html">专家爬虫</a>中开启。
>Scheduler Steering<==>调度器控制<
The scheduler on crawls can be changed or removed using the <a href="Table_API_p.html">API Steering</a>.==可以使用<a href="Table_API_p.html">API控制</a>改变或删除爬虫调度器。
>With this file it is possible to find locations in Germany using the location (city) name, a zip code, a car sign or a telephone pre-dial number.<==>使用此插件, 则能通过查询城市名, 邮编, 车牌号或者电话区号得到德国任何地点的位置信息.<
Extend media search results (images, videos or applications specific) to pages including such medias (provides generally more results, but eventually less relevant).==将媒体搜索结果(特定于图像、视频或应用程序)扩展到包含此类媒体的页面(通常提供更多结果,但最终相关性较低)。
> Extended==> 拓展
Strictly limit media search results (images, videos or applications specific) to indexed documents matching exactly the desired content domain.==严格将媒体搜索结果(特定于图像、视频或应用程序)限制为与所需内容域完全匹配的索引文档。
> Strict==> 严格
>Query Operators<==>查询运算符<
>restrictions<==>限制<
only urls with the <phrase> in the url==仅包含词组<phrase>的网址的结果
only urls with the <phrase> within outbound links of the document==仅在文档的出站链接中包含带有词组<phrase>的网址
only urls with extension <ext>==仅包含拓展名为<ext>的网址
only urls from host <host>==仅服务器为<host>的网址
only pages with as-author-anotated <author>==仅包含作者为<author>的页面
only pages from top-level-domains <tld>==仅来自顶级域<tld>的页面
only pages with <date> in content==仅内容包含<date>的页面
only pages with a date between <date1> and <date2> in content==内容中只有日期介于<date1>和<date2>之间的页面
only pages with keyword anotation containing <phrase>==仅包含包含<phrase>的关键字注释的页面
only resources from http or https servers==仅限来自http或https服务器的资源
only resources from ftp servers (they are rare, <a href="CrawlStartSite.html">crawl them yourself</a>==只有来自ftp服务器的资源(它们很少见,请<a href="CrawlStartSite.html">自己抓取</a>)
only resources from smb servers (<a href="ConfigBasic.html">Intranet Indexing</a> must be selected)==仅限来自smb服务器的资源(必须选择<a href="ConfigBasic.html">内网索引</a>)
only files from a local file system (<a href="ConfigBasic.html">Intranet Indexing</a> must be selected)==仅来自本地文件系统的文件(必须选择<a href="ConfigBasic.html">内网索引</a>)
>spatial restrictions<==>空间限制<
only documents having location metadata (geographical coordinates)==仅包含位置元数据(地理坐标)的文档
only documents within a square zone embracing a circle of given radius (in decimal degrees) around the specified latitude and longitude (in decimal degrees)==仅限于包含指定经纬度(十进制度数)周围给定半径(十进制度数)圆圈的正方形区域内的文档
>ranking modifier<==>排名修饰符<
sort by date (latest first)==按日期排序(最新优先)
multiple words shall appear near==多个单词应出现在附近
"" (doublequotes)=="" (双引号)
/language/<lang>==/language/<语言>
prefer given language (an <a href="http://www.loc.gov/standards/iso639-2/php/English_list.php" title="Reference alpha-2 language codes list">ISO 639-1</a> 2-letter code)==首选给定语言(<a href="http://www.loc.gov/standards/iso639-2/php/English_list.php" title="Reference alpha-2 language codes list">ISO 639-1</a>的2字母代码)
>heuristics<==>启发式<
>add search results from external opensearch systems<==>从外部开放搜索系统添加搜索结果<
>Search Navigation<==>搜索导航<
>keyboard shortcuts<==>键盘快捷键<
>Access key<==>访问键<
> modifier + n<==> 修饰语 + n<
>next result page<==>下页结果<
> modifier + p<==> 修饰语 + p<
>previous result page<==>上页结果<
>automatic result retrieval<==>自动结果检索<
>browser integration<==>浏览器集成<
after searching, click-open on the default search engine in the upper right search field of your browser and select 'Add "YaCy Search.."'==搜索完成后,单击浏览器右上角搜索字段中默认搜索引擎上的“打开”,然后选择'添加YaCy搜索..'
>search as rss feed<==>作为rss源搜索<
click on the red icon in the upper right after a search. this works good in combination with the '/date' ranking modifier. See an <a href="yacysearch.rss?query=news+%2Fdate&Enter=Search&verify=cacheonly&contentdom=text&nav=hosts%2Cauthors%2Cnamespace%2Ctopics%2Cfiletype%2Cprotocol&startRecord=0&indexof=off&meanCount=5&maximumRecords=10&resource=global&prefermaskfilter=">example</a>.==搜索后点击右上角的红色图标。这与“/date”排名修饰符结合使用效果很好。看一个<a href="yacysearch.rss?query=news+%2Fdate&Enter=Search&verify=cacheonly&contentdom=text&nav=hosts%2Cauthors%2Cnamespace%2Ctopics%2Cfiletype%2Cprotocol&startRecord=0&indexof=off&meanCount=5&maximumRecords=10&resource=global&prefermaskfilter=">例子</a>。
>json search results<==>json搜索结果<
for ajax developers: get the search rss feed and replace the '.rss' extension in the search result url with '.json'==对于ajax开发人员:获取搜索rss提要并替换搜索结果地址'.rss'扩展名为'.json'
this may produce unresolved references at other word indexes but they do not harm==这可能和其他关键字产生未解析关联, 但是这并不影响系统性能
"Delete URL and remove all references from words"=="删除地址并从关键字中删除所有关联"
delete the reference to this url at every other word where the reference exists (very extensive, but prevents unresolved references)==删除指向此链接的关联字,(很多, 但是会阻止未解析关联的产生)
The search index contains #[doccount]# documents. You can delete them here.==搜索索引包含#[doccount]#篇文档。你可以在这儿删除它们。
Deletions are made concurrently which can cause that recently deleted documents are not yet reflected in the document count.==删除是同步进行的,这可能导致最近删除的文档还没有反映在文档计数中。
Index deletion will not immediately reduce the storage size on disk because entries are only marked as deleted in a first step.==索引删除不会立即减少磁盘上的存储大小,因为条目仅在第一步中被标记为已删除。
The storage size will later on shrink by itself if new documents are indexed or you can force a shrinking by <a href="/IndexControlURLs_p.html">performing an "Optimize Solr" procedure.</a>==如果新文档被索引,存储大小将在稍后自行缩小,或者你可以通过执行<a href="/IndexControlURLs_p.html">优化Solr</a>过程强制缩小。
Delete all documents within a sub-path of the given urls. That means all documents must start with one of the url stubs as given here.==删除给定网址的子路径中的所有文档. 这意味着所有文档必须以此处给出的其中一个url存根开头.
One URL stub, a list of URL stubs<br/>or a regular expression==一个URL存根, 一个URL存根列表<br/> 或一条正则表达式
Matching Method<==匹配方法<
sub-path of given URLs==给定URL的子路径
matching with regular expression==与正则表达式匹配
"Simulate Deletion"=="模拟删除"
"no actual deletion, generates only a deletion count"=="没有实际删除,只生成删除计数"
"Engage Deletion"=="真正删除"
"simulate a deletion first to calculate the deletion count"=="首先请模拟删除以计算删除数量"
"engaged"=="删除了"
selected #[count]# documents for deletion==选择 #[count]# 篇文档以删除
deleted #[count]# documents==删除了 #[count]# 篇文档
Delete by Age<==按年龄删除<
Delete all documents which are older than a given time period.==删除所有超过给定时间段的文档.
YaCy supports multiple index storage locations.==YaCy支持多地索引储存。
As an internal indexing database a deep-embedded multi-core Solr is used and it is possible to attach also a remote Solr.==内部索引数据库使用了深度嵌入式多核Solr,并且还可以附加远端Solr。
>Solr Search Index<==>Solr搜索索引<
Solr stores the main search index. It is the home of two cores, the default 'collection1' core for documents and the 'webgraph' core for a web structure graph. Detailed information about the used Solr fields can be edited in the <a href="IndexSchema_p.html">Schema Editor</a>.==Solr存储主搜索索引。它是两个核心的所在地,默认的'collection1'核心用于文档,'webgraph'核心用于网络结构图。可以在<a href="IndexSchema_p.html">模式编辑器</a>中编辑有关已用Solr字段的详细信息。
>Lazy Value Initialization <==>惰性值初始化 <
If checked, only non-zero values and non-empty strings are written to Solr fields.==如果选中,则仅将非零值和非空字符串写入 Solr 字段。
>Use deep-embedded local Solr <==>使用深度嵌入的本地Solr <
This will write the YaCy-embedded Solr index which stored within the YaCy DATA directory.==这将写入存储在YaCy的DATA目录下的 YaCy嵌入式Solr索引。
write-enabled (if unchecked, the remote server(s) will only be used as search peers)==启用写入(如果未选中,远程服务器将仅用作搜索节点)
value="Set"==value="设置"
Web Structure Index==网络结构图索引
The web structure index is used for host browsing (to discover the internal file/folder structure), ranking (counting the number of references) and file search (there are about fourty times more links from loaded pages as in documents of the main search index). ==网页结构索引用于服务器浏览(发现内部文件/文件夹结构)、排名(计算引用次数)和文件搜索(加载页面的链接大约是主搜索索引的文档中的40倍)。
use citation reference index (lightweight and fast)==使用引文参考索引(轻量且快速)
use webgraph search index (rich information in second Solr core)==使用网图搜索索引(第二个Solr核心中的丰富信息)
Peer-to-Peer Operation==P2P运行
The 'RWI' (Reverse Word Index) is necessary for index transmission in distributed mode. For portal or intranet mode this must be switched off.=='RWI'(反向词索引)对于分布式模式下的索引传输是必需的。在门户或内网模式下,必须将其关闭。
support peer-to-peer index transmission (DHT RWI index)==支持点对点索引传输(DHT RWI索引)
You can import <a href="https://dumps.wikimedia.org/backup-index-bydb.html" target="_blank">MediaWiki dumps</a> here. An example is the file==你可以在这导入<a href="https://dumps.wikimedia.org/backup-index-bydb.html" target="_blank">MediaWiki副本</a>副本. 示例
Dumps must be in XML format and may be compressed in gz or bz2. Place the file in the YaCy folder or in one of its sub-folders.==副本文件必须是XML格式并用bz2压缩的.将其放进YaCy目录或其子目录中.
When the import is started, the following happens:==:开始导入时, 会进行以下工作
The dump is extracted on the fly and wiki entries are translated into Dublin Core data format. The output looks like this:==备份文件即时被解压, 并被译为Dublin核心元数据格式:
Each 10000 wiki records are combined in one output file which is written to /DATA/SURROGATES/in into a temporary file.==每个输出文件都含有10000个百科记录, 并都被保存在 /DATA/SURROGATES/in 的临时目录中.
Each time a xml surrogate file appears in /DATA/SURROGATES/in, the YaCy indexer fetches the file and indexes the record entries.==只要 /DATA/SURROGATES/in 中含有 xml文件, YaCy索引器就会读取它们并为其中的词条制作索引.
You can recycle processed surrogate files by moving them from /DATA/SURROGATES/out to /DATA/SURROGATES/in==你可以将文件从/DATA/SURROGATES/out 移动到 /DATA/SURROGATES/in 以重复索引.
Results from the import can be monitored in the <a href="CrawlResults.html?process=7">indexing results for surrogates==导入结果<a href="CrawlResults.html?process=7">监控
In case that an index schema of the embedded/local index has changed, all documents with missing field entries can be indexed again with a reindex job.==如果嵌入式/本地索引的索引架构发生更改,则可以使用重新索引作业再次索引所有缺少字段条目的文档。
"refresh page"=="刷新页面"
Documents in current queue<==当前队列中的文档<
Documents processed<==已处理的文档<
current select query==当前选择查询
"start reindex job now"=="立即开始重新索引作业"
"stop reindexing"=="停止重新索引"
Remaining field list==剩余字段列表
reindex documents containing these fields:==重新索引包含这些字段的文档:
Re-Crawl Index Documents==重新抓取索引文档
Searches the local index and selects documents to add to the crawler (recrawl the document).==搜索本地索引并选择要添加到爬虫的文档(重新爬取文档)。
This runs transparent as background job.==这作为后台作业透明运行。
Documents are added to the crawler only if no other crawls are active==仅当没有其他爬取处于活动状态时,才会将文档添加到爬虫中
and are added in small chunks.==并以小块添加。
"start recrawl job now"=="立即开始重新抓取作业"
"stop recrawl job"=="停止重新抓取作业"
Re-Crawl Query Details==重新抓取查询详情
Documents to process==待处理的文档
Current Query==当前查询
Edit Solr Query==编辑Solr查询
update==更新
to re-crawl documents selected with the given query.==重新抓取使用给定查询选择的文档。
Include failed URLs==包含失败的地址
>Field<==>字段<
>count<==>计数<
Re-crawl works only with an embedded local Solr index!==重新抓取仅适用于嵌入的本地Solr索引!
Simulate==模拟
Check only how many documents would be selected for recrawl==仅检查将选择多少文档进行重新抓取
"Browse metadata of the #[rows]# first selected documents"=="浏览 #[rows]# 个第一个选定文档的元数据"
Last #(/jobStatus)#Re-Crawl job report==最近的#(/jobStatus)#重新抓取作业报告
Automatically refreshing==自动刷新
An error occurred while trying to refresh automatically==尝试自动刷新时出错
The job terminated early due to an error when requesting the Solr index.==由于请求Solr索引时出错,作业提前终止。
>Status<==>状态<
"Running"=="运行中"
"Shutdown in progress"=="正在关闭"
"Terminated"=="已终止"
Running::Shutdown in progress::Terminated==运行中::正在关闭:已终止
>Query<==>查询<
>Start time<==>开启时间<
>End time<==>结束时间<
URLs added to the crawler queue for recrawl==添加到爬虫队列以进行重新爬取的地址
>Recrawled URLs<==>已重新爬取的地址<
URLs rejected for some reason by the crawl stacker or the crawler queue. Please check the logs for more details.==由于某种原因在抓取堆栈器或抓取器队列中被拒绝的地址。请检查日志以获取更多详细信息。
If you use a custom Solr schema you may enter a different field name in the column 'Custom Solr Field Name' of the YaCy default attribute name==如果您使用自定义 Solr 架构,您可以在YaCy默认属性名称的'自定义Solr字段名称'列中输入不同的字段名称
Select a core:==选择核心:
the core can be searched at==核心可以在以下位置搜索
Active==激活
Attribute==属性
Custom Solr Field Name==自定义Solr字段名称
Comment==注释
show active==显示激活
show all available==显示全部可用
show disabled==显示未激活
"Set"=="设置"
"reset selection to default"=="将选择值重置为默认值"
>Reindex documents<==>重新索引文档<
If you unselected some fields, old documents in the index still contain the unselected fields.==如果您取消选择某些字段,但索引中的旧文档仍包含取消选择的字段。
To physically remove them from the index you need to reindex the documents.==要从索引中实际删除它们,您需要重新索引文档。
Here you can reindex all documents with inactive fields.==在这里,您可以重新索引所有具有非活动字段的文档。
"reindex Solr"=="重新索引Solr"
You may monitor progress (or stop the job) under <a href="IndexReIndexMonitor_p.html">IndexReIndexMonitor_p.html</a>==您可以在<a href="IndexReIndexMonitor_p.html">IndexReIndexMonitor_p.html</a>下监控进度(或停止工作)
The local index currently consists of (at least) #[wcount]# reverse word indexes and #[ucount]# URL references==本地索引目前包含(至少) #[wcount]# 反向词索引和 #[ucount]# 地址引用
Index: ==索引:
distribute ==分发
receive grant default: == 接受准许默认值:
receive==接受
for each remote peer == 对每个远端节点
It is much better to retrieve the forum postings directly from the database.==所以, 直接从数据库中获取帖子内容效果更好.
This will cause that YaCy is able to offer nice navigation features after searches.==这会使得YaCy在每次搜索后提供较好引导特性.
YaCy has a phpBB3 extraction feature, please go to the <a href="ContentIntegrationPHPBB3_p.html">phpBB3 content integration</a> servlet for direct database imports.==YaCy能够解析phpBB3关键字, 参见 <a href="ContentIntegrationPHPBB3_p.html">phpBB3内容集成</a> 直接导入数据库方法.
Inserting a Search Window to phpBB3==在phpBB3中添加搜索框
To integrate a search window into phpBB3, you must insert some code into a forum template.==在论坛模板中添加以下代码以将搜索框集成到phpBB3中.
There are several templates that can be used for phpBB3, but in this guide we consider that==phpBB3中有多种模板,
you are using the default template, 'prosilver'==在此我们使用默认模板 'prosilver'.
open styles/prosilver/template/overall_header.html==打开 styles/prosilver/template/overall_header.html
find the line where the default search window is displayed, thats right behind the <pre><div id="search-box"></pre> statement==找到搜索框显示代码部分, 它们在 <pre><div id="search-box"></pre> 下面
Insert the following code right behind the div tag==在div标签后插入以下代码
RSS feeds can be loaded into the YaCy search index.==YaCy能够读取RSS饲料.
This does not load the rss file as such into the index but all the messages inside the RSS feeds as individual documents.==但不是直接读取RSS文件, 而是将RSS饲料中的所有信息分别当作单独的文件来读取.
The peer does not respond. It was now removed from the peer-list.==远端节点未响应, 将从节点列表中删除.
The peer <b>==peer <b>
is alive and responded:==可用:
You are allowed to send me a message==你现在可以给我发送消息
kb and an==kb和一个
attachment ≤==附件 ≤
Your Message==你的短消息
Subject:==主题:
Text:==内容:
"Enter"=="发送"
"Preview"=="预览"
You can use==你可以在这使用
Wiki Code</a> here.==Wiki Code </a>.
Preview message==预览消息
The message has not been sent yet!==短消息未发送!
The peer is alive but cannot respond. Sorry.==节点属于活动状态但是无响应.
Your message has been sent. The target peer responded:==你的短消息已发送. 接收节点返回:
The target peer is alive but did not receive your message. Sorry.==抱歉, 接收节点属于活动状态但是没有接收到你的消息.
Here is a copy of your message, so you can copy it to save it for further attempts:==这是你的消息副本, 可被保存已备用:
You cannot call this page directly. Instead, use a link on the <a href="Network.html">Network</a> page.==你不能直接使用此页面. 请使用 <a href="Network.html">网络</a> 页面的对应功能.
You can process these news with a button on the page to remove their appearance from the IndexCreate and Network page==你可以使用'创建首页'和'网络'页面的设置隐藏它们.
<strong>Processed News (#[prsize]#)</strong>: this is simply an archive of incoming news that you removed by processing.==<strong>处理的新闻(#[prsize]#)</strong>: 此页面显示你已删除的传入新闻存档.
<strong>Outgoing News (#[ousize]#)</strong>: here your can see news entries that you have created. These news are currently broadcasted to other peers.==<strong>传出的新闻(#[ousize]#)</strong>: 此页面显示你节点创建的新闻词条, 正在发布给其他节点.
<strong>Published News (#[pusize]#)</strong>: your news that have been broadcasted sufficiently or that you have removed from the broadcast list.==<strong>发布的新闻(#[pusize]#)</strong>: 显示已经完全发布出去的新闻或者从传出列表中删除的新闻.
"#(page)#::Process All News::Delete All News::Abort Publication of All News::Delete All News#(/page)#"=="#(page)#::处理所有新闻::删除所有新闻::停止发布所有新闻::删除所有新闻#(/page)#"
Memory reserved for <abbr title="Java Virtual Machine">JVM</abbr>==为<abbr title="Java Virtual Machine">JVM</abbr>保留的内存
"Set"=="设置"
#Resource Observer
Resource Observer==资源查看器
Memory state==内存状态
>proper<==>合适<
>exhausted<==>耗尽<
Reset state==重置状态
Manually reset to 'proper' state==手动设置到'合适'状态
Enough memory is available for proper operation.==有足够内存保证正常运行.
Within the last eleven minutes, at least four operations have tried to request memory that would have reduced free space within the minimum required.==在过去的11分钟内,至少有四次操作尝试请求内存,这将减少所需的最低可用空间。
Minimum required==最低要求
Amount of memory (in Mebibytes) that should at least be free for proper operation==为保证正常运行的最低内存量(以MB为单位)
The autoregulation task performs the following sequence of operations, stopping once free space disk is over the steady-state value :==自动调节任务执行以下操作序列,一旦硬盘可用空间超过稳态值就停止:
The autoregulation task performs the following sequence of operations, stopping once used space disk is below the steady-state value:==自动调节任务执行以下操作序列,一旦使用的硬盘空间低于稳态值就停止:
The crawl balancer tries to avoid that domains are==crawl平衡器能够避免频繁地访问同一域名,
accessed too often, but if the balancer fails (i.e. if there are only links left from the same domain), then these minimum==如果平衡器失效(比如相同域名下只剩链接了), 则此有此间歇
grey -> the search target hash order position(s) (more targets if a dht partition is used)<==灰色 -> 搜索目标hash序列位置(如果使用dht会产生更多目标)<
"Search event picture"=="搜索时间图况"
#-----------------------------
#File: ProxyIndexingMonitor_p.html
#---------------------------
Indexing with Proxy==代理索引
YaCy can be used to 'scrape' content from pages that pass the integrated caching HTTP proxy.==YaCy能够通过集成缓存HTTP代理进行搜索.
When scraping proxy pages then <strong>no personal or protected page is indexed</strong>;==当通过代理进行搜索时不会索引<strong>私有或者受保护页面</strong>;
# This is the control page for web pages that your peer has indexed during the current application run-time==Dies ist die Kontrollseite für Internetseiten, die Ihr Peer während der aktuellen Sitzung
# as result of proxy fetch/prefetch.==durch Besuchen einer Seite indexiert.
# No personal or protected page is indexed==Persönliche Seiten und geschütze Seiten werden nicht indexiert
those pages are detected by properties in the HTTP header (like Cookie-Use, or HTTP Authorization)==通过检测HTTP头部属性(比如cookie用途或者http认证)
this controls the proxy auto configuration script for browsers at http://localhost:8090/autoconfig.pac==这会影响浏览器代理自动配置脚本 http://localhost:8090/autoconfig.pac
.yacy-domains only==仅 .yacy 域名
whether the proxy should only be used for .yacy-Domains==代理是否只对 .yacy 域名有效.
Proxy pre-fetch setting:==代理预读设置:
this is an automated html page loading procedure that takes actual proxy-requested==这是一个自动预读网页的过程
It is almost always recommended to set this on. The only exception is that you have another caching proxy running as secondary proxy and YaCy is configured to used that proxy in proxy-proxy - mode.==推荐打开此项设置. 唯一的例外是你有另一个缓存代理作为二级代理并且YaCy设置为使用'代理到代理'模式.
Simply drag and drop the link shown below to your Browsers Toolbar/Link-Bar.==仅需拖动以下链接至浏览器工具栏/书签栏.
If you click on it while browsing, the currently viewed website will be inserted into the YaCy crawling queue for indexing.==如果在浏览网页时点击, 当前查看页面会被插入到crawl队列已用于索引
The document ranking influences the order of the search result entities.==文档排名会影响实际搜索结果的顺序。
A ranking is computed using a number of attributes from the documents that match with the search word.==排名计算使用到与搜索词匹配的文档中的多个属性。
The attributes are first normalized over all search results and then the normalized attribute is multiplied with the ranking coefficient computed from this list.==在所有搜索结果基础上,先对属性进行归一化,然后将归一化的属性与相应的排名系数相乘。
The ranking coefficient grows exponentially with the ranking levels given in the following table.==排名系数随着下表中给出的排名水平呈指数增长。
If you increase a single value by one, then the strength of the parameter doubles.==如果将单个值增加1,则参数的影响效果加倍。
</body>==<script>window.onload = function () {$("label:contains('Appearance In Emphasized Text')").text('出现在强调的文本中');$("label:contains('Appearance In URL')").text('出现在地址中'); $("label:contains('Appearance In Author')").text('出现在作者中'); $("label:contains('Appearance In Reference/Anchor Name')").text('出现在参考/锚点名称中'); $("label:contains('Appearance In Tags')").text('出现在标签中'); $("label:contains('Appearance In Title')").text('出现在标题中'); $("label:contains('Authority of Domain')").text('域名权威'); $("label:contains('Category App, Appearance')").text('类别:出现在应用中'); $("label:contains('Category Audio Appearance')").text('类别:出现在音频中'); $("label:contains('Category Image Appearance')").text('类别:出现在图片中'); $("label:contains('Category Video Appearance')").text('类别:出现在视频中'); $("label:contains('Category Index Page')").text('类别:索引页面'); $("label:contains('Date')").text('日期'); $("label:contains('Domain Length')").text('域名长度'); $("label:contains('Hit Count')").text('命中数'); $("label:contains('Preferred Language')").text('倾向的语言'); $("label:contains('Links To Local Domain')").text('本地域名链接'); $("label:contains('Links To Other Domain')").text('其他域名链接'); $("label:contains('Phrases In Text')").text('文本中短语');$("label:contains('Position In Phrase')").text('在短语中位置');$("label:contains('Position In Text')").text('在文本中位置');$("label:contains('Position Of Phrase')").text('短语的位置'); $("label:contains('Term Frequency')").text('术语频率'); $("label:contains('URL Components')").text('地址组件'); $("label:contains('Term Frequency')").text('术语频率'); $("label:contains('URL Length')").text('地址长度'); $("label:contains('Word Distance')").text('词汇距离'); $("label:contains('Words In Text')").text('文本词汇'); $("label:contains('Words In Title')").text('标题词汇');}</script></body>
There are two ranking stages: first all results are ranked using the pre-ranking and from the resulting list the documents are ranked again with a post-ranking.==有两个排名阶段:首先对搜索结果进行一次排名, 然后再对首次排名结果进行二次排名。
The two stages are separated because they need statistical information from the result of the pre-ranking.==两个结果是分开的, 因为它们都需要上次排名的统计结果。
These are ranking attributes for Solr. This ranking applies for internal and remote (P2P or shard) Solr access.==这些是 Solr 的排名属性。 此排名适用于内部和远端(P2P或分片)的Solr访问。
A Boost Function can combine numeric values from the result document to produce a number which is multiplied with the score value from the query result.==提升函数可以组合结果文档中的数值以生成一个数字,该数字与查询结果中的得分值相乘。
To see all available fields, see the <a href="IndexSchema_p.html">YaCy Solr Schema</a> and look for numeric values (these are names with suffix '_i').==要查看所有可用字段,请参阅<a href="IndexSchema_p.html">YaCy Solr架构</a>并查找数值(它们都是带有后缀“_i”的名称)。
To find out which kind of operations are possible, see the <a href="https://lucene.apache.org/solr/guide/6_6/function-queries.html" target="_blank">Solr Function Query</a> documentation.==要了解可能的操作类型,请参阅<a href="https://lucene.apache.org/solr/guide/6_6/function-queries.html" target="_blank">Solr函数查询</a>文档。
Example: to order by date, use "recip(ms(NOW,last_modified),3.16e-11,1,1)", to order by crawldepth, use "div(100,add(crawldepth_i,1))".==示例:要按日期排序,使用"recip(ms(NOW,last_modified),3.16e-11,1,1)";要按爬虫深度排序,使用"div(100,add(crawldepth_i,1))"。
You can boost with vocabularies, use the occurrence counters #[vocabulariesvoccount]# and #[vocabulariesvoclogcount]#.==你可以使用出现次数计数器#[vocabulariesvoccount]#和#[vocabulariesvoclogcount]#来提升词汇量。
The Boost Query is attached to every query. Use this to statically boost specific content in the index.==提升查询附加到每个查询。使用它来静态提升索引中的特定内容。
Example: "fuzzy_signature_unique_b:true^100000.0f" means that documents, identified as 'double' are ranked very bad and appended to the end of all results (because the unique are ranked high).==示例:“fuzzy_signature_unique_b:true^100000.0f”表示被标识为“double”的文档排名很差,并附加到所有结果的末尾(因为唯一的排名很高)。
To find appropriate fields for this query, see the <a href="IndexSchema_p.html">YaCy Solr Schema</a> and look for boolean values (with suffix '_b') or tags inside string fields (with suffix '_s' or '_sxt').==要为此查询找到适当的字段,请参阅<a href="IndexSchema_p.html">YaCy Solr架构</a>并查找布尔值(带有后缀“_b”)或字符串字段中的标签(带有后缀“_s”或“_sxt”)。
You can boost with vocabularies, use the field '#[vocabulariesfield]#' with values #[vocabulariesavailable]#. You can also boost on logarithmic occurrence counters of the fields #[vocabulariesvoclogcounts]#.==你可以使用词汇表进行提升,使用值为#[vocabulariesavailable]#的字段'#[vocabulariesfield]#'。你还可以提高字段#[vocabulariesvoclogcounts]#的对数出现计数器。
The Filter Query is attached to every query. Use this to statically add a selection criteria to reduce the set of results.==过滤器查询附加到每个查询。使用它静态添加选择标准以减少结果集。
Example: "http_unique_b:true AND www_unique_b:true" will filter out all results where urls appear also with/without http(s) and/or with/without 'www.' prefix.==示例:"http_unique_b:true AND www_unique_b:true"将过滤掉URL包含/不包含http(s) 和/或 包含/不包含“www”的结果。
To find appropriate fields for this query, see the <a href="IndexSchema_p.html">YaCy Solr Schema</a>. Warning: bad expressions here will cause that you don't have any search result!==要寻找此查询的适当字段,请参阅<a href="IndexSchema_p.html">YaCy Solr架构</a>。警告:此处的错误表达式将导致你没有任何搜索结果!
This is the set of searchable fields (see <a href="IndexSchema_p.html">YaCy Solr Schema</a>). Entries without a boost value are not searched. Boost values make hits inside the corresponding field more important.==这是一组可搜索字段(请参阅 <a href="IndexSchema_p.html">YaCy Solr架构</a>)。没有提升值的条目不会被搜索。提升值使相应字段内的命中更加重要。
This is a <a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">Java Pattern</a>==这是一种<a href="https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html" target="_blank">Java模式</a>
You can configure here limitations on access rate to this peer search interface by unauthenticated users and users without extended search right==你可以在此处配置未经验证的用户和没有扩展搜索权限的用户对该节点搜索界面的访问速率限制
(see the <a href="ConfigAccounts_p.html">Accounts</a> configuration page for details on users rights).==(有关用户权限详情请参见<a href="ConfigAccounts_p.html">账户</a>配置页面)。
YaCy search==YaCy搜索
Access rate limitations to this peer search interface.==本节点搜索界面访问率限制。
When a user with limited rights (unauthenticated or without extended search right) exceeds a limit, the search is blocked.==当具有有限权限的用户(未经验证或没有扩展搜索权限)超过限制时,搜索阻塞。
Access rate limitations to the peer-to-peer search mode.==P2P搜索模式下访问率限制。
When a user with limited rights (unauthenticated or without extended search right) exceeds a limit, the search scope falls back to only this local peer index.==当具有有限权限的用户(未经验证或没有扩展搜索权限)超过限制时,搜索范围缩小为本地索引。
Peer-to-peer search with JavaScript results resorting==带有结果排序的P2P搜索
Access rate limitations to the peer-to-peer search mode with browser-side JavaScript results resorting enabled==对启用了浏览器端JavaScript结果重新排序的P2P搜索模式的访问率限制
(check the 'Remote results resorting' section in the <a href="ConfigPortal_p.html">Search Portal</a> configuration page).==(在<a href="ConfigPortal_p.html">搜索门户</a>配置页面勾选'远端结果重排序')。
When a user with limited rights (unauthenticated or without extended search right) exceeds a limit, results resorting becomes only applicable on demand, server-side.==当具有有限权限的用户(未经验证或没有扩展搜索权限)超过限制时,搜索结果重排仅可用于服务器侧搜索。
Limitations on snippet loading from remote websites.==对从远程网站加载摘录的限制。
When a user with limited rights (unauthenticated or without extended search right) exceeds a limit, the snippets fetch strategy falls back to 'CACHEONLY'==当具有有限权限的用户(未经验证或没有扩展搜索权限)超过限制时,摘录获取策略缩小为'CACHEONLY'。
(check the default Snippet Fetch Strategy on the <a href="ConfigPortal_p.html">Search Portal</a> configuration page).==(在<a href="ConfigPortal_p.html">搜索门户</a>配置页面勾选默认的摘录获取策略)。
If you want to restore all settings to the default values,==你如果要恢复所有设置到默认值,
but <strong>forgot your administration password</strong>, you must stop the proxy,==但是<strong>忘记了管理员密码</strong>, 则你首先必须停止代理,
delete the file 'DATA/SETTINGS/yacy.conf' in the YaCy application root folder and start YaCy again.==然后删除YaCy应用根目录下的 'DATA/SETTINGS/yacy.conf' 文件,最后再次启动YaCy。
The sshd port of the host, like '22'==ssh端口, 比如'22'
Path</label>==路径</label>
The remote path on the server, like '~/yacy/seed.txt'. Missing sub-directories are NOT created automatically.==ssh服务器上传路径, 比如'~/yacy/seed.txt'. 不会自动创建缺少的子目录.
Filter have to be entered as IP, IP range or using CIDR notation separated by comma (e.g. 192.168.1.1,2001:db8::ff00:42:8329,192.168.1.10-192.168.1.20,192.168.1.30-40,192.168.2.0/24)==过滤器必须输入使用逗号分隔的IP、IP范围或CIDR符号 (比如 192.168.1.1,2001:db8::ff00:42:8329,192.168.1.10-192.168.1.20,192.168.1.30-40,192.168.2.0/24)
further details on format see Jetty==关于格式的进一步细节参见Jetty
Set this to avoid error-messages like 'proxy use not allowed / granted' on accessing your Peer by its hostname.==设置此选项可避免在通过服务器名访问对等服务器时出现‘代理使用不允许/已授权’等错误消息。
Virtual host for httpdFileServlet access for example http://FILEHOST/ shall access the file servlet and==用于 httpdFileServlet 访问的虚拟主机,
return the defaultFile at rootPath either way, http://FILEHOST/ denotes the same as http://localhost:<port>/==例如 http://FILEHOST/ 应访问文件服务器并以任一方式返回根路径下的默认文件,对预值'localpeer'而言,http://FILEHOST/ 与 http://localhost:<port>/表示相同,
for the preconfigured value 'localpeer', the URL is: http://localpeer/.==地址为:http://localpeer/。
This is the local port on the loopback address (127.0.0.1 or :1) to listen for a shutdown signal to stop the YaCy server (-1 disables the shutdown port, recommended default is 8005). A change requires a restart.==这是环地址(127.0.0.1 或:1)上的本地端口,用于侦听关闭信号以停止YaCy服务器(-1禁用关闭端口,推荐默认值为8005)。更改需要重新启动。
When checked (default), HTTP responses can be compressed using gzip.==选中时(默认),可以使用gzip压缩HTTP响应。
The requesting user-agent (a web browser, another YaCy peer or any other tool) uses the header 'Accept-Encoding' to tell whether it accepts gzip compression or not.==请求用户代理(网页浏览器、另一个YaCy节点或任何其他工具)使用标头'Accept-Encoding'来判断它是否接受gzip压缩。
This adds some processing overhead, but can significantly reduce the amount of bytes transmitted over the network.==这增加了一些处理开销,但可以显着减少通过网络传输的字节量。
With this settings you can activate or deactivate URL proxy.==With this settings you can activate or deactivate URL proxy.
Service call: ==Service call:
, where parameter is the url of an external web page.==, where parameter is the url of an external web page.
>URL proxy:<==>URL proxy:<
>Enabled<==>开启<
Globally enables or disables URL proxy via ==Globally enables or disables URL proxy via
Show search results via URL proxy:==Show search results via URL proxy:
Enables or disables URL proxy for all search results. If enabled, all search results will be tunneled through URL proxy.==Enables or disables URL proxy for all search results. If enabled, all search results will be tunneled through URL proxy.
Alternatively you may add this javascript to your browser favorites/short-cuts, which will reload the current browser address==Alternatively you may add this javascript to your browser favorites/short-cuts, which will reload the current browser address
via the YaCy proxy servlet.==via the YaCy proxy servlet.
or right-click this link and add to favorites:==or right-click this link and add to favorites:
Define URL substitution rules which allow navigating in proxy environment. Possible values: all, domainlist. Default: domainlist.==Define URL substitution rules which allow navigating in proxy environment. Possible values: all, domainlist. Default: domainlist.
The password redundancy check failed. You have probably misstyped your password.==密码冗余检查错误.
Shutting down.</strong><br />Application will terminate after working off all crawling tasks.==正在关闭</strong><br />所有crawl任务完成后程序会关闭.
Your administration account setting has been made.==已创建管理账户设置.
Your new administration account name is #[user]#. The password has been accepted.<br />If you go back to the Settings page, you must log-in again.==新帐户名是 #[user]#. 密码输入正确.<br />如果返回设置页面, 需要再次输入密码.
Your proxy access setting has been changed.==代理访问设置已改变.
Your proxy account check has been disabled, since you did not supply a password.==不能进行代理账户检查, 密码不正确.
The new proxy IP filter is set to==代理IP过滤设置为
The proxy port is:==代理端口号:
Port rebinding will be done in a few seconds.==端口在几秒后绑定完成.
You can reach your YaCy server under the new location==可以通过新位置访问YaCy服务器:
Your proxy access setting has been changed.==代理访问设置已改变.
Your server access filter is now set to==服务器访问过滤为
Auto pop-up of the Status page is now <strong>disabled</strong>==自动弹出状态页面<strong>关闭.</strong>
Auto pop-up of the Status page is now <strong>enabled</strong>==自动弹出状态页面<strong>打开.</strong>
You are now permanently <strong>online</strong>.==你现在处于永久<strong>在线状态</strong>.
After a short while you should see the effect on the====一会儿可以在
status</a> page.==Status</a> 页面看到变化.
The Peer Name is:==节点名:
Your static Ip(or DynDns) is:==静态IP(或DynDns)为:
Seed Settings changed.#(success)#::You are now a principal peer.==seed设置已改变.#(success)#::本地节点已成为主要节点.
Seed Settings changed, but something is wrong.==seed设置已改变, 但是未完全成功.
Seed Uploading was deactivated automatically.==seed上传自动关闭.
Please return to the settings page and modify the data.==请返回设置页面修改参数.
The remote-proxy setting has been changed==远端代理设置已改变.
The new setting is effective immediately, you don't need to re-start.==新设置立即生效.
The submitted peer name is already used by another peer. Please choose a different name.</strong> The Peer name has not been changed.==提交的节点名已存在, 请更改.</strong> 节点名未改变.
Your Peer Language is:==节点语言:
The submitted peer name is not well-formed. Please choose a different name.</strong> The Peer name has not been changed.
Peer names must not contain characters other than (a-z, A-Z, 0-9, '-', '_') and must not be longer than 80 characters.
#The new parser settings where changed successfully.==Die neuen Parser Einstellungen wurden erfolgreich gespeichert.
Parsing of the following mime-types was enabled:
Seed Upload method was changed successfully.==seed上传方式改变成功.
You are now a principal peer.==本地节点已成为主要节点.
Seed Upload Method:==seed上传方式:
Seed File URL:==seed文件URL:
Your proxy networking settings have been changed.==代理网络设置已改变.
Transparent Proxy Support is:==透明代理支持:
Connection Keep-Alive Support is:==连接保持支持:
Your message forwarding settings have been changed.==消息发送设置已改变.
Message Forwarding Support is:==消息发送支持:
Message Forwarding Command:==消息:
Recipient Address:==收件人地址:
Please return to the settings page and modify the data.==请返回设置页面修改参数.
You are now <strong>event-based online</strong>.==你现在处于<strong>事件驱动在线</strong>.
After a short while you should see the effect on the==查看变化
You are now in <strong>Cache Mode</strong>.==你现在处于<strong>Cache模式</strong>.
Only Proxy-cache ist available in this mode.==此模式下仅代理缓存可用.
After a short while you should see the effect on the==查看变化
You can now go back to the==现在可返回
Settings</a> page if you want to make more changes.==设置</a> 页面, 如果需要更改更多参数的话.
You can reach your YaCy server under the new location==现在可以通过新位置访问YaCy服务器:
If you do this without YaCy running on a desktop-pc or without Java 6 installed, this will possibly break startup.==如果你不是在台式机上或者已安装Java6的机器上运行, 可能会破坏开机程序.
In this case, you will have to edit the configuration manually in DATA/SETTINGS/yacy.conf==在此情况下, 你需要手动修改配置文件 DATA/SETTINGS/yacy.conf
Log-in as administrator to see full status==登录管理用户以查看完整状态
Welcome to YaCy!==欢迎使用YaCy!
Your settings are _not_ protected!</strong>==你的设置 _未_ 受保护!</strong>
Please open the <a href="ConfigAccounts_p.html">accounts configuration</a> page <strong>immediately</strong>==请打开<a href="ConfigAccounts_p.html">账户设置</a> <strong>页面</strong>
and set an administration password.==并设置管理密码.
Access is unrestricted from localhost (this includes administration features).==访问权限在localhost不受限制(这包括管理功能)。
Please check the <a href="ConfigAccounts_p.html">accounts configuration</a> page to ensure that the settings match the security level you need.==请检查<a href="ConfigAccounts_p.html">帐户配置</a>页面,确保设置符合你所需的安全级别。
You can download a more recent version of YaCy. Click here to install this update and restart YaCy:==你可以下载最新版本YaCy, 点此进行升级并重启:
Install YaCy==安装YaCy
You can download the latest releases here:==你可以在此处下载最新版本:
You are running a server in senior mode and you support the global internet index,==服务器运行在高级模式, 并支持全球索引,
which you can also <a href="index.html">search yourself</a>.==你也能进行<a href="index.html">本地搜索</a>.
You have a principal peer because you publish your seed-list to a public accessible server==你是一个骨干节点, 因为你向公共服务器公布了你的种子列表,
where it can be retrieved using the URL==可使用此URL进行接收:
Your Web Page Indexer is idle. You can start your own web crawl <a href="CrawlStartSite.html">here</a>==网页索引器当前空闲. 可以点击<a href="CrawlStartSite.html">这里</a>开始爬取网页
Your Web Page Indexer is busy. You can <a href="Crawler_p.html">monitor your web crawl</a> here==网页索引器当前忙碌. 点击<a href="Crawler_p.html">这里</a>查看状态
If you need professional support, please write to==如果你需要专业级支持, 请EMAIL来信
For community support, please visit our==如果只是社区支持, 请访问我们的
Please go to the <a href="ConfigAccounts_p.html">User Administration</a> page and set an administration password.==请在<a href="ConfigAccounts_p.html">用户管理</a>页面设置管理密码.
You don't have the correct access right to perform this task.==无执行此任务权限.
You can now go back to the <a href="Settings_p.html">Settings</a> page if you want to make more changes.==你现在可以返回<a href="Settings_p.html">设置</a>页面进行详细设置.
Application will terminate after working off all scheduled tasks.==程序在所有任务完成后将停止.
Please send us feed-back!==可以给我们一个反馈嘛!
We don't track YaCy users, YaCy does not send 'home-pings', we do not even know how many people use YaCy as their private search engine.==我们不跟踪YAY用户,YaCy不发送“回家Ping”,我们甚至不知道有多少人使用Yyas作为他们的私人搜索引擎。
Therefore we like to ask you: do you like YaCy?==所以我们想问你:你喜欢YaCy吗?
Will you use it again... if not, why?==你会再次使用它吗?如果不是,为什么?
If you are a professional user and you would like to use YaCy in your company in combination with consulting services by YaCy specialists, please see==如果你是专业用户,并且希望在公司中使用YaCy并获得YaCy专家的咨询服务,请参阅
The information that is presented on this page can also be retrieved as XML.==The information that is presented on this page can also be retrieved as XML.
Click the API icon to see the XML.==Click the API icon to see the XML.
To see a list of all APIs, please visit the ==To see a list of all APIs, please visit the
To see a list of all APIs, please visit the <a href="https://wiki.yacy.net/index.php/Dev:API">API wiki page</a>.==查看所有API, 请访问<a href="https://wiki.yacy.net/index.php/Dev:API">API Wiki</a>.
Vocabularies can be used to produce a search navigation.==词汇表可用于生成搜索导航.
A vocabulary must be created before content is indexed.==必须在索引内容之前创建词汇.
The vocabulary is used to annotate the indexed content with a reference to the object that is denoted by the term of the vocabulary.==词汇用于通过引用由词汇的术语表示的对象来注释索引的内容.
The object can be denoted by a url stub that, combined with the term, becomes the url for the object.==该对象可以用地址存根表示,该存根与该术语一起成为该对象的地址.
>Vocabulary Selection<==>词汇选择<
>Vocabulary Name<==>词汇名<
"View"=="查看"
>Vocabulary Production<==>词汇生成<
Empty Vocabulary== 空词汇
>Auto-Discover<==>自动发现<
> from file name==> 来自文件名
> from page title (splitted)==> 来自页面标题(拆分)
> from page title==> 来自页面标题
> from page author==> 来自页面作者
>Objectspace<==>对象空间<
It is possible to produce a vocabulary out of the existing search index.==可以从现有搜索索引中生成词汇表.
The data that is visualized here can also be retrieved in a XML file, which lists the reference relation between the domains.==此页面数据显示域之间的关联关系, 能以XML文件形式查看.
With a GET-property 'about' you get only reference relations about the host that you give in the argument field for 'about'.==使用GET属性'about'仅能获得带有'about'参数的域关联关系.
With a GET-property 'latest' you get a list of references that had been computed during the current run-time of YaCy, and with each next call only an update to the next list of references.==使用GET属性'latest'能获得当前的关联关系列表, 并且每一次调用都只能更新下一级关联关系列表.
Click the API icon to see the XML file.==点击API图标查看XML文件.
To see a list of all APIs, please visit the <a href="https://wiki.yacy.net/index.php/Dev:API">API wiki page</a>.==查看所有API, 请访问<a href="https://wiki.yacy.net/index.php/Dev:API">API Wiki</a>.
Text will be displayed <span class="underline">underlined</span>.==文本要显示<span class =“underline”>下划线</ span>.
Code==代码
This tag displays a Youtube or Vimeo video with the id specified and fixed width 425 pixels and height 350 pixels.==这个标签显示一个425像素和350像素的Youtube或Vimeo视频.
i.e. use==比如用
Wiki Help==Wiki帮助
Wiki-Code==Wiki代码
This table contains a short description of the tags that can be used in the Wiki and several other servlets==此表列出了用于Wiki和几个插件代码标签简述,
of YaCy. For a more detailed description visit the==详情请见
#YaCy Wiki==YaCy Wiki
Description==描述
#=headline===headline
These tags create headlines. If a page has three or more headlines, a table of content will be created automatically.==此标记标识标题内容. 如果页面有多于三个标题, 则会自动创建一个表格.
Headlines of level 1 will be ignored in the table of content.==一级标题.
#text==Text
These tags create stressed texts. The first pair emphasizes the text (most browsers will display it in italics),==这些标记标识文本内容. 第一对中为强调内容(多数浏览器用斜体表示),
the second one emphazises it more strongly (i.e. bold) and the last tags create a combination of both.==第二对用粗体表示, 第三对为两者的联合.
Text will be displayed <span class="strike">stricken through</span>.==文本内容以<span class="strike">删除线</span>表示.
Lines will be indented. This tag is supposed to mark citations, but may as well be used for styling purposes.==缩进内容, 此标记主要用于引用, 也能用于标识样式.
#point==point
These tags create a numbered list.==此标记用于有序列表.
#something<==something<
#another thing==another thing
#and yet another==and yet another
#something else==something else
These tags create an unnumbered list.==用于创建无序列表.
#word==word
#:definition==:definition
These tags create a definition list.==用于创建定义列表.
This tag creates a horizontal line.==创建水平线.
#pagename==pagename
#description]]==description]]
This tag creates links to other pages of the wiki.==创建到其他wiki页面的链接.
This tag displays an image, it can be aligned left, right or center.==显示图片, 可设置左对齐, 右对齐和居中.
These tags create a table, whereas the first marks the beginning of the table, the second starts==用于创建表格, 第一个标记为表格开头, 第二个为换行,
a new line, the third and fourth each create a new cell in the line. The last displayed tag==第三个与第四个创建列.
closes the table.==最后一个为表格结尾.
#The escape tags will cause all tags in the text between the starting and the closing tag to not be treated as wiki-code.==Durch diesen Tag wird der Text, der zwischen den Klammern steht, nicht interpretiert und unformatiert als normaler Text ausgegeben.
A text between these tags will keep all the spaces and linebreaks in it. Great for ASCII-art and program code.==此标记之间的文本会保留所有空格和换行, 主要用于ASCII艺术图片和编程代码.
If a line starts with a space, it will be displayed in a non-proportional font.==如果一行以空格开头, 则会以非比例形式显示.
url description==URL描述
This tag creates links to external websites.==此标记创建外部网站链接.
This search result can also be retrieved as RSS/<a href="http://www.opensearch.org" target="_blank">opensearch</a> output.==此搜索结果能以RSS/<a href="http://www.opensearch.org" target="_blank">opensearch</a>被形式检索。
The query format is similar to <a href="http://www.loc.gov/standards/sru/" target="_blank">SRU</a>.==请求的格式与<a href="http://www.loc.gov/standards/sru/" target="_blank">SRU</a>相似。
Click the API icon to see an example call to the search rss API.==点击API图标查看调用rss API的示例。
To see a list of all APIs, please visit the <a href="https://wiki.yacy.net/index.php/Dev:API" target="_blank">API wiki page</a>.==查看所有API, 请访问<a href="https://wiki.yacy.net/index.php/Dev:API" target="_blank">API百科页面</a>。
To see a list of all APIs, please visit the <a href="https://wiki.yacy.net/index.php/Dev:API" target="_blank">API wiki page</a>.==要查看所有 API 的列表,请访问<a href="https://wiki.yacy.net/index.php/Dev:API" target="_blank">API wiki</a>页面。
# Do not translate id="search" and rel="search" which only have technical html semantics
Search Page==搜索页面
This search result can also be retrieved as RSS/<a href="http://www.opensearch.org" target="_blank">opensearch</a> output.==此搜索结果能以RSS/<a href="http://www.opensearch.org" target="_blank">opensearch</a>形式表示.
In Mozilla Firefox, you can the Search-Plugin via the search box on the toolbar.<br />In Mozilla (Seamonkey) you can access the Search-Plugin via the Sidebar or the Location Bar.==在Mozilla Firefox中,你可以通过工具栏上的搜索框打开搜索插件。<br />在Mozilla(Seamonkey)中,你可以通过侧栏或位置栏访问搜索插件。
show search results for "#[query]#" on map==在地图上显示 "#[query]#" 的搜索结果
Your search is done using peers in the YaCy P2P network.==你的搜索是靠YaCy P2P网络中的节点完成的。
You can switch to 'Stealth Mode' which will switch off P2P, giving you full privacy. Expect less results then, because then only your own search index is used.==你可以切换到'隐形模式',这将关闭P2P,给你完全的隐私。期待较少的结果,因为那时只有你自己的搜索索引被使用。
Your search is done using only your own peer, locally.==你的搜索是靠在本地的YaCy节点完成的。
You can switch to 'Peer-to-Peer Mode' which will cause that your search is done using the other peers in the YaCy network.==你可以切换到'P2P',这将让你的搜索使用YaCy网络中的YaCy节点。