From af63aa1d0e4e95ac36c77d11d023e2ec95b20b12 Mon Sep 17 00:00:00 2001 From: orbiter Date: Tue, 31 May 2011 13:33:04 +0000 Subject: [PATCH] added fresh links to java regular expression api-doc git-svn-id: https://svn.berlios.de/svnroot/repos/yacy/trunk@7763 6c8d7289-2bf4-0310-a012-ef5d649a1542 --- htroot/CrawlStartExpert_p.html | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/htroot/CrawlStartExpert_p.html b/htroot/CrawlStartExpert_p.html index d80b2076d..5000e0fd8 100644 --- a/htroot/CrawlStartExpert_p.html +++ b/htroot/CrawlStartExpert_p.html @@ -144,8 +144,8 @@ Restrict to sub-path - The filter is a regular expression - that must match with the URLs which are used to be crawled; default is 'catch all'. + The filter is a regular expression + that must match with the URLs which are used to be crawled; default is 'catch all'. Example: to allow only urls that contain the word 'science', set the filter to '.*science.*'. You can also use an automatic domain-restriction to fully crawl a single domain. @@ -156,7 +156,8 @@ - This filter must not match to allow that the page is accepted for crawling. + The filter is a regular expression + that must not match to allow that the page is accepted for crawling. The empty string is a never-match filter which should do well for most cases. If you don't know what this means, please leave this field empty.