@ -75,9 +75,13 @@ import de.anomic.http.client.Cache;
import de.anomic.server.serverObjects ;
import de.anomic.server.serverSwitch ;
public class IndexControlRWIs_p {
public class IndexControlRWIs_p
{
public static serverObjects respond ( final RequestHeader header , final serverObjects post , final serverSwitch env ) throws IOException {
public static serverObjects respond (
final RequestHeader header ,
final serverObjects post ,
final serverSwitch env ) throws IOException {
// return variable that accumulates replacements
final Switchboard sb = ( Switchboard ) env ;
final serverObjects prop = new serverObjects ( ) ;
@ -87,7 +91,8 @@ public class IndexControlRWIs_p {
prop . put ( "keyhash" , "" ) ;
prop . put ( "result" , "" ) ;
prop . put ( "cleanup" , post = = null | | post . containsKey ( "maxReferencesLimit" ) ? 1 : 0 ) ;
prop . put ( "cleanup_solr" , sb . indexSegments . segment ( Segments . Process . LOCALCRAWLING ) . getSolr ( ) = = null | | ! sb . getConfigBool ( "federated.service.solr.indexing.enabled" , false ) ? 0 : 1 ) ;
prop . put ( "cleanup_solr" , sb . indexSegments . segment ( Segments . Process . LOCALCRAWLING ) . getSolr ( ) = = null
| | ! sb . getConfigBool ( "federated.service.solr.indexing.enabled" , false ) ? 0 : 1 ) ;
String segmentName = sb . getConfig ( SwitchboardConstants . SEGMENT_PUBLIC , "default" ) ;
int i = 0 ;
@ -126,8 +131,20 @@ public class IndexControlRWIs_p {
// read values from checkboxes
final String [ ] urls = post . getAll ( "urlhx.*" ) ;
HandleSet urlb = new HandleSet ( URIMetadataRow . rowdef . primaryKeyLength , URIMetadataRow . rowdef . objectOrder , urls . length ) ;
if ( urls ! = null ) for ( final String s : urls ) try { urlb . put ( s . getBytes ( ) ) ; } catch ( final RowSpaceExceededException e ) { Log . logException ( e ) ; }
HandleSet urlb =
new HandleSet (
URIMetadataRow . rowdef . primaryKeyLength ,
URIMetadataRow . rowdef . objectOrder ,
urls . length ) ;
if ( urls ! = null ) {
for ( final String s : urls ) {
try {
urlb . put ( s . getBytes ( ) ) ;
} catch ( final RowSpaceExceededException e ) {
Log . logException ( e ) ;
}
}
}
final boolean delurl = post . containsKey ( "delurl" ) ;
final boolean delurlref = post . containsKey ( "delurlref" ) ;
@ -157,11 +174,14 @@ public class IndexControlRWIs_p {
if ( post . get ( "deleteIndex" , "" ) . equals ( "on" ) ) {
segment . clear ( ) ;
}
if ( post . get ( "deleteSolr" , "" ) . equals ( "on" ) & & sb . getConfigBool ( "federated.service.solr.indexing.enabled" , false ) ) try {
if ( post . get ( "deleteSolr" , "" ) . equals ( "on" )
& & sb . getConfigBool ( "federated.service.solr.indexing.enabled" , false ) ) {
try {
sb . indexSegments . segment ( Segments . Process . LOCALCRAWLING ) . getSolr ( ) . clear ( ) ;
} catch ( final Exception e ) {
Log . logException ( e ) ;
}
}
if ( post . get ( "deleteCrawlQueues" , "" ) . equals ( "on" ) ) {
sb . crawlQueues . clear ( ) ;
sb . crawlStacker . clear ( ) ;
@ -190,14 +210,25 @@ public class IndexControlRWIs_p {
}
// delete word
if ( post . containsKey ( "keyhashdeleteall" ) ) try {
if ( post . containsKey ( "keyhashdeleteall" ) ) {
try {
if ( delurl | | delurlref ) {
// generate urlx: an array of url hashes to be deleted
ReferenceContainer < WordReference > index = null ;
index = segment . termIndex ( ) . get ( keyhash , null ) ;
final Iterator < WordReference > en = index . entries ( ) ;
urlb = new HandleSet ( URIMetadataRow . rowdef . primaryKeyLength , URIMetadataRow . rowdef . objectOrder , index . size ( ) ) ;
while ( en . hasNext ( ) ) try { urlb . put ( en . next ( ) . urlhash ( ) ) ; } catch ( final RowSpaceExceededException e ) { Log . logException ( e ) ; }
urlb =
new HandleSet (
URIMetadataRow . rowdef . primaryKeyLength ,
URIMetadataRow . rowdef . objectOrder ,
index . size ( ) ) ;
while ( en . hasNext ( ) ) {
try {
urlb . put ( en . next ( ) . urlhash ( ) ) ;
} catch ( final RowSpaceExceededException e ) {
Log . logException ( e ) ;
}
}
index = null ;
}
if ( delurlref ) {
@ -207,24 +238,40 @@ public class IndexControlRWIs_p {
segment . termIndex ( ) . delete ( keyhash ) ;
// now delete all urls if demanded
if ( delurl | | delurlref ) {
for ( final byte [ ] b : urlb ) sb . urlRemove ( segment , b ) ;
for ( final byte [ ] b : urlb ) {
sb . urlRemove ( segment , b ) ;
}
}
post . remove ( "keyhashdeleteall" ) ;
post . put ( "urllist" , "generated" ) ;
} catch ( final IOException e ) {
Log . logException ( e ) ;
}
}
// delete selected URLs
if ( post . containsKey ( "keyhashdelete" ) ) try {
if ( post . containsKey ( "keyhashdelete" ) ) {
try {
if ( delurlref ) {
segment . removeAllUrlReferences ( urlb , sb . loader , CacheStrategy . IFEXIST ) ;
}
if ( delurl | | delurlref ) {
for ( final byte [ ] b : urlb ) sb . urlRemove ( segment , b ) ;
for ( final byte [ ] b : urlb ) {
sb . urlRemove ( segment , b ) ;
}
}
final HandleSet urlHashes =
new HandleSet (
URIMetadataRow . rowdef . primaryKeyLength ,
URIMetadataRow . rowdef . objectOrder ,
0 ) ;
for ( final byte [ ] b : urlb ) {
try {
urlHashes . put ( b ) ;
} catch ( final RowSpaceExceededException e ) {
Log . logException ( e ) ;
}
}
final HandleSet urlHashes = new HandleSet ( URIMetadataRow . rowdef . primaryKeyLength , URIMetadataRow . rowdef . objectOrder , 0 ) ;
for ( final byte [ ] b : urlb ) try { urlHashes . put ( b ) ; } catch ( final RowSpaceExceededException e ) { Log . logException ( e ) ; }
segment . termIndex ( ) . remove ( keyhash , urlHashes ) ;
// this shall lead to a presentation of the list; so handle that the remaining program
// thinks that it was called for a list presentation
@ -233,6 +280,7 @@ public class IndexControlRWIs_p {
} catch ( final IOException e ) {
Log . logException ( e ) ;
}
}
if ( post . containsKey ( "urllist" ) ) {
if ( keystring . length ( ) = = 0 | | ! ByteBuffer . equals ( Word . word2hash ( keystring ) , keyhash ) ) {
@ -245,7 +293,8 @@ public class IndexControlRWIs_p {
}
// transfer to other peer
if ( post . containsKey ( "keyhashtransfer" ) ) try {
if ( post . containsKey ( "keyhashtransfer" ) ) {
try {
if ( keystring . length ( ) = = 0 | | ! ByteBuffer . equals ( Word . word2hash ( keystring ) , keyhash ) ) {
prop . put ( "keystring" , "<not possible to compute word from hash>" ) ;
}
@ -272,8 +321,13 @@ public class IndexControlRWIs_p {
index = segment . termIndex ( ) . get ( keyhash , null ) ;
// built urlCache
final Iterator < WordReference > urlIter = index . entries ( ) ;
final TreeMap < byte [ ] , URIMetadataRow > knownURLs = new TreeMap < byte [ ] , URIMetadataRow > ( Base64Order . enhancedCoder ) ;
final HandleSet unknownURLEntries = new HandleSet ( WordReferenceRow . urlEntryRow . primaryKeyLength , WordReferenceRow . urlEntryRow . objectOrder , index . size ( ) ) ;
final TreeMap < byte [ ] , URIMetadataRow > knownURLs =
new TreeMap < byte [ ] , URIMetadataRow > ( Base64Order . enhancedCoder ) ;
final HandleSet unknownURLEntries =
new HandleSet (
WordReferenceRow . urlEntryRow . primaryKeyLength ,
WordReferenceRow . urlEntryRow . objectOrder ,
index . size ( ) ) ;
Reference iEntry ;
URIMetadataRow lurl ;
while ( urlIter . hasNext ( ) ) {
@ -292,7 +346,11 @@ public class IndexControlRWIs_p {
}
// make an indexContainerCache
final ReferenceContainerCache < WordReference > icc = new ReferenceContainerCache < WordReference > ( Segment . wordReferenceFactory , Segment . wordOrder , Word . commonHashLength ) ;
final ReferenceContainerCache < WordReference > icc =
new ReferenceContainerCache < WordReference > (
Segment . wordReferenceFactory ,
Segment . wordOrder ,
Word . commonHashLength ) ;
try {
icc . add ( index ) ;
} catch ( final RowSpaceExceededException e ) {
@ -302,28 +360,33 @@ public class IndexControlRWIs_p {
// transport to other peer
final boolean gzipBody = sb . getConfigBool ( "indexControl.gzipBody" , false ) ;
final int timeout = ( int ) sb . getConfigLong ( "indexControl.timeout" , 60000 ) ;
final String error = Protocol . transferIndex (
seed ,
icc ,
knownURLs ,
gzipBody ,
timeout ) ;
prop . put ( "result" , ( error = = null ) ? ( "Successfully transferred " + knownURLs . size ( ) + " words in " + ( ( System . currentTimeMillis ( ) - starttime ) / 1000 ) + " seconds, " + unknownURLEntries . size ( ) + " URL not found" ) : "error: " + error ) ;
final String error = Protocol . transferIndex ( seed , icc , knownURLs , gzipBody , timeout ) ;
prop . put ( "result" , ( error = = null ) ? ( "Successfully transferred "
+ knownURLs . size ( )
+ " words in "
+ ( ( System . currentTimeMillis ( ) - starttime ) / 1000 )
+ " seconds, "
+ unknownURLEntries . size ( ) + " URL not found" ) : "error: " + error ) ;
index = null ;
} catch ( final IOException e ) {
Log . logException ( e ) ;
}
}
// generate list
if ( post . containsKey ( "keyhashsimilar" ) ) try {
final Iterator < ReferenceContainer < WordReference > > containerIt = segment . termIndex ( ) . referenceContainer ( keyhash , true , 256 , false ) . iterator ( ) ;
if ( post . containsKey ( "keyhashsimilar" ) ) {
try {
final Iterator < ReferenceContainer < WordReference > > containerIt =
segment . termIndex ( ) . referenceContainer ( keyhash , true , 256 , false ) . iterator ( ) ;
ReferenceContainer < WordReference > container ;
i = 0 ;
int rows = 0 , cols = 0 ;
prop . put ( "keyhashsimilar" , "1" ) ;
while ( containerIt . hasNext ( ) & & i < 256 ) {
container = containerIt . next ( ) ;
prop . put ( "keyhashsimilar_rows_" + rows + "_cols_" + cols + "_wordHash" , container . getTermHash ( ) ) ;
prop . put (
"keyhashsimilar_rows_" + rows + "_cols_" + cols + "_wordHash" ,
container . getTermHash ( ) ) ;
cols + + ;
if ( cols = = 8 ) {
prop . put ( "keyhashsimilar_rows_" + rows + "_cols" , cols ) ;
@ -338,15 +401,22 @@ public class IndexControlRWIs_p {
} catch ( final IOException e ) {
Log . logException ( e ) ;
}
}
if ( post . containsKey ( "blacklist" ) ) {
final String blacklist = post . get ( "blacklist" , "" ) ;
final HandleSet urlHashes = new HandleSet ( URIMetadataRow . rowdef . primaryKeyLength , URIMetadataRow . rowdef . objectOrder , urlb . size ( ) ) ;
final HandleSet urlHashes =
new HandleSet (
URIMetadataRow . rowdef . primaryKeyLength ,
URIMetadataRow . rowdef . objectOrder ,
urlb . size ( ) ) ;
if ( post . containsKey ( "blacklisturls" ) ) {
PrintWriter pw ;
try {
final String [ ] supportedBlacklistTypes = env . getConfig ( "BlackLists.types" , "" ) . split ( "," ) ;
pw = new PrintWriter ( new FileWriter ( new File ( ListManager . listsPath , blacklist ) , true ) ) ;
final String [ ] supportedBlacklistTypes =
env . getConfig ( "BlackLists.types" , "" ) . split ( "," ) ;
pw =
new PrintWriter ( new FileWriter ( new File ( ListManager . listsPath , blacklist ) , true ) ) ;
DigestURI url ;
for ( final byte [ ] b : urlb ) {
try {
@ -360,7 +430,9 @@ public class IndexControlRWIs_p {
url = e . metadata ( ) . url ( ) ;
pw . println ( url . getHost ( ) + "/" + url . getFile ( ) ) ;
for ( final String supportedBlacklistType : supportedBlacklistTypes ) {
if ( ListManager . listSetContains ( supportedBlacklistType + ".BlackLists" , blacklist ) ) {
if ( ListManager . listSetContains (
supportedBlacklistType + ".BlackLists" ,
blacklist ) ) {
Switchboard . urlBlacklist . add (
supportedBlacklistType ,
url . getHost ( ) ,
@ -379,7 +451,8 @@ public class IndexControlRWIs_p {
PrintWriter pw ;
try {
final String [ ] supportedBlacklistTypes = Blacklist . BLACKLIST_TYPES_STRING . split ( "," ) ;
pw = new PrintWriter ( new FileWriter ( new File ( ListManager . listsPath , blacklist ) , true ) ) ;
pw =
new PrintWriter ( new FileWriter ( new File ( ListManager . listsPath , blacklist ) , true ) ) ;
DigestURI url ;
for ( final byte [ ] b : urlb ) {
try {
@ -393,10 +466,13 @@ public class IndexControlRWIs_p {
url = e . metadata ( ) . url ( ) ;
pw . println ( url . getHost ( ) + "/.*" ) ;
for ( final String supportedBlacklistType : supportedBlacklistTypes ) {
if ( ListManager . listSetContains ( supportedBlacklistType + ".BlackLists" , blacklist ) ) {
if ( ListManager . listSetContains (
supportedBlacklistType + ".BlackLists" ,
blacklist ) ) {
Switchboard . urlBlacklist . add (
supportedBlacklistType ,
url . getHost ( ) , ".*" ) ;
url . getHost ( ) ,
".*" ) ;
}
}
}
@ -412,20 +488,29 @@ public class IndexControlRWIs_p {
}
}
if ( prop . getInt ( "searchresult" , 0 ) = = 3 ) listHosts ( prop , keyhash , sb ) ;
if ( prop . getInt ( "searchresult" , 0 ) = = 3 ) {
listHosts ( prop , keyhash , sb ) ;
}
}
// insert constants
prop . putNum ( "wcount" , segment . termIndex ( ) . sizesMax ( ) ) ;
prop . put ( "cleanup_maxReferencesRadioChecked" , ReferenceContainer . maxReferences > 0 ? 1 : 0 ) ;
prop . put ( "cleanup_maxReferences" , ReferenceContainer . maxReferences > 0 ? ReferenceContainer . maxReferences : 100000 ) ;
prop . put ( "cleanup_maxReferences" , ReferenceContainer . maxReferences > 0
? ReferenceContainer . maxReferences
: 100000 ) ;
// return rewrite properties
return prop ;
}
public static void genURLList ( final serverObjects prop , final byte [ ] keyhash , final String keystring , final RWIProcess ranked , final Bitfield flags , final int maxlines ) {
public static void genURLList (
final serverObjects prop ,
final byte [ ] keyhash ,
final String keystring ,
final RWIProcess ranked ,
final Bitfield flags ,
final int maxlines ) {
// search for a word hash and generate a list of url links
final String keyhashs = ASCII . String ( keyhash ) ;
prop . put ( "genUrlList_keyHash" , keyhashs ) ;
@ -445,27 +530,48 @@ public class IndexControlRWIs_p {
String us ;
long rn = - 1 ;
while ( ! ranked . isEmpty ( ) & & ( entry = ranked . takeURL ( false , 1000 ) ) ! = null ) {
if ( ( entry = = null ) | | ( entry . metadata ( ) = = null ) ) continue ;
if ( ( entry = = null ) | | ( entry . metadata ( ) = = null ) ) {
continue ;
}
url = entry . metadata ( ) . url ( ) ;
if ( url = = null ) continue ;
if ( url = = null ) {
continue ;
}
us = url . toNormalform ( false , false ) ;
if ( rn = = - 1 ) rn = entry . ranking ( ) ;
if ( rn = = - 1 ) {
rn = entry . ranking ( ) ;
}
prop . put ( "genUrlList_urlList_" + i + "_urlExists" , "1" ) ;
prop . put ( "genUrlList_urlList_" + i + "_urlExists_urlhxCount" , i ) ;
prop . putHTML ( "genUrlList_urlList_" + i + "_urlExists_urlhxValue" , entry . word ( ) . urlhash ( ) ) ;
prop . putHTML ( "genUrlList_urlList_" + i + "_urlExists_keyString" , keystring ) ;
prop . put ( "genUrlList_urlList_" + i + "_urlExists_keyHash" , keyhashs ) ;
prop . putHTML ( "genUrlList_urlList_" + i + "_urlExists_urlString" , us ) ;
prop . put ( "genUrlList_urlList_" + i + "_urlExists_urlStringShort" , ( us . length ( ) > 40 ) ? ( us . substring ( 0 , 20 ) + "<br>" + us . substring ( 20 , 40 ) + "..." ) : ( ( us . length ( ) > 30 ) ? ( us . substring ( 0 , 20 ) + "<br>" + us . substring ( 20 ) ) : us ) ) ;
prop . put (
"genUrlList_urlList_" + i + "_urlExists_urlStringShort" ,
( us . length ( ) > 40 ) ? ( us . substring ( 0 , 20 ) + "<br>" + us . substring ( 20 , 40 ) + "..." ) : ( ( us
. length ( ) > 30 ) ? ( us . substring ( 0 , 20 ) + "<br>" + us . substring ( 20 ) ) : us ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_ranking" , ( entry . ranking ( ) - rn ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_domlength" , DigestURI . domLengthEstimation ( entry . hash ( ) ) ) ;
prop . putNum (
"genUrlList_urlList_" + i + "_urlExists_domlength" ,
DigestURI . domLengthEstimation ( entry . hash ( ) ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_ybr" , BlockRank . ranking ( entry . hash ( ) ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_tf" , 1000.0 * entry . word ( ) . termFrequency ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_authority" , ( ranked . getOrder ( ) = = null ) ? - 1 : ranked . getOrder ( ) . authority ( ASCII . String ( entry . hash ( ) , 6 , 6 ) ) ) ;
prop . put ( "genUrlList_urlList_" + i + "_urlExists_date" , GenericFormatter . SHORT_DAY_FORMATTER . format ( new Date ( entry . word ( ) . lastModified ( ) ) ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_wordsintitle" , entry . word ( ) . wordsintitle ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_tf" , 1000.0 * entry
. word ( )
. termFrequency ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_authority" , ( ranked . getOrder ( ) = = null )
? - 1
: ranked . getOrder ( ) . authority ( ASCII . String ( entry . hash ( ) , 6 , 6 ) ) ) ;
prop . put (
"genUrlList_urlList_" + i + "_urlExists_date" ,
GenericFormatter . SHORT_DAY_FORMATTER . format ( new Date ( entry . word ( ) . lastModified ( ) ) ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_wordsintitle" , entry
. word ( )
. wordsintitle ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_wordsintext" , entry . word ( ) . wordsintext ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_phrasesintext" , entry . word ( ) . phrasesintext ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_phrasesintext" , entry
. word ( )
. phrasesintext ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_llocal" , entry . word ( ) . llocal ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_lother" , entry . word ( ) . lother ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_hitcount" , entry . word ( ) . hitcount ( ) ) ;
@ -476,25 +582,50 @@ public class IndexControlRWIs_p {
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_posinphrase" , entry . word ( ) . posinphrase ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_urlcomps" , entry . word ( ) . urlcomps ( ) ) ;
prop . putNum ( "genUrlList_urlList_" + i + "_urlExists_urllength" , entry . word ( ) . urllength ( ) ) ;
prop . put ( "genUrlList_urlList_" + i + "_urlExists_props" ,
( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_indexof ) ) ? "appears on index page, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_hasimage ) ) ? "contains images, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_hasaudio ) ) ? "contains audio, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_hasvideo ) ) ? "contains video, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_hasapp ) ) ? "contains applications, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_identifier ) ) ? "appears in url, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_title ) ) ? "appears in title, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_creator ) ) ? "appears in author, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_subject ) ) ? "appears in subject, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_description ) ) ? "appears in description, " : "" ) +
( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_emphasized ) ) ? "appears emphasized, " : "" ) +
( ( DigestURI . probablyRootURL ( entry . word ( ) . urlhash ( ) ) ) ? "probably root url" : "" )
) ;
prop
. put (
"genUrlList_urlList_" + i + "_urlExists_props" ,
( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_indexof ) )
? "appears on index page, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_hasimage ) )
? "contains images, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_hasaudio ) )
? "contains audio, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_hasvideo ) )
? "contains video, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( Condenser . flag_cat_hasapp ) )
? "contains applications, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_identifier ) )
? "appears in url, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_title ) )
? "appears in title, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_creator ) )
? "appears in author, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_subject ) )
? "appears in subject, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_dc_description ) )
? "appears in description, "
: "" )
+ ( ( entry . word ( ) . flags ( ) . get ( WordReferenceRow . flag_app_emphasized ) )
? "appears emphasized, "
: "" )
+ ( ( DigestURI . probablyRootURL ( entry . word ( ) . urlhash ( ) ) ) ? "probably root url" : "" ) ) ;
if ( Switchboard . urlBlacklist . isListed ( Blacklist . BLACKLIST_DHT , url ) ) {
prop . put ( "genUrlList_urlList_" + i + "_urlExists_urlhxChecked" , "1" ) ;
}
i + + ;
if ( ( maxlines > = 0 ) & & ( i > = maxlines ) ) break ;
if ( ( maxlines > = 0 ) & & ( i > = maxlines ) ) {
break ;
}
}
final Iterator < byte [ ] > iter = ranked . miss ( ) ; // iterates url hash strings
byte [ ] b ;
@ -515,28 +646,55 @@ public class IndexControlRWIs_p {
public static void putBlacklists ( final serverObjects prop , final List < String > lists ) {
prop . put ( "genUrlList_blacklists" , lists . size ( ) ) ;
int i = 0 ;
for ( final String list : lists )
for ( final String list : lists ) {
prop . put ( "genUrlList_blacklists_" + i + + + "_name" , list ) ;
}
}
public static Bitfield compileFlags ( final serverObjects post ) {
final Bitfield b = new Bitfield ( 4 ) ;
if ( post . get ( "allurl" , "" ) . equals ( "on" ) ) return null ;
if ( post . get ( "allurl" , "" ) . equals ( "on" ) ) {
return null ;
}
if ( post . get ( "flags" ) ! = null ) {
if ( post . get ( "flags" , "" ) . length ( ) = = 0 ) return null ;
if ( post . get ( "flags" , "" ) . length ( ) = = 0 ) {
return null ;
}
return new Bitfield ( 4 , post . get ( "flags" ) ) ;
}
if ( post . get ( "description" , "" ) . equals ( "on" ) ) b . set ( WordReferenceRow . flag_app_dc_description , true ) ;
if ( post . get ( "title" , "" ) . equals ( "on" ) ) b . set ( WordReferenceRow . flag_app_dc_title , true ) ;
if ( post . get ( "creator" , "" ) . equals ( "on" ) ) b . set ( WordReferenceRow . flag_app_dc_creator , true ) ;
if ( post . get ( "subject" , "" ) . equals ( "on" ) ) b . set ( WordReferenceRow . flag_app_dc_subject , true ) ;
if ( post . get ( "url" , "" ) . equals ( "on" ) ) b . set ( WordReferenceRow . flag_app_dc_identifier , true ) ;
if ( post . get ( "emphasized" , "" ) . equals ( "on" ) ) b . set ( WordReferenceRow . flag_app_emphasized , true ) ;
if ( post . get ( "image" , "" ) . equals ( "on" ) ) b . set ( Condenser . flag_cat_hasimage , true ) ;
if ( post . get ( "audio" , "" ) . equals ( "on" ) ) b . set ( Condenser . flag_cat_hasaudio , true ) ;
if ( post . get ( "video" , "" ) . equals ( "on" ) ) b . set ( Condenser . flag_cat_hasvideo , true ) ;
if ( post . get ( "app" , "" ) . equals ( "on" ) ) b . set ( Condenser . flag_cat_hasapp , true ) ;
if ( post . get ( "indexof" , "" ) . equals ( "on" ) ) b . set ( Condenser . flag_cat_indexof , true ) ;
if ( post . get ( "description" , "" ) . equals ( "on" ) ) {
b . set ( WordReferenceRow . flag_app_dc_description , true ) ;
}
if ( post . get ( "title" , "" ) . equals ( "on" ) ) {
b . set ( WordReferenceRow . flag_app_dc_title , true ) ;
}
if ( post . get ( "creator" , "" ) . equals ( "on" ) ) {
b . set ( WordReferenceRow . flag_app_dc_creator , true ) ;
}
if ( post . get ( "subject" , "" ) . equals ( "on" ) ) {
b . set ( WordReferenceRow . flag_app_dc_subject , true ) ;
}
if ( post . get ( "url" , "" ) . equals ( "on" ) ) {
b . set ( WordReferenceRow . flag_app_dc_identifier , true ) ;
}
if ( post . get ( "emphasized" , "" ) . equals ( "on" ) ) {
b . set ( WordReferenceRow . flag_app_emphasized , true ) ;
}
if ( post . get ( "image" , "" ) . equals ( "on" ) ) {
b . set ( Condenser . flag_cat_hasimage , true ) ;
}
if ( post . get ( "audio" , "" ) . equals ( "on" ) ) {
b . set ( Condenser . flag_cat_hasaudio , true ) ;
}
if ( post . get ( "video" , "" ) . equals ( "on" ) ) {
b . set ( Condenser . flag_cat_hasvideo , true ) ;
}
if ( post . get ( "app" , "" ) . equals ( "on" ) ) {
b . set ( Condenser . flag_cat_hasapp , true ) ;
}
if ( post . get ( "indexof" , "" ) . equals ( "on" ) ) {
b . set ( Condenser . flag_cat_indexof , true ) ;
}
return b ;
}
@ -545,20 +703,29 @@ public class IndexControlRWIs_p {
Seed seed ;
int hc = 0 ;
prop . put ( "searchresult_keyhash" , startHash ) ;
final Iterator < Seed > e = PeerSelection . getAcceptRemoteIndexSeeds ( sb . peers , startHash , sb . peers . sizeConnected ( ) , true ) ;
final Iterator < Seed > e =
PeerSelection . getAcceptRemoteIndexSeeds ( sb . peers , startHash , sb . peers . sizeConnected ( ) , true ) ;
while ( e . hasNext ( ) ) {
seed = e . next ( ) ;
if ( seed ! = null ) {
prop . put ( "searchresult_hosts_" + hc + "_hosthash" , seed . hash ) ;
prop . putHTML ( "searchresult_hosts_" + hc + "_hostname" , seed . hash + " " + seed . get ( Seed . NAME , "nameless" ) ) ;
prop . putHTML (
"searchresult_hosts_" + hc + "_hostname" ,
seed . hash + " " + seed . get ( Seed . NAME , "nameless" ) ) ;
hc + + ;
}
}
prop . put ( "searchresult_hosts" , hc ) ;
}
public static RWIProcess genSearchresult ( final serverObjects prop , final Switchboard sb , final Segment segment , final byte [ ] keyhash , final Bitfield filter ) {
final QueryParams query = new QueryParams ( ASCII . String ( keyhash ) , - 1 , filter , segment , sb . getRanking ( ) , "IndexControlRWIs_p" ) ;
public static RWIProcess genSearchresult (
final serverObjects prop ,
final Switchboard sb ,
final Segment segment ,
final byte [ ] keyhash ,
final Bitfield filter ) {
final QueryParams query =
new QueryParams ( ASCII . String ( keyhash ) , - 1 , filter , segment , sb . getRanking ( ) , "IndexControlRWIs_p" ) ;
final ReferenceOrder order = new ReferenceOrder ( query . ranking , UTF8 . getBytes ( query . targetlang ) ) ;
final RWIProcess ranked = new RWIProcess ( query , order , Integer . MAX_VALUE ) ;
ranked . run ( ) ;
@ -569,7 +736,8 @@ public class IndexControlRWIs_p {
} else {
prop . put ( "searchresult" , 3 ) ;
prop . put ( "searchresult_allurl" , ranked . filteredCount ( ) ) ;
prop . put ( "searchresult_description" , ranked . flagCount ( ) [ WordReferenceRow . flag_app_dc_description ] ) ;
prop
. put ( "searchresult_description" , ranked . flagCount ( ) [ WordReferenceRow . flag_app_dc_description ] ) ;
prop . put ( "searchresult_title" , ranked . flagCount ( ) [ WordReferenceRow . flag_app_dc_title ] ) ;
prop . put ( "searchresult_creator" , ranked . flagCount ( ) [ WordReferenceRow . flag_app_dc_creator ] ) ;
prop . put ( "searchresult_subject" , ranked . flagCount ( ) [ WordReferenceRow . flag_app_dc_subject ] ) ;