text
stringlengths 30
1.67M
|
|---|
<s> package com . senseidb . examples ; import java . io . BufferedReader ; import java . io . File ; import java . io . FileReader ; import java . util . HashSet ; import java . util . Set ; import org . apache . lucene . analysis . standard . StandardAnalyzer ; import org . apache . lucene . document . Document ; import org . apache . lucene . index . IndexWriter ; import org . apache . lucene . index . IndexWriter . MaxFieldLength ; import org . apache . lucene . store . SimpleFSDirectory ; import org . apache . lucene . util . Version ; import org . json . JSONObject ; import proj . zoie . api . indexing . ZoieIndexable ; import proj . zoie . api . indexing . ZoieIndexable . IndexingReq ; import com . senseidb . conf . SenseiSchema ; import com . senseidb . conf . SenseiServerBuilder ; import com . senseidb . indexing . DefaultJsonSchemaInterpreter ; import com . senseidb . search . plugin . PluggableSearchEngineManager ; public class SchemaIndexingExample { public static void main ( String [ ] args ) throws Exception { File confDir = new File ( args [ <NUM_LIT:0> ] ) ; File dataFile = new File ( args [ <NUM_LIT:1> ] ) ; File idxDir = new File ( args [ <NUM_LIT:2> ] ) ; JSONObject schemaData = SenseiServerBuilder . loadSchema ( confDir ) ; SenseiSchema schema = SenseiSchema . build ( schemaData ) ; DefaultJsonSchemaInterpreter defaultInterpreter = new DefaultJsonSchemaInterpreter ( schema , new PluggableSearchEngineManager ( ) { @ Override public Set < String > getFieldNames ( ) { return new HashSet < String > ( ) ; } } ) ; FileReader freader = new FileReader ( dataFile ) ; BufferedReader br = new BufferedReader ( freader ) ; IndexWriter idxWriter = new IndexWriter ( SimpleFSDirectory . open ( idxDir ) , new StandardAnalyzer ( Version . LUCENE_CURRENT ) , MaxFieldLength . UNLIMITED ) ; while ( true ) { String line = br . readLine ( ) ; if ( line == null ) break ; JSONObject obj = new JSONObject ( line ) ; ZoieIndexable indexable = defaultInterpreter . convertAndInterpret ( obj ) ; IndexingReq [ ] idxReqs = indexable . buildIndexingReqs ( ) ; for ( IndexingReq req : idxReqs ) { Document doc = req . getDocument ( ) ; idxWriter . addDocument ( doc ) ; } idxWriter . commit ( ) ; idxWriter . optimize ( ) ; idxWriter . close ( ) ; } freader . close ( ) ; } } </s>
|
<s> package com . senseidb . cluster . client ; public interface ClientBoostrapFactory { public static class DefaultClientBoostrapFactory implements ClientBoostrapFactory { } } </s>
|
<s> package com . senseidb . cluster . client ; import java . util . Set ; import java . util . concurrent . Future ; import com . google . protobuf . Message ; import com . linkedin . norbert . cluster . ClusterDisconnectedException ; import com . linkedin . norbert . cluster . InvalidClusterException ; import com . linkedin . norbert . cluster . InvalidNodeException ; import com . linkedin . norbert . javacompat . cluster . Node ; import com . linkedin . norbert . javacompat . network . * ; import com . linkedin . norbert . javacompat . network . IntegerConsistentHashPartitionedLoadBalancerFactory ; import com . linkedin . norbert . javacompat . network . NettyPartitionedNetworkClient ; import com . linkedin . norbert . network . NoNodesAvailableException ; import com . linkedin . norbert . network . ResponseIterator ; import com . linkedin . norbert . network . Serializer ; import com . senseidb . cluster . routing . SenseiPartitionedLoadBalancerFactory ; public class SenseiNetworkClient implements PartitionedNetworkClient < String > { private final PartitionedNetworkClient < String > _networkClient ; public SenseiNetworkClient ( NetworkClientConfig netConfig , PartitionedLoadBalancerFactory < String > routerFactory ) { if ( routerFactory != null ) { _networkClient = new NettyPartitionedNetworkClient < String > ( netConfig , routerFactory ) ; } else { SenseiPartitionedLoadBalancerFactory lbf = new SenseiPartitionedLoadBalancerFactory ( <NUM_LIT> ) ; _networkClient = new NettyPartitionedNetworkClient < String > ( netConfig , lbf ) ; } } @ Override public < RequestMsg , ResponseMsg > Future < ResponseMsg > sendRequest ( String partitionedId , RequestMsg request , Serializer < RequestMsg , ResponseMsg > serializer ) throws InvalidClusterException , NoNodesAvailableException , ClusterDisconnectedException { return _networkClient . sendRequest ( partitionedId , request , serializer ) ; } @ Override public < RequestMsg , ResponseMsg > ResponseIterator < ResponseMsg > sendRequest ( Set < String > partitionedIds , RequestMsg request , Serializer < RequestMsg , ResponseMsg > serializer ) throws InvalidClusterException , NoNodesAvailableException , ClusterDisconnectedException { return _networkClient . sendRequest ( partitionedIds , request , serializer ) ; } @ Override public < RequestMsg , ResponseMsg > ResponseIterator < ResponseMsg > sendRequest ( Set < String > partitionedIds , RequestBuilder < String , RequestMsg > requestBuilder , Serializer < RequestMsg , ResponseMsg > serializer ) throws Exception { return _networkClient . sendRequest ( partitionedIds , requestBuilder , serializer ) ; } @ Override public < RequestMsg , ResponseMsg , T > T sendRequest ( Set < String > partitionedIds , RequestBuilder < String , RequestMsg > requestBuilder , ScatterGatherHandler < RequestMsg , ResponseMsg , T , String > scatterGatherHandler , Serializer < RequestMsg , ResponseMsg > serializer ) throws Exception { return _networkClient . sendRequest ( partitionedIds , requestBuilder , scatterGatherHandler , serializer ) ; } @ Override public < RequestMsg , ResponseMsg > ResponseIterator < ResponseMsg > sendRequestToOneReplica ( String partitionedId , RequestBuilder < Integer , RequestMsg > requestBuilder , Serializer < RequestMsg , ResponseMsg > serializer ) throws InvalidClusterException , NoNodesAvailableException , ClusterDisconnectedException { return _networkClient . sendRequestToOneReplica ( partitionedId , requestBuilder , serializer ) ; } @ Override public < RequestMsg , ResponseMsg > ResponseIterator < ResponseMsg > sendRequestToReplicas ( String s , RequestMsg requestMsg , Serializer < RequestMsg , ResponseMsg > serializer ) throws InvalidClusterException , NoNodesAvailableException , ClusterDisconnectedException { return _networkClient . sendRequestToReplicas ( s , requestMsg , serializer ) ; } @ Override public < RequestMsg , ResponseMsg > ResponseIterator < ResponseMsg > sendRequestToOneReplica ( String partitionedId , RequestMsg request , Serializer < RequestMsg , ResponseMsg > serializer ) throws InvalidClusterException , NoNodesAvailableException , ClusterDisconnectedException { return _networkClient . sendRequestToOneReplica ( partitionedId , request , serializer ) ; } @ Override public < RequestMsg , ResponseMsg > Future < ResponseMsg > sendRequestToNode ( RequestMsg request , Node node , Serializer < RequestMsg , ResponseMsg > serializer ) throws InvalidNodeException , ClusterDisconnectedException { return _networkClient . sendRequestToNode ( request , node , serializer ) ; } @ Override public < RequestMsg , ResponseMsg > ResponseIterator < ResponseMsg > broadcastMessage ( RequestMsg request , Serializer < RequestMsg , ResponseMsg > serializer ) throws ClusterDisconnectedException { return _networkClient . broadcastMessage ( request , serializer ) ; } @ Override public < RequestMsg , ResponseMsg > ResponseIterator < ResponseMsg > sendRequestToPartitions ( String partitionedId , Set < Integer > partitions , RequestBuilder < Integer , RequestMsg > requestBuilder , Serializer < RequestMsg , ResponseMsg > serializer ) throws InvalidClusterException , NoNodesAvailableException , ClusterDisconnectedException { return _networkClient . sendRequestToPartitions ( partitionedId , partitions , requestBuilder , serializer ) ; } public void shutdown ( ) { _networkClient . shutdown ( ) ; } } </s>
|
<s> package com . senseidb . cluster . client ; import java . io . UnsupportedEncodingException ; import org . apache . lucene . search . SortField ; import org . json . JSONObject ; import com . browseengine . bobo . api . BrowseSelection ; import com . browseengine . bobo . api . FacetSpec ; import com . browseengine . bobo . api . FacetSpec . FacetSortSpec ; import com . senseidb . search . req . SenseiJSONQuery ; import com . senseidb . search . req . SenseiQuery ; import com . senseidb . search . req . SenseiRequest ; public class BrowseRequestBuilder { private SenseiRequest _req ; public BrowseRequestBuilder ( ) { clear ( ) ; } public void addSelection ( String name , String val , boolean isNot ) { BrowseSelection sel = _req . getSelection ( name ) ; if ( sel == null ) { sel = new BrowseSelection ( name ) ; } if ( isNot ) { sel . addNotValue ( val ) ; } else { sel . addValue ( val ) ; } _req . addSelection ( sel ) ; } public void clearSelection ( String name ) { _req . removeSelection ( name ) ; } public void applyFacetSpec ( String name , int minHitCount , int maxCount , boolean expand , FacetSortSpec orderBy ) { FacetSpec fspec = new FacetSpec ( ) ; fspec . setMinHitCount ( minHitCount ) ; fspec . setMaxCount ( maxCount ) ; fspec . setExpandSelection ( expand ) ; fspec . setOrderBy ( orderBy ) ; _req . setFacetSpec ( name , fspec ) ; } public void applySort ( SortField [ ] sorts ) { if ( sorts == null ) { _req . clearSort ( ) ; } else { _req . setSort ( sorts ) ; } } public void clearFacetSpecs ( ) { _req . getFacetSpecs ( ) . clear ( ) ; } public void clearFacetSpec ( String name ) { _req . getFacetSpecs ( ) . remove ( name ) ; } public void setOffset ( int offset ) { _req . setOffset ( offset ) ; } public void setCount ( int count ) { _req . setCount ( count ) ; } public void setQuery ( String qString ) { JSONObject qObj = new JSONObject ( ) ; if ( qString != null ) { try { qObj . put ( "<STR_LIT:query>" , qString ) ; } catch ( Exception e ) { e . printStackTrace ( ) ; } } _req . setQuery ( new SenseiJSONQuery ( qObj ) ) ; } public void clear ( ) { _req = new SenseiRequest ( ) ; _req . setOffset ( <NUM_LIT:0> ) ; _req . setCount ( <NUM_LIT:5> ) ; _req . setFetchStoredFields ( true ) ; } public void clearSelections ( ) { _req . clearSelections ( ) ; } public SenseiRequest getRequest ( ) { return _req ; } public String getQueryString ( ) { SenseiQuery q = _req . getQuery ( ) ; if ( q != null ) { try { return new String ( q . toBytes ( ) , "<STR_LIT:UTF-8>" ) ; } catch ( UnsupportedEncodingException e ) { e . printStackTrace ( ) ; return null ; } } return null ; } } </s>
|
<s> package com . senseidb . cluster . client ; import java . util . Arrays ; import java . util . List ; import java . util . Map ; import java . util . Map . Entry ; import java . util . Set ; import com . browseengine . bobo . api . BrowseFacet ; import com . browseengine . bobo . api . BrowseHit ; import com . browseengine . bobo . api . BrowseResult ; import com . browseengine . bobo . api . FacetAccessible ; public class BrowseResultFormatter { public static String formatResults ( BrowseResult res ) { StringBuffer sb = new StringBuffer ( ) ; sb . append ( res . getNumHits ( ) ) ; sb . append ( "<STR_LIT>" ) ; sb . append ( res . getTotalDocs ( ) ) ; if ( res . getNumGroups ( ) > <NUM_LIT:0> ) { sb . append ( "<STR_LIT>" ) ; sb . append ( res . getNumGroups ( ) ) ; sb . append ( "<STR_LIT>" ) ; } else sb . append ( "<STR_LIT>" ) ; BrowseHit [ ] hits = res . getHits ( ) ; Map < String , FacetAccessible > map = res . getFacetMap ( ) ; for ( Entry < String , FacetAccessible > entry : map . entrySet ( ) ) { String key = entry . getKey ( ) ; FacetAccessible fa = entry . getValue ( ) ; sb . append ( key ) . append ( "<STR_LIT:n>" ) ; List < BrowseFacet > lf = fa . getFacets ( ) ; for ( BrowseFacet bf : lf ) { sb . append ( "<STR_LIT:t>" ) . append ( bf ) . append ( "<STR_LIT:n>" ) ; } } for ( BrowseHit hit : hits ) { sb . append ( "<STR_LIT>" ) ; sb . append ( formatHit ( hit ) ) ; sb . append ( "<STR_LIT:n>" ) ; } sb . append ( "<STR_LIT>" ) ; return sb . toString ( ) ; } static StringBuffer formatHit ( BrowseHit hit ) { StringBuffer sb = new StringBuffer ( ) ; if ( hit . getGroupHitsCount ( ) > <NUM_LIT:0> ) { sb . append ( "<STR_LIT>" ) ; sb . append ( hit . getGroupValue ( ) ) ; sb . append ( "<STR_LIT>" ) ; sb . append ( hit . getGroupHitsCount ( ) ) ; sb . append ( '<STR_LIT:\n>' ) ; } Map < String , String [ ] > fields = hit . getFieldValues ( ) ; if ( fields != null ) { for ( Entry < String , String [ ] > entry : fields . entrySet ( ) ) { String key = entry . getKey ( ) ; sb . append ( "<STR_LIT:t>" ) . append ( key ) . append ( "<STR_LIT>" ) ; String [ ] values = entry . getValue ( ) ; sb . append ( Arrays . toString ( values ) ) ; sb . append ( "<STR_LIT:n>" ) ; } } return sb ; } } </s>
|
<s> package com . senseidb . cluster . routing ; import com . linkedin . norbert . cluster . InvalidClusterException ; import com . linkedin . norbert . javacompat . cluster . Node ; import com . linkedin . norbert . javacompat . network . * ; import java . util . Map ; import java . util . Set ; public class SenseiPartitionedLoadBalancerFactory implements PartitionedLoadBalancerFactory < String > { private final ConsistentHashPartitionedLoadBalancerFactory < String > lbf ; public SenseiPartitionedLoadBalancerFactory ( int bucketCount ) { HashFunction . MD5HashFunction hashFn = new HashFunction . MD5HashFunction ( ) ; MultiRingConsistentHashPartitionedLoadBalancerFactory < String > fallThroughLbf = new MultiRingConsistentHashPartitionedLoadBalancerFactory < String > ( - <NUM_LIT:1> , bucketCount , hashFn , hashFn , true ) ; this . lbf = new ConsistentHashPartitionedLoadBalancerFactory < String > ( bucketCount , hashFn , fallThroughLbf ) ; } @ Override public PartitionedLoadBalancer < String > newLoadBalancer ( Set < Endpoint > endpoints ) throws InvalidClusterException { return lbf . newLoadBalancer ( endpoints ) ; } @ Override public Integer getNumPartitions ( Set < Endpoint > endpoints ) { return lbf . getNumPartitions ( endpoints ) ; } } </s>
|
<s> package com . senseidb . cluster . routing ; public interface HashProvider { public long hash ( String key ) ; } </s>
|
<s> package com . senseidb . cluster . routing ; import java . util . Arrays ; import java . util . List ; import com . linkedin . norbert . javacompat . cluster . Node ; public class RoutingInfo { public final int [ ] partitions ; public final int [ ] nodegroup ; public final List < Node > [ ] nodelist ; public RoutingInfo ( final List < Node > [ ] nodelist , int [ ] partitions , int [ ] nodegroup ) { this . partitions = partitions ; this . nodelist = nodelist ; this . nodegroup = nodegroup ; } public String toString ( ) { StringBuffer sb = new StringBuffer ( ) ; sb . append ( "<STR_LIT>" ) . append ( Arrays . toString ( nodegroup ) ) . append ( "<STR_LIT>" ) . append ( Arrays . toString ( partitions ) ) ; return sb . toString ( ) ; } } </s>
|
<s> package com . senseidb . cluster . routing ; import java . security . MessageDigest ; import java . security . NoSuchAlgorithmException ; import org . apache . log4j . Logger ; public class MD5HashProvider implements HashProvider { private final static Logger logger = Logger . getLogger ( MD5HashProvider . class ) ; private final ThreadLocal < MessageDigest > _md = new ThreadLocal < MessageDigest > ( ) { protected MessageDigest initialValue ( ) { try { return MessageDigest . getInstance ( "<STR_LIT>" ) ; } catch ( NoSuchAlgorithmException e ) { logger . error ( e ) ; } return null ; } } ; public long hash ( String key ) { byte [ ] kbytes = _md . get ( ) . digest ( key . getBytes ( ) ) ; long hc = ( ( long ) ( kbytes [ <NUM_LIT:3> ] & <NUM_LIT> ) << <NUM_LIT:24> ) | ( ( long ) ( kbytes [ <NUM_LIT:2> ] & <NUM_LIT> ) << <NUM_LIT:16> ) | ( ( long ) ( kbytes [ <NUM_LIT:1> ] & <NUM_LIT> ) << <NUM_LIT:8> ) | ( long ) ( kbytes [ <NUM_LIT:0> ] & <NUM_LIT> ) ; _md . get ( ) . reset ( ) ; return Math . abs ( hc ) ; } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . search . ConstantScoreQuery ; import org . apache . lucene . search . Filter ; import org . apache . lucene . search . Query ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . search . query . filters . FilterConstructor ; public class UIDQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { JSONObject filterJson = new JSONObject ( ) ; filterJson . put ( QUERY_TYPE , jsonQuery ) ; Filter filter = null ; try { filter = FilterConstructor . constructFilter ( filterJson , null ) ; } catch ( Exception e ) { throw new JSONException ( e ) ; } ConstantScoreQuery query = new ConstantScoreQuery ( filter ) ; query . setBoost ( ( float ) jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; return query ; } } </s>
|
<s> package com . senseidb . search . query ; import java . util . Iterator ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . PrefixQuery ; import org . apache . lucene . search . Query ; import org . json . JSONException ; import org . json . JSONObject ; public class PrefixQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject json ) throws JSONException { Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; String field = iter . next ( ) ; String prefix = null ; Object obj = json . get ( field ) ; if ( obj instanceof JSONObject ) { prefix = ( ( JSONObject ) obj ) . optString ( VALUE_PARAM , "<STR_LIT>" ) ; float boost = ( float ) ( ( JSONObject ) obj ) . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ; Query query = new PrefixQuery ( new Term ( field , prefix ) ) ; query . setBoost ( boost ) ; return query ; } else { prefix = String . valueOf ( obj ) ; return new PrefixQuery ( new Term ( field , prefix ) ) ; } } } </s>
|
<s> package com . senseidb . search . query ; import java . util . Iterator ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . TermQuery ; import org . json . JSONException ; import org . json . JSONObject ; public class TermQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject json ) throws JSONException { Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; String field = iter . next ( ) ; Object obj = json . get ( field ) ; String txt ; float boost ; if ( obj instanceof JSONObject ) { txt = ( ( JSONObject ) obj ) . optString ( TERM_PARAM ) ; if ( txt == null || txt . length ( ) == <NUM_LIT:0> ) txt = ( ( JSONObject ) obj ) . getString ( VALUE_PARAM ) ; boost = ( float ) ( ( JSONObject ) obj ) . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ; } else { txt = ( String ) obj ; boost = <NUM_LIT:1.0f> ; } Query q = new TermQuery ( new Term ( field , txt ) ) ; q . setBoost ( boost ) ; return q ; } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . search . Filter ; public abstract class RetentionFilterFactory { public RetentionFilterFactory ( ) { } public abstract Filter buildRetentionFilter ( int nDays ) ; } </s>
|
<s> package com . senseidb . search . query ; import java . util . ArrayList ; import org . apache . lucene . search . DisjunctionMaxQuery ; import org . apache . lucene . search . Query ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; public class DistMaxQueryConstructor extends QueryConstructor { private TermQueryConstructor termQueryConstructor = new TermQueryConstructor ( ) ; public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { JSONArray jsonArray = jsonQuery . getJSONArray ( QUERIES_PARAM ) ; ArrayList < Query > ar = new ArrayList < Query > ( ) ; for ( int i = <NUM_LIT:0> ; i < jsonArray . length ( ) ; i ++ ) { JSONObject json = jsonArray . getJSONObject ( i ) . getJSONObject ( TERM_PARAM ) ; ar . add ( termQueryConstructor . doConstructQuery ( json ) ) ; } float tieBreakerMultiplier = ( float ) jsonQuery . optDouble ( TIE_BREAKER_PARAM , <NUM_LIT> ) ; float boost = ( float ) jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ; Query dmq = new DisjunctionMaxQuery ( ar , tieBreakerMultiplier ) ; dmq . setBoost ( boost ) ; return dmq ; } } </s>
|
<s> package com . senseidb . search . query ; import java . util . ArrayList ; import java . util . List ; import org . apache . lucene . queryParser . ParseException ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . BooleanClause ; import org . apache . lucene . search . BooleanQuery ; import org . apache . lucene . search . DisjunctionMaxQuery ; import org . apache . lucene . search . MatchAllDocsQuery ; import org . apache . lucene . search . Query ; import org . apache . lucene . util . Version ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; public class QueryStringQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; private QueryParser _qparser ; public QueryStringQueryConstructor ( QueryParser qparser ) { _qparser = qparser ; } @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { String queryText = jsonQuery . optString ( QUERY_PARAM , null ) ; if ( queryText == null || queryText . length ( ) == <NUM_LIT:0> ) { return new MatchAllDocsQuery ( ) ; } try { for ( String name : JSONObject . getNames ( jsonQuery ) ) { if ( QUERY_PARAM . equals ( name ) ) continue ; if ( BOOST_PARAM . equals ( name ) ) continue ; List < String > fields = new ArrayList < String > ( ) ; String defaultField = jsonQuery . optString ( DEFAULT_FIELD_PARAM , null ) ; if ( defaultField != null && defaultField . length ( ) != <NUM_LIT:0> ) fields . add ( defaultField ) ; JSONArray fieldsArray = jsonQuery . optJSONArray ( FIELDS_PARAM ) ; if ( fieldsArray != null ) { for ( int i = <NUM_LIT:0> ; i < fieldsArray . length ( ) ; ++ i ) { String field = fieldsArray . optString ( i , null ) ; if ( field != null && field . length ( ) != <NUM_LIT:0> ) fields . add ( field ) ; } } if ( fields . size ( ) == <NUM_LIT:0> ) fields . add ( "<STR_LIT>" ) ; String default_operator = jsonQuery . optString ( DEFAULT_OPERATOR_PARAM , "<STR_LIT>" ) ; boolean allow_leading_wildcard = jsonQuery . optBoolean ( ALLOW_LEADING_WILDCARD_PARAM , true ) ; boolean lowercase_expanded_terms = jsonQuery . optBoolean ( LOWERCASE_EXPANDED_TERMS_PARAM , true ) ; boolean enable_position_increments = jsonQuery . optBoolean ( ENABLE_POSITION_INCREMENTS_PARAM , true ) ; int fuzzy_prefix_length = jsonQuery . optInt ( FUZZY_PREFIX_LENGTH_PARAM , <NUM_LIT:0> ) ; float fuzzy_min_sim = ( float ) jsonQuery . optDouble ( FUZZY_MIN_SIM_PARAM , <NUM_LIT> ) ; int phrase_slop = jsonQuery . optInt ( PHRASE_SLOP_PARAM , <NUM_LIT:0> ) ; boolean use_dis_max = jsonQuery . optBoolean ( USE_DIS_MAX_PARAM , true ) ; float tie_breaker = ( float ) jsonQuery . optDouble ( TIE_BREAKER_PARAM , <NUM_LIT> ) ; List < Query > queries = new ArrayList < Query > ( fields . size ( ) ) ; for ( String field : fields ) { QueryParser qparser = new QueryParser ( Version . LUCENE_CURRENT , field , _qparser . getAnalyzer ( ) ) ; qparser . setAllowLeadingWildcard ( allow_leading_wildcard ) ; qparser . setEnablePositionIncrements ( enable_position_increments ) ; qparser . setFuzzyMinSim ( fuzzy_min_sim ) ; qparser . setFuzzyPrefixLength ( fuzzy_prefix_length ) ; qparser . setLowercaseExpandedTerms ( lowercase_expanded_terms ) ; qparser . setPhraseSlop ( phrase_slop ) ; queries . add ( qparser . parse ( queryText ) ) ; } if ( fields . size ( ) == <NUM_LIT:1> ) { Query query = queries . get ( <NUM_LIT:0> ) ; query . setBoost ( ( float ) jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; return query ; } if ( use_dis_max ) { Query query = new DisjunctionMaxQuery ( queries , tie_breaker ) ; query . setBoost ( ( float ) jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; return query ; } else { BooleanQuery query = new BooleanQuery ( ) ; if ( AND_PARAM . equals ( default_operator ) ) { for ( Query q : queries ) { query . add ( q , BooleanClause . Occur . MUST ) ; } } else { for ( Query q : queries ) { query . add ( q , BooleanClause . Occur . SHOULD ) ; } } query . setBoost ( ( float ) jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; return query ; } } synchronized ( _qparser ) { Query query = _qparser . parse ( queryText ) ; query . setBoost ( ( float ) jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; return query ; } } catch ( ParseException e ) { throw new JSONException ( e ) ; } } } </s>
|
<s> package com . senseidb . search . query ; import java . io . IOException ; import org . apache . log4j . Logger ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . Explanation ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . Scorer ; import org . json . JSONException ; import org . json . JSONObject ; import com . browseengine . bobo . api . BoboIndexReader ; public class ScoreAugmentQuery extends AbstractScoreAdjuster { private static final long serialVersionUID = <NUM_LIT:1L> ; public static interface ScoreAugmentFunction { public void initializeReader ( BoboIndexReader reader , JSONObject jsonParams ) throws IOException ; public void initializeGlobal ( JSONObject jsonParams ) throws JSONException ; public boolean useInnerScore ( ) ; public float newScore ( float rawScore , int docID ) ; public float newScore ( int docID ) ; public String getExplainString ( float rawscore , int docID ) ; public ScoreAugmentFunction getCopy ( ) ; } private static class AugmentScorer extends Scorer { private static Logger logger = Logger . getLogger ( AugmentScorer . class ) ; private final ScoreAugmentFunction _func ; private final Scorer _innerScorer ; protected AugmentScorer ( BoboIndexReader reader , Scorer innerScorer , ScoreAugmentFunction func , JSONObject jsonParms ) throws IOException { super ( innerScorer . getSimilarity ( ) ) ; _innerScorer = innerScorer ; _func = func ; _func . initializeReader ( reader , jsonParms ) ; } @ Override public float score ( ) throws IOException { return ( _func . useInnerScore ( ) ) ? _func . newScore ( _innerScorer . score ( ) , _innerScorer . docID ( ) ) : _func . newScore ( _innerScorer . docID ( ) ) ; } @ Override public int advance ( int target ) throws IOException { return _innerScorer . advance ( target ) ; } @ Override public int docID ( ) { return _innerScorer . docID ( ) ; } @ Override public int nextDoc ( ) throws IOException { return _innerScorer . nextDoc ( ) ; } } private transient ScoreAugmentFunction _func ; private transient JSONObject _jsonParam ; public ScoreAugmentQuery ( Query query , ScoreAugmentFunction func , JSONObject jsonParam ) throws JSONException { super ( query ) ; _func = func ; _func . initializeGlobal ( jsonParam ) ; _jsonParam = jsonParam ; if ( _func == null ) throw new IllegalArgumentException ( "<STR_LIT>" ) ; } @ Override protected Scorer createScorer ( Scorer innerScorer , IndexReader reader , boolean scoreDocsInOrder , boolean topScorer ) throws IOException { if ( reader instanceof BoboIndexReader ) { return new AugmentScorer ( ( BoboIndexReader ) reader , innerScorer , _func . getCopy ( ) , _jsonParam ) ; } else { throw new IllegalStateException ( "<STR_LIT>" + BoboIndexReader . class ) ; } } @ Override protected Explanation createExplain ( Explanation innerExplain , IndexReader reader , int doc ) throws IOException { if ( reader instanceof BoboIndexReader ) { Explanation finalExpl = new Explanation ( ) ; finalExpl . addDetail ( innerExplain ) ; _func . initializeReader ( ( BoboIndexReader ) reader , _jsonParam ) ; float innerValue = innerExplain . getValue ( ) ; float value = <NUM_LIT:0> ; if ( _func . useInnerScore ( ) ) value = _func . newScore ( innerValue , doc ) ; else value = _func . newScore ( doc ) ; finalExpl . setValue ( value ) ; finalExpl . setDescription ( "<STR_LIT>" + _func . getExplainString ( innerValue , doc ) ) ; return finalExpl ; } else { throw new IllegalStateException ( "<STR_LIT>" + BoboIndexReader . class ) ; } } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . search . ConstantScoreQuery ; import org . apache . lucene . search . Filter ; import org . apache . lucene . search . Query ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . search . query . filters . FilterConstructor ; public class RangeQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { Filter filter = null ; try { JSONObject newJson = new JSONObject ( ) ; newJson . put ( QUERY_TYPE , jsonQuery ) ; filter = FilterConstructor . constructFilter ( newJson , null ) ; } catch ( Exception e ) { throw new JSONException ( e ) ; } ConstantScoreQuery query = new ConstantScoreQuery ( filter ) ; query . setBoost ( ( float ) jsonQuery . getJSONObject ( ( String ) jsonQuery . keys ( ) . next ( ) ) . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; return query ; } } </s>
|
<s> package com . senseidb . search . query ; import java . util . ArrayList ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . spans . SpanOrQuery ; import org . apache . lucene . search . spans . SpanQuery ; import org . apache . lucene . search . spans . SpanTermQuery ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; public class SpanOrQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { String fieldCheck = null ; JSONArray jsonArray = jsonQuery . getJSONArray ( CLAUSES_PARAM ) ; ArrayList < SpanTermQuery > clausesList = new ArrayList < SpanTermQuery > ( ) ; for ( int i = <NUM_LIT:0> ; i < jsonArray . length ( ) ; i ++ ) { JSONObject json = jsonArray . getJSONObject ( i ) . getJSONObject ( SPAN_TERM_PARAM ) ; String field = ( String ) ( json . keys ( ) . next ( ) ) ; if ( fieldCheck == null ) fieldCheck = field ; else if ( ! fieldCheck . equals ( field ) ) throw new IllegalArgumentException ( "<STR_LIT>" + jsonQuery ) ; String value = ( String ) json . get ( field ) ; clausesList . add ( new SpanTermQuery ( new Term ( field , value ) ) ) ; } SpanQuery [ ] clauses = clausesList . toArray ( new SpanQuery [ clausesList . size ( ) ] ) ; SpanOrQuery soq = new SpanOrQuery ( clauses ) ; return soq ; } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . search . Query ; import org . json . JSONException ; import org . json . JSONObject ; public class CustomQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { try { String className = jsonQuery . getString ( CLASS_PARAM ) ; Class queryClass = Class . forName ( className ) ; Object q = queryClass . newInstance ( ) ; ( ( Query ) q ) . setBoost ( ( float ) jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; return ( Query ) q ; } catch ( Throwable t ) { throw new IllegalArgumentException ( t . getMessage ( ) ) ; } } } </s>
|
<s> package com . senseidb . search . query ; import java . io . IOException ; import java . io . StringReader ; import java . util . Iterator ; import org . apache . log4j . Logger ; import org . apache . lucene . analysis . Analyzer ; import org . apache . lucene . analysis . tokenattributes . TermAttribute ; import org . apache . lucene . analysis . TokenStream ; import org . apache . lucene . index . Term ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . BooleanClause ; import org . apache . lucene . search . BooleanQuery ; import org . apache . lucene . search . MultiPhraseQuery ; import org . apache . lucene . search . PhraseQuery ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . TermQuery ; import org . json . JSONException ; import org . json . JSONObject ; public class TextQueryConstructor extends QueryConstructor { private static final Logger logger = Logger . getLogger ( TextQueryConstructor . class ) ; public static final String QUERY_TYPE = "<STR_LIT:text>" ; private Analyzer _analyzer ; public TextQueryConstructor ( QueryParser qparser ) { _analyzer = qparser . getAnalyzer ( ) ; } @ Override protected Query doConstructQuery ( JSONObject json ) throws JSONException { String op = null ; String type = null ; String field = null ; String text = null ; Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) throw new IllegalArgumentException ( "<STR_LIT>" ) ; field = iter . next ( ) ; Object obj = json . get ( field ) ; if ( obj instanceof JSONObject ) { text = ( ( JSONObject ) obj ) . getString ( VALUE_PARAM ) ; op = ( ( JSONObject ) obj ) . optString ( OPERATOR_PARAM ) ; type = ( ( JSONObject ) obj ) . optString ( TYPE_PARAM ) ; } else { text = String . valueOf ( obj ) ; } TokenStream tokenStream = _analyzer . tokenStream ( field , new StringReader ( text ) ) ; TermAttribute termAttribute = tokenStream . getAttribute ( TermAttribute . class ) ; try { if ( PHRASE_PARAM . equals ( type ) ) { PhraseQuery q = new PhraseQuery ( ) ; while ( tokenStream . incrementToken ( ) ) { q . add ( new Term ( field , termAttribute . term ( ) ) ) ; } return q ; } else if ( PHRASE_PREFIX_PARAM . equals ( type ) ) { MultiPhraseQuery q = new MultiPhraseQuery ( ) ; while ( tokenStream . incrementToken ( ) ) { q . add ( new Term ( field , termAttribute . term ( ) ) ) ; } return q ; } else { BooleanQuery q = new BooleanQuery ( ) ; if ( AND_PARAM . equals ( op ) ) { while ( tokenStream . incrementToken ( ) ) { q . add ( new TermQuery ( new Term ( field , termAttribute . term ( ) ) ) , BooleanClause . Occur . MUST ) ; } } else { while ( tokenStream . incrementToken ( ) ) { q . add ( new TermQuery ( new Term ( field , termAttribute . term ( ) ) ) , BooleanClause . Occur . SHOULD ) ; } } return q ; } } catch ( IOException ioe ) { logger . error ( ioe . getMessage ( ) , ioe ) ; return null ; } } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . ConstantScoreQuery ; import org . apache . lucene . search . Filter ; import org . apache . lucene . search . FilteredQuery ; import org . apache . lucene . search . Query ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . search . query . filters . FilterConstructor ; public class FilteredQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; private QueryParser _qparser ; public FilteredQueryConstructor ( QueryParser qparser ) { _qparser = qparser ; } @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { JSONObject queryJson = jsonQuery . getJSONObject ( QUERY_PARAM ) ; JSONObject filterJson = jsonQuery . getJSONObject ( FILTER_PARAM ) ; if ( queryJson == null || filterJson == null ) throw new IllegalArgumentException ( "<STR_LIT>" + jsonQuery ) ; Query query = null ; try { query = QueryConstructor . constructQuery ( queryJson , _qparser ) ; } catch ( Exception e ) { throw new JSONException ( e ) ; } Filter filter = null ; try { filter = FilterConstructor . constructFilter ( filterJson , _qparser ) ; } catch ( Exception e ) { throw new JSONException ( e ) ; } return new FilteredQuery ( query , filter ) ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . io . IOException ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . DocIdSet ; import org . apache . lucene . search . DocIdSetIterator ; import org . apache . lucene . search . Filter ; import org . json . JSONObject ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . facets . data . FacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . filter . FacetFilter ; import com . browseengine . bobo . facets . filter . FacetFilter . FacetDocIdSetIterator ; import com . browseengine . bobo . util . BigNestedIntArray ; public class NullFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; @ Override protected Filter doConstructFilter ( Object json ) throws Exception { final String fieldName = json instanceof String ? ( String ) json : ( ( JSONObject ) json ) . getString ( "<STR_LIT:field>" ) ; return new Filter ( ) { @ Override public DocIdSet getDocIdSet ( IndexReader reader ) throws IOException { final Object data = ( ( BoboIndexReader ) reader ) . getFacetData ( fieldName ) ; if ( data instanceof MultiValueFacetDataCache ) { return new DocIdSet ( ) { @ Override public DocIdSetIterator iterator ( ) throws IOException { return new MultiValueFacetDocIdSetIterator ( ( MultiValueFacetDataCache ) data , <NUM_LIT:0> ) ; } } ; } else if ( data instanceof FacetDataCache ) { return new DocIdSet ( ) { @ Override public DocIdSetIterator iterator ( ) throws IOException { return new FacetFilter . FacetDocIdSetIterator ( ( FacetDataCache ) data , <NUM_LIT:0> ) ; } } ; } throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } } ; } public final static class MultiValueFacetDocIdSetIterator extends FacetDocIdSetIterator { private final BigNestedIntArray _nestedArray ; public MultiValueFacetDocIdSetIterator ( MultiValueFacetDataCache dataCache , int index ) { super ( dataCache , index ) ; _nestedArray = dataCache . _nestedArray ; } @ Override final public int nextDoc ( ) throws IOException { return ( _doc = ( _doc < _maxID ? _nestedArray . findValue ( _index , ( _doc + <NUM_LIT:1> ) , _maxID , true ) : NO_MORE_DOCS ) ) ; } @ Override final public int advance ( int id ) throws IOException { if ( _doc < id ) { return ( _doc = ( id <= _maxID ? _nestedArray . findValue ( _index , id , _maxID , true ) : NO_MORE_DOCS ) ) ; } return nextDoc ( ) ; } } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . io . IOException ; import java . util . ArrayList ; import java . util . Iterator ; import java . util . Map ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . DocIdSet ; import org . apache . lucene . search . Filter ; import org . json . JSONObject ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . api . BrowseSelection ; import com . browseengine . bobo . api . BrowseSelection . ValueOperation ; import com . browseengine . bobo . facets . FacetHandler ; import com . kamikaze . docidset . impl . AndDocIdSet ; import com . senseidb . util . RequestConverter2 ; public class FacetSelectionFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; public static BrowseSelection buildFacetSelection ( String name , JSONObject json ) throws Exception { BrowseSelection sel = new BrowseSelection ( name ) ; String [ ] vals = RequestConverter2 . getStrings ( json . optJSONArray ( VALUES_PARAM ) ) ; String [ ] nots = RequestConverter2 . getStrings ( json . optJSONArray ( EXCLUDES_PARAM ) ) ; sel . setValues ( vals ) ; sel . setNotValues ( nots ) ; String op = json . optString ( OPERATOR_PARAM , OR_PARAM ) ; if ( OR_PARAM . equalsIgnoreCase ( op ) ) { sel . setSelectionOperation ( ValueOperation . ValueOperationOr ) ; } else { sel . setSelectionOperation ( ValueOperation . ValueOperationAnd ) ; } JSONObject paramsObj = json . optJSONObject ( PARAMS_PARAM ) ; if ( paramsObj != null ) { Map < String , String > paramMap = convertParams ( paramsObj ) ; if ( paramMap != null && paramMap . size ( ) > <NUM_LIT:0> ) { sel . setSelectionProperties ( paramMap ) ; } } return sel ; } @ Override protected Filter doConstructFilter ( Object obj ) throws Exception { final JSONObject json = ( JSONObject ) obj ; return new Filter ( ) { @ Override public DocIdSet getDocIdSet ( IndexReader reader ) throws IOException { if ( reader instanceof BoboIndexReader ) { BoboIndexReader boboReader = ( BoboIndexReader ) reader ; Iterator < String > iter = json . keys ( ) ; ArrayList < DocIdSet > docSets = new ArrayList < DocIdSet > ( ) ; while ( iter . hasNext ( ) ) { String key = iter . next ( ) ; FacetHandler facetHandler = boboReader . getFacetHandler ( key ) ; if ( facetHandler != null ) { try { JSONObject jsonObj = json . getJSONObject ( key ) ; BrowseSelection sel = buildFacetSelection ( key , jsonObj ) ; docSets . add ( facetHandler . buildFilter ( sel ) . getDocIdSet ( boboReader ) ) ; } catch ( Exception e ) { throw new IOException ( e . getMessage ( ) ) ; } } else { throw new IOException ( key + "<STR_LIT>" ) ; } } if ( docSets . size ( ) == <NUM_LIT:0> ) return null ; else if ( docSets . size ( ) == <NUM_LIT:1> ) return docSets . get ( <NUM_LIT:0> ) ; return new AndDocIdSet ( docSets ) ; } else { throw new IllegalStateException ( "<STR_LIT>" + BoboIndexReader . class ) ; } } } ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import org . apache . lucene . search . Filter ; import org . json . JSONObject ; public class CustomFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; @ Override protected Filter doConstructFilter ( Object json ) throws Exception { try { String className = ( ( JSONObject ) json ) . getString ( CLASS_PARAM ) ; Class filterClass = Class . forName ( className ) ; Filter f = ( Filter ) filterClass . newInstance ( ) ; return f ; } catch ( Throwable t ) { throw new IllegalArgumentException ( t . getMessage ( ) ) ; } } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . util . HashMap ; import java . util . Iterator ; import java . util . Map ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . Filter ; import org . json . JSONObject ; public abstract class FilterConstructor { public static final String VALUES_PARAM = "<STR_LIT>" ; public static final String VALUE_PARAM = "<STR_LIT:value>" ; public static final String EXCLUDES_PARAM = "<STR_LIT>" ; public static final String OPERATOR_PARAM = "<STR_LIT>" ; public static final String PARAMS_PARAM = "<STR_LIT>" ; public static final String MUST_PARAM = "<STR_LIT>" ; public static final String MUST_NOT_PARAM = "<STR_LIT>" ; public static final String SHOULD_PARAM = "<STR_LIT>" ; public static final String FROM_PARAM = "<STR_LIT>" ; public static final String TO_PARAM = "<STR_LIT>" ; public static final String NOOPTIMIZE_PARAM = "<STR_LIT>" ; public static final String RANGE_FIELD_TYPE = "<STR_LIT>" ; public static final String RANGE_DATE_FORMAT = "<STR_LIT>" ; public static final String QUERY_PARAM = "<STR_LIT:query>" ; public static final String OR_PARAM = "<STR_LIT>" ; public static final String DEPTH_PARAM = "<STR_LIT>" ; public static final String STRICT_PARAM = "<STR_LIT>" ; public static final String INCLUDE_LOWER_PARAM = "<STR_LIT>" ; public static final String INCLUDE_UPPER_PARAM = "<STR_LIT>" ; public static final String GT_PARAM = "<STR_LIT>" ; public static final String GTE_PARAM = "<STR_LIT>" ; public static final String LT_PARAM = "<STR_LIT>" ; public static final String LTE_PARAM = "<STR_LIT>" ; public static final String CLASS_PARAM = "<STR_LIT:class>" ; private static final Map < String , FilterConstructor > FILTER_CONSTRUCTOR_MAP = new HashMap < String , FilterConstructor > ( ) ; static { FILTER_CONSTRUCTOR_MAP . put ( UIDFilterConstructor . FILTER_TYPE , new UIDFilterConstructor ( ) ) ; FILTER_CONSTRUCTOR_MAP . put ( FacetSelectionFilterConstructor . FILTER_TYPE , new FacetSelectionFilterConstructor ( ) ) ; FILTER_CONSTRUCTOR_MAP . put ( RangeFilterConstructor . FILTER_TYPE , new RangeFilterConstructor ( ) ) ; FILTER_CONSTRUCTOR_MAP . put ( TermFilterConstructor . FILTER_TYPE , new TermFilterConstructor ( ) ) ; FILTER_CONSTRUCTOR_MAP . put ( PathFilterConstructor . FILTER_TYPE , new PathFilterConstructor ( ) ) ; FILTER_CONSTRUCTOR_MAP . put ( TermsFilterConstructor . FILTER_TYPE , new TermsFilterConstructor ( ) ) ; FILTER_CONSTRUCTOR_MAP . put ( NullFilterConstructor . FILTER_TYPE , new NullFilterConstructor ( ) ) ; FILTER_CONSTRUCTOR_MAP . put ( CustomFilterConstructor . FILTER_TYPE , new CustomFilterConstructor ( ) ) ; } public static FilterConstructor getFilterConstructor ( String type , QueryParser qparser ) { FilterConstructor filterConstructor = FILTER_CONSTRUCTOR_MAP . get ( type ) ; if ( filterConstructor == null ) { if ( QueryFilterConstructor . FILTER_TYPE . equals ( type ) ) filterConstructor = new QueryFilterConstructor ( qparser ) ; else if ( AndFilterConstructor . FILTER_TYPE . equals ( type ) ) filterConstructor = new AndFilterConstructor ( qparser ) ; else if ( OrFilterConstructor . FILTER_TYPE . equals ( type ) ) filterConstructor = new OrFilterConstructor ( qparser ) ; else if ( BooleanFilterConstructor . FILTER_TYPE . equals ( type ) ) filterConstructor = new BooleanFilterConstructor ( qparser ) ; } return filterConstructor ; } public static Map < String , String > convertParams ( JSONObject obj ) { Map < String , String > paramMap = new HashMap < String , String > ( ) ; String [ ] names = JSONObject . getNames ( obj ) ; if ( names != null ) { for ( String name : names ) { String val = obj . optString ( name , null ) ; if ( val != null ) { paramMap . put ( name , val ) ; } } } return paramMap ; } public static Filter constructFilter ( JSONObject json , QueryParser qparser ) throws Exception { if ( json == null ) return null ; Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; String type = iter . next ( ) ; FilterConstructor filterConstructor = FilterConstructor . getFilterConstructor ( type , qparser ) ; if ( filterConstructor == null ) throw new IllegalArgumentException ( "<STR_LIT>" + type + "<STR_LIT>" ) ; return filterConstructor . doConstructFilter ( json . get ( type ) ) ; } abstract protected Filter doConstructFilter ( Object json ) throws Exception ; } </s>
|
<s> package com . senseidb . search . query . filters ; import java . io . IOException ; import java . util . Iterator ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . DocIdSet ; import org . apache . lucene . search . DocIdSetIterator ; import org . apache . lucene . search . Filter ; import org . json . JSONObject ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . api . BrowseSelection ; import com . browseengine . bobo . facets . impl . PathFacetHandler ; import com . browseengine . bobo . facets . FacetHandler ; import com . browseengine . bobo . query . MatchAllDocIdSetIterator ; public class PathFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT:path>" ; @ Override protected Filter doConstructFilter ( Object param ) throws Exception { JSONObject json = ( JSONObject ) param ; Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) return null ; final String field = iter . next ( ) ; final String path ; final int depth ; final boolean strict ; Object obj = json . get ( field ) ; if ( obj instanceof JSONObject ) { path = ( ( JSONObject ) obj ) . getString ( VALUE_PARAM ) ; depth = ( ( JSONObject ) obj ) . optInt ( DEPTH_PARAM , <NUM_LIT:0> ) ; strict = ( ( JSONObject ) obj ) . optBoolean ( STRICT_PARAM , false ) ; } else { path = String . valueOf ( obj ) ; depth = <NUM_LIT:0> ; strict = false ; } return new Filter ( ) { @ Override public DocIdSet getDocIdSet ( final IndexReader reader ) throws IOException { if ( reader instanceof BoboIndexReader ) { BoboIndexReader boboReader = ( BoboIndexReader ) reader ; FacetHandler facetHandler = boboReader . getFacetHandler ( field ) ; if ( facetHandler != null && facetHandler instanceof PathFacetHandler ) { BrowseSelection sel = new BrowseSelection ( field ) ; sel . setValues ( new String [ ] { path } ) ; sel . setSelectionProperty ( PathFacetHandler . SEL_PROP_NAME_DEPTH , String . valueOf ( depth ) ) ; sel . setSelectionProperty ( PathFacetHandler . SEL_PROP_NAME_STRICT , String . valueOf ( strict ) ) ; Filter filter = ( ( PathFacetHandler ) facetHandler ) . buildFilter ( sel ) ; if ( filter == null ) return new DocIdSet ( ) { @ Override public boolean isCacheable ( ) { return false ; } @ Override public DocIdSetIterator iterator ( ) throws IOException { return new MatchAllDocIdSetIterator ( reader ) ; } } ; return filter . getDocIdSet ( reader ) ; } } throw new UnsupportedOperationException ( "<STR_LIT>" + field ) ; } } ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . Filter ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . QueryWrapperFilter ; import org . json . JSONObject ; import com . senseidb . search . query . QueryConstructor ; public class QueryFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT:query>" ; private QueryParser _qparser ; public QueryFilterConstructor ( QueryParser qparser ) { _qparser = qparser ; } @ Override protected Filter doConstructFilter ( Object json ) throws Exception { Query q = QueryConstructor . constructQuery ( ( JSONObject ) json , _qparser ) ; if ( q == null ) return null ; return new QueryWrapperFilter ( q ) ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . io . IOException ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . DocIdSet ; import org . apache . lucene . search . Filter ; import org . json . JSONObject ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . api . BrowseSelection ; import com . browseengine . bobo . facets . FacetHandler ; import com . senseidb . conf . SenseiFacetHandlerBuilder ; import com . senseidb . search . facet . UIDFacetHandler ; import com . senseidb . util . RequestConverter2 ; public class UIDFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; @ Override protected Filter doConstructFilter ( Object obj ) throws Exception { final JSONObject json = ( JSONObject ) obj ; return new Filter ( ) { @ Override public DocIdSet getDocIdSet ( IndexReader reader ) throws IOException { if ( reader instanceof BoboIndexReader ) { BoboIndexReader boboReader = ( BoboIndexReader ) reader ; FacetHandler uidHandler = boboReader . getFacetHandler ( SenseiFacetHandlerBuilder . UID_FACET_NAME ) ; if ( uidHandler != null && uidHandler instanceof UIDFacetHandler ) { UIDFacetHandler uidFacet = ( UIDFacetHandler ) uidHandler ; try { String [ ] vals = RequestConverter2 . getStrings ( json . optJSONArray ( VALUES_PARAM ) ) ; String [ ] nots = RequestConverter2 . getStrings ( json . optJSONArray ( EXCLUDES_PARAM ) ) ; BrowseSelection uidSel = new BrowseSelection ( SenseiFacetHandlerBuilder . UID_FACET_NAME ) ; if ( vals != null ) uidSel . setValues ( vals ) ; if ( nots != null ) uidSel . setNotValues ( nots ) ; return uidFacet . buildFilter ( uidSel ) . getDocIdSet ( boboReader ) ; } catch ( Exception e ) { throw new IOException ( e ) ; } } else { throw new IllegalStateException ( "<STR_LIT>" + uidHandler ) ; } } else { throw new IllegalStateException ( "<STR_LIT>" + BoboIndexReader . class ) ; } } } ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . util . Iterator ; import org . apache . lucene . search . Filter ; import org . json . JSONArray ; import org . json . JSONObject ; import com . senseidb . util . RequestConverter2 ; public class TermsFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; @ Override protected Filter doConstructFilter ( Object obj ) throws Exception { JSONObject json = ( JSONObject ) obj ; Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) return null ; String field = iter . next ( ) ; boolean noOptimize = false ; obj = json . get ( field ) ; if ( obj == null ) { throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; } if ( obj instanceof JSONArray ) { JSONArray jsonArray = ( JSONArray ) obj ; String [ ] vals = RequestConverter2 . getStrings ( jsonArray ) ; return new SenseiTermFilter ( field , vals , null , false , noOptimize ) ; } else if ( obj instanceof JSONObject ) { JSONObject jsonObj = ( JSONObject ) obj ; String [ ] vals = RequestConverter2 . getStrings ( jsonObj , VALUES_PARAM ) ; String [ ] notVals = RequestConverter2 . getStrings ( jsonObj , EXCLUDES_PARAM ) ; String op = jsonObj . optString ( OPERATOR_PARAM , OR_PARAM ) ; noOptimize = jsonObj . optBoolean ( NOOPTIMIZE_PARAM , false ) ; boolean isAnd = false ; if ( ! OR_PARAM . equals ( op ) ) { isAnd = true ; } return new SenseiTermFilter ( field , vals , notVals , isAnd , noOptimize ) ; } else { throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; } } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . io . IOException ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . index . Term ; import org . apache . lucene . index . TermDocs ; import org . apache . lucene . search . DocIdSetIterator ; public class TermDocIdSetIterator extends DocIdSetIterator { private final TermDocs termDocs ; private int doc = - <NUM_LIT:1> ; private final int [ ] docs ; private final int [ ] freqs ; private int pointer ; private int pointerMax ; public TermDocIdSetIterator ( Term term , IndexReader reader , int buffer ) throws IOException { termDocs = reader . termDocs ( term ) ; docs = new int [ buffer ] ; freqs = new int [ buffer ] ; } public TermDocIdSetIterator ( Term term , IndexReader reader ) throws IOException { this ( term , reader , <NUM_LIT:32> ) ; } @ Override public int docID ( ) { return doc ; } @ Override public int nextDoc ( ) throws IOException { pointer ++ ; if ( pointer >= pointerMax ) { pointerMax = termDocs . read ( docs , freqs ) ; if ( pointerMax != <NUM_LIT:0> ) { pointer = <NUM_LIT:0> ; } else { termDocs . close ( ) ; return doc = NO_MORE_DOCS ; } } doc = docs [ pointer ] ; return doc ; } @ Override public int advance ( int target ) throws IOException { for ( pointer ++ ; pointer < pointerMax ; pointer ++ ) { if ( docs [ pointer ] >= target ) { return doc = docs [ pointer ] ; } } boolean result = termDocs . skipTo ( target ) ; if ( result ) { pointerMax = <NUM_LIT:1> ; pointer = <NUM_LIT:0> ; docs [ pointer ] = doc = termDocs . doc ( ) ; } else { doc = NO_MORE_DOCS ; } return doc ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . util . Iterator ; import org . apache . lucene . search . Filter ; import org . json . JSONObject ; public class TermFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; @ Override protected Filter doConstructFilter ( Object param ) throws Exception { JSONObject json = ( JSONObject ) param ; Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) return null ; String field = iter . next ( ) ; String text ; boolean noOptimize = false ; Object obj = json . get ( field ) ; if ( obj instanceof JSONObject ) { text = ( ( JSONObject ) obj ) . getString ( VALUE_PARAM ) ; noOptimize = json . optBoolean ( NOOPTIMIZE_PARAM , false ) ; } else { text = String . valueOf ( obj ) ; } return new SenseiTermFilter ( field , new String [ ] { text } , null , false , noOptimize ) ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . io . IOException ; import java . util . ArrayList ; import java . util . Arrays ; import org . apache . log4j . Logger ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . DocIdSet ; import org . apache . lucene . search . DocIdSetIterator ; import org . apache . lucene . search . Filter ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . api . BrowseSelection ; import com . browseengine . bobo . api . BrowseSelection . ValueOperation ; import com . browseengine . bobo . facets . FacetHandler ; import com . browseengine . bobo . facets . data . FacetDataCache ; import com . browseengine . bobo . facets . data . TermValueList ; import com . browseengine . bobo . facets . filter . EmptyFilter ; import com . browseengine . bobo . query . MatchAllDocIdSetIterator ; import com . kamikaze . docidset . impl . AndDocIdSet ; import com . kamikaze . docidset . impl . NotDocIdSet ; import com . kamikaze . docidset . impl . OrDocIdSet ; public class SenseiTermFilter extends Filter { private static final long serialVersionUID = <NUM_LIT:1L> ; private static Logger logger = Logger . getLogger ( SenseiTermFilter . class ) ; private final String _name ; private final String [ ] _vals ; private final String [ ] _not ; private final boolean _isAnd ; private final boolean _noAutoOptimize ; public SenseiTermFilter ( String name , String vals [ ] , String [ ] not , boolean isAnd , boolean noAutoOptimize ) { _name = name ; _vals = vals != null ? vals : new String [ <NUM_LIT:0> ] ; _not = not != null ? not : new String [ <NUM_LIT:0> ] ; _isAnd = isAnd ; _noAutoOptimize = noAutoOptimize ; } private static DocIdSet buildDefaultDocIdSets ( final BoboIndexReader reader , final String name , final String [ ] vals , boolean isAnd ) { if ( vals == null ) return null ; ArrayList < DocIdSet > docSetList = new ArrayList < DocIdSet > ( vals . length ) ; for ( final String val : vals ) { docSetList . add ( new DocIdSet ( ) { @ Override public DocIdSetIterator iterator ( ) throws IOException { return new TermDocIdSetIterator ( new Term ( name , val ) , reader ) ; } } ) ; } if ( docSetList . size ( ) == <NUM_LIT:1> ) { return docSetList . get ( <NUM_LIT:0> ) ; } else if ( docSetList . size ( ) == <NUM_LIT:0> ) return null ; else { if ( isAnd ) { return new AndDocIdSet ( docSetList ) ; } else { return new OrDocIdSet ( docSetList ) ; } } } private static DocIdSet buildLuceneDefaultDocIdSet ( final BoboIndexReader reader , final String name , final String [ ] vals , String [ ] nots , boolean isAnd ) throws IOException { if ( reader . getRuntimeFacetHandlerFactoryMap ( ) . containsKey ( name ) ) { return new DocIdSet ( ) { @ Override public boolean isCacheable ( ) { return false ; } @ Override public DocIdSetIterator iterator ( ) throws IOException { return new MatchAllDocIdSetIterator ( reader ) ; } } ; } DocIdSet positiveSet = null ; DocIdSet negativeSet = null ; if ( vals != null && vals . length > <NUM_LIT:0> ) positiveSet = buildDefaultDocIdSets ( reader , name , vals , isAnd ) ; if ( nots != null && nots . length > <NUM_LIT:0> ) negativeSet = buildDefaultDocIdSets ( reader , name , nots , false ) ; if ( positiveSet != null ) { if ( negativeSet == null ) { return positiveSet ; } else { DocIdSet [ ] sets = new DocIdSet [ ] { positiveSet , new NotDocIdSet ( negativeSet , reader . maxDoc ( ) ) } ; return new AndDocIdSet ( Arrays . asList ( sets ) ) ; } } else { if ( negativeSet == null ) { return EmptyFilter . getInstance ( ) . getRandomAccessDocIdSet ( reader ) ; } else { return new NotDocIdSet ( negativeSet , reader . maxDoc ( ) ) ; } } } @ Override public DocIdSet getDocIdSet ( IndexReader reader ) throws IOException { if ( reader instanceof BoboIndexReader ) { BoboIndexReader boboReader = ( BoboIndexReader ) reader ; FacetHandler facetHandler = ( FacetHandler ) boboReader . getFacetHandler ( _name ) ; Object obj = null ; if ( facetHandler != null ) { obj = facetHandler . getFacetData ( boboReader ) ; if ( _noAutoOptimize && obj != null && obj instanceof FacetDataCache ) { FacetDataCache facetData = ( FacetDataCache ) obj ; TermValueList valArray = facetData . valArray ; ArrayList < String > validVals = new ArrayList < String > ( _vals . length ) ; for ( String val : _vals ) { int idx = valArray . indexOf ( val ) ; if ( idx >= <NUM_LIT:0> ) { validVals . add ( valArray . get ( idx ) ) ; } } return buildLuceneDefaultDocIdSet ( boboReader , _name , validVals . toArray ( new String [ <NUM_LIT:0> ] ) , _not , _isAnd ) ; } BrowseSelection sel = new BrowseSelection ( _name ) ; sel . setValues ( _vals ) ; if ( _not != null ) sel . setNotValues ( _not ) ; if ( _isAnd ) sel . setSelectionOperation ( ValueOperation . ValueOperationAnd ) ; else sel . setSelectionOperation ( ValueOperation . ValueOperationOr ) ; Filter filter = facetHandler . buildFilter ( sel ) ; if ( filter == null ) filter = EmptyFilter . getInstance ( ) ; return filter . getDocIdSet ( boboReader ) ; } else { if ( logger . isDebugEnabled ( ) ) { logger . debug ( "<STR_LIT>" + _name ) ; } return buildLuceneDefaultDocIdSet ( boboReader , _name , _vals , _not , _isAnd ) ; } } else { throw new IllegalStateException ( "<STR_LIT>" + BoboIndexReader . class ) ; } } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . io . IOException ; import java . text . DecimalFormat ; import java . text . DecimalFormatSymbols ; import java . text . SimpleDateFormat ; import java . util . Date ; import java . util . Iterator ; import java . util . Locale ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . DocIdSet ; import org . apache . lucene . search . DocIdSetIterator ; import org . apache . lucene . search . Filter ; import org . apache . lucene . search . TermRangeFilter ; import org . json . JSONObject ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . facets . FacetHandler ; import com . browseengine . bobo . facets . filter . FacetRangeFilter ; import com . browseengine . bobo . query . MatchAllDocIdSetIterator ; import com . senseidb . indexing . DefaultSenseiInterpreter ; import com . senseidb . indexing . MetaType ; public class RangeFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; @ Override protected Filter doConstructFilter ( Object obj ) throws Exception { final JSONObject json = ( JSONObject ) obj ; Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) return null ; final String field = iter . next ( ) ; JSONObject jsonObj = json . getJSONObject ( field ) ; final String from , to ; final boolean include_lower , include_upper ; final boolean noOptimize = jsonObj . optBoolean ( NOOPTIMIZE_PARAM , false ) ; final String type ; final String dateFormat ; String gt = jsonObj . optString ( GT_PARAM , null ) ; String gte = jsonObj . optString ( GTE_PARAM , null ) ; String lt = jsonObj . optString ( LT_PARAM , null ) ; String lte = jsonObj . optString ( LTE_PARAM , null ) ; type = jsonObj . optString ( RANGE_FIELD_TYPE , null ) ; dateFormat = jsonObj . optString ( RANGE_DATE_FORMAT , null ) ; if ( gt != null && gt . length ( ) != <NUM_LIT:0> ) { from = gt ; include_lower = false ; } else if ( gte != null && gte . length ( ) != <NUM_LIT:0> ) { from = gte ; include_lower = true ; } else { from = jsonObj . optString ( FROM_PARAM , null ) ; include_lower = jsonObj . optBoolean ( INCLUDE_LOWER_PARAM , true ) ; } if ( lt != null && lt . length ( ) != <NUM_LIT:0> ) { to = lt ; include_upper = false ; } else if ( lte != null && lte . length ( ) != <NUM_LIT:0> ) { to = lte ; include_upper = true ; } else { to = jsonObj . optString ( TO_PARAM , null ) ; include_upper = jsonObj . optBoolean ( INCLUDE_UPPER_PARAM , true ) ; } return new Filter ( ) { @ Override public DocIdSet getDocIdSet ( final IndexReader reader ) throws IOException { String fromPadded = from , toPadded = to ; if ( ! noOptimize ) { if ( reader instanceof BoboIndexReader ) { BoboIndexReader boboReader = ( BoboIndexReader ) reader ; FacetHandler facetHandler = boboReader . getFacetHandler ( field ) ; if ( facetHandler != null ) { StringBuilder sb = new StringBuilder ( ) ; if ( include_lower && from != null && from . length ( ) != <NUM_LIT:0> ) sb . append ( "<STR_LIT:[>" ) ; else sb . append ( "<STR_LIT:(>" ) ; if ( from == null || from . length ( ) == <NUM_LIT:0> ) sb . append ( "<STR_LIT:*>" ) ; else sb . append ( from ) ; sb . append ( "<STR_LIT>" ) ; if ( to == null || to . length ( ) == <NUM_LIT:0> ) sb . append ( "<STR_LIT:*>" ) ; else sb . append ( to ) ; if ( include_upper && to != null && to . length ( ) != <NUM_LIT:0> ) sb . append ( "<STR_LIT:]>" ) ; else sb . append ( "<STR_LIT:)>" ) ; FacetRangeFilter filter = new FacetRangeFilter ( facetHandler , sb . toString ( ) ) ; return filter . getDocIdSet ( reader ) ; } } } if ( type == null ) throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; if ( "<STR_LIT:int>" . equals ( type ) ) { MetaType metaType = DefaultSenseiInterpreter . CLASS_METATYPE_MAP . get ( int . class ) ; String formatString = DefaultSenseiInterpreter . DEFAULT_FORMAT_STRING_MAP . get ( metaType ) ; DecimalFormat formatter = new DecimalFormat ( formatString , new DecimalFormatSymbols ( Locale . US ) ) ; fromPadded = formatter . format ( Integer . parseInt ( from ) ) ; toPadded = formatter . format ( Integer . parseInt ( to ) ) ; } else if ( "<STR_LIT>" . equals ( type ) ) { MetaType metaType = DefaultSenseiInterpreter . CLASS_METATYPE_MAP . get ( short . class ) ; String formatString = DefaultSenseiInterpreter . DEFAULT_FORMAT_STRING_MAP . get ( metaType ) ; DecimalFormat formatter = new DecimalFormat ( formatString , new DecimalFormatSymbols ( Locale . US ) ) ; fromPadded = formatter . format ( Short . parseShort ( from ) ) ; toPadded = formatter . format ( Short . parseShort ( to ) ) ; } else if ( "<STR_LIT:long>" . equals ( type ) ) { MetaType metaType = DefaultSenseiInterpreter . CLASS_METATYPE_MAP . get ( long . class ) ; String formatString = DefaultSenseiInterpreter . DEFAULT_FORMAT_STRING_MAP . get ( metaType ) ; DecimalFormat formatter = new DecimalFormat ( formatString , new DecimalFormatSymbols ( Locale . US ) ) ; fromPadded = formatter . format ( Long . parseLong ( from ) ) ; toPadded = formatter . format ( Long . parseLong ( to ) ) ; } else if ( "<STR_LIT:date>" . equals ( type ) ) { if ( dateFormat == null ) throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; else { SimpleDateFormat formatter = new SimpleDateFormat ( dateFormat ) ; fromPadded = formatter . format ( new Date ( Long . parseLong ( from ) ) ) ; toPadded = formatter . format ( new Date ( Long . parseLong ( to ) ) ) ; } } else if ( "<STR_LIT:float>" . equals ( type ) ) { MetaType metaType = DefaultSenseiInterpreter . CLASS_METATYPE_MAP . get ( float . class ) ; String formatString = DefaultSenseiInterpreter . DEFAULT_FORMAT_STRING_MAP . get ( metaType ) ; DecimalFormat formatter = new DecimalFormat ( formatString , new DecimalFormatSymbols ( Locale . US ) ) ; fromPadded = formatter . format ( Float . parseFloat ( from ) ) ; toPadded = formatter . format ( Float . parseFloat ( to ) ) ; } else if ( "<STR_LIT:double>" . equals ( type ) ) { MetaType metaType = DefaultSenseiInterpreter . CLASS_METATYPE_MAP . get ( double . class ) ; String formatString = DefaultSenseiInterpreter . DEFAULT_FORMAT_STRING_MAP . get ( metaType ) ; DecimalFormat formatter = new DecimalFormat ( formatString , new DecimalFormatSymbols ( Locale . US ) ) ; fromPadded = formatter . format ( Double . parseDouble ( from ) ) ; toPadded = formatter . format ( Double . parseDouble ( to ) ) ; } if ( fromPadded == null || fromPadded . length ( ) == <NUM_LIT:0> ) if ( toPadded == null || toPadded . length ( ) == <NUM_LIT:0> ) return new DocIdSet ( ) { @ Override public boolean isCacheable ( ) { return false ; } @ Override public DocIdSetIterator iterator ( ) throws IOException { return new MatchAllDocIdSetIterator ( reader ) ; } } ; else return new TermRangeFilter ( field , fromPadded , toPadded , false , include_upper ) . getDocIdSet ( reader ) ; else if ( toPadded == null || toPadded . length ( ) == <NUM_LIT:0> ) return new TermRangeFilter ( field , fromPadded , toPadded , include_lower , false ) . getDocIdSet ( reader ) ; return new TermRangeFilter ( field , fromPadded , toPadded , include_lower , include_upper ) . getDocIdSet ( reader ) ; } } ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . util . ArrayList ; import java . util . List ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . Filter ; import org . json . JSONArray ; import com . browseengine . bobo . facets . filter . AndFilter ; public class AndFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; private QueryParser _qparser ; public AndFilterConstructor ( QueryParser qparser ) { _qparser = qparser ; } @ Override protected Filter doConstructFilter ( Object obj ) throws Exception { JSONArray filterArray = ( JSONArray ) obj ; List < Filter > filters = new ArrayList < Filter > ( filterArray . length ( ) ) ; for ( int i = <NUM_LIT:0> ; i < filterArray . length ( ) ; ++ i ) { Filter filter = FilterConstructor . constructFilter ( filterArray . getJSONObject ( i ) , _qparser ) ; if ( filter != null ) filters . add ( filter ) ; } return new AndFilter ( filters ) ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . util . ArrayList ; import java . util . List ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . Filter ; import org . json . JSONArray ; import org . json . JSONObject ; import com . browseengine . bobo . facets . filter . AndFilter ; import com . browseengine . bobo . facets . filter . NotFilter ; import com . browseengine . bobo . facets . filter . OrFilter ; public class BooleanFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; private QueryParser _qparser ; public BooleanFilterConstructor ( QueryParser qparser ) { _qparser = qparser ; } @ Override protected Filter doConstructFilter ( Object param ) throws Exception { JSONObject json = ( JSONObject ) param ; Object obj = json . opt ( MUST_PARAM ) ; List < Filter > andFilters = new ArrayList < Filter > ( ) ; if ( obj != null ) { if ( obj instanceof JSONArray ) { for ( int i = <NUM_LIT:0> ; i < ( ( JSONArray ) obj ) . length ( ) ; ++ i ) { andFilters . add ( FilterConstructor . constructFilter ( ( ( JSONArray ) obj ) . getJSONObject ( i ) , _qparser ) ) ; } } else if ( obj instanceof JSONObject ) { andFilters . add ( FilterConstructor . constructFilter ( ( JSONObject ) obj , _qparser ) ) ; } } obj = json . opt ( MUST_NOT_PARAM ) ; if ( obj != null ) { if ( obj instanceof JSONArray ) { for ( int i = <NUM_LIT:0> ; i < ( ( JSONArray ) obj ) . length ( ) ; ++ i ) { andFilters . add ( new NotFilter ( FilterConstructor . constructFilter ( ( ( JSONArray ) obj ) . getJSONObject ( i ) , _qparser ) ) ) ; } } else if ( obj instanceof JSONObject ) { andFilters . add ( new NotFilter ( FilterConstructor . constructFilter ( ( JSONObject ) obj , _qparser ) ) ) ; } } JSONArray array = json . optJSONArray ( SHOULD_PARAM ) ; if ( array != null ) { List < Filter > orFilters = new ArrayList < Filter > ( array . length ( ) ) ; for ( int i = <NUM_LIT:0> ; i < array . length ( ) ; ++ i ) { orFilters . add ( FilterConstructor . constructFilter ( array . getJSONObject ( i ) , _qparser ) ) ; } if ( orFilters . size ( ) > <NUM_LIT:0> ) andFilters . add ( new OrFilter ( orFilters ) ) ; } if ( andFilters . size ( ) > <NUM_LIT:0> ) return new AndFilter ( andFilters ) ; else return null ; } } </s>
|
<s> package com . senseidb . search . query . filters ; import java . util . ArrayList ; import java . util . List ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . Filter ; import org . json . JSONArray ; import org . json . JSONObject ; import com . browseengine . bobo . facets . filter . OrFilter ; public class OrFilterConstructor extends FilterConstructor { public static final String FILTER_TYPE = "<STR_LIT>" ; private QueryParser _qparser ; public OrFilterConstructor ( QueryParser qparser ) { _qparser = qparser ; } @ Override protected Filter doConstructFilter ( Object obj ) throws Exception { JSONArray filterArray = ( JSONArray ) obj ; List < Filter > filters = new ArrayList < Filter > ( filterArray . length ( ) ) ; for ( int i = <NUM_LIT:0> ; i < filterArray . length ( ) ; ++ i ) { Filter filter = FilterConstructor . constructFilter ( filterArray . getJSONObject ( i ) , _qparser ) ; if ( filter != null ) filters . add ( filter ) ; } return new OrFilter ( filters ) ; } } </s>
|
<s> package com . senseidb . search . query ; import java . io . IOException ; import java . text . DecimalFormat ; import java . util . ArrayList ; import java . util . List ; import java . util . concurrent . TimeUnit ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . DocIdSet ; import org . apache . lucene . search . Filter ; import proj . zoie . api . ZoieIndexReader ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . api . BrowseSelection ; import com . browseengine . bobo . facets . FacetHandler ; import com . kamikaze . docidset . impl . OrDocIdSet ; import com . senseidb . indexing . DefaultSenseiInterpreter ; import com . senseidb . indexing . MetaType ; public class TimeRetentionFilter extends Filter { private final String _column ; private final int _nDays ; private final TimeUnit _dataUnit ; static { DefaultSenseiInterpreter . DEFAULT_FORMAT_STRING_MAP . get ( MetaType . Long ) ; } public TimeRetentionFilter ( String column , int nDays , TimeUnit dataUnit ) { _column = column ; _nDays = nDays ; _dataUnit = dataUnit ; } private DocIdSet buildFilterSet ( BoboIndexReader boboReader ) throws IOException { FacetHandler facetHandler = boboReader . getFacetHandler ( _column ) ; if ( facetHandler != null ) { DecimalFormat formatter = new DecimalFormat ( DefaultSenseiInterpreter . DEFAULT_FORMAT_STRING_MAP . get ( MetaType . Long ) ) ; BrowseSelection sel = new BrowseSelection ( _column ) ; long duration = _dataUnit . convert ( _nDays , TimeUnit . DAYS ) ; long now = _dataUnit . convert ( System . currentTimeMillis ( ) , TimeUnit . MILLISECONDS ) ; long from = now - duration ; sel . addValue ( "<STR_LIT:[>" + formatter . format ( from ) + "<STR_LIT>" ) ; return facetHandler . buildFilter ( sel ) . getDocIdSet ( boboReader ) ; } throw new IllegalStateException ( "<STR_LIT>" + _column ) ; } @ Override public DocIdSet getDocIdSet ( IndexReader reader ) throws IOException { if ( reader instanceof ZoieIndexReader ) { ZoieIndexReader < BoboIndexReader > zoieReader = ( ZoieIndexReader < BoboIndexReader > ) reader ; List < BoboIndexReader > decorated = zoieReader . getDecoratedReaders ( ) ; List < DocIdSet > docIdSetList = new ArrayList < DocIdSet > ( decorated . size ( ) ) ; for ( BoboIndexReader bobo : decorated ) { docIdSetList . add ( buildFilterSet ( bobo ) ) ; } return new OrDocIdSet ( docIdSetList ) ; } else { throw new IllegalStateException ( "<STR_LIT>" + ZoieIndexReader . class ) ; } } } </s>
|
<s> package com . senseidb . search . query ; import java . util . ArrayList ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . spans . SpanNearQuery ; import org . apache . lucene . search . spans . SpanQuery ; import org . apache . lucene . search . spans . SpanTermQuery ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; public class SpanNearQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { String fieldCheck = null ; JSONArray jsonArray = jsonQuery . getJSONArray ( CLAUSES_PARAM ) ; ArrayList < SpanTermQuery > clausesList = new ArrayList < SpanTermQuery > ( ) ; for ( int i = <NUM_LIT:0> ; i < jsonArray . length ( ) ; i ++ ) { JSONObject json = jsonArray . getJSONObject ( i ) . getJSONObject ( SPAN_TERM_PARAM ) ; String field = ( String ) ( json . keys ( ) . next ( ) ) ; if ( fieldCheck == null ) fieldCheck = field ; else if ( ! fieldCheck . equals ( field ) ) throw new IllegalArgumentException ( "<STR_LIT>" + jsonQuery ) ; String value = ( String ) json . get ( field ) ; clausesList . add ( new SpanTermQuery ( new Term ( field , value ) ) ) ; } SpanQuery [ ] clauses = clausesList . toArray ( new SpanQuery [ clausesList . size ( ) ] ) ; int slop = jsonQuery . optInt ( SLOP_PARAM , <NUM_LIT:12> ) ; boolean inOrder = jsonQuery . optBoolean ( IN_ORDER_PARAM , false ) ; boolean collectPayloads = jsonQuery . optBoolean ( COLLECT_PAYLOADS_PARAM , false ) ; return new SpanNearQuery ( clauses , slop , inOrder , collectPayloads ) ; } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . spans . SpanNotQuery ; import org . apache . lucene . search . spans . SpanQuery ; import org . apache . lucene . search . spans . SpanTermQuery ; import org . json . JSONException ; import org . json . JSONObject ; public class SpanNotQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { JSONObject jsonInclude = jsonQuery . getJSONObject ( INCLUDE_PARAM ) ; JSONObject jsonExclude = jsonQuery . getJSONObject ( EXCLUDE_PARAM ) ; JSONObject jsonInc = jsonInclude . getJSONObject ( SPAN_TERM_PARAM ) ; String fieldInc = ( String ) ( jsonInc . keys ( ) . next ( ) ) ; String valueInc = ( String ) jsonInc . get ( fieldInc ) ; SpanQuery sInc = new SpanTermQuery ( new Term ( fieldInc , valueInc ) ) ; JSONObject jsonExc = jsonExclude . getJSONObject ( SPAN_TERM_PARAM ) ; String fieldExc = ( String ) ( jsonExc . keys ( ) . next ( ) ) ; String valueExc = ( String ) jsonExc . get ( fieldExc ) ; SpanQuery sExc = new SpanTermQuery ( new Term ( fieldExc , valueExc ) ) ; if ( ! fieldInc . equals ( fieldExc ) ) throw new IllegalArgumentException ( "<STR_LIT>" + jsonQuery ) ; SpanNotQuery snq = new SpanNotQuery ( sInc , sExc ) ; return snq ; } } </s>
|
<s> package com . senseidb . search . query ; import java . util . Iterator ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . spans . SpanTermQuery ; import org . json . JSONException ; import org . json . JSONObject ; public class SpanTermQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject json ) throws JSONException { Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; String field = iter . next ( ) ; String spanterm = null ; Object obj = json . get ( field ) ; if ( obj instanceof JSONObject ) { spanterm = ( ( JSONObject ) obj ) . optString ( VALUE_PARAM , "<STR_LIT>" ) ; float boost = ( float ) ( ( JSONObject ) obj ) . optDouble ( BOOST_PARAM , <NUM_LIT> ) ; Query query = new SpanTermQuery ( new Term ( field , spanterm ) ) ; query . setBoost ( boost ) ; return query ; } else { spanterm = String . valueOf ( obj ) ; return new SpanTermQuery ( new Term ( field , spanterm ) ) ; } } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . BooleanClause ; import org . apache . lucene . search . BooleanQuery ; import org . apache . lucene . search . Query ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; public class BooleanQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; private QueryParser _qparser ; public BooleanQueryConstructor ( QueryParser qparser ) { _qparser = qparser ; } @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { BooleanQuery query = new BooleanQuery ( jsonQuery . optBoolean ( DISABLE_COORD_PARAM , false ) ) ; Object obj = jsonQuery . opt ( MUST_PARAM ) ; if ( obj != null ) { if ( obj instanceof JSONArray ) { for ( int i = <NUM_LIT:0> ; i < ( ( JSONArray ) obj ) . length ( ) ; ++ i ) { query . add ( QueryConstructor . constructQuery ( ( ( JSONArray ) obj ) . getJSONObject ( i ) , _qparser ) , BooleanClause . Occur . MUST ) ; } } else if ( obj instanceof JSONObject ) { query . add ( QueryConstructor . constructQuery ( ( JSONObject ) obj , _qparser ) , BooleanClause . Occur . MUST ) ; } } obj = jsonQuery . opt ( MUST_NOT_PARAM ) ; if ( obj != null ) { if ( obj instanceof JSONArray ) { for ( int i = <NUM_LIT:0> ; i < ( ( JSONArray ) obj ) . length ( ) ; ++ i ) { query . add ( QueryConstructor . constructQuery ( ( ( JSONArray ) obj ) . getJSONObject ( i ) , _qparser ) , BooleanClause . Occur . MUST_NOT ) ; } } else if ( obj instanceof JSONObject ) { query . add ( QueryConstructor . constructQuery ( ( JSONObject ) obj , _qparser ) , BooleanClause . Occur . MUST_NOT ) ; } } JSONArray array = jsonQuery . optJSONArray ( SHOULD_PARAM ) ; if ( array != null && array . length ( ) > <NUM_LIT:0> ) { for ( int i = <NUM_LIT:0> ; i < array . length ( ) ; ++ i ) { query . add ( QueryConstructor . constructQuery ( array . getJSONObject ( i ) , _qparser ) , BooleanClause . Occur . SHOULD ) ; } query . setMinimumNumberShouldMatch ( jsonQuery . optInt ( MINIMUM_NUMBER_SHOULD_MATCH_PARAM , <NUM_LIT:0> ) ) ; } query . setBoost ( ( float ) jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; return query ; } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . search . ConstantScoreQuery ; import org . apache . lucene . search . Filter ; import org . apache . lucene . search . Query ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . search . query . filters . FilterConstructor ; public class PathQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT:path>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { Filter filter = null ; try { JSONObject newJson = new JSONObject ( ) ; newJson . put ( QUERY_TYPE , jsonQuery ) ; filter = FilterConstructor . constructFilter ( newJson , null ) ; } catch ( Exception e ) { throw new JSONException ( e ) ; } ConstantScoreQuery query = new ConstantScoreQuery ( filter ) ; Object obj = jsonQuery . get ( ( String ) jsonQuery . keys ( ) . next ( ) ) ; if ( obj instanceof JSONObject ) { query . setBoost ( ( float ) ( ( JSONObject ) obj ) . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ) ; } return query ; } } </s>
|
<s> package com . senseidb . search . query ; import java . util . Iterator ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . BooleanClause ; import org . apache . lucene . search . BooleanQuery ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . TermQuery ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; public class TermsQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject json ) throws JSONException { String field = null ; JSONArray values = null , excludes = null ; String op = null ; int minimum_match = <NUM_LIT:1> ; float boost = <NUM_LIT:1.0f> ; Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; field = iter . next ( ) ; Object obj = json . get ( field ) ; if ( obj instanceof JSONObject ) { values = ( ( JSONObject ) obj ) . optJSONArray ( VALUES_PARAM ) ; excludes = ( ( JSONObject ) obj ) . optJSONArray ( EXCLUDES_PARAM ) ; op = ( ( JSONObject ) obj ) . optString ( OPERATOR_PARAM ) ; minimum_match = ( ( JSONObject ) obj ) . optInt ( MINIMUM_MATCH_PARAM , <NUM_LIT:0> ) ; boost = ( float ) ( ( JSONObject ) obj ) . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ; } else if ( obj instanceof JSONArray ) { values = ( JSONArray ) obj ; } BooleanQuery query = new BooleanQuery ( ) ; if ( values != null && values . length ( ) > <NUM_LIT:0> ) { for ( int i = <NUM_LIT:0> ; i < values . length ( ) ; ++ i ) { if ( AND_PARAM . equals ( op ) ) { query . add ( new TermQuery ( new Term ( field , values . getString ( i ) ) ) , BooleanClause . Occur . MUST ) ; } else { query . add ( new TermQuery ( new Term ( field , values . getString ( i ) ) ) , BooleanClause . Occur . SHOULD ) ; } } if ( ! AND_PARAM . equals ( op ) ) query . setMinimumNumberShouldMatch ( minimum_match ) ; } if ( excludes != null ) { for ( int i = <NUM_LIT:0> ; i < excludes . length ( ) ; ++ i ) { query . add ( new TermQuery ( new Term ( field , excludes . getString ( i ) ) ) , BooleanClause . Occur . MUST_NOT ) ; } } query . setBoost ( boost ) ; return query ; } } </s>
|
<s> package com . senseidb . search . query ; import java . io . IOException ; import java . util . Set ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . Explanation ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . Scorer ; import org . apache . lucene . search . Searcher ; import org . apache . lucene . search . Weight ; public abstract class AbstractScoreAdjuster extends Query { private static final long serialVersionUID = <NUM_LIT:1L> ; public class ScoreAdjusterWeight extends Weight { private static final long serialVersionUID = <NUM_LIT:1L> ; Weight _innerWeight ; public ScoreAdjusterWeight ( Weight innerWeight ) throws IOException { _innerWeight = innerWeight ; } public String toString ( ) { return "<STR_LIT>" + AbstractScoreAdjuster . this + "<STR_LIT:)>" ; } public Query getQuery ( ) { return _innerWeight . getQuery ( ) ; } public float getValue ( ) { return _innerWeight . getValue ( ) ; } public float sumOfSquaredWeights ( ) throws IOException { return _innerWeight . sumOfSquaredWeights ( ) ; } public void normalize ( float queryNorm ) { _innerWeight . normalize ( queryNorm ) ; } @ Override public Scorer scorer ( IndexReader reader , boolean scoreDocsInOrder , boolean topScorer ) throws IOException { Scorer innerScorer = _innerWeight . scorer ( reader , scoreDocsInOrder , topScorer ) ; return createScorer ( innerScorer , reader , scoreDocsInOrder , topScorer ) ; } public Explanation explain ( IndexReader reader , int doc ) throws IOException { Explanation innerExplain = _innerWeight . explain ( reader , doc ) ; return createExplain ( innerExplain , reader , doc ) ; } } protected final Query _query ; public AbstractScoreAdjuster ( Query query ) { _query = query ; } protected abstract Scorer createScorer ( Scorer innerScorer , IndexReader reader , boolean scoreDocsInOrder , boolean topScorer ) throws IOException ; protected Explanation createExplain ( Explanation innerExplain , IndexReader reader , int doc ) throws IOException { return innerExplain ; } @ Override public Weight createWeight ( Searcher searcher ) throws IOException { return new ScoreAdjusterWeight ( _query . createWeight ( searcher ) ) ; } @ Override public Query rewrite ( IndexReader reader ) throws IOException { _query . rewrite ( reader ) ; return this ; } @ Override public String toString ( String field ) { return _query . toString ( field ) ; } @ Override public void extractTerms ( Set terms ) { _query . extractTerms ( terms ) ; } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . search . MatchAllDocsQuery ; import org . apache . lucene . search . Query ; import org . json . JSONException ; import org . json . JSONObject ; public class MatchAllQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { double boost = jsonQuery . optDouble ( BOOST_PARAM , <NUM_LIT:1.0> ) ; MatchAllDocsQuery q = new MatchAllDocsQuery ( ) ; q . setBoost ( ( float ) boost ) ; return q ; } } </s>
|
<s> package com . senseidb . search . query ; import java . util . Iterator ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . WildcardQuery ; import org . json . JSONException ; import org . json . JSONObject ; public class WildcardQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject json ) throws JSONException { Iterator < String > iter = json . keys ( ) ; if ( ! iter . hasNext ( ) ) throw new IllegalArgumentException ( "<STR_LIT>" + json ) ; String field = iter . next ( ) ; String wildcard = null ; Object obj = json . get ( field ) ; if ( obj instanceof JSONObject ) { wildcard = ( ( JSONObject ) obj ) . optString ( VALUE_PARAM , "<STR_LIT>" ) ; float boost = ( float ) ( ( JSONObject ) obj ) . optDouble ( BOOST_PARAM , <NUM_LIT> ) ; Query query = new WildcardQuery ( new Term ( field , wildcard ) ) ; query . setBoost ( boost ) ; return query ; } else { wildcard = String . valueOf ( obj ) ; return new WildcardQuery ( new Term ( field , wildcard ) ) ; } } } </s>
|
<s> package com . senseidb . search . query ; import org . apache . lucene . index . Term ; import org . apache . lucene . search . Query ; import org . apache . lucene . search . spans . SpanFirstQuery ; import org . apache . lucene . search . spans . SpanQuery ; import org . apache . lucene . search . spans . SpanTermQuery ; import org . json . JSONException ; import org . json . JSONObject ; public class SpanFirstQueryConstructor extends QueryConstructor { public static final String QUERY_TYPE = "<STR_LIT>" ; @ Override protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException { JSONObject spanJson = jsonQuery . getJSONObject ( MATCH_PARAM ) . getJSONObject ( SPAN_TERM_PARAM ) ; String field = ( String ) ( spanJson . keys ( ) . next ( ) ) ; String spanterm = ( String ) spanJson . getString ( field ) ; SpanQuery sq = new SpanTermQuery ( new Term ( field , spanterm ) ) ; int end = jsonQuery . optInt ( END_PARAM , <NUM_LIT:3> ) ; return new SpanFirstQuery ( sq , end ) ; } } </s>
|
<s> package com . senseidb . search . query ; import java . util . HashMap ; import java . util . Iterator ; import java . util . Map ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . Query ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . search . query . ScoreAugmentQuery . ScoreAugmentFunction ; import com . senseidb . search . relevance . RelevanceFunctionBuilder ; import com . senseidb . search . relevance . impl . RelevanceJSONConstants ; public abstract class QueryConstructor { public static final String VALUES_PARAM = "<STR_LIT>" ; public static final String EXCLUDES_PARAM = "<STR_LIT>" ; public static final String OPERATOR_PARAM = "<STR_LIT>" ; public static final String PARAMS_PARAM = "<STR_LIT>" ; public static final String MUST_PARAM = "<STR_LIT>" ; public static final String MUST_NOT_PARAM = "<STR_LIT>" ; public static final String SHOULD_PARAM = "<STR_LIT>" ; public static final String BOOST_PARAM = "<STR_LIT>" ; public static final String DISABLE_COORD_PARAM = "<STR_LIT>" ; public static final String MINIMUM_NUMBER_SHOULD_MATCH_PARAM = "<STR_LIT>" ; public static final String TYPE_PARAM = "<STR_LIT:type>" ; public static final String PHRASE_PARAM = "<STR_LIT>" ; public static final String PHRASE_PREFIX_PARAM = "<STR_LIT>" ; public static final String AND_PARAM = "<STR_LIT>" ; public static final String INCLUDE_PARAM = "<STR_LIT>" ; public static final String EXCLUDE_PARAM = "<STR_LIT>" ; public static final String SPAN_TERM_PARAM = "<STR_LIT>" ; public static final String MATCH_PARAM = "<STR_LIT>" ; public static final String END_PARAM = "<STR_LIT>" ; public static final String CLASS_PARAM = "<STR_LIT:class>" ; public static final String MINIMUM_MATCH_PARAM = "<STR_LIT>" ; public static final String CLAUSES_PARAM = "<STR_LIT>" ; public static final String SLOP_PARAM = "<STR_LIT>" ; public static final String IN_ORDER_PARAM = "<STR_LIT>" ; public static final String COLLECT_PAYLOADS_PARAM = "<STR_LIT>" ; public static final String VALUE_PARAM = "<STR_LIT:value>" ; public static final String QUERY_PARAM = "<STR_LIT:query>" ; public static final String FILTER_PARAM = "<STR_LIT>" ; public static final String QUERIES_PARAM = "<STR_LIT>" ; public static final String TERM_PARAM = "<STR_LIT>" ; public static final String TIE_BREAKER_PARAM = "<STR_LIT>" ; public static final String DEFAULT_FIELD_PARAM = "<STR_LIT>" ; public static final String FIELDS_PARAM = "<STR_LIT>" ; public static final String DEFAULT_OPERATOR_PARAM = "<STR_LIT>" ; public static final String ALLOW_LEADING_WILDCARD_PARAM = "<STR_LIT>" ; public static final String LOWERCASE_EXPANDED_TERMS_PARAM = "<STR_LIT>" ; public static final String ENABLE_POSITION_INCREMENTS_PARAM = "<STR_LIT>" ; public static final String FUZZY_PREFIX_LENGTH_PARAM = "<STR_LIT>" ; public static final String FUZZY_MIN_SIM_PARAM = "<STR_LIT>" ; public static final String PHRASE_SLOP_PARAM = "<STR_LIT>" ; public static final String ANALYZE_WILDCARD_PARAM = "<STR_LIT>" ; public static final String AUTO_GENERATE_PHRASE_QUERIES_PARAM = "<STR_LIT>" ; public static final String USE_DIS_MAX_PARAM = "<STR_LIT>" ; public static final String RELEVANCE = "<STR_LIT>" ; private static final Map < String , QueryConstructor > QUERY_CONSTRUCTOR_MAP = new HashMap < String , QueryConstructor > ( ) ; static { QUERY_CONSTRUCTOR_MAP . put ( DistMaxQueryConstructor . QUERY_TYPE , new DistMaxQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( PrefixQueryConstructor . QUERY_TYPE , new PrefixQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( WildcardQueryConstructor . QUERY_TYPE , new WildcardQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( RangeQueryConstructor . QUERY_TYPE , new RangeQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( PathQueryConstructor . QUERY_TYPE , new PathQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( SpanFirstQueryConstructor . QUERY_TYPE , new SpanFirstQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( SpanNearQueryConstructor . QUERY_TYPE , new SpanNearQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( SpanNotQueryConstructor . QUERY_TYPE , new SpanNotQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( SpanOrQueryConstructor . QUERY_TYPE , new SpanOrQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( SpanTermQueryConstructor . QUERY_TYPE , new SpanTermQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( MatchAllQueryConstructor . QUERY_TYPE , new MatchAllQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( TermQueryConstructor . QUERY_TYPE , new TermQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( TermsQueryConstructor . QUERY_TYPE , new TermsQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( UIDQueryConstructor . QUERY_TYPE , new UIDQueryConstructor ( ) ) ; QUERY_CONSTRUCTOR_MAP . put ( CustomQueryConstructor . QUERY_TYPE , new CustomQueryConstructor ( ) ) ; } public static QueryConstructor getQueryConstructor ( String type , QueryParser qparser ) { QueryConstructor queryConstructor = QUERY_CONSTRUCTOR_MAP . get ( type ) ; if ( queryConstructor == null ) { if ( QueryStringQueryConstructor . QUERY_TYPE . equals ( type ) ) queryConstructor = new QueryStringQueryConstructor ( qparser ) ; else if ( TextQueryConstructor . QUERY_TYPE . equals ( type ) ) queryConstructor = new TextQueryConstructor ( qparser ) ; else if ( BooleanQueryConstructor . QUERY_TYPE . equals ( type ) ) queryConstructor = new BooleanQueryConstructor ( qparser ) ; else if ( FilteredQueryConstructor . QUERY_TYPE . equals ( type ) ) queryConstructor = new FilteredQueryConstructor ( qparser ) ; } return queryConstructor ; } public static Query constructQuery ( JSONObject jsonQuery , QueryParser qparser ) throws JSONException { if ( jsonQuery == null ) return null ; Iterator < String > iter = jsonQuery . keys ( ) ; if ( ! iter . hasNext ( ) ) return null ; String type = iter . next ( ) ; QueryConstructor queryConstructor = QueryConstructor . getQueryConstructor ( type , qparser ) ; if ( queryConstructor == null ) throw new IllegalArgumentException ( "<STR_LIT>" + type + "<STR_LIT>" ) ; JSONObject jsonValue = jsonQuery . getJSONObject ( type ) ; Query baseQuery = queryConstructor . doConstructQuery ( jsonValue ) ; JSONObject jsonRelevance = null ; if ( jsonValue . has ( RELEVANCE ) ) { jsonRelevance = jsonValue . optJSONObject ( RELEVANCE ) ; if ( jsonRelevance == null ) throw new JSONException ( "<STR_LIT>" ) ; } if ( jsonRelevance == null ) { return baseQuery ; } else { ScoreAugmentFunction func = RelevanceFunctionBuilder . build ( jsonRelevance ) ; JSONObject valuesJson = jsonRelevance . optJSONObject ( RelevanceJSONConstants . KW_VALUES ) ; if ( func == null ) throw new JSONException ( "<STR_LIT>" ) ; return new ScoreAugmentQuery ( baseQuery , func , valuesJson ) ; } } abstract protected Query doConstructQuery ( JSONObject jsonQuery ) throws JSONException ; } </s>
|
<s> package com . senseidb . search . plugin ; import java . util . Comparator ; import java . util . List ; import java . util . Set ; import org . apache . lucene . index . IndexReader ; import org . json . JSONObject ; import com . browseengine . bobo . facets . FacetHandler ; import com . senseidb . conf . SenseiSchema ; import com . senseidb . indexing . ShardingStrategy ; import com . senseidb . plugin . SenseiPluginRegistry ; import com . senseidb . search . node . SenseiCore ; public interface PluggableSearchEngine { public void init ( String indexDirectory , int nodeId , SenseiSchema senseiSchema , Comparator < String > versionComparator , SenseiPluginRegistry pluginRegistry , ShardingStrategy shardingStrategy ) ; public String getVersion ( ) ; public JSONObject acceptEvent ( JSONObject event , String version ) ; public boolean acceptEventsForAllPartitions ( ) ; public Set < String > getFieldNames ( ) ; public Set < String > getFacetNames ( ) ; public List < FacetHandler < ? > > createFacetHandlers ( ) ; public void onDelete ( IndexReader indexReader , long ... uids ) ; public void start ( SenseiCore senseiCore ) ; public void stop ( ) ; } </s>
|
<s> package com . senseidb . search . plugin ; import java . util . ArrayList ; import java . util . Arrays ; import java . util . Comparator ; import java . util . HashSet ; import java . util . List ; import java . util . Set ; import org . apache . log4j . Logger ; import org . apache . lucene . index . IndexReader ; import org . json . JSONException ; import org . json . JSONObject ; import proj . zoie . api . Zoie ; import proj . zoie . api . ZoieIndexReader ; import proj . zoie . api . ZoieMultiReader ; import proj . zoie . api . ZoieSegmentReader ; import proj . zoie . hourglass . impl . HourglassListener ; import com . browseengine . bobo . facets . FacetHandler ; import com . senseidb . conf . SenseiSchema ; import com . senseidb . indexing . ShardingStrategy ; import com . senseidb . indexing . activity . CompositeActivityManager ; import com . senseidb . indexing . activity . deletion . DeletionListener ; import com . senseidb . plugin . SenseiPluginRegistry ; import com . senseidb . search . node . SenseiCore ; public class PluggableSearchEngineManager implements DeletionListener , HourglassListener < IndexReader , IndexReader > { private final static Logger logger = Logger . getLogger ( PluggableSearchEngineManager . class ) ; private ShardingStrategy shardingStrategy ; private SenseiCore senseiCore ; private String version ; private Comparator < String > versionComparator ; private SenseiSchema senseiSchema ; private SenseiPluginRegistry pluginRegistry ; private int nodeId ; private List < PluggableSearchEngine > pluggableEngines = new ArrayList < PluggableSearchEngine > ( ) ; private int maxPartition ; private boolean acceptEventsForAllPartitions ; public PluggableSearchEngineManager ( ) { } public String getOldestVersion ( ) { return version ; } public boolean acceptEventsForAllPartitions ( ) { return acceptEventsForAllPartitions ; } public final void init ( String indexDirectory , int nodeId , SenseiSchema senseiSchema , Comparator < String > versionComparator , SenseiPluginRegistry pluginRegistry , ShardingStrategy shardingStrategy ) { this . nodeId = nodeId ; this . senseiSchema = senseiSchema ; this . versionComparator = versionComparator ; this . pluginRegistry = pluginRegistry ; this . shardingStrategy = shardingStrategy ; maxPartition = pluginRegistry . getConfiguration ( ) . getInt ( "<STR_LIT>" ) + <NUM_LIT:1> ; pluggableEngines = new ArrayList < PluggableSearchEngine > ( pluginRegistry . resolveBeansByListKey ( "<STR_LIT>" , PluggableSearchEngine . class ) ) ; if ( CompositeActivityManager . activitiesPresent ( senseiSchema ) ) { pluggableEngines . add ( new CompositeActivityManager ( ) ) ; } acceptEventsForAllPartitions = false ; for ( PluggableSearchEngine engine : pluggableEngines ) { engine . init ( indexDirectory , nodeId , senseiSchema , versionComparator , pluginRegistry , shardingStrategy ) ; if ( engine . acceptEventsForAllPartitions ( ) ) { acceptEventsForAllPartitions = true ; } } initVersion ( versionComparator ) ; } public void initVersion ( Comparator < String > versionComparator ) { List < String > versions = new ArrayList < String > ( ) ; for ( PluggableSearchEngine engine : pluggableEngines ) { if ( engine . getVersion ( ) != null && ! "<STR_LIT>" . equals ( engine . getVersion ( ) ) ) { versions . add ( engine . getVersion ( ) ) ; } } if ( versions . size ( ) > <NUM_LIT:0> ) { String version = versions . get ( <NUM_LIT:0> ) ; for ( String ver : versions ) { if ( versionComparator . compare ( version , ver ) > <NUM_LIT:0> ) { version = ver ; } } } } public JSONObject update ( JSONObject event , String version ) { if ( this . version != null && versionComparator . compare ( this . version , version ) > <NUM_LIT:0> ) { return event ; } else { this . version = version ; } boolean validForCurrentNode ; try { validForCurrentNode = Arrays . binarySearch ( senseiCore . getPartitions ( ) , shardingStrategy . caculateShard ( maxPartition , event ) ) >= <NUM_LIT:0> ; } catch ( JSONException e ) { throw new RuntimeException ( e ) ; } for ( PluggableSearchEngine pluggableSearchEngine : pluggableEngines ) { if ( pluggableSearchEngine . acceptEventsForAllPartitions ( ) || validForCurrentNode ) { try { event = pluggableSearchEngine . acceptEvent ( event , version ) ; } catch ( Exception ex ) { logger . error ( ex . getMessage ( ) , ex ) ; } } } return event ; } public void close ( ) { for ( PluggableSearchEngine pluggableSearchEngine : pluggableEngines ) { pluggableSearchEngine . stop ( ) ; } } @ Override public void onDelete ( IndexReader indexReader , long ... uids ) { for ( PluggableSearchEngine pluggableSearchEngine : pluggableEngines ) { pluggableSearchEngine . onDelete ( indexReader , uids ) ; } } @ Override public void onNewZoie ( Zoie < IndexReader , IndexReader > zoie ) { } @ Override public void onRetiredZoie ( Zoie < IndexReader , IndexReader > zoie ) { } @ Override public void onIndexReaderCleanUp ( ZoieIndexReader < IndexReader > indexReader ) { if ( indexReader instanceof ZoieMultiReader ) { ZoieSegmentReader [ ] segments = ( ZoieSegmentReader [ ] ) ( ( ZoieMultiReader ) indexReader ) . getSequentialSubReaders ( ) ; for ( ZoieSegmentReader segmentReader : segments ) { handleSegment ( segmentReader ) ; } } else if ( indexReader instanceof ZoieSegmentReader ) { handleSegment ( ( ZoieSegmentReader ) indexReader ) ; } else { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } } private void handleSegment ( ZoieSegmentReader segmentReader ) { onDelete ( segmentReader , segmentReader . getUIDArray ( ) ) ; } public void start ( SenseiCore senseiCore ) { this . senseiCore = senseiCore ; for ( PluggableSearchEngine pluggableSearchEngine : pluggableEngines ) { pluggableSearchEngine . start ( senseiCore ) ; } } public Set < String > getFieldNames ( ) { Set < String > ret = new HashSet < String > ( ) ; for ( PluggableSearchEngine pluggableSearchEngine : pluggableEngines ) { ret . addAll ( pluggableSearchEngine . getFieldNames ( ) ) ; } return ret ; } public List < FacetHandler < ? > > createFacetHandlers ( ) { List < FacetHandler < ? > > ret = new ArrayList < FacetHandler < ? > > ( ) ; for ( PluggableSearchEngine pluggableSearchEngine : pluggableEngines ) { ret . addAll ( pluggableSearchEngine . createFacetHandlers ( ) ) ; } return ret ; } public Set < String > getFacetNames ( ) { Set < String > ret = new HashSet < String > ( ) ; for ( PluggableSearchEngine pluggableSearchEngine : pluggableEngines ) { ret . addAll ( pluggableSearchEngine . getFacetNames ( ) ) ; } return ret ; } public List < PluggableSearchEngine > getPluggableEngines ( ) { return pluggableEngines ; } } </s>
|
<s> package com . senseidb . search . relevance ; import org . apache . log4j . Logger ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . search . query . ScoreAugmentQuery . ScoreAugmentFunction ; import com . senseidb . search . relevance . CustomRelevanceFunction . CustomRelevanceFunctionFactory ; import com . senseidb . search . relevance . RuntimeRelevanceFunction . RuntimeRelevanceFunctionFactory ; import com . senseidb . search . relevance . impl . CompilationHelper ; import com . senseidb . search . relevance . impl . CompilationHelper . DataTable ; import com . senseidb . search . relevance . impl . CustomMathModel ; import com . senseidb . search . relevance . impl . RelevanceException ; import com . senseidb . search . relevance . impl . RelevanceJSONConstants ; import com . senseidb . search . req . ErrorType ; public class RelevanceFunctionBuilder { private static Logger logger = Logger . getLogger ( RelevanceFunctionBuilder . class ) ; public static ScoreAugmentFunction build ( JSONObject jsonRelevance ) throws JSONException { if ( jsonRelevance . has ( RelevanceJSONConstants . KW_PREDEFINED ) ) { String modelName = jsonRelevance . getString ( RelevanceJSONConstants . KW_PREDEFINED ) ; if ( ModelStorage . hasPreloadedModel ( modelName ) ) { CustomRelevanceFunctionFactory crfFactory = ModelStorage . getPreloadedModel ( modelName ) ; return crfFactory . build ( ) ; } else if ( ModelStorage . hasRuntimeModel ( modelName ) ) { RuntimeRelevanceFunctionFactory rrfFactory = ModelStorage . getRuntimeModel ( modelName ) ; return rrfFactory . build ( ) ; } else { throw new JSONException ( "<STR_LIT>" + modelName ) ; } } else if ( jsonRelevance . has ( RelevanceJSONConstants . KW_MODEL ) ) { JSONObject modelJson = jsonRelevance . optJSONObject ( RelevanceJSONConstants . KW_MODEL ) ; RuntimeRelevanceFunctionFactory rrfFactory = ( RuntimeRelevanceFunctionFactory ) buildModelFactoryFromModelJSON ( modelJson ) ; RuntimeRelevanceFunction sm = ( RuntimeRelevanceFunction ) rrfFactory . build ( ) ; if ( modelJson . has ( RelevanceJSONConstants . KW_SAVE_AS ) ) { JSONObject jsonSaveAS = modelJson . getJSONObject ( RelevanceJSONConstants . KW_SAVE_AS ) ; String newRuntimeName = jsonSaveAS . getString ( RelevanceJSONConstants . KW_NAME_AS ) ; boolean overwrite = false ; if ( jsonSaveAS . has ( RelevanceJSONConstants . KW_OVERWRITE ) ) overwrite = jsonSaveAS . getBoolean ( RelevanceJSONConstants . KW_OVERWRITE ) ; if ( ( ModelStorage . hasRuntimeModel ( newRuntimeName ) || ModelStorage . hasPreloadedModel ( newRuntimeName ) ) && ( overwrite == false ) ) throw new IllegalArgumentException ( "<STR_LIT>" + newRuntimeName + "<STR_LIT>" ) ; ModelStorage . injectRuntimeModel ( newRuntimeName , rrfFactory ) ; } return sm ; } else { throw new IllegalArgumentException ( "<STR_LIT>" ) ; } } public static CustomRelevanceFunctionFactory buildModelFactoryFromModelJSON ( JSONObject modelJson ) throws RelevanceException { try { DataTable _dt = new DataTable ( ) ; CustomMathModel _cModel = CompilationHelper . createCustomMathScorer ( modelJson , _dt ) ; RuntimeRelevanceFunction sm = new RuntimeRelevanceFunction ( _cModel , _dt ) ; RuntimeRelevanceFunctionFactory rrfFactory = new RuntimeRelevanceFunctionFactory ( sm ) ; return rrfFactory ; } catch ( JSONException e ) { if ( e instanceof RelevanceException ) { RelevanceException re = ( RelevanceException ) e ; throw ( RelevanceException ) e ; } else throw new RelevanceException ( ErrorType . JsonParsingError , "<STR_LIT>" , e ) ; } } } </s>
|
<s> package com . senseidb . search . relevance ; import java . io . IOException ; import java . util . ArrayList ; import java . util . Iterator ; import java . util . List ; import java . util . Map ; import java . util . Set ; import org . apache . log4j . Logger ; import org . json . JSONException ; import org . json . JSONObject ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . facets . FacetHandler ; import com . browseengine . bobo . facets . data . FacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermDoubleList ; import com . browseengine . bobo . facets . data . TermFloatList ; import com . browseengine . bobo . facets . data . TermIntList ; import com . browseengine . bobo . facets . data . TermLongList ; import com . browseengine . bobo . facets . data . TermShortList ; import com . browseengine . bobo . facets . data . TermStringList ; import com . browseengine . bobo . facets . data . TermValueList ; import com . browseengine . bobo . util . BigSegmentedArray ; import com . senseidb . indexing . activity . facet . ActivityRangeFacetHandler ; import com . senseidb . search . query . ScoreAugmentQuery . ScoreAugmentFunction ; import com . senseidb . search . relevance . impl . CompilationHelper ; import com . senseidb . search . relevance . impl . CustomMathModel ; import com . senseidb . search . relevance . impl . RelevanceJSONConstants ; import com . senseidb . search . relevance . impl . MFacetDouble ; import com . senseidb . search . relevance . impl . MFacetFloat ; import com . senseidb . search . relevance . impl . MFacetInt ; import com . senseidb . search . relevance . impl . MFacetLong ; import com . senseidb . search . relevance . impl . MFacetShort ; import com . senseidb . search . relevance . impl . MFacetString ; import com . senseidb . search . relevance . impl . WeightedMFacetDouble ; import com . senseidb . search . relevance . impl . WeightedMFacetFloat ; import com . senseidb . search . relevance . impl . WeightedMFacetInt ; import com . senseidb . search . relevance . impl . WeightedMFacetLong ; import com . senseidb . search . relevance . impl . WeightedMFacetShort ; import com . senseidb . search . relevance . impl . WeightedMFacetString ; import com . senseidb . search . relevance . impl . CompilationHelper . DataTable ; public class RuntimeRelevanceFunction extends CustomRelevanceFunction { private static Logger logger = Logger . getLogger ( RuntimeRelevanceFunction . class ) ; private DataTable _dt ; private CustomMathModel _cModel ; private BigSegmentedArray [ ] _orderArrays ; private TermValueList [ ] _termLists ; private MultiValueFacetDataCache [ ] _mDataCaches ; private TermValueList [ ] _mTermLists ; private ActivityRangeFacetHandler [ ] _aHandlers ; private int [ ] [ ] _aData ; private int [ ] _types ; private int [ ] _facetIndex ; private int [ ] _arrayIndex ; private int [ ] _mFacetIndex ; private int [ ] _mArrayIndex ; private int [ ] _aFacetIndex ; private int _paramSize ; private short [ ] shorts ; private int [ ] ints ; private long [ ] longs ; private float [ ] floats ; private double [ ] doubles ; private boolean [ ] booleans ; private String [ ] strings ; private Set [ ] sets ; private Map [ ] maps ; private Object [ ] objs ; private MFacetInt [ ] mFacetInts ; private MFacetLong [ ] mFacetLongs ; private MFacetShort [ ] mFacetShorts ; private MFacetFloat [ ] mFacetFloats ; private MFacetDouble [ ] mFacetDoubles ; private MFacetString [ ] mFacetStrings ; private int [ ] dynamicAR ; public RuntimeRelevanceFunction ( CustomMathModel cModel , DataTable dt ) { _cModel = cModel ; _dt = dt ; } private void initialRunningData ( BoboIndexReader boboReader , CustomMathModel cModel , DataTable _dt ) throws IOException { int numFacet = _dt . hm_symbol_facet . keySet ( ) . size ( ) ; final BigSegmentedArray [ ] orderArrays = new BigSegmentedArray [ numFacet ] ; final TermValueList [ ] termLists = new TermValueList [ numFacet ] ; Iterator < String > iter_facet = _dt . hm_facet_index . keySet ( ) . iterator ( ) ; while ( iter_facet . hasNext ( ) ) { String facetName = iter_facet . next ( ) ; Object dataObj = boboReader . getFacetData ( facetName ) ; if ( ! ( dataObj instanceof FacetDataCache < ? > ) ) throw new IllegalArgumentException ( "<STR_LIT>" + facetName + "<STR_LIT>" ) ; int index = _dt . hm_facet_index . get ( facetName ) ; orderArrays [ index ] = ( ( FacetDataCache ) ( boboReader . getFacetData ( facetName ) ) ) . orderArray ; termLists [ index ] = ( ( FacetDataCache ) ( boboReader . getFacetData ( facetName ) ) ) . valArray ; } int numMultiFacet = _dt . hm_symbol_mfacet . keySet ( ) . size ( ) ; final MultiValueFacetDataCache [ ] mDataCaches = new MultiValueFacetDataCache [ numMultiFacet ] ; final TermValueList [ ] mTermLists = new TermValueList [ numMultiFacet ] ; Iterator < String > iter_mfacet = _dt . hm_mfacet_index . keySet ( ) . iterator ( ) ; while ( iter_mfacet . hasNext ( ) ) { String mFacetName = iter_mfacet . next ( ) ; Object dataObj = boboReader . getFacetData ( mFacetName ) ; if ( ! ( dataObj instanceof FacetDataCache < ? > ) ) throw new IllegalArgumentException ( "<STR_LIT>" + mFacetName + "<STR_LIT>" ) ; int index = _dt . hm_mfacet_index . get ( mFacetName ) ; mDataCaches [ index ] = ( MultiValueFacetDataCache ) ( boboReader . getFacetData ( mFacetName ) ) ; mTermLists [ index ] = ( ( MultiValueFacetDataCache ) ( boboReader . getFacetData ( mFacetName ) ) ) . valArray ; } int numAFacet = _dt . hm_symbol_afacet . keySet ( ) . size ( ) ; final int [ ] [ ] aData = new int [ numAFacet ] [ ] ; final ActivityRangeFacetHandler [ ] aHandlers = new ActivityRangeFacetHandler [ numAFacet ] ; Iterator < String > iter_afacet = _dt . hm_afacet_index . keySet ( ) . iterator ( ) ; while ( iter_afacet . hasNext ( ) ) { String afacetName = iter_afacet . next ( ) ; FacetHandler arHandler = boboReader . getFacetHandler ( afacetName ) ; Object dataObj = boboReader . getFacetData ( afacetName ) ; if ( ! ( dataObj instanceof int [ ] ) ) throw new IllegalArgumentException ( "<STR_LIT>" + afacetName + "<STR_LIT>" ) ; if ( ! ( arHandler instanceof ActivityRangeFacetHandler ) ) throw new IllegalArgumentException ( "<STR_LIT>" + afacetName + "<STR_LIT>" ) ; int index = _dt . hm_afacet_index . get ( afacetName ) ; aData [ index ] = ( int [ ] ) ( boboReader . getFacetData ( afacetName ) ) ; aHandlers [ index ] = ( ActivityRangeFacetHandler ) arHandler ; } final int paramSize = _dt . lls_params . size ( ) ; final int [ ] types = new int [ paramSize ] ; final int [ ] facetIndex = new int [ paramSize ] ; final int [ ] arrayIndex = new int [ paramSize ] ; final int [ ] mFacetIndex = new int [ paramSize ] ; final int [ ] mArrayIndex = new int [ paramSize ] ; final int [ ] aFacetIndex = new int [ paramSize ] ; updateArrayIndex ( _dt , paramSize , types , facetIndex , arrayIndex , mFacetIndex , mArrayIndex , aFacetIndex ) ; _cModel = cModel ; _orderArrays = orderArrays ; _termLists = termLists ; _types = types ; _facetIndex = facetIndex ; _arrayIndex = arrayIndex ; _mDataCaches = mDataCaches ; _mTermLists = mTermLists ; _mFacetIndex = mFacetIndex ; _mArrayIndex = mArrayIndex ; _aHandlers = aHandlers ; _aData = aData ; _aFacetIndex = aFacetIndex ; _paramSize = paramSize ; shorts = new short [ _paramSize ] ; ints = new int [ _paramSize ] ; longs = new long [ _paramSize ] ; floats = new float [ _paramSize ] ; doubles = new double [ _paramSize ] ; booleans = new boolean [ _paramSize ] ; strings = new String [ _paramSize ] ; sets = new Set [ _paramSize ] ; maps = new Map [ _paramSize ] ; objs = new Object [ _paramSize ] ; mFacetInts = new MFacetInt [ _paramSize ] ; mFacetLongs = new MFacetLong [ _paramSize ] ; mFacetShorts = new MFacetShort [ _paramSize ] ; mFacetFloats = new MFacetFloat [ _paramSize ] ; mFacetDoubles = new MFacetDouble [ _paramSize ] ; mFacetStrings = new MFacetString [ _paramSize ] ; ArrayList < Integer > arDynamic = new ArrayList < Integer > ( ) ; for ( int i = <NUM_LIT:0> ; i < _paramSize ; i ++ ) { switch ( _types [ i ] ) { case RelevanceJSONConstants . TYPENUMBER_NOW : longs [ _arrayIndex [ i ] ] = ( ( Long ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ) . longValue ( ) ; break ; case RelevanceJSONConstants . TYPENUMBER_INT : ints [ _arrayIndex [ i ] ] = ( ( Integer ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ) . intValue ( ) ; break ; case RelevanceJSONConstants . TYPENUMBER_LONG : longs [ _arrayIndex [ i ] ] = ( ( Long ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ) . longValue ( ) ; break ; case RelevanceJSONConstants . TYPENUMBER_DOUBLE : doubles [ _arrayIndex [ i ] ] = ( ( Double ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ) . doubleValue ( ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FLOAT : floats [ _arrayIndex [ i ] ] = ( ( Float ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ) . floatValue ( ) ; break ; case RelevanceJSONConstants . TYPENUMBER_BOOLEAN : booleans [ _arrayIndex [ i ] ] = ( ( Boolean ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ) . booleanValue ( ) ; break ; case RelevanceJSONConstants . TYPENUMBER_STRING : strings [ _arrayIndex [ i ] ] = ( String ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_SET : sets [ _arrayIndex [ i ] ] = ( Set ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_MAP : maps [ _arrayIndex [ i ] ] = ( Map ) _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_CUSTOM_OBJ : objs [ _arrayIndex [ i ] ] = _dt . hm_var . get ( _dt . lls_params . get ( i ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_INT : mFacetInts [ _mArrayIndex [ i ] ] = new MFacetInt ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_LONG : mFacetLongs [ _mArrayIndex [ i ] ] = new MFacetLong ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_DOUBLE : mFacetDoubles [ _mArrayIndex [ i ] ] = new MFacetDouble ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_FLOAT : mFacetFloats [ _mArrayIndex [ i ] ] = new MFacetFloat ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_SHORT : mFacetShorts [ _mArrayIndex [ i ] ] = new MFacetShort ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_STRING : mFacetStrings [ _mArrayIndex [ i ] ] = new MFacetString ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_INT : mFacetInts [ _mArrayIndex [ i ] ] = new WeightedMFacetInt ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_LONG : mFacetLongs [ _mArrayIndex [ i ] ] = new WeightedMFacetLong ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_DOUBLE : mFacetDoubles [ _mArrayIndex [ i ] ] = new WeightedMFacetDouble ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_FLOAT : mFacetFloats [ _mArrayIndex [ i ] ] = new WeightedMFacetFloat ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_SHORT : mFacetShorts [ _mArrayIndex [ i ] ] = new WeightedMFacetShort ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_STRING : mFacetStrings [ _mArrayIndex [ i ] ] = new WeightedMFacetString ( _mDataCaches [ _mFacetIndex [ i ] ] ) ; arDynamic . add ( i ) ; break ; default : arDynamic . add ( i ) ; } } dynamicAR = convertIntegers ( arDynamic ) ; } private int [ ] convertIntegers ( List < Integer > integers ) { int [ ] ret = new int [ integers . size ( ) ] ; Iterator < Integer > iterator = integers . iterator ( ) ; for ( int i = <NUM_LIT:0> ; i < ret . length ; i ++ ) { ret [ i ] = iterator . next ( ) . intValue ( ) ; } return ret ; } private void updateArrayIndex ( DataTable _dt , int paramSize , int [ ] types , int [ ] facetIndex , int [ ] arrayIndex , int [ ] mFacetIndex , int [ ] mArrayIndex , int [ ] aFacetIndex ) { int short_index = <NUM_LIT:0> , m_short_index = <NUM_LIT:0> ; int int_index = <NUM_LIT:0> , m_int_index = <NUM_LIT:0> ; int long_index = <NUM_LIT:0> , m_long_index = <NUM_LIT:0> ; int float_index = <NUM_LIT:0> , m_float_index = <NUM_LIT:0> ; int double_index = <NUM_LIT:0> , m_double_index = <NUM_LIT:0> ; int string_index = <NUM_LIT:0> , m_string_index = <NUM_LIT:0> ; int boolean_index = <NUM_LIT:0> ; int set_index = <NUM_LIT:0> ; int map_index = <NUM_LIT:0> ; int obj_index = <NUM_LIT:0> ; for ( int i = <NUM_LIT:0> ; i < paramSize ; i ++ ) { String symbol = _dt . lls_params . get ( i ) ; int typeNum = _dt . hm_type . get ( symbol ) ; types [ i ] = typeNum ; String facetName = null ; int index ; switch ( typeNum ) { case RelevanceJSONConstants . TYPENUMBER_INNER_SCORE : arrayIndex [ i ] = float_index ; float_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_NOW : arrayIndex [ i ] = long_index ; long_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_CUSTOM_OBJ : arrayIndex [ i ] = obj_index ; obj_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_INT : facetName = _dt . hm_symbol_facet . get ( symbol ) ; index = _dt . hm_facet_index . get ( facetName ) ; facetIndex [ i ] = index ; arrayIndex [ i ] = int_index ; int_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_LONG : facetName = _dt . hm_symbol_facet . get ( symbol ) ; index = _dt . hm_facet_index . get ( facetName ) ; facetIndex [ i ] = index ; arrayIndex [ i ] = long_index ; long_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_FLOAT : facetName = _dt . hm_symbol_facet . get ( symbol ) ; index = _dt . hm_facet_index . get ( facetName ) ; facetIndex [ i ] = index ; arrayIndex [ i ] = float_index ; float_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_SHORT : facetName = _dt . hm_symbol_facet . get ( symbol ) ; index = _dt . hm_facet_index . get ( facetName ) ; facetIndex [ i ] = index ; arrayIndex [ i ] = short_index ; short_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_STRING : facetName = _dt . hm_symbol_facet . get ( symbol ) ; index = _dt . hm_facet_index . get ( facetName ) ; facetIndex [ i ] = index ; arrayIndex [ i ] = string_index ; string_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_A_INT : facetName = _dt . hm_symbol_afacet . get ( symbol ) ; index = _dt . hm_afacet_index . get ( facetName ) ; aFacetIndex [ i ] = index ; arrayIndex [ i ] = int_index ; int_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_INT : case RelevanceJSONConstants . TYPENUMBER_FACET_WM_INT : facetName = _dt . hm_symbol_mfacet . get ( symbol ) ; index = _dt . hm_mfacet_index . get ( facetName ) ; mFacetIndex [ i ] = index ; mArrayIndex [ i ] = m_int_index ; m_int_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_LONG : case RelevanceJSONConstants . TYPENUMBER_FACET_WM_LONG : facetName = _dt . hm_symbol_mfacet . get ( symbol ) ; index = _dt . hm_mfacet_index . get ( facetName ) ; mFacetIndex [ i ] = index ; mArrayIndex [ i ] = m_long_index ; m_long_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_DOUBLE : case RelevanceJSONConstants . TYPENUMBER_FACET_WM_DOUBLE : facetName = _dt . hm_symbol_mfacet . get ( symbol ) ; index = _dt . hm_mfacet_index . get ( facetName ) ; mFacetIndex [ i ] = index ; mArrayIndex [ i ] = m_double_index ; m_double_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_FLOAT : case RelevanceJSONConstants . TYPENUMBER_FACET_WM_FLOAT : facetName = _dt . hm_symbol_mfacet . get ( symbol ) ; index = _dt . hm_mfacet_index . get ( facetName ) ; mFacetIndex [ i ] = index ; mArrayIndex [ i ] = m_float_index ; m_float_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_SHORT : case RelevanceJSONConstants . TYPENUMBER_FACET_WM_SHORT : facetName = _dt . hm_symbol_mfacet . get ( symbol ) ; index = _dt . hm_mfacet_index . get ( facetName ) ; mFacetIndex [ i ] = index ; mArrayIndex [ i ] = m_short_index ; m_short_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_STRING : case RelevanceJSONConstants . TYPENUMBER_FACET_WM_STRING : facetName = _dt . hm_symbol_mfacet . get ( symbol ) ; index = _dt . hm_mfacet_index . get ( facetName ) ; mFacetIndex [ i ] = index ; mArrayIndex [ i ] = m_string_index ; m_string_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_INT : arrayIndex [ i ] = int_index ; int_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_LONG : arrayIndex [ i ] = long_index ; long_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_DOUBLE : arrayIndex [ i ] = double_index ; double_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_FLOAT : arrayIndex [ i ] = float_index ; float_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_BOOLEAN : arrayIndex [ i ] = boolean_index ; boolean_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_STRING : arrayIndex [ i ] = string_index ; string_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_SET_INT : case RelevanceJSONConstants . TYPENUMBER_SET_LONG : case RelevanceJSONConstants . TYPENUMBER_SET_DOUBLE : case RelevanceJSONConstants . TYPENUMBER_SET_FLOAT : case RelevanceJSONConstants . TYPENUMBER_SET_STRING : types [ i ] = RelevanceJSONConstants . TYPENUMBER_SET ; arrayIndex [ i ] = set_index ; set_index ++ ; break ; case RelevanceJSONConstants . TYPENUMBER_MAP_INT_INT : case RelevanceJSONConstants . TYPENUMBER_MAP_INT_LONG : case RelevanceJSONConstants . TYPENUMBER_MAP_INT_DOUBLE : case RelevanceJSONConstants . TYPENUMBER_MAP_INT_FLOAT : case RelevanceJSONConstants . TYPENUMBER_MAP_INT_STRING : case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_INT : case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_LONG : case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_DOUBLE : case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_FLOAT : case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_STRING : types [ i ] = RelevanceJSONConstants . TYPENUMBER_MAP ; arrayIndex [ i ] = map_index ; map_index ++ ; break ; } } } @ Override public float newScore ( float innerScore , int docID ) { for ( int j = <NUM_LIT:0> ; j < dynamicAR . length ; j ++ ) { switch ( _types [ dynamicAR [ j ] ] ) { case RelevanceJSONConstants . TYPENUMBER_INNER_SCORE : floats [ _arrayIndex [ dynamicAR [ j ] ] ] = innerScore ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_INT : ints [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermIntList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_LONG : longs [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermLongList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_DOUBLE : doubles [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermDoubleList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_FLOAT : floats [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermFloatList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_SHORT : shorts [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermShortList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_STRING : strings [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermStringList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . get ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_INT : mFacetInts [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_LONG : mFacetLongs [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_DOUBLE : mFacetDoubles [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_FLOAT : mFacetFloats [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_SHORT : mFacetShorts [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_STRING : mFacetStrings [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_INT : ( ( WeightedMFacetInt ) mFacetInts [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_LONG : ( ( WeightedMFacetLong ) mFacetLongs [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_DOUBLE : ( ( WeightedMFacetDouble ) mFacetDoubles [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_FLOAT : ( ( WeightedMFacetFloat ) mFacetFloats [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_SHORT : ( ( WeightedMFacetShort ) mFacetShorts [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_STRING : ( ( WeightedMFacetString ) mFacetStrings [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_A_INT : ints [ _arrayIndex [ dynamicAR [ j ] ] ] = _aHandlers [ _aFacetIndex [ dynamicAR [ j ] ] ] . getIntActivityValue ( ( int [ ] ) _aData [ _aFacetIndex [ dynamicAR [ j ] ] ] , docID ) ; break ; default : break ; } } return _cModel . score ( shorts , ints , longs , floats , doubles , booleans , strings , sets , maps , mFacetInts , mFacetLongs , mFacetFloats , mFacetDoubles , mFacetShorts , mFacetStrings , objs ) ; } @ Override public float newScore ( int docID ) { for ( int j = <NUM_LIT:0> ; j < dynamicAR . length ; j ++ ) { switch ( _types [ dynamicAR [ j ] ] ) { case RelevanceJSONConstants . TYPENUMBER_FACET_INT : ints [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermIntList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_LONG : longs [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermLongList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_DOUBLE : doubles [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermDoubleList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_FLOAT : floats [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermFloatList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_SHORT : shorts [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermShortList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . getPrimitiveValue ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_STRING : strings [ _arrayIndex [ dynamicAR [ j ] ] ] = ( ( TermStringList ) _termLists [ _facetIndex [ dynamicAR [ j ] ] ] ) . get ( _orderArrays [ _facetIndex [ dynamicAR [ j ] ] ] . get ( docID ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_INT : mFacetInts [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_LONG : mFacetLongs [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_DOUBLE : mFacetDoubles [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_FLOAT : mFacetFloats [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_SHORT : mFacetShorts [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_M_STRING : mFacetStrings [ _mArrayIndex [ dynamicAR [ j ] ] ] . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_INT : ( ( WeightedMFacetInt ) mFacetInts [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_LONG : ( ( WeightedMFacetLong ) mFacetLongs [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_DOUBLE : ( ( WeightedMFacetDouble ) mFacetDoubles [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_FLOAT : ( ( WeightedMFacetFloat ) mFacetFloats [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_SHORT : ( ( WeightedMFacetShort ) mFacetShorts [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_WM_STRING : ( ( WeightedMFacetString ) mFacetStrings [ _mArrayIndex [ dynamicAR [ j ] ] ] ) . refresh ( docID ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FACET_A_INT : ints [ _arrayIndex [ dynamicAR [ j ] ] ] = _aHandlers [ _aFacetIndex [ dynamicAR [ j ] ] ] . getIntActivityValue ( ( int [ ] ) _aData [ _aFacetIndex [ dynamicAR [ j ] ] ] , docID ) ; break ; default : break ; } } return _cModel . score ( shorts , ints , longs , floats , doubles , booleans , strings , sets , maps , mFacetInts , mFacetLongs , mFacetFloats , mFacetDoubles , mFacetShorts , mFacetStrings , objs ) ; } @ Override public String getExplainString ( float innerScore , int doc ) { return _dt . funcBody ; } @ Override public void initializeReader ( BoboIndexReader reader , JSONObject jsonParams ) throws IOException { initialRunningData ( reader , _cModel , _dt ) ; } @ Override public void initializeGlobal ( JSONObject jsonValues ) throws JSONException { CompilationHelper . initializeValues ( jsonValues , _dt ) ; } @ Override public ScoreAugmentFunction getCopy ( ) { return new RuntimeRelevanceFunction ( this . _cModel , this . _dt ) ; } @ Override public boolean useInnerScore ( ) { return this . _dt . useInnerScore ; } public static class RuntimeRelevanceFunctionFactory extends CustomRelevanceFunctionFactory { RuntimeRelevanceFunction _rrf ; public RuntimeRelevanceFunctionFactory ( RuntimeRelevanceFunction rrf ) { _rrf = rrf ; } @ Override public CustomRelevanceFunction build ( ) { return ( CustomRelevanceFunction ) _rrf . getCopy ( ) ; } } } </s>
|
<s> package com . senseidb . search . relevance ; import java . util . HashMap ; import java . util . Map ; import com . senseidb . search . relevance . CustomRelevanceFunction . CustomRelevanceFunctionFactory ; import com . senseidb . search . relevance . RuntimeRelevanceFunction . RuntimeRelevanceFunctionFactory ; public class ModelStorage { private static Map < String , CustomRelevanceFunctionFactory > preloadedCRFMap = new HashMap < String , CustomRelevanceFunctionFactory > ( ) ; private static Map < String , RuntimeRelevanceFunctionFactory > runtimeCRFMap = new HashMap < String , RuntimeRelevanceFunctionFactory > ( ) ; public static void injectPreloadedModel ( String name , CustomRelevanceFunctionFactory crf ) { preloadedCRFMap . put ( name , crf ) ; } public static void injectRuntimeModel ( String name , RuntimeRelevanceFunctionFactory rrf ) { runtimeCRFMap . put ( name , rrf ) ; } public static boolean hasRuntimeModel ( String modelName ) { if ( runtimeCRFMap . containsKey ( modelName ) ) return true ; else return false ; } public static boolean hasPreloadedModel ( String modelName ) { if ( preloadedCRFMap . containsKey ( modelName ) ) return true ; else return false ; } public static RuntimeRelevanceFunctionFactory getRuntimeModel ( String modelName ) { if ( runtimeCRFMap . containsKey ( modelName ) ) return runtimeCRFMap . get ( modelName ) ; else return null ; } public static CustomRelevanceFunctionFactory getPreloadedModel ( String modelName ) { if ( preloadedCRFMap . containsKey ( modelName ) ) return preloadedCRFMap . get ( modelName ) ; else return null ; } } </s>
|
<s> package com . senseidb . search . relevance ; import java . util . HashMap ; import java . util . HashSet ; import com . senseidb . plugin . SenseiPlugin ; public class ExternalRelevanceDataStorage { private static HashMap < String , RelevanceObjPlugin > hmObjs ; private static HashMap < String , String > hmClsNames ; static { hmObjs = new HashMap < String , RelevanceObjPlugin > ( ) ; hmClsNames = new HashMap < String , String > ( ) ; ExternalRelevanceDataExample example = new ExternalRelevanceDataExample ( ) ; HashSet < String > hsColor = new HashSet < String > ( ) ; hsColor . add ( "<STR_LIT>" ) ; example . setHashSet ( hsColor ) ; example . setName ( "<STR_LIT>" ) ; putObj ( example ) ; } public static Object getObj ( String name ) { return hmObjs . get ( name ) ; } public static String getObjClsName ( String name ) { return hmClsNames . get ( name ) ; } public static void putObj ( RelevanceObjPlugin obj ) { String name = obj . getName ( ) ; hmObjs . put ( name , obj ) ; hmClsNames . put ( name , obj . getClass ( ) . getName ( ) ) ; } public interface RelevanceObjPlugin extends SenseiPlugin { public String getName ( ) ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import it . unimi . dsi . fastutil . objects . ObjectOpenHashSet ; import java . util . Set ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermStringList ; public class MFacetString extends MFacet { public MFacetString ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; } @ Override public boolean containsAll ( Set set ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } public boolean containsAll ( String [ ] target ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } @ Override public boolean containsAny ( Object set ) { ObjectOpenHashSet setString = ( ObjectOpenHashSet ) set ; for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( setString . contains ( ( ( TermStringList ) _mTermList ) . get ( _buf [ i ] ) ) ) return true ; return false ; } public boolean contains ( String target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermStringList ) _mTermList ) . get ( _buf [ i ] ) . equals ( target ) ) return true ; return false ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import java . util . Map ; import java . util . Set ; public interface CustomMathModel { float score ( short [ ] shorts , int [ ] ints , long [ ] longs , float [ ] floats , double [ ] doubles , boolean [ ] booleans , String [ ] strings , Set [ ] sets , Map [ ] maps , MFacetInt [ ] mFacetInts , MFacetLong [ ] mFacetLongs , MFacetFloat [ ] mFacetFloats , MFacetDouble [ ] mFacetDoubles , MFacetShort [ ] mFacetShorts , MFacetString [ ] mFacetStrings , Object [ ] objs ) ; } </s>
|
<s> package com . senseidb . search . relevance . impl ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueWithWeightFacetDataCache ; import com . browseengine . bobo . facets . data . TermDoubleList ; public class WeightedMFacetDouble extends MFacetDouble implements WeightedMFacet { public WeightedMFacetDouble ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; MultiValueWithWeightFacetDataCache wmDataCaches = ( MultiValueWithWeightFacetDataCache ) mDataCaches ; _weightArray = wmDataCaches . _weightArray ; weightBuf = new int [ <NUM_LIT> ] ; } @ Override public void refresh ( int id ) { _length = _nestedArray . getData ( id , _buf ) ; _weightArray . getData ( id , weightBuf ) ; } public boolean hasWeight ( double target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermDoubleList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) { _weight [ <NUM_LIT:0> ] = weightBuf [ i ] ; return true ; } return false ; } @ Override public int getWeight ( ) { return _weight [ <NUM_LIT:0> ] ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import it . unimi . dsi . fastutil . ints . IntOpenHashSet ; import java . util . Set ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermIntList ; public class MFacetInt extends MFacet { public MFacetInt ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; } @ Override public boolean containsAll ( Set set ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } public boolean containsAll ( int [ ] target ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } @ Override public boolean containsAny ( Object set ) { IntOpenHashSet setInt = ( IntOpenHashSet ) set ; for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( setInt . contains ( ( ( TermIntList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) ) ) return true ; return false ; } public boolean contains ( int target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermIntList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) return true ; return false ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import java . util . Set ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermValueList ; import com . browseengine . bobo . util . BigNestedIntArray ; public abstract class MFacet { protected MultiValueFacetDataCache _mDataCaches = null ; protected BigNestedIntArray _nestedArray = null ; protected TermValueList _mTermList = null ; protected int [ ] _buf = new int [ <NUM_LIT> ] ; protected int _length = <NUM_LIT:0> ; protected int [ ] _weight = new int [ <NUM_LIT:1> ] ; protected BigNestedIntArray _weightArray = null ; protected int [ ] weightBuf = new int [ <NUM_LIT> ] ; public MFacet ( MultiValueFacetDataCache mDataCaches ) { _mDataCaches = mDataCaches ; _mTermList = _mDataCaches . valArray ; _nestedArray = _mDataCaches . _nestedArray ; } public void refresh ( int id ) { _length = _nestedArray . getData ( id , _buf ) ; } public int size ( ) { return _length ; } abstract public boolean containsAll ( Set set ) ; abstract public boolean containsAny ( Object set ) ; } </s>
|
<s> package com . senseidb . search . relevance . impl ; import java . util . Map ; import java . util . HashMap ; public class RelevanceJSONConstants { public static final String KW_MODEL = "<STR_LIT>" ; public static final String KW_PREDEFINED = "<STR_LIT>" ; public static final String KW_KEY = "<STR_LIT:key>" ; public static final String KW_VALUE = "<STR_LIT:value>" ; public static final String KW_VALUES = "<STR_LIT>" ; public static final String KW_VARIABLES = "<STR_LIT>" ; public static final String KW_FACETS = "<STR_LIT>" ; public static final String KW_FUNC_PARAMETERS = "<STR_LIT>" ; public static final String KW_FUNCTION = "<STR_LIT>" ; public static final String KW_SAVE_AS = "<STR_LIT>" ; public static final String KW_NAME_AS = "<STR_LIT:name>" ; public static final String KW_OVERWRITE = "<STR_LIT>" ; public static final String KW_TYPE_SET_INT = "<STR_LIT>" ; public static final String KW_TYPE_SET_FLOAT = "<STR_LIT>" ; public static final String KW_TYPE_SET_STRING = "<STR_LIT>" ; public static final String KW_TYPE_SET_DOUBLE = "<STR_LIT>" ; public static final String KW_TYPE_SET_LONG = "<STR_LIT>" ; public static final String KW_TYPE_MAP_INT_INT = "<STR_LIT>" ; public static final String KW_TYPE_MAP_INT_FLOAT = "<STR_LIT>" ; public static final String KW_TYPE_MAP_INT_STRING = "<STR_LIT>" ; public static final String KW_TYPE_MAP_INT_DOUBLE = "<STR_LIT>" ; public static final String KW_TYPE_MAP_INT_LONG = "<STR_LIT>" ; public static final String KW_TYPE_MAP_STRING_INT = "<STR_LIT>" ; public static final String KW_TYPE_MAP_STRING_FLOAT = "<STR_LIT>" ; public static final String KW_TYPE_MAP_STRING_STRING = "<STR_LIT>" ; public static final String KW_TYPE_MAP_STRING_DOUBLE = "<STR_LIT>" ; public static final String KW_TYPE_MAP_STRING_LONG = "<STR_LIT>" ; public static final String KW_TYPE_INT = "<STR_LIT:int>" ; public static final String KW_TYPE_FLOAT = "<STR_LIT:float>" ; public static final String KW_TYPE_STRING = "<STR_LIT:string>" ; public static final String KW_TYPE_DOUBLE = "<STR_LIT:double>" ; public static final String KW_TYPE_LONG = "<STR_LIT:long>" ; public static final String KW_TYPE_BOOL = "<STR_LIT>" ; public static final String KW_TYPE_CUSTOM = "<STR_LIT>" ; public static final String KW_TYPE_FACET_INT = "<STR_LIT:int>" ; public static final String KW_TYPE_FACET_FLOAT = "<STR_LIT:float>" ; public static final String KW_TYPE_FACET_STRING = "<STR_LIT:string>" ; public static final String KW_TYPE_FACET_DOUBLE = "<STR_LIT:double>" ; public static final String KW_TYPE_FACET_LONG = "<STR_LIT:long>" ; public static final String KW_TYPE_FACET_SHORT = "<STR_LIT>" ; public static final String KW_TYPE_FACET_M_INT = "<STR_LIT>" ; public static final String KW_TYPE_FACET_M_FLOAT = "<STR_LIT>" ; public static final String KW_TYPE_FACET_M_STRING = "<STR_LIT>" ; public static final String KW_TYPE_FACET_M_DOUBLE = "<STR_LIT>" ; public static final String KW_TYPE_FACET_M_LONG = "<STR_LIT>" ; public static final String KW_TYPE_FACET_M_SHORT = "<STR_LIT>" ; public static final String KW_TYPE_FACET_WM_INT = "<STR_LIT>" ; public static final String KW_TYPE_FACET_WM_FLOAT = "<STR_LIT>" ; public static final String KW_TYPE_FACET_WM_STRING = "<STR_LIT>" ; public static final String KW_TYPE_FACET_WM_DOUBLE = "<STR_LIT>" ; public static final String KW_TYPE_FACET_WM_LONG = "<STR_LIT>" ; public static final String KW_TYPE_FACET_WM_SHORT = "<STR_LIT>" ; public static final String KW_TYPE_FACET_A_INT = "<STR_LIT>" ; public static final String KW_INNER_SCORE = "<STR_LIT>" ; public static final String KW_NOW = "<STR_LIT>" ; public static final int TYPENUMBER_INNER_SCORE = <NUM_LIT:0> ; public static final int TYPENUMBER_NOW = <NUM_LIT:1> ; public static final int TYPENUMBER_CUSTOM_OBJ = <NUM_LIT:2> ; public static final int TYPENUMBER_INT = <NUM_LIT:10> ; public static final int TYPENUMBER_LONG = <NUM_LIT:20> ; public static final int TYPENUMBER_DOUBLE = <NUM_LIT:30> ; public static final int TYPENUMBER_FLOAT = <NUM_LIT> ; public static final int TYPENUMBER_BOOLEAN = <NUM_LIT> ; public static final int TYPENUMBER_STRING = <NUM_LIT> ; public static final int TYPENUMBER_SET = <NUM_LIT:7> ; public static final int TYPENUMBER_MAP = <NUM_LIT:8> ; public static final int TYPENUMBER_SET_INT = <NUM_LIT> ; public static final int TYPENUMBER_SET_LONG = <NUM_LIT> ; public static final int TYPENUMBER_SET_DOUBLE = <NUM_LIT> ; public static final int TYPENUMBER_SET_FLOAT = <NUM_LIT> ; public static final int TYPENUMBER_SET_STRING = <NUM_LIT> ; public static final int TYPENUMBER_MAP_INT_INT = <NUM_LIT> ; public static final int TYPENUMBER_MAP_INT_LONG = <NUM_LIT> ; public static final int TYPENUMBER_MAP_INT_DOUBLE = <NUM_LIT> ; public static final int TYPENUMBER_MAP_INT_FLOAT = <NUM_LIT> ; public static final int TYPENUMBER_MAP_INT_STRING = <NUM_LIT> ; public static final int TYPENUMBER_MAP_STRING_INT = <NUM_LIT> ; public static final int TYPENUMBER_MAP_STRING_LONG = <NUM_LIT> ; public static final int TYPENUMBER_MAP_STRING_DOUBLE = <NUM_LIT> ; public static final int TYPENUMBER_MAP_STRING_FLOAT = <NUM_LIT> ; public static final int TYPENUMBER_MAP_STRING_STRING = <NUM_LIT> ; public static final int TYPENUMBER_FACET_INT = <NUM_LIT:100> ; public static final int TYPENUMBER_FACET_LONG = <NUM_LIT> ; public static final int TYPENUMBER_FACET_DOUBLE = <NUM_LIT> ; public static final int TYPENUMBER_FACET_FLOAT = <NUM_LIT> ; public static final int TYPENUMBER_FACET_SHORT = <NUM_LIT> ; public static final int TYPENUMBER_FACET_STRING = <NUM_LIT> ; public static final int TYPENUMBER_FACET_M_INT = <NUM_LIT> ; public static final int TYPENUMBER_FACET_M_LONG = <NUM_LIT> ; public static final int TYPENUMBER_FACET_M_DOUBLE = <NUM_LIT> ; public static final int TYPENUMBER_FACET_M_FLOAT = <NUM_LIT> ; public static final int TYPENUMBER_FACET_M_SHORT = <NUM_LIT> ; public static final int TYPENUMBER_FACET_M_STRING = <NUM_LIT> ; public static final int TYPENUMBER_FACET_WM_INT = <NUM_LIT> ; public static final int TYPENUMBER_FACET_WM_LONG = <NUM_LIT> ; public static final int TYPENUMBER_FACET_WM_DOUBLE = <NUM_LIT> ; public static final int TYPENUMBER_FACET_WM_FLOAT = <NUM_LIT> ; public static final int TYPENUMBER_FACET_WM_SHORT = <NUM_LIT> ; public static final int TYPENUMBER_FACET_WM_STRING = <NUM_LIT> ; public static final int TYPENUMBER_FACET_A_INT = <NUM_LIT> ; public static Map < String , Integer [ ] > FACET_INFO_MAP = new HashMap < String , Integer [ ] > ( ) ; public static Map < String , Integer > VARIABLE_INFO_MAP = new HashMap < String , Integer > ( ) ; static { FACET_INFO_MAP . put ( KW_TYPE_FACET_INT , new Integer [ ] { TYPENUMBER_FACET_INT , <NUM_LIT:0> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_SHORT , new Integer [ ] { TYPENUMBER_FACET_SHORT , <NUM_LIT:0> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_DOUBLE , new Integer [ ] { TYPENUMBER_FACET_DOUBLE , <NUM_LIT:0> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_FLOAT , new Integer [ ] { TYPENUMBER_FACET_FLOAT , <NUM_LIT:0> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_LONG , new Integer [ ] { TYPENUMBER_FACET_LONG , <NUM_LIT:0> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_STRING , new Integer [ ] { TYPENUMBER_FACET_STRING , <NUM_LIT:0> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_M_INT , new Integer [ ] { TYPENUMBER_FACET_M_INT , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_M_SHORT , new Integer [ ] { TYPENUMBER_FACET_M_SHORT , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_M_DOUBLE , new Integer [ ] { TYPENUMBER_FACET_M_DOUBLE , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_M_FLOAT , new Integer [ ] { TYPENUMBER_FACET_M_FLOAT , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_M_LONG , new Integer [ ] { TYPENUMBER_FACET_M_LONG , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_M_STRING , new Integer [ ] { TYPENUMBER_FACET_M_STRING , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_WM_INT , new Integer [ ] { TYPENUMBER_FACET_WM_INT , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_WM_SHORT , new Integer [ ] { TYPENUMBER_FACET_WM_SHORT , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_WM_DOUBLE , new Integer [ ] { TYPENUMBER_FACET_WM_DOUBLE , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_WM_FLOAT , new Integer [ ] { TYPENUMBER_FACET_WM_FLOAT , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_WM_LONG , new Integer [ ] { TYPENUMBER_FACET_WM_LONG , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_WM_STRING , new Integer [ ] { TYPENUMBER_FACET_WM_STRING , <NUM_LIT:1> } ) ; FACET_INFO_MAP . put ( KW_TYPE_FACET_A_INT , new Integer [ ] { TYPENUMBER_FACET_A_INT , <NUM_LIT:2> } ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_INT , TYPENUMBER_INT ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_LONG , TYPENUMBER_LONG ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_DOUBLE , TYPENUMBER_DOUBLE ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_FLOAT , TYPENUMBER_FLOAT ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_BOOL , TYPENUMBER_BOOLEAN ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_STRING , TYPENUMBER_STRING ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_SET_INT , TYPENUMBER_SET_INT ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_SET_LONG , TYPENUMBER_SET_LONG ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_SET_DOUBLE , TYPENUMBER_SET_DOUBLE ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_SET_FLOAT , TYPENUMBER_SET_FLOAT ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_SET_STRING , TYPENUMBER_SET_STRING ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_INT_INT , TYPENUMBER_MAP_INT_INT ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_INT_LONG , TYPENUMBER_MAP_INT_LONG ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_INT_DOUBLE , TYPENUMBER_MAP_INT_DOUBLE ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_INT_FLOAT , TYPENUMBER_MAP_INT_FLOAT ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_INT_STRING , TYPENUMBER_MAP_INT_STRING ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_STRING_INT , TYPENUMBER_MAP_STRING_INT ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_STRING_LONG , TYPENUMBER_MAP_STRING_LONG ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_STRING_DOUBLE , TYPENUMBER_MAP_STRING_DOUBLE ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_STRING_FLOAT , TYPENUMBER_MAP_STRING_FLOAT ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_MAP_STRING_STRING , TYPENUMBER_MAP_STRING_STRING ) ; VARIABLE_INFO_MAP . put ( KW_TYPE_CUSTOM , TYPENUMBER_CUSTOM_OBJ ) ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import org . json . JSONException ; import com . senseidb . search . req . ErrorType ; public class RelevanceException extends JSONException { private static final long serialVersionUID = <NUM_LIT:1L> ; private Throwable myCause ; public Throwable getCause ( ) { return ( myCause == this ? null : myCause ) ; } public synchronized Throwable initCause ( Throwable cause ) { myCause = cause ; return this ; } private String message ; private int errorcode = Integer . MIN_VALUE ; public String getMessage ( ) { if ( message != null ) return message ; else return this . toString ( ) ; } public int getErrorCode ( ) { return this . errorcode ; } public RelevanceException ( String msg ) { super ( msg ) ; errorcode = ErrorType . UnknownError . getDefaultErrorCode ( ) ; message = msg ; initCause ( null ) ; } public RelevanceException ( ErrorType errorType , String message ) { super ( message ) ; this . message = message ; this . errorcode = errorType . getDefaultErrorCode ( ) ; initCause ( null ) ; } public RelevanceException ( ErrorType errorType , String message , Exception e ) { super ( message ) ; this . message = message ; this . errorcode = errorType . getDefaultErrorCode ( ) ; initCause ( e ) ; } public RelevanceException ( Throwable e ) { super ( "<STR_LIT>" + e . toString ( ) ) ; message = null ; errorcode = ErrorType . UnknownError . getDefaultErrorCode ( ) ; initCause ( e ) ; } public RelevanceException ( String msg , Throwable e ) { this ( msg ) ; errorcode = ErrorType . UnknownError . getDefaultErrorCode ( ) ; initCause ( e ) ; } public RelevanceException ( Exception e ) { this ( "<STR_LIT>" + e . getMessage ( ) , e ) ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import java . io . IOException ; import java . util . ArrayList ; import java . util . Iterator ; import java . util . List ; import java . util . Map ; import java . util . Set ; import org . apache . log4j . Logger ; import org . apache . lucene . search . Scorer ; import org . json . JSONException ; import org . json . JSONObject ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . facets . data . FacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermDoubleList ; import com . browseengine . bobo . facets . data . TermFloatList ; import com . browseengine . bobo . facets . data . TermIntList ; import com . browseengine . bobo . facets . data . TermLongList ; import com . browseengine . bobo . facets . data . TermShortList ; import com . browseengine . bobo . facets . data . TermStringList ; import com . browseengine . bobo . facets . data . TermValueList ; import com . browseengine . bobo . util . BigSegmentedArray ; import com . senseidb . search . query . ScoreAugmentQuery . ScoreAugmentFunction ; import com . senseidb . search . relevance . CustomRelevanceFunction ; import com . senseidb . search . relevance . RuntimeRelevanceFunction ; import com . senseidb . search . relevance . impl . CompilationHelper . DataTable ; public class CustomScorer extends Scorer { private static Logger logger = Logger . getLogger ( CustomScorer . class ) ; final Scorer _innerScorer ; private RuntimeRelevanceFunction _sModifier = null ; public CustomScorer ( Scorer innerScorer , BoboIndexReader boboReader , CustomMathModel cModel , DataTable _dt , JSONObject _valueJson ) throws Exception { super ( innerScorer . getSimilarity ( ) ) ; _innerScorer = innerScorer ; _sModifier = new RuntimeRelevanceFunction ( cModel , _dt ) ; _sModifier . initializeGlobal ( _valueJson ) ; _sModifier . initializeReader ( boboReader , _valueJson ) ; } @ Override public float score ( ) throws IOException { return _sModifier . newScore ( _innerScorer . score ( ) , docID ( ) ) ; } @ Override public int advance ( int target ) throws IOException { return _innerScorer . advance ( target ) ; } @ Override public int docID ( ) { return _innerScorer . docID ( ) ; } @ Override public int nextDoc ( ) throws IOException { return _innerScorer . nextDoc ( ) ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; public interface WeightedMFacet { public int getWeight ( ) ; } </s>
|
<s> package com . senseidb . search . relevance . impl ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueWithWeightFacetDataCache ; import com . browseengine . bobo . facets . data . TermStringList ; public class WeightedMFacetString extends MFacetString implements WeightedMFacet { public WeightedMFacetString ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; MultiValueWithWeightFacetDataCache wmDataCaches = ( MultiValueWithWeightFacetDataCache ) mDataCaches ; _weightArray = wmDataCaches . _weightArray ; weightBuf = new int [ <NUM_LIT> ] ; } @ Override public void refresh ( int id ) { _length = _nestedArray . getData ( id , _buf ) ; _weightArray . getData ( id , weightBuf ) ; } public boolean hasWeight ( String target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermStringList ) _mTermList ) . get ( _buf [ i ] ) . equals ( target ) ) { _weight [ <NUM_LIT:0> ] = weightBuf [ i ] ; return true ; } return false ; } @ Override public int getWeight ( ) { return _weight [ <NUM_LIT:0> ] ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import it . unimi . dsi . fastutil . longs . LongOpenHashSet ; import java . util . Set ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermLongList ; public class MFacetLong extends MFacet { public MFacetLong ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; } @ Override public boolean containsAll ( Set set ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } public boolean containsAll ( long [ ] target ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } @ Override public boolean containsAny ( Object set ) { LongOpenHashSet setLong = ( LongOpenHashSet ) set ; for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( setLong . contains ( ( ( TermLongList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) ) ) return true ; return false ; } public boolean contains ( long target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermLongList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) return true ; return false ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import it . unimi . dsi . fastutil . shorts . ShortOpenHashSet ; import java . util . Set ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermShortList ; public class MFacetShort extends MFacet { public MFacetShort ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; } @ Override public boolean containsAll ( Set set ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } public boolean containsAll ( short [ ] target ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } @ Override public boolean containsAny ( Object set ) { ShortOpenHashSet setShort = ( ShortOpenHashSet ) set ; for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( setShort . contains ( ( ( TermShortList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) ) ) return true ; return false ; } public boolean contains ( short target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermShortList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) return true ; return false ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import it . unimi . dsi . fastutil . doubles . DoubleOpenHashSet ; import it . unimi . dsi . fastutil . floats . FloatOpenHashSet ; import it . unimi . dsi . fastutil . ints . Int2DoubleOpenHashMap ; import it . unimi . dsi . fastutil . ints . Int2FloatOpenHashMap ; import it . unimi . dsi . fastutil . ints . Int2IntOpenHashMap ; import it . unimi . dsi . fastutil . ints . Int2LongOpenHashMap ; import it . unimi . dsi . fastutil . ints . Int2ObjectOpenHashMap ; import it . unimi . dsi . fastutil . ints . IntOpenHashSet ; import it . unimi . dsi . fastutil . longs . LongOpenHashSet ; import it . unimi . dsi . fastutil . objects . Object2DoubleOpenHashMap ; import it . unimi . dsi . fastutil . objects . Object2FloatOpenHashMap ; import it . unimi . dsi . fastutil . objects . Object2IntOpenHashMap ; import it . unimi . dsi . fastutil . objects . Object2LongOpenHashMap ; import it . unimi . dsi . fastutil . objects . Object2ObjectOpenHashMap ; import it . unimi . dsi . fastutil . objects . ObjectOpenHashSet ; import java . util . HashMap ; import java . util . HashSet ; import java . util . Iterator ; import java . util . LinkedList ; import java . util . Map ; import java . util . Set ; import org . apache . log4j . Logger ; import org . json . JSONArray ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . search . relevance . ExternalRelevanceDataStorage ; import com . senseidb . search . req . ErrorType ; import javassist . CannotCompileException ; import javassist . ClassClassPath ; import javassist . ClassPool ; import javassist . CtClass ; import javassist . CtMethod ; import javassist . CtNewMethod ; import javassist . NotFoundException ; public class CompilationHelper { private static Logger logger = Logger . getLogger ( CompilationHelper . class ) ; private static ClassPool pool = ClassPool . getDefault ( ) ; private static HashSet < String > hs_safe = new HashSet < String > ( ) ; private static String [ ] PARAM_FORMAT_STRINGS = new String [ ] { "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" } ; private static Map < Integer , int [ ] > PARAM_INIT_MAP = new HashMap < Integer , int [ ] > ( ) ; private static int TOTAL_INPUT_DATA_ARRAYS = <NUM_LIT:16> ; private static String EXP_INT_METHOD = "<STR_LIT>" ; private static String EXP_DOUBLE_METHOD = "<STR_LIT>" ; private static String EXP_FLOAT_METHOD = "<STR_LIT>" ; private static String SCORE_METHOD_HEADER = "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" + "<STR_LIT>" ; static { hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; pool . importPackage ( "<STR_LIT>" ) ; for ( String cls : hs_safe ) { pool . importPackage ( cls ) ; } pool . insertClassPath ( new ClassClassPath ( CompilationHelper . class ) ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; hs_safe . add ( "<STR_LIT>" ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_INNER_SCORE , new int [ ] { <NUM_LIT:3> , <NUM_LIT:3> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_NOW , new int [ ] { <NUM_LIT:1> , <NUM_LIT:1> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_INT , new int [ ] { <NUM_LIT:0> , <NUM_LIT:0> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_LONG , new int [ ] { <NUM_LIT:1> , <NUM_LIT:1> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_DOUBLE , new int [ ] { <NUM_LIT:2> , <NUM_LIT:2> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FLOAT , new int [ ] { <NUM_LIT:3> , <NUM_LIT:3> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_STRING , new int [ ] { <NUM_LIT:4> , <NUM_LIT:4> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_BOOLEAN , new int [ ] { <NUM_LIT:6> , <NUM_LIT:12> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_INT , new int [ ] { <NUM_LIT:0> , <NUM_LIT:0> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_LONG , new int [ ] { <NUM_LIT:1> , <NUM_LIT:1> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_DOUBLE , new int [ ] { <NUM_LIT:2> , <NUM_LIT:2> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_FLOAT , new int [ ] { <NUM_LIT:3> , <NUM_LIT:3> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_STRING , new int [ ] { <NUM_LIT:4> , <NUM_LIT:4> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_SHORT , new int [ ] { <NUM_LIT:5> , <NUM_LIT:5> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_SET_INT , new int [ ] { <NUM_LIT:7> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_SET_LONG , new int [ ] { <NUM_LIT:8> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_SET_DOUBLE , new int [ ] { <NUM_LIT:9> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_SET_FLOAT , new int [ ] { <NUM_LIT:10> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_SET_STRING , new int [ ] { <NUM_LIT:11> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_INT_INT , new int [ ] { <NUM_LIT:12> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_INT_LONG , new int [ ] { <NUM_LIT> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_INT_DOUBLE , new int [ ] { <NUM_LIT> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_INT_FLOAT , new int [ ] { <NUM_LIT:15> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_INT_STRING , new int [ ] { <NUM_LIT:16> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_STRING_INT , new int [ ] { <NUM_LIT> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_STRING_LONG , new int [ ] { <NUM_LIT> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_STRING_DOUBLE , new int [ ] { <NUM_LIT> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_STRING_FLOAT , new int [ ] { <NUM_LIT:20> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_MAP_STRING_STRING , new int [ ] { <NUM_LIT> , <NUM_LIT> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_M_INT , new int [ ] { <NUM_LIT> , <NUM_LIT:6> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_M_LONG , new int [ ] { <NUM_LIT> , <NUM_LIT:7> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_M_DOUBLE , new int [ ] { <NUM_LIT:24> , <NUM_LIT:8> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_M_FLOAT , new int [ ] { <NUM_LIT> , <NUM_LIT:9> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_M_STRING , new int [ ] { <NUM_LIT> , <NUM_LIT:10> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_M_SHORT , new int [ ] { <NUM_LIT> , <NUM_LIT:11> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_WM_INT , new int [ ] { <NUM_LIT> , <NUM_LIT:6> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_WM_LONG , new int [ ] { <NUM_LIT> , <NUM_LIT:7> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_WM_DOUBLE , new int [ ] { <NUM_LIT:30> , <NUM_LIT:8> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_WM_FLOAT , new int [ ] { <NUM_LIT:31> , <NUM_LIT:9> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_WM_STRING , new int [ ] { <NUM_LIT:32> , <NUM_LIT:10> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_WM_SHORT , new int [ ] { <NUM_LIT> , <NUM_LIT:11> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_FACET_A_INT , new int [ ] { <NUM_LIT:0> , <NUM_LIT:0> } ) ; PARAM_INIT_MAP . put ( RelevanceJSONConstants . TYPENUMBER_CUSTOM_OBJ , new int [ ] { <NUM_LIT> , <NUM_LIT:15> } ) ; } private static int MAX_NUM_MODELS = <NUM_LIT> ; static HashMap < String , CustomMathModel > hmModels = new HashMap < String , CustomMathModel > ( ) ; public static CustomMathModel createCustomMathScorer ( JSONObject jsonModel , DataTable dataTable ) throws RelevanceException , JSONException { CustomMathModel cMathModel = null ; if ( jsonModel == null ) throw new RelevanceException ( ErrorType . JsonParsingError , "<STR_LIT>" ) ; JSONObject jsonVariables = jsonModel . optJSONObject ( RelevanceJSONConstants . KW_VARIABLES ) ; JSONObject jsonFacets = jsonModel . optJSONObject ( RelevanceJSONConstants . KW_FACETS ) ; JSONArray jsonFuncParameter = jsonModel . optJSONArray ( RelevanceJSONConstants . KW_FUNC_PARAMETERS ) ; for ( int j = <NUM_LIT:0> ; j < jsonFuncParameter . length ( ) ; j ++ ) { String paramName = jsonFuncParameter . optString ( j ) ; dataTable . lls_params . add ( paramName ) ; } dataTable . funcBody = jsonModel . optString ( RelevanceJSONConstants . KW_FUNCTION ) ; int [ ] facetIndice = new int [ ] { <NUM_LIT:0> , <NUM_LIT:0> , <NUM_LIT:0> } ; Iterator < String > it_facet = jsonFacets . keys ( ) ; while ( it_facet . hasNext ( ) ) { String facetType = it_facet . next ( ) ; JSONArray facetArray = jsonFacets . getJSONArray ( facetType ) ; handleFacetSymbols ( facetType , facetArray , facetIndice , dataTable ) ; } Iterator < String > it_var = jsonVariables . keys ( ) ; while ( it_var . hasNext ( ) ) { String type = it_var . next ( ) ; JSONArray varArray = jsonVariables . getJSONArray ( type ) ; for ( int i = <NUM_LIT:0> ; i < varArray . length ( ) ; ++ i ) { String symbol = varArray . getString ( i ) ; if ( symbol . equals ( RelevanceJSONConstants . KW_INNER_SCORE ) || symbol . equals ( RelevanceJSONConstants . KW_NOW ) ) { throw new RelevanceException ( ErrorType . JsonParsingError , "<STR_LIT>" + symbol + "<STR_LIT>" ) ; } Integer typeNum = RelevanceJSONConstants . VARIABLE_INFO_MAP . get ( type ) ; if ( typeNum == null ) { throw new RelevanceException ( ErrorType . JsonParsingError , "<STR_LIT>" + type + "<STR_LIT>" ) ; } dataTable . hm_type . put ( symbol , typeNum ) ; } } String symbolNow = RelevanceJSONConstants . KW_NOW ; long now = System . currentTimeMillis ( ) ; dataTable . hm_var . put ( symbolNow , now ) ; dataTable . hm_type . put ( symbolNow , RelevanceJSONConstants . TYPENUMBER_NOW ) ; String symbolInnerScore = RelevanceJSONConstants . KW_INNER_SCORE ; dataTable . hm_var . put ( symbolInnerScore , symbolInnerScore ) ; dataTable . hm_type . put ( symbolInnerScore , RelevanceJSONConstants . TYPENUMBER_INNER_SCORE ) ; if ( dataTable . funcBody == null || dataTable . funcBody . length ( ) == <NUM_LIT:0> ) throw new RelevanceException ( ErrorType . JsonParsingError , "<STR_LIT>" ) ; if ( dataTable . funcBody . indexOf ( "<STR_LIT>" ) == - <NUM_LIT:1> ) throw new RelevanceException ( ErrorType . JsonParsingError , "<STR_LIT>" ) ; for ( int i = <NUM_LIT:0> ; i < dataTable . lls_params . size ( ) ; i ++ ) { String symbol = dataTable . lls_params . get ( i ) ; if ( ! dataTable . hm_type . containsKey ( symbol ) ) throw new RelevanceException ( ErrorType . JsonParsingError , "<STR_LIT>" + symbol + "<STR_LIT>" ) ; Integer typeNum = dataTable . hm_type . get ( symbol ) ; if ( typeNum >= RelevanceJSONConstants . TYPENUMBER_FACET_INT && typeNum <= RelevanceJSONConstants . TYPENUMBER_FACET_WM_STRING ) { if ( ( ! dataTable . hm_symbol_facet . containsKey ( symbol ) ) && ( ! dataTable . hm_symbol_mfacet . containsKey ( symbol ) ) ) throw new RelevanceException ( ErrorType . JsonParsingError , "<STR_LIT>" + symbol + "<STR_LIT>" ) ; } } dataTable . lls_params = filterParameters ( dataTable ) ; String paramString = getParamString ( dataTable ) ; dataTable . classIDString = dataTable . funcBody + paramString ; String className = "<STR_LIT>" + dataTable . classIDString . hashCode ( ) ; logger . info ( "<STR_LIT>" + className ) ; if ( hmModels . containsKey ( className ) ) { cMathModel = hmModels . get ( className ) ; logger . info ( "<STR_LIT>" + className ) ; return cMathModel ; } else { synchronized ( CompilationHelper . class ) { if ( hmModels . containsKey ( className ) ) { cMathModel = hmModels . get ( className ) ; logger . info ( "<STR_LIT>" + className ) ; return cMathModel ; } CtClass ch = CompilationHelper . pool . makeClass ( className ) ; CtClass ci ; try { ci = CompilationHelper . pool . get ( "<STR_LIT>" ) ; } catch ( NotFoundException e ) { logger . info ( e . getMessage ( ) ) ; throw new RelevanceException ( e ) ; } ch . addInterface ( ci ) ; String functionString = makeFuncString ( dataTable ) ; addFacilityMethods ( ch ) ; CtMethod m ; try { m = CtNewMethod . make ( functionString , ch ) ; ch . addMethod ( m ) ; } catch ( CannotCompileException e ) { logger . info ( e . getMessage ( ) ) ; throw new RelevanceException ( ErrorType . JsonCompilationError , "<STR_LIT>" , e ) ; } Class h ; try { h = CompilationHelper . pool . toClass ( ch , new CompilationHelper . CustomLoader ( CompilationHelper . class . getClassLoader ( ) , className ) ) ; } catch ( CannotCompileException e ) { if ( hmModels . containsKey ( className ) ) { cMathModel = hmModels . get ( className ) ; logger . info ( "<STR_LIT>" + className ) ; return cMathModel ; } else { logger . info ( e . getMessage ( ) ) ; throw new RelevanceException ( ErrorType . JsonCompilationError , "<STR_LIT>" , e ) ; } } try { cMathModel = ( CustomMathModel ) h . newInstance ( ) ; } catch ( InstantiationException e ) { logger . info ( e . getMessage ( ) ) ; throw new RelevanceException ( ErrorType . JsonCompilationError , "<STR_LIT>" , e ) ; } catch ( IllegalAccessException e ) { logger . info ( e . getMessage ( ) ) ; throw new RelevanceException ( ErrorType . JsonCompilationError , "<STR_LIT>" , e ) ; } if ( hmModels . size ( ) > MAX_NUM_MODELS ) hmModels = new HashMap < String , CustomMathModel > ( ) ; hmModels . put ( className , cMathModel ) ; logger . info ( "<STR_LIT>" + className ) ; return cMathModel ; } } } public static void initializeValues ( JSONObject jsonValues , DataTable dataTable ) throws JSONException { HashMap < String , Integer > hm_type = dataTable . hm_type ; Iterator it = hm_type . keySet ( ) . iterator ( ) ; while ( it . hasNext ( ) ) { String symbol = ( String ) it . next ( ) ; Integer typeNum = dataTable . hm_type . get ( symbol ) ; if ( typeNum == RelevanceJSONConstants . TYPENUMBER_INNER_SCORE || typeNum == RelevanceJSONConstants . TYPENUMBER_NOW ) continue ; if ( typeNum >= RelevanceJSONConstants . TYPENUMBER_SET_INT && typeNum <= RelevanceJSONConstants . TYPENUMBER_SET_STRING ) { Set hs = null ; JSONArray values = jsonValues . optJSONArray ( symbol ) ; if ( values == null ) throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; switch ( typeNum ) { case RelevanceJSONConstants . TYPENUMBER_SET_INT : hs = new IntOpenHashSet ( ) ; for ( int k = <NUM_LIT:0> ; k < values . length ( ) ; k ++ ) { hs . add ( values . getInt ( k ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_SET_DOUBLE : hs = new DoubleOpenHashSet ( ) ; for ( int k = <NUM_LIT:0> ; k < values . length ( ) ; k ++ ) { hs . add ( values . getDouble ( k ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_SET_FLOAT : hs = new FloatOpenHashSet ( ) ; for ( int k = <NUM_LIT:0> ; k < values . length ( ) ; k ++ ) { hs . add ( ( float ) values . getDouble ( k ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_SET_LONG : hs = new LongOpenHashSet ( ) ; for ( int k = <NUM_LIT:0> ; k < values . length ( ) ; k ++ ) { hs . add ( values . getLong ( k ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_SET_STRING : hs = new ObjectOpenHashSet ( ) ; for ( int k = <NUM_LIT:0> ; k < values . length ( ) ; k ++ ) { hs . add ( values . getString ( k ) ) ; } break ; } dataTable . hm_var . put ( symbol , hs ) ; } else if ( typeNum >= RelevanceJSONConstants . TYPENUMBER_MAP_INT_INT && typeNum <= RelevanceJSONConstants . TYPENUMBER_MAP_STRING_STRING ) { JSONObject values = jsonValues . optJSONObject ( symbol ) ; if ( values == null ) throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; JSONArray keysList = values . names ( ) ; int keySize = keysList . length ( ) ; boolean isKeyValue = isKeyValueArrayMethod ( values ) ; JSONArray keysArrayList = null , valuesArrayList = null ; int keyArraySize , valueArraySize ; if ( isKeyValue ) { keysArrayList = values . optJSONArray ( RelevanceJSONConstants . KW_KEY ) ; valuesArrayList = values . optJSONArray ( RelevanceJSONConstants . KW_VALUE ) ; if ( keysArrayList == null ) throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; if ( valuesArrayList == null ) throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; keyArraySize = keysArrayList . length ( ) ; valueArraySize = valuesArrayList . length ( ) ; if ( keyArraySize != valueArraySize ) throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; keySize = keysArrayList . length ( ) ; } Map hm = null ; switch ( typeNum ) { case RelevanceJSONConstants . TYPENUMBER_MAP_INT_INT : hm = new Int2IntOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Int2IntOpenHashMap ) hm ) . put ( keysArrayList . getInt ( j ) , valuesArrayList . getInt ( j ) ) ; else ( ( Int2IntOpenHashMap ) hm ) . put ( keysList . getInt ( j ) , values . getInt ( keysList . getString ( j ) ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_INT_DOUBLE : hm = new Int2DoubleOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Int2DoubleOpenHashMap ) hm ) . put ( keysArrayList . getInt ( j ) , valuesArrayList . getDouble ( j ) ) ; else ( ( Int2DoubleOpenHashMap ) hm ) . put ( keysList . getInt ( j ) , values . getDouble ( keysList . getString ( j ) ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_INT_FLOAT : hm = new Int2FloatOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Int2FloatOpenHashMap ) hm ) . put ( keysArrayList . getInt ( j ) , ( float ) valuesArrayList . getDouble ( j ) ) ; else ( ( Int2FloatOpenHashMap ) hm ) . put ( keysList . getInt ( j ) , ( float ) values . getDouble ( keysList . getString ( j ) ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_INT_LONG : hm = new Int2LongOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Int2LongOpenHashMap ) hm ) . put ( keysArrayList . getInt ( j ) , Long . parseLong ( valuesArrayList . getString ( j ) ) ) ; else ( ( Int2LongOpenHashMap ) hm ) . put ( keysList . getInt ( j ) , Long . parseLong ( values . getString ( keysList . getString ( j ) ) ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_INT_STRING : hm = new Int2ObjectOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Int2ObjectOpenHashMap ) hm ) . put ( keysArrayList . getInt ( j ) , valuesArrayList . getString ( j ) ) ; else ( ( Int2ObjectOpenHashMap ) hm ) . put ( keysList . getInt ( j ) , values . getString ( keysList . getString ( j ) ) ) ; } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_INT : hm = new Object2IntOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Object2IntOpenHashMap ) hm ) . put ( keysArrayList . getString ( j ) , valuesArrayList . getInt ( j ) ) ; else { String key = keysList . getString ( j ) ; ( ( Object2IntOpenHashMap ) hm ) . put ( key , values . getInt ( key ) ) ; } } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_DOUBLE : hm = new Object2DoubleOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Object2DoubleOpenHashMap ) hm ) . put ( keysArrayList . getString ( j ) , valuesArrayList . getDouble ( j ) ) ; else { String key = keysList . getString ( j ) ; ( ( Object2DoubleOpenHashMap ) hm ) . put ( key , values . getDouble ( key ) ) ; } } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_FLOAT : hm = new Object2FloatOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Object2FloatOpenHashMap ) hm ) . put ( keysArrayList . getString ( j ) , ( float ) valuesArrayList . getDouble ( j ) ) ; else { String key = keysList . getString ( j ) ; ( ( Object2FloatOpenHashMap ) hm ) . put ( key , ( float ) values . getDouble ( key ) ) ; } } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_LONG : hm = new Object2LongOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Object2LongOpenHashMap ) hm ) . put ( keysArrayList . getString ( j ) , Long . parseLong ( valuesArrayList . getString ( j ) ) ) ; else { String key = keysList . getString ( j ) ; ( ( Object2LongOpenHashMap ) hm ) . put ( key , Long . parseLong ( values . getString ( keysList . getString ( j ) ) ) ) ; } } break ; case RelevanceJSONConstants . TYPENUMBER_MAP_STRING_STRING : hm = new Object2ObjectOpenHashMap ( ) ; for ( int j = <NUM_LIT:0> ; j < keySize ; j ++ ) { if ( isKeyValue ) ( ( Object2ObjectOpenHashMap ) hm ) . put ( keysArrayList . getString ( j ) , valuesArrayList . getString ( j ) ) ; else { String key = keysList . getString ( j ) ; ( ( Object2ObjectOpenHashMap ) hm ) . put ( key , values . getString ( key ) ) ; } } break ; } dataTable . hm_var . put ( symbol , hm ) ; } else if ( typeNum >= RelevanceJSONConstants . TYPENUMBER_INT && typeNum <= RelevanceJSONConstants . TYPENUMBER_STRING ) { if ( ! jsonValues . has ( symbol ) ) throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; switch ( typeNum ) { case RelevanceJSONConstants . TYPENUMBER_INT : dataTable . hm_var . put ( symbol , jsonValues . getInt ( symbol ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_DOUBLE : dataTable . hm_var . put ( symbol , jsonValues . getDouble ( symbol ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_FLOAT : dataTable . hm_var . put ( symbol , ( float ) jsonValues . getDouble ( symbol ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_LONG : dataTable . hm_var . put ( symbol , Long . parseLong ( jsonValues . getString ( symbol ) ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_STRING : dataTable . hm_var . put ( symbol , jsonValues . getString ( symbol ) ) ; break ; case RelevanceJSONConstants . TYPENUMBER_BOOLEAN : dataTable . hm_var . put ( symbol , jsonValues . getBoolean ( symbol ) ) ; break ; } } else if ( typeNum == RelevanceJSONConstants . TYPENUMBER_CUSTOM_OBJ ) { Object obj = ExternalRelevanceDataStorage . getObj ( symbol ) ; if ( obj != null ) dataTable . hm_var . put ( symbol , obj ) ; else throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; } } for ( int i = <NUM_LIT:0> ; i < dataTable . lls_params . size ( ) ; i ++ ) { String symbol = dataTable . lls_params . get ( i ) ; Integer typeNum = dataTable . hm_type . get ( symbol ) ; if ( typeNum < RelevanceJSONConstants . TYPENUMBER_FACET_INT || typeNum > RelevanceJSONConstants . TYPENUMBER_FACET_A_INT ) { if ( ! dataTable . hm_var . containsKey ( symbol ) ) throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; } } } private static boolean isKeyValueArrayMethod ( JSONObject mapJSON ) { if ( mapJSON . has ( RelevanceJSONConstants . KW_KEY ) && mapJSON . has ( RelevanceJSONConstants . KW_VALUE ) ) { JSONArray keysList = mapJSON . optJSONArray ( RelevanceJSONConstants . KW_KEY ) ; JSONArray valuesList = mapJSON . optJSONArray ( RelevanceJSONConstants . KW_VALUE ) ; if ( keysList != null && valuesList != null ) return true ; } return false ; } private static void addFacilityMethods ( CtClass ch ) throws JSONException { addMethod ( EXP_INT_METHOD , ch ) ; addMethod ( EXP_DOUBLE_METHOD , ch ) ; addMethod ( EXP_FLOAT_METHOD , ch ) ; } private static void addMethod ( String expStr , CtClass ch ) throws JSONException { CtMethod m_exp ; try { m_exp = CtNewMethod . make ( expStr , ch ) ; ch . addMethod ( m_exp ) ; } catch ( CannotCompileException e ) { logger . info ( e . getMessage ( ) ) ; throw new JSONException ( e ) ; } } private static String getParamString ( DataTable dataTable ) { StringBuilder sb = new StringBuilder ( ) ; for ( String param : dataTable . lls_params ) { sb . append ( param ) ; sb . append ( "<STR_LIT:#>" ) ; sb . append ( dataTable . hm_type . get ( param ) ) ; sb . append ( "<STR_LIT:#>" ) ; } return sb . toString ( ) ; } private static LinkedList < String > filterParameters ( DataTable dataTable ) { LinkedList < String > lls_new = new LinkedList < String > ( ) ; for ( String param : dataTable . lls_params ) { if ( ! ( dataTable . funcBody . indexOf ( param ) == - <NUM_LIT:1> ) ) lls_new . add ( param ) ; } return lls_new ; } private static void handleFacetSymbols ( String facetType , JSONArray facetArray , int [ ] facetIndice , DataTable dataTable ) throws JSONException { Integer [ ] facetInfo = RelevanceJSONConstants . FACET_INFO_MAP . get ( facetType ) ; if ( facetInfo == null ) { throw new JSONException ( "<STR_LIT>" + facetType ) ; } Integer type = facetInfo [ <NUM_LIT:0> ] ; for ( int i = <NUM_LIT:0> ; i < facetArray . length ( ) ; i ++ ) { String facetName = facetArray . getString ( i ) ; String symbol = facetName ; if ( dataTable . hm_symbol_facet . containsKey ( symbol ) || dataTable . hm_symbol_mfacet . containsKey ( symbol ) || dataTable . hm_symbol_afacet . containsKey ( symbol ) ) throw new JSONException ( "<STR_LIT>" + symbol + "<STR_LIT>" ) ; if ( dataTable . hm_facet_index . containsKey ( facetName ) || dataTable . hm_mfacet_index . containsKey ( facetName ) || dataTable . hm_afacet_index . containsKey ( facetName ) ) throw new JSONException ( "<STR_LIT>" + facetName + "<STR_LIT>" ) ; if ( facetInfo [ <NUM_LIT:1> ] == <NUM_LIT:0> ) { dataTable . hm_symbol_facet . put ( symbol , facetName ) ; dataTable . hm_facet_index . put ( facetName , facetIndice [ <NUM_LIT:0> ] ) ; facetIndice [ <NUM_LIT:0> ] = facetIndice [ <NUM_LIT:0> ] + <NUM_LIT:1> ; } else if ( facetInfo [ <NUM_LIT:1> ] == <NUM_LIT:1> ) { dataTable . hm_symbol_mfacet . put ( symbol , facetName ) ; dataTable . hm_mfacet_index . put ( facetName , facetIndice [ <NUM_LIT:1> ] ) ; facetIndice [ <NUM_LIT:1> ] = facetIndice [ <NUM_LIT:1> ] + <NUM_LIT:1> ; } else if ( facetInfo [ <NUM_LIT:1> ] == <NUM_LIT:2> ) { dataTable . hm_symbol_afacet . put ( symbol , facetName ) ; dataTable . hm_afacet_index . put ( facetName , facetIndice [ <NUM_LIT:2> ] ) ; facetIndice [ <NUM_LIT:2> ] = facetIndice [ <NUM_LIT:2> ] + <NUM_LIT:1> ; } dataTable . hm_type . put ( symbol , type ) ; } } private static String makeFuncString ( DataTable dataTable ) throws JSONException { int [ ] paramIndices = new int [ TOTAL_INPUT_DATA_ARRAYS ] ; for ( int i = <NUM_LIT:0> ; i < TOTAL_INPUT_DATA_ARRAYS ; i ++ ) { paramIndices [ i ] = <NUM_LIT:0> ; } StringBuffer sb = new StringBuffer ( ) ; sb . append ( SCORE_METHOD_HEADER ) . append ( "<STR_LIT>" ) ; dataTable . useInnerScore = false ; for ( int i = <NUM_LIT:0> ; i < dataTable . lls_params . size ( ) ; i ++ ) { String paramName = dataTable . lls_params . get ( i ) ; if ( ! dataTable . hm_type . containsKey ( paramName ) || ( dataTable . hm_type . get ( paramName ) == null ) ) throw new JSONException ( "<STR_LIT>" + paramName + "<STR_LIT>" ) ; Integer paramType = dataTable . hm_type . get ( paramName ) ; int [ ] paramInfo = PARAM_INIT_MAP . get ( paramType ) ; if ( paramType . intValue ( ) == RelevanceJSONConstants . TYPENUMBER_CUSTOM_OBJ ) { String className = ExternalRelevanceDataStorage . getObjClsName ( paramName ) ; if ( className == null ) throw new JSONException ( "<STR_LIT>" + paramName + "<STR_LIT>" ) ; String className2 = className . replace ( '<CHAR_LIT>' , '<CHAR_LIT:.>' ) ; hs_safe . add ( className ) ; pool . importPackage ( className ) ; sb . append ( String . format ( PARAM_FORMAT_STRINGS [ paramInfo [ <NUM_LIT:0> ] ] , className2 , paramName , className2 , paramIndices [ paramInfo [ <NUM_LIT:1> ] ] ++ ) ) ; } else sb . append ( String . format ( PARAM_FORMAT_STRINGS [ paramInfo [ <NUM_LIT:0> ] ] , paramName , paramIndices [ paramInfo [ <NUM_LIT:1> ] ] ++ ) ) ; if ( paramType == RelevanceJSONConstants . TYPENUMBER_INNER_SCORE ) { dataTable . useInnerScore = true ; } } sb . append ( dataTable . funcBody ) ; sb . append ( "<STR_LIT:}>" ) ; return sb . toString ( ) ; } public static class CustomLoader extends ClassLoader { private ClassLoader _cl ; private String _target ; @ Override public Class < ? > loadClass ( String name ) throws ClassNotFoundException { if ( hs_safe . contains ( name ) || name . equals ( _target ) ) return _cl . loadClass ( name ) ; else throw new ClassNotFoundException ( ) ; } public CustomLoader ( ClassLoader cl , String target ) { _cl = cl ; _target = target ; } } public static class DataTable { public HashMap < String , Object > hm_var ; public HashMap < String , Integer > hm_type ; public HashMap < String , String > hm_symbol_facet ; public HashMap < String , Integer > hm_facet_index ; public HashMap < String , String > hm_symbol_mfacet ; public HashMap < String , Integer > hm_mfacet_index ; public HashMap < String , String > hm_symbol_afacet ; public HashMap < String , Integer > hm_afacet_index ; public LinkedList < String > lls_params ; public String funcBody = null ; public String classIDString = null ; public boolean useInnerScore = true ; public DataTable ( ) { hm_var = new HashMap < String , Object > ( ) ; hm_type = new HashMap < String , Integer > ( ) ; hm_symbol_facet = new HashMap < String , String > ( ) ; hm_facet_index = new HashMap < String , Integer > ( ) ; hm_symbol_mfacet = new HashMap < String , String > ( ) ; hm_mfacet_index = new HashMap < String , Integer > ( ) ; hm_symbol_afacet = new HashMap < String , String > ( ) ; hm_afacet_index = new HashMap < String , Integer > ( ) ; lls_params = new LinkedList < String > ( ) ; useInnerScore = true ; } } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import it . unimi . dsi . fastutil . floats . FloatOpenHashSet ; import java . util . Set ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermFloatList ; public class MFacetFloat extends MFacet { public MFacetFloat ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; } @ Override public boolean containsAll ( Set set ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } public boolean containsAll ( float [ ] target ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } @ Override public boolean containsAny ( Object set ) { FloatOpenHashSet setFloat = ( FloatOpenHashSet ) set ; for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( setFloat . contains ( ( ( TermFloatList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) ) ) return true ; return false ; } public boolean contains ( float target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermFloatList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) return true ; return false ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueWithWeightFacetDataCache ; import com . browseengine . bobo . facets . data . TermLongList ; public class WeightedMFacetLong extends MFacetLong implements WeightedMFacet { public WeightedMFacetLong ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; MultiValueWithWeightFacetDataCache wmDataCaches = ( MultiValueWithWeightFacetDataCache ) mDataCaches ; _weightArray = wmDataCaches . _weightArray ; weightBuf = new int [ <NUM_LIT> ] ; } @ Override public void refresh ( int id ) { _length = _nestedArray . getData ( id , _buf ) ; _weightArray . getData ( id , weightBuf ) ; } public boolean hasWeight ( long target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermLongList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) { _weight [ <NUM_LIT:0> ] = weightBuf [ i ] ; return true ; } return false ; } @ Override public int getWeight ( ) { return _weight [ <NUM_LIT:0> ] ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueWithWeightFacetDataCache ; import com . browseengine . bobo . facets . data . TermIntList ; public class WeightedMFacetInt extends MFacetInt implements WeightedMFacet { public WeightedMFacetInt ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; MultiValueWithWeightFacetDataCache wmDataCaches = ( MultiValueWithWeightFacetDataCache ) mDataCaches ; _weightArray = wmDataCaches . _weightArray ; weightBuf = new int [ <NUM_LIT> ] ; } @ Override public void refresh ( int id ) { _length = _nestedArray . getData ( id , _buf ) ; _weightArray . getData ( id , weightBuf ) ; } public boolean hasWeight ( int target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermIntList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) { _weight [ <NUM_LIT:0> ] = weightBuf [ i ] ; return true ; } return false ; } @ Override public int getWeight ( ) { return _weight [ <NUM_LIT:0> ] ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueWithWeightFacetDataCache ; import com . browseengine . bobo . facets . data . TermShortList ; public class WeightedMFacetShort extends MFacetShort implements WeightedMFacet { public WeightedMFacetShort ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; MultiValueWithWeightFacetDataCache wmDataCaches = ( MultiValueWithWeightFacetDataCache ) mDataCaches ; _weightArray = wmDataCaches . _weightArray ; weightBuf = new int [ <NUM_LIT> ] ; } @ Override public void refresh ( int id ) { _length = _nestedArray . getData ( id , _buf ) ; _weightArray . getData ( id , weightBuf ) ; } public boolean hasWeight ( short target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermShortList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) { _weight [ <NUM_LIT:0> ] = weightBuf [ i ] ; return true ; } return false ; } @ Override public int getWeight ( ) { return _weight [ <NUM_LIT:0> ] ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueWithWeightFacetDataCache ; import com . browseengine . bobo . facets . data . TermFloatList ; public class WeightedMFacetFloat extends MFacetFloat implements WeightedMFacet { public WeightedMFacetFloat ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; MultiValueWithWeightFacetDataCache wmDataCaches = ( MultiValueWithWeightFacetDataCache ) mDataCaches ; _weightArray = wmDataCaches . _weightArray ; weightBuf = new int [ <NUM_LIT> ] ; } @ Override public void refresh ( int id ) { _length = _nestedArray . getData ( id , _buf ) ; _weightArray . getData ( id , weightBuf ) ; } public boolean hasWeight ( float target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermFloatList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) { _weight [ <NUM_LIT:0> ] = weightBuf [ i ] ; return true ; } return false ; } @ Override public int getWeight ( ) { return _weight [ <NUM_LIT:0> ] ; } } </s>
|
<s> package com . senseidb . search . relevance . impl ; import it . unimi . dsi . fastutil . doubles . DoubleOpenHashSet ; import java . util . Set ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; import com . browseengine . bobo . facets . data . TermDoubleList ; public class MFacetDouble extends MFacet { public MFacetDouble ( MultiValueFacetDataCache mDataCaches ) { super ( mDataCaches ) ; } @ Override public boolean containsAll ( Set set ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } public boolean containsAll ( double [ ] target ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } @ Override public boolean containsAny ( Object set ) { DoubleOpenHashSet setDouble = ( DoubleOpenHashSet ) set ; for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( setDouble . contains ( ( ( TermDoubleList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) ) ) return true ; return false ; } public boolean contains ( double target ) { for ( int i = <NUM_LIT:0> ; i < this . _length ; i ++ ) if ( ( ( TermDoubleList ) _mTermList ) . getPrimitiveValue ( _buf [ i ] ) == target ) return true ; return false ; } } </s>
|
<s> package com . senseidb . search . relevance ; import com . senseidb . search . query . ScoreAugmentQuery . ScoreAugmentFunction ; public abstract class CustomRelevanceFunction implements ScoreAugmentFunction { public static abstract class CustomRelevanceFunctionFactory { public abstract CustomRelevanceFunction build ( ) ; } } </s>
|
<s> package com . senseidb . search . relevance ; import java . util . HashSet ; import java . util . Map ; import com . senseidb . plugin . SenseiPluginRegistry ; import com . senseidb . search . relevance . ExternalRelevanceDataStorage . RelevanceObjPlugin ; public class ExternalRelevanceDataExample implements RelevanceObjPlugin { String name = "<STR_LIT>" ; HashSet < String > hsColor = ExampleExternalExternalObj . hs ; public boolean contains ( String color ) { return hsColor . contains ( color ) ; } public void setName ( String testName ) { name = testName ; } public void setHashSet ( HashSet < String > hs ) { hsColor = hs ; } @ Override public String getName ( ) { return name ; } @ Override public void init ( Map < String , String > config , SenseiPluginRegistry pluginRegistry ) { } @ Override public void start ( ) { } @ Override public void stop ( ) { } public static class ExampleExternalExternalObj { public static HashSet < String > hs = new HashSet < String > ( ) ; static { hs . add ( "<STR_LIT>" ) ; } } } </s>
|
<s> package com . senseidb . search . facet ; import it . unimi . dsi . fastutil . ints . IntArrayList ; import it . unimi . dsi . fastutil . longs . LongIterator ; import it . unimi . dsi . fastutil . longs . LongOpenHashSet ; import it . unimi . dsi . fastutil . longs . LongSet ; import java . io . IOException ; import java . util . Arrays ; import java . util . Collections ; import java . util . Properties ; import org . apache . log4j . Logger ; import org . apache . lucene . index . IndexReader ; import org . apache . lucene . search . DocIdSetIterator ; import org . apache . lucene . search . ScoreDoc ; import proj . zoie . api . DocIDMapper ; import proj . zoie . api . ZoieIndexReader ; import proj . zoie . api . ZoieSegmentReader ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . api . BrowseSelection ; import com . browseengine . bobo . api . FacetSpec ; import com . browseengine . bobo . docidset . EmptyDocIdSet ; import com . browseengine . bobo . docidset . RandomAccessDocIdSet ; import com . browseengine . bobo . facets . FacetCountCollectorSource ; import com . browseengine . bobo . facets . FacetHandler ; import com . browseengine . bobo . facets . filter . EmptyFilter ; import com . browseengine . bobo . facets . filter . RandomAccessFilter ; import com . browseengine . bobo . facets . filter . RandomAccessNotFilter ; import com . browseengine . bobo . sort . DocComparator ; import com . browseengine . bobo . sort . DocComparatorSource ; import com . kamikaze . docidset . impl . IntArrayDocIdSet ; public class UIDFacetHandler extends FacetHandler < long [ ] > { private static Logger logger = Logger . getLogger ( UIDFacetHandler . class ) ; public UIDFacetHandler ( String name ) { super ( name ) ; } private static class SingleDocRandmAccessDocIdSet extends RandomAccessDocIdSet { final int docid ; SingleDocRandmAccessDocIdSet ( int doc ) { docid = doc ; } @ Override public DocIdSetIterator iterator ( ) throws IOException { return new DocIdSetIterator ( ) { protected int _doc = - <NUM_LIT:1> ; @ Override public int advance ( int id ) throws IOException { _doc = id - <NUM_LIT:1> ; return nextDoc ( ) ; } @ Override public int docID ( ) { return _doc ; } @ Override public int nextDoc ( ) throws IOException { if ( _doc < docid ) { return _doc = docid ; } return _doc = DocIdSetIterator . NO_MORE_DOCS ; } } ; } @ Override public boolean get ( int doc ) { return doc == docid ; } } private RandomAccessFilter buildRandomAccessFilter ( final long val ) throws IOException { return new RandomAccessFilter ( ) { private static final long serialVersionUID = <NUM_LIT:1L> ; @ Override public RandomAccessDocIdSet getRandomAccessDocIdSet ( BoboIndexReader reader ) throws IOException { ZoieIndexReader < ? > zoieReader = ( ZoieIndexReader < ? > ) ( reader . getInnerReader ( ) ) ; DocIDMapper < ? > docidMapper = zoieReader . getDocIDMaper ( ) ; final int docid = docidMapper . getDocID ( val ) ; if ( docid == DocIDMapper . NOT_FOUND ) { return EmptyDocIdSet . getInstance ( ) ; } return new SingleDocRandmAccessDocIdSet ( docid ) ; } } ; } private RandomAccessFilter buildRandomAccessFilter ( final LongSet valSet ) throws IOException { return new RandomAccessFilter ( ) { private static final long serialVersionUID = <NUM_LIT:1L> ; @ Override public RandomAccessDocIdSet getRandomAccessDocIdSet ( BoboIndexReader reader ) throws IOException { ZoieSegmentReader < ? > zoieReader = ( ZoieSegmentReader < ? > ) ( reader . getInnerReader ( ) ) ; DocIDMapper < ? > docidMapper = zoieReader . getDocIDMaper ( ) ; final IntArrayList docidList = new IntArrayList ( valSet . size ( ) ) ; LongIterator iter = valSet . iterator ( ) ; while ( iter . hasNext ( ) ) { int docid = docidMapper . getDocID ( iter . nextLong ( ) ) ; if ( docid != DocIDMapper . NOT_FOUND ) { docidList . add ( docid ) ; } } if ( docidList . size ( ) == <NUM_LIT:0> ) return EmptyDocIdSet . getInstance ( ) ; int [ ] delDocIds = zoieReader . getDelDocIds ( ) ; if ( docidList . size ( ) == <NUM_LIT:1> ) { int docId = docidList . getInt ( <NUM_LIT:0> ) ; if ( delDocIds == null || delDocIds . length == <NUM_LIT:0> || Arrays . binarySearch ( delDocIds , docId ) < <NUM_LIT:0> ) { return new SingleDocRandmAccessDocIdSet ( docidList . getInt ( <NUM_LIT:0> ) ) ; } else { return EmptyDocIdSet . getInstance ( ) ; } } Collections . sort ( docidList ) ; final IntArrayDocIdSet intArraySet = new IntArrayDocIdSet ( docidList . size ( ) ) ; boolean deletesPresent = delDocIds != null && delDocIds . length > <NUM_LIT:0> ; for ( int docid : docidList ) { if ( ! deletesPresent || Arrays . binarySearch ( delDocIds , docid ) < <NUM_LIT:0> ) { intArraySet . addDoc ( docid ) ; } } return new RandomAccessDocIdSet ( ) { @ Override public boolean get ( int docid ) { return docidList . contains ( docid ) ; } @ Override public DocIdSetIterator iterator ( ) throws IOException { return intArraySet . iterator ( ) ; } } ; } } ; } @ Override public RandomAccessFilter buildRandomAccessFilter ( String value , Properties selectionProperty ) throws IOException { try { long val = Long . parseLong ( value ) ; return buildRandomAccessFilter ( val ) ; } catch ( Exception e ) { throw new IOException ( e . getMessage ( ) ) ; } } @ Override public RandomAccessFilter buildRandomAccessAndFilter ( String [ ] vals , Properties prop ) throws IOException { LongSet longSet = new LongOpenHashSet ( ) ; for ( String val : vals ) { try { longSet . add ( Long . parseLong ( val ) ) ; } catch ( Exception e ) { throw new IOException ( e . getMessage ( ) ) ; } } if ( longSet . size ( ) != <NUM_LIT:1> ) { return EmptyFilter . getInstance ( ) ; } else { return buildRandomAccessFilter ( longSet . iterator ( ) . nextLong ( ) ) ; } } @ Override public RandomAccessFilter buildRandomAccessOrFilter ( String [ ] vals , Properties prop , boolean isNot ) throws IOException { LongSet longSet = new LongOpenHashSet ( ) ; for ( String val : vals ) { try { longSet . add ( Long . parseLong ( val ) ) ; } catch ( Exception e ) { throw new IOException ( e . getMessage ( ) ) ; } } RandomAccessFilter filter ; if ( longSet . size ( ) == <NUM_LIT:1> ) { filter = buildRandomAccessFilter ( longSet . iterator ( ) . nextLong ( ) ) ; } else { filter = buildRandomAccessFilter ( longSet ) ; } if ( filter == null ) return filter ; if ( isNot ) { filter = new RandomAccessNotFilter ( filter ) ; } return filter ; } @ Override public DocComparatorSource getDocComparatorSource ( ) { return new DocComparatorSource ( ) { @ Override public DocComparator getComparator ( IndexReader reader , int docbase ) throws IOException { final UIDFacetHandler uidFacetHandler = UIDFacetHandler . this ; if ( reader instanceof BoboIndexReader ) { BoboIndexReader boboReader = ( BoboIndexReader ) reader ; final long [ ] uidArray = uidFacetHandler . getFacetData ( boboReader ) ; return new DocComparator ( ) { @ Override public Comparable value ( ScoreDoc doc ) { int docid = doc . doc ; return Long . valueOf ( uidArray [ docid ] ) ; } @ Override public int compare ( ScoreDoc doc1 , ScoreDoc doc2 ) { long uid1 = uidArray [ doc1 . doc ] ; long uid2 = uidArray [ doc2 . doc ] ; if ( uid1 == uid2 ) { return <NUM_LIT:0> ; } else { if ( uid1 < uid2 ) return - <NUM_LIT:1> ; return <NUM_LIT:1> ; } } } ; } else { throw new IOException ( "<STR_LIT>" + BoboIndexReader . class ) ; } } } ; } @ Override public FacetCountCollectorSource getFacetCountCollectorSource ( BrowseSelection sel , FacetSpec fspec ) { throw new UnsupportedOperationException ( "<STR_LIT>" ) ; } @ Override public String [ ] getFieldValues ( BoboIndexReader reader , int id ) { long [ ] uidArray = getFacetData ( reader ) ; return new String [ ] { String . valueOf ( uidArray [ id ] ) } ; } @ Override public Object [ ] getRawFieldValues ( BoboIndexReader reader , int id ) { long [ ] uidArray = getFacetData ( reader ) ; return new Long [ ] { uidArray [ id ] } ; } @ Override public long [ ] load ( BoboIndexReader reader ) throws IOException { IndexReader innerReader = reader . getInnerReader ( ) ; if ( innerReader instanceof ZoieSegmentReader ) { ZoieSegmentReader zoieReader = ( ZoieSegmentReader ) innerReader ; return zoieReader . getUIDArray ( ) ; } else { throw new IOException ( "<STR_LIT>" + ZoieSegmentReader . class ) ; } } } </s>
|
<s> package com . senseidb . search . facet . attribute ; import java . io . IOException ; import org . apache . lucene . search . DocIdSetIterator ; import com . browseengine . bobo . facets . data . FacetDataCache ; @ SuppressWarnings ( "<STR_LIT:rawtypes>" ) public class PredicateDocIdIterator extends DocIdSetIterator { private final int startDocId ; private final int endDocId ; private final FacetPredicate facetPredicate ; private int docId = - <NUM_LIT:1> ; private final FacetDataCache facetDataCache ; public PredicateDocIdIterator ( int startDocId , int endDocId , FacetPredicate facetPredicate , FacetDataCache facetDataCache ) { this . startDocId = startDocId ; this . endDocId = endDocId ; if ( startDocId > endDocId ) { throw new IllegalStateException ( "<STR_LIT>" ) ; } this . facetPredicate = facetPredicate ; this . facetDataCache = facetDataCache ; } @ Override public int docID ( ) { return docId ; } @ Override public int nextDoc ( ) throws IOException { if ( docId == - <NUM_LIT:1> ) { docId = startDocId - <NUM_LIT:1> ; } if ( docId == NO_MORE_DOCS ) { return docId ; } docId ++ ; while ( ! facetPredicate . evaluate ( facetDataCache , docId ) || docId > endDocId ) { if ( docId > endDocId ) { docId = NO_MORE_DOCS ; return docId ; } docId ++ ; } return docId ; } @ Override public int advance ( int target ) throws IOException { if ( target > endDocId ) { return NO_MORE_DOCS ; } docId = target - <NUM_LIT:1> ; return nextDoc ( ) ; } } </s>
|
<s> package com . senseidb . search . facet . attribute ; import com . browseengine . bobo . facets . data . FacetDataCache ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; public class RangePredicate implements FacetPredicate { private final String value ; private final char separator ; private MultiValueFacetDataCache lastDataCache ; private Range range ; private int [ ] buffer ; private final Range getRange ( FacetDataCache cache ) { if ( cache == lastDataCache ) { return range ; } lastDataCache = ( MultiValueFacetDataCache ) cache ; range = Range . getRange ( lastDataCache , value , separator ) ; buffer = new int [ lastDataCache . _nestedArray . getMaxItems ( ) ] ; return range ; } public RangePredicate ( String val , char separator ) { value = val ; this . separator = separator ; } @ Override public boolean evaluate ( FacetDataCache cache , int docId ) { if ( cache != lastDataCache ) { getRange ( cache ) ; } return lastDataCache . _nestedArray . containsValueInRange ( docId , range . start , range . end ) ; } @ Override public boolean evaluateValue ( FacetDataCache cache , int valId ) { if ( cache != lastDataCache ) { getRange ( cache ) ; } return valId >= range . start && valId < range . end ; } @ Override public int valueStartIndex ( FacetDataCache cache ) { if ( cache != lastDataCache ) { getRange ( cache ) ; } return range . start ; } @ Override public int valueEndIndex ( FacetDataCache cache ) { return range . end ; } } </s>
|
<s> package com . senseidb . search . facet . attribute ; import com . browseengine . bobo . facets . data . MultiValueFacetDataCache ; public class Range { int start ; int end ; public Range ( int start , int end ) { this . start = start ; this . end = end ; } public boolean inRange ( int val ) { return val >= start && val < end ; } public static Range [ ] getRanges ( MultiValueFacetDataCache cache , String [ ] vals , char separator ) { Range [ ] ret = new Range [ vals . length ] ; for ( int i = <NUM_LIT:0> ; i < vals . length ; i ++ ) { ret [ i ] = getRange ( cache , vals [ i ] , separator ) ; } return ret ; } public static Range getRange ( MultiValueFacetDataCache cache , String val , char separator ) { int start = cache . valArray . indexOf ( val + separator ) ; if ( start < <NUM_LIT:0> ) { start = ( - <NUM_LIT:1> ) * start - <NUM_LIT:1> ; } int end = cache . valArray . indexOf ( val + ( char ) ( separator + <NUM_LIT:1> ) ) ; if ( end < <NUM_LIT:0> ) { end = ( - <NUM_LIT:1> ) * end - <NUM_LIT:1> ; } return new Range ( start , end ) ; } @ Override public String toString ( ) { return "<STR_LIT>" + start + "<STR_LIT>" + end + "<STR_LIT:]>" ; } } </s>
|
<s> package com . senseidb . search . facet . attribute ; import java . io . IOException ; import org . apache . lucene . search . DocIdSetIterator ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . docidset . EmptyDocIdSet ; import com . browseengine . bobo . docidset . RandomAccessDocIdSet ; import com . browseengine . bobo . facets . data . FacetDataCache ; import com . browseengine . bobo . facets . filter . RandomAccessFilter ; import com . browseengine . bobo . facets . range . SimpleDataCacheBuilder ; public class PredicateFacetFilter extends RandomAccessFilter { private final SimpleDataCacheBuilder dataCacheBuilder ; private final FacetPredicate facetPredicate ; public PredicateFacetFilter ( SimpleDataCacheBuilder dataCacheBuilder , FacetPredicate facetPredicate ) { this . dataCacheBuilder = dataCacheBuilder ; this . facetPredicate = facetPredicate ; } @ Override public RandomAccessDocIdSet getRandomAccessDocIdSet ( BoboIndexReader reader ) throws IOException { final FacetDataCache facetDataCache = dataCacheBuilder . build ( reader ) ; int startDocIdTemp = Integer . MAX_VALUE ; int endDocIdTemp = - <NUM_LIT:1> ; for ( int i = facetPredicate . valueStartIndex ( facetDataCache ) ; i < facetPredicate . valueEndIndex ( facetDataCache ) ; i ++ ) { if ( facetPredicate . evaluateValue ( facetDataCache , i ) ) { if ( ! facetPredicate . evaluateValue ( facetDataCache , i ) ) { continue ; } if ( startDocIdTemp > facetDataCache . minIDs [ i ] ) { startDocIdTemp = facetDataCache . minIDs [ i ] ; } if ( endDocIdTemp < facetDataCache . maxIDs [ i ] ) { endDocIdTemp = facetDataCache . maxIDs [ i ] ; } } } final int startDocId = startDocIdTemp ; final int endDocId = endDocIdTemp ; if ( startDocId > endDocId ) { return EmptyDocIdSet . getInstance ( ) ; } return new RandomAccessDocIdSet ( ) { @ Override public boolean get ( int docId ) { return facetPredicate . evaluate ( facetDataCache , docId ) ; } @ Override public DocIdSetIterator iterator ( ) throws IOException { return new PredicateDocIdIterator ( startDocId , endDocId , facetPredicate , facetDataCache ) ; } } ; } @ Override public double getFacetSelectivity ( BoboIndexReader reader ) { FacetDataCache dataCache = dataCacheBuilder . build ( reader ) ; int [ ] frequencies = dataCache . freqs ; double selectivity = <NUM_LIT:0> ; int accumFreq = <NUM_LIT:0> ; int total = reader . maxDoc ( ) ; for ( int i = facetPredicate . valueStartIndex ( dataCache ) ; i < facetPredicate . valueEndIndex ( dataCache ) ; i ++ ) { if ( ! facetPredicate . evaluateValue ( dataCache , i ) ) { continue ; } accumFreq += frequencies [ i ] ; } selectivity = ( double ) accumFreq / ( double ) total ; if ( selectivity > <NUM_LIT> ) { selectivity = <NUM_LIT:1.0> ; } return selectivity ; } } </s>
|
<s> package com . senseidb . search . facet . attribute ; import com . browseengine . bobo . facets . data . FacetDataCache ; public interface FacetPredicate { public boolean evaluate ( FacetDataCache cache , int docId ) ; public boolean evaluateValue ( FacetDataCache cache , int valId ) ; public int valueStartIndex ( FacetDataCache cache ) ; public int valueEndIndex ( FacetDataCache cache ) ; } </s>
|
<s> package com . senseidb . search . node ; import com . senseidb . svc . api . SenseiException ; public interface Broker < REQUEST , RESULT > { public RESULT browse ( final REQUEST req ) throws SenseiException ; } </s>
|
<s> package com . senseidb . search . node . broker ; import java . util . List ; import com . senseidb . search . req . SenseiRequest ; public class AllClustersPruner implements LayeredClusterPruner { @ Override public List < String > pruneClusters ( SenseiRequest request , List < String > clusters ) { return clusters ; } @ Override public boolean clusterPrioritiesEqual ( SenseiRequest request ) { return false ; } } </s>
|
<s> package com . senseidb . search . node . broker ; import java . util . Map ; import org . apache . commons . configuration . Configuration ; import com . linkedin . norbert . javacompat . network . PartitionedLoadBalancerFactory ; import com . senseidb . conf . SenseiConfParams ; import com . senseidb . servlet . SenseiConfigServletContextListener ; public class CompoundBrokerConfig extends BrokerConfig { public CompoundBrokerConfig ( Configuration senseiConf , PartitionedLoadBalancerFactory < String > loadBalancerFactory , Map < String , String > config , String clusterName ) { super ( senseiConf , loadBalancerFactory ) ; this . clusterName = clusterName ; zkurl = getStrParam ( clusterName , config , SenseiConfParams . SENSEI_CLUSTER_URL , zkurl ) ; zkTimeout = getIntParam ( clusterName , config , SenseiConfParams . SENSEI_CLUSTER_TIMEOUT , zkTimeout ) ; connectTimeoutMillis = getIntParam ( clusterName , config , SenseiConfigServletContextListener . SENSEI_CONF_NC_CONN_TIMEOUT , connectTimeoutMillis ) ; writeTimeoutMillis = getIntParam ( clusterName , config , SenseiConfigServletContextListener . SENSEI_CONF_NC_WRITE_TIMEOUT , writeTimeoutMillis ) ; maxConnectionsPerNode = getIntParam ( clusterName , config , SenseiConfigServletContextListener . SENSEI_CONF_NC_MAX_CONN_PER_NODE , maxConnectionsPerNode ) ; staleRequestTimeoutMins = getIntParam ( clusterName , config , SenseiConfigServletContextListener . SENSEI_CONF_NC_STALE_TIMEOUT_MINS , staleRequestTimeoutMins ) ; staleRequestCleanupFrequencyMins = getIntParam ( clusterName , config , SenseiConfigServletContextListener . SENSEI_CONF_NC_STALE_CLEANUP_FREQ_MINS , staleRequestCleanupFrequencyMins ) ; } private String getStrParam ( String clusterName , Map < String , String > config , String paramName , String defaultParam ) { return config . containsKey ( clusterName + "<STR_LIT:.>" + paramName ) ? config . get ( clusterName + "<STR_LIT:.>" + paramName . substring ( paramName . lastIndexOf ( "<STR_LIT:.>" ) + <NUM_LIT:1> ) ) : defaultParam ; } private Integer getIntParam ( String clusterName , Map < String , String > config , String paramName , int defaultParam ) { return config . containsKey ( clusterName + "<STR_LIT:.>" + paramName ) ? Integer . parseInt ( config . get ( clusterName + "<STR_LIT:.>" + paramName . substring ( paramName . lastIndexOf ( "<STR_LIT:.>" ) + <NUM_LIT:1> ) ) ) : defaultParam ; } } </s>
|
<s> package com . senseidb . search . node . broker ; import java . util . Comparator ; import org . apache . commons . configuration . Configuration ; import com . linkedin . norbert . javacompat . cluster . ZooKeeperClusterClient ; import com . linkedin . norbert . javacompat . network . NetworkClientConfig ; import com . linkedin . norbert . javacompat . network . PartitionedLoadBalancerFactory ; import com . senseidb . cluster . client . SenseiNetworkClient ; import com . senseidb . conf . SenseiConfParams ; import com . senseidb . search . node . SenseiBroker ; import com . senseidb . search . node . SenseiSysBroker ; import com . senseidb . search . req . SenseiRequest ; import com . senseidb . servlet . SenseiConfigServletContextListener ; import com . senseidb . svc . api . SenseiException ; public class BrokerConfig { protected String clusterName ; protected String zkurl ; protected int zkTimeout ; protected int writeTimeoutMillis ; protected int connectTimeoutMillis ; protected int maxConnectionsPerNode ; protected int staleRequestTimeoutMins ; protected int staleRequestCleanupFrequencyMins ; protected PartitionedLoadBalancerFactory < String > loadBalancerFactory ; private final NetworkClientConfig networkClientConfig = new NetworkClientConfig ( ) ; protected boolean allowPartialMerge ; private ZooKeeperClusterClient clusterClient ; private SenseiNetworkClient networkClient ; private SenseiBroker senseiBroker ; private SenseiSysBroker senseiSysBroker ; private long brokerTimeout ; public BrokerConfig ( Configuration senseiConf , PartitionedLoadBalancerFactory < String > loadBalancerFactory ) { this . loadBalancerFactory = loadBalancerFactory ; clusterName = senseiConf . getString ( SenseiConfParams . SENSEI_CLUSTER_NAME ) ; zkurl = senseiConf . getString ( SenseiConfParams . SENSEI_CLUSTER_URL ) ; zkTimeout = senseiConf . getInt ( SenseiConfParams . SENSEI_CLUSTER_TIMEOUT , <NUM_LIT> ) ; zkurl = senseiConf . getString ( SenseiConfigServletContextListener . SENSEI_CONF_ZKURL , zkurl ) ; clusterName = senseiConf . getString ( SenseiConfigServletContextListener . SENSEI_CONF_CLUSTER_NAME , clusterName ) ; zkTimeout = senseiConf . getInt ( SenseiConfigServletContextListener . SENSEI_CONF_ZKTIMEOUT , zkTimeout ) ; connectTimeoutMillis = senseiConf . getInt ( SenseiConfigServletContextListener . SENSEI_CONF_NC_CONN_TIMEOUT , <NUM_LIT:1000> ) ; writeTimeoutMillis = senseiConf . getInt ( SenseiConfigServletContextListener . SENSEI_CONF_NC_WRITE_TIMEOUT , <NUM_LIT> ) ; maxConnectionsPerNode = senseiConf . getInt ( SenseiConfigServletContextListener . SENSEI_CONF_NC_MAX_CONN_PER_NODE , <NUM_LIT:5> ) ; staleRequestTimeoutMins = senseiConf . getInt ( SenseiConfigServletContextListener . SENSEI_CONF_NC_STALE_TIMEOUT_MINS , <NUM_LIT:10> ) ; staleRequestCleanupFrequencyMins = senseiConf . getInt ( SenseiConfigServletContextListener . SENSEI_CONF_NC_STALE_CLEANUP_FREQ_MINS , <NUM_LIT:10> ) ; allowPartialMerge = senseiConf . getBoolean ( SenseiConfParams . ALLOW_PARTIAL_MERGE , true ) ; brokerTimeout = senseiConf . getLong ( SenseiConfParams . SERVER_BROKER_TIMEOUT , <NUM_LIT> ) ; } public void init ( ) { networkClientConfig . setServiceName ( clusterName ) ; networkClientConfig . setZooKeeperConnectString ( zkurl ) ; networkClientConfig . setZooKeeperSessionTimeoutMillis ( zkTimeout ) ; networkClientConfig . setConnectTimeoutMillis ( connectTimeoutMillis ) ; networkClientConfig . setWriteTimeoutMillis ( writeTimeoutMillis ) ; networkClientConfig . setMaxConnectionsPerNode ( maxConnectionsPerNode ) ; networkClientConfig . setStaleRequestTimeoutMins ( staleRequestTimeoutMins ) ; networkClientConfig . setStaleRequestCleanupFrequencyMins ( staleRequestCleanupFrequencyMins ) ; clusterClient = new ZooKeeperClusterClient ( clusterName , zkurl , zkTimeout ) ; networkClientConfig . setClusterClient ( clusterClient ) ; networkClient = new SenseiNetworkClient ( networkClientConfig , this . loadBalancerFactory ) ; clusterClient . awaitConnectionUninterruptibly ( ) ; } public SenseiBroker buildSenseiBroker ( ) { senseiBroker = new SenseiBroker ( networkClient , clusterClient , allowPartialMerge ) ; senseiBroker . setTimeout ( brokerTimeout ) ; return senseiBroker ; } public SenseiSysBroker buildSysSenseiBroker ( Comparator < String > versionComparator ) { senseiSysBroker = new SenseiSysBroker ( networkClient , clusterClient , versionComparator , allowPartialMerge ) ; return senseiSysBroker ; } public ZooKeeperClusterClient getClusterClient ( ) { return clusterClient ; } public SenseiNetworkClient getNetworkClient ( ) { return networkClient ; } public SenseiBroker getSenseiBroker ( ) { return senseiBroker ; } public SenseiSysBroker getSenseiSysBroker ( ) { return senseiSysBroker ; } public String getClusterName ( ) { return clusterName ; } public void setClusterName ( String clusterName ) { this . clusterName = clusterName ; } public void setZkurl ( String zkurl ) { this . zkurl = zkurl ; } public void setZkTimeout ( int zkTimeout ) { this . zkTimeout = zkTimeout ; } public void setWriteTimeoutMillis ( int writeTimeoutMillis ) { this . writeTimeoutMillis = writeTimeoutMillis ; } public void setConnectTimeoutMillis ( int connectTimeoutMillis ) { this . connectTimeoutMillis = connectTimeoutMillis ; } public void setMaxConnectionsPerNode ( int maxConnectionsPerNode ) { this . maxConnectionsPerNode = maxConnectionsPerNode ; } public void setStaleRequestTimeoutMins ( int staleRequestTimeoutMins ) { this . staleRequestTimeoutMins = staleRequestTimeoutMins ; } public void setStaleRequestCleanupFrequencyMins ( int staleRequestCleanupFrequencyMins ) { this . staleRequestCleanupFrequencyMins = staleRequestCleanupFrequencyMins ; } public void setLoadBalancerFactory ( PartitionedLoadBalancerFactory < String > loadBalancerFactory ) { this . loadBalancerFactory = loadBalancerFactory ; } public void setAllowPartialMerge ( boolean allowPartialMerge ) { this . allowPartialMerge = allowPartialMerge ; } } </s>
|
<s> package com . senseidb . search . node . broker ; import java . util . ArrayList ; import java . util . HashMap ; import java . util . List ; import java . util . Map ; import com . browseengine . bobo . api . BrowseSelection ; import com . browseengine . bobo . facets . filter . FacetRangeFilter ; import com . senseidb . indexing . activity . time . Clock ; import com . senseidb . plugin . SenseiPlugin ; import com . senseidb . plugin . SenseiPluginRegistry ; import com . senseidb . search . req . SenseiRequest ; public class LayeredTimeBasedPruner implements LayeredClusterPruner , SenseiPlugin { private static final String CLUSTERS = "<STR_LIT>" ; private List < String > clusters = new ArrayList < String > ( ) ; private Map < String , DateRange > clusterRanges = new HashMap < String , DateRange > ( ) ; private static final String TIME_COLUMN = "<STR_LIT>" ; private String timeColumn ; private static class DateRange { private long startTime ; private long endTime ; public static DateRange valueOf ( String dateStr ) { String startTimeStr = dateStr . split ( "<STR_LIT:->" ) [ <NUM_LIT:0> ] ; String endTimeStr = dateStr . split ( "<STR_LIT:->" ) [ <NUM_LIT:1> ] ; DateRange ret = new DateRange ( ) ; ret . startTime = Long . parseLong ( startTimeStr ) * <NUM_LIT:24> * <NUM_LIT> * <NUM_LIT> * <NUM_LIT:1000> ; ret . endTime = Long . parseLong ( endTimeStr ) * <NUM_LIT:24> * <NUM_LIT> * <NUM_LIT> * <NUM_LIT:1000> ; return ret ; } } @ Override public void init ( Map < String , String > config , SenseiPluginRegistry pluginRegistry ) { String clustersConfig = config . get ( CLUSTERS ) ; timeColumn = config . get ( TIME_COLUMN ) ; if ( timeColumn == null ) { throw new IllegalArgumentException ( TIME_COLUMN + "<STR_LIT>" ) ; } if ( clustersConfig == null ) { throw new IllegalArgumentException ( CLUSTERS + "<STR_LIT>" ) ; } for ( String cluster : clustersConfig . split ( "<STR_LIT:U+002C>" ) ) { String trimmed = cluster . trim ( ) ; if ( trimmed . length ( ) > <NUM_LIT:0> ) { clusters . add ( trimmed ) ; String dayRange = config . get ( "<STR_LIT>" + trimmed ) ; if ( dayRange == null || dayRange . contains ( "<STR_LIT:->" ) ) { throw new IllegalStateException ( "<STR_LIT>" + trimmed + "<STR_LIT>" ) ; } clusterRanges . put ( trimmed , DateRange . valueOf ( dayRange ) ) ; } } } @ Override public void start ( ) { } @ Override public void stop ( ) { } @ Override public List < String > pruneClusters ( SenseiRequest request , List < String > clusters ) { BrowseSelection selection = request . getSelection ( timeColumn ) ; if ( selection == null ) { return clusters ; } String [ ] rangeStrings = FacetRangeFilter . getRangeStrings ( selection . getValues ( ) [ <NUM_LIT:0> ] ) ; final long start = getStartTime ( rangeStrings ) ; final long end = getEndTime ( rangeStrings ) ; List < String > ret = new ArrayList < String > ( ) ; for ( String cluster : clusters ) { DateRange clusterRange = clusterRanges . get ( cluster ) ; if ( Clock . getTime ( ) + clusterRange . startTime > end || Clock . getTime ( ) + clusterRange . endTime < start ) { } else { ret . add ( cluster ) ; } } return ret ; } @ Override public boolean clusterPrioritiesEqual ( SenseiRequest request ) { return false ; } public static long getStartTime ( String [ ] rangeStrings ) { long start ; if ( "<STR_LIT:*>" . equals ( rangeStrings [ <NUM_LIT:0> ] ) ) { start = Long . MIN_VALUE ; } else { start = Long . parseLong ( rangeStrings [ <NUM_LIT:0> ] ) ; if ( "<STR_LIT:true>" . equals ( rangeStrings [ <NUM_LIT:2> ] ) ) { start -- ; } } return start ; } public static long getEndTime ( String [ ] rangeStrings ) { long end ; if ( "<STR_LIT:*>" . equals ( rangeStrings [ <NUM_LIT:1> ] ) ) { end = Long . MAX_VALUE ; } else { end = Long . parseLong ( rangeStrings [ <NUM_LIT:1> ] ) ; if ( "<STR_LIT:true>" . equals ( rangeStrings [ <NUM_LIT:3> ] ) ) { end ++ ; } } return end ; } } </s>
|
<s> package com . senseidb . search . node . broker ; import java . util . List ; import com . senseidb . search . req . SenseiRequest ; public interface LayeredClusterPruner { public List < String > pruneClusters ( SenseiRequest request , List < String > clusters ) ; public boolean clusterPrioritiesEqual ( SenseiRequest request ) ; } </s>
|
<s> package com . senseidb . search . node . broker ; import java . util . ArrayList ; import java . util . HashMap ; import java . util . List ; import java . util . Map ; import com . linkedin . norbert . javacompat . network . PartitionedLoadBalancerFactory ; import com . senseidb . cluster . routing . SenseiPartitionedLoadBalancerFactory ; import com . senseidb . conf . SenseiConfParams ; import com . senseidb . plugin . SenseiPlugin ; import com . senseidb . plugin . SenseiPluginRegistry ; import com . senseidb . search . node . Broker ; import com . senseidb . search . node . ResultMerger ; import com . senseidb . search . node . SenseiBroker ; import com . senseidb . search . req . SenseiRequest ; import com . senseidb . search . req . SenseiResult ; import com . senseidb . svc . api . SenseiException ; public class LayeredBroker implements SenseiPlugin , Broker < SenseiRequest , SenseiResult > { private static final String CLUSTERS = "<STR_LIT>" ; private List < String > clusters = new ArrayList < String > ( ) ; private Map < String , CompoundBrokerConfig > clusterBrokerConfig = new HashMap < String , CompoundBrokerConfig > ( ) ; private Map < String , SenseiBroker > brokers = new HashMap < String , SenseiBroker > ( ) ; private LayeredClusterPruner federatedPruner ; @ Override public void init ( Map < String , String > config , SenseiPluginRegistry pluginRegistry ) { String clustersConfig = config . get ( CLUSTERS ) ; if ( clustersConfig == null ) { throw new IllegalArgumentException ( "<STR_LIT>" ) ; } federatedPruner = pluginRegistry . getBeanByFullPrefix ( SenseiConfParams . SENSEI_FEDERATED_BROKER_PRUNER , LayeredClusterPruner . class ) ; if ( federatedPruner == null ) { federatedPruner = new AllClustersPruner ( ) ; } PartitionedLoadBalancerFactory < String > routerFactory = pluginRegistry . getBeanByFullPrefix ( SenseiConfParams . SERVER_SEARCH_ROUTER_FACTORY , PartitionedLoadBalancerFactory . class ) ; if ( routerFactory == null ) { routerFactory = new SenseiPartitionedLoadBalancerFactory ( <NUM_LIT> ) ; } for ( String cluster : clustersConfig . split ( "<STR_LIT:U+002C>" ) ) { String trimmed = cluster . trim ( ) ; if ( trimmed . length ( ) > <NUM_LIT:0> ) { clusters . add ( trimmed ) ; clusterBrokerConfig . put ( trimmed , new CompoundBrokerConfig ( pluginRegistry . getConfiguration ( ) , routerFactory , config , trimmed ) ) ; } } } @ Override public void start ( ) { for ( String cluster : clusters ) { CompoundBrokerConfig brokerConfig = clusterBrokerConfig . get ( cluster ) ; brokerConfig . init ( ) ; brokers . put ( cluster , brokerConfig . buildSenseiBroker ( ) ) ; } } @ Override public void stop ( ) { for ( CompoundBrokerConfig brokerConfig : clusterBrokerConfig . values ( ) ) { brokerConfig . getSenseiBroker ( ) . shutdown ( ) ; brokerConfig . getNetworkClient ( ) . shutdown ( ) ; brokerConfig . getClusterClient ( ) . shutdown ( ) ; } } public void warmUp ( ) { for ( SenseiBroker broker : brokers . values ( ) ) { try { broker . browse ( new SenseiRequest ( ) ) ; } catch ( SenseiException e ) { throw new RuntimeException ( e ) ; } } } public SenseiResult browse ( final SenseiRequest req ) throws SenseiException { List < String > prunedClusters = federatedPruner . pruneClusters ( req , clusters ) ; int count = req . getCount ( ) ; int offset = req . getOffset ( ) ; if ( count == <NUM_LIT:0> ) { return new SenseiResult ( ) ; } List < SenseiResult > results = new ArrayList < SenseiResult > ( ) ; if ( ! federatedPruner . clusterPrioritiesEqual ( req ) ) { for ( String cluster : prunedClusters ) { if ( count <= <NUM_LIT:0> ) { break ; } SenseiRequest request = req . clone ( ) ; request . setCount ( count ) ; request . setOffset ( offset ) ; SenseiResult currentResult = brokers . get ( cluster ) . browse ( request ) ; int numHits = currentResult . getNumHits ( ) ; if ( offset >= numHits ) { offset -= numHits ; continue ; } else { numHits -= offset ; offset = <NUM_LIT:0> ; count -= numHits ; results . add ( currentResult ) ; } } } else { for ( String cluster : prunedClusters ) { SenseiRequest request = req . clone ( ) ; SenseiResult currentResult = brokers . get ( cluster ) . browse ( request ) ; results . add ( currentResult ) ; } } SenseiResult res = ResultMerger . merge ( req , results , false ) ; return res ; } } </s>
|
<s> package com . senseidb . search . node ; import proj . zoie . api . indexing . AbstractZoieIndexableInterpreter ; import proj . zoie . api . indexing . ZoieIndexable ; public class NoOpIndexableInterpreter extends AbstractZoieIndexableInterpreter < ZoieIndexable > { @ Override public ZoieIndexable convertAndInterpret ( ZoieIndexable obj ) { return obj ; } } </s>
|
<s> package com . senseidb . search . node ; import it . unimi . dsi . fastutil . ints . IntSet ; import org . apache . log4j . Logger ; import com . google . protobuf . Message ; import com . linkedin . norbert . NorbertException ; import com . linkedin . norbert . javacompat . cluster . ClusterListener ; import com . linkedin . norbert . javacompat . network . PartitionedNetworkClient ; import com . senseidb . search . req . AbstractSenseiRequest ; import com . senseidb . search . req . AbstractSenseiResult ; import com . senseidb . svc . api . SenseiException ; public abstract class AbstractSenseiBroker < REQUEST extends AbstractSenseiRequest , RESULT extends AbstractSenseiResult > implements ClusterListener , Broker < REQUEST , RESULT > { private final static Logger logger = Logger . getLogger ( AbstractSenseiBroker . class ) ; protected final PartitionedNetworkClient < String > _networkClient ; protected volatile IntSet _partitions = null ; public AbstractSenseiBroker ( PartitionedNetworkClient < String > networkClient ) throws NorbertException { _networkClient = networkClient ; } public abstract RESULT getEmptyResultInstance ( ) ; public RESULT browse ( final REQUEST req ) throws SenseiException { if ( _partitions == null ) throw new SenseiException ( "<STR_LIT>" ) ; try { return doBrowse ( _networkClient , req , _partitions ) ; } catch ( Exception e ) { throw new SenseiException ( e . getMessage ( ) , e ) ; } } protected abstract RESULT doBrowse ( PartitionedNetworkClient < String > networkClient , REQUEST req , IntSet partitions ) throws Exception ; public void shutdown ( ) { logger . info ( "<STR_LIT>" ) ; } public abstract void setTimeout ( long timeoutMillis ) ; public abstract long getTimeout ( ) ; } </s>
|
<s> package com . senseidb . search . node ; public interface ServerBootstrapFactory { public static class DefaultServerBootstrapFactory implements ServerBootstrapFactory { } } </s>
|
<s> package com . senseidb . search . node ; import java . util . Comparator ; import java . util . List ; import java . util . Set ; import org . apache . log4j . Logger ; import com . google . protobuf . Message ; import com . linkedin . norbert . javacompat . cluster . Node ; import com . senseidb . search . req . SenseiRequest ; import com . senseidb . search . req . SenseiSystemInfo ; public class SenseiSysScatterGatherHandler extends AbstractSenseiScatterGatherHandler < SenseiRequest , SenseiSystemInfo > { private final static Logger logger = Logger . getLogger ( SenseiSysScatterGatherHandler . class ) ; private final Comparator < String > _versionComparator ; public SenseiSysScatterGatherHandler ( SenseiRequest request , Comparator < String > versionComparator ) { super ( request ) ; _versionComparator = versionComparator ; } @ Override public SenseiRequest customizeRequest ( SenseiRequest senseiRequest , Node node , Set < Integer > partitions ) { senseiRequest . setPartitions ( partitions ) ; return senseiRequest ; } @ Override public SenseiSystemInfo mergeResults ( SenseiRequest request , List < SenseiSystemInfo > resultList ) { SenseiSystemInfo result = new SenseiSystemInfo ( ) ; if ( resultList == null ) return result ; for ( SenseiSystemInfo res : resultList ) { result . setNumDocs ( result . getNumDocs ( ) + res . getNumDocs ( ) ) ; if ( result . getLastModified ( ) < res . getLastModified ( ) ) result . setLastModified ( res . getLastModified ( ) ) ; if ( result . getVersion ( ) == null || _versionComparator . compare ( result . getVersion ( ) , res . getVersion ( ) ) < <NUM_LIT:0> ) result . setVersion ( res . getVersion ( ) ) ; if ( res . getFacetInfos ( ) != null ) result . setFacetInfos ( res . getFacetInfos ( ) ) ; if ( res . getClusterInfo ( ) != null ) { if ( result . getClusterInfo ( ) != null ) result . getClusterInfo ( ) . addAll ( res . getClusterInfo ( ) ) ; else result . setClusterInfo ( res . getClusterInfo ( ) ) ; } } return result ; } } </s>
|
<s> package com . senseidb . search . node ; import java . util . Date ; import java . util . HashMap ; import java . util . HashSet ; import java . util . Map ; import java . util . Set ; import org . apache . log4j . Logger ; import proj . zoie . api . DataProvider ; import proj . zoie . api . IndexReaderFactory ; import proj . zoie . api . Zoie ; import proj . zoie . api . ZoieException ; import proj . zoie . api . ZoieIndexReader ; import com . browseengine . bobo . api . BoboIndexReader ; import com . browseengine . bobo . facets . FacetHandler ; import com . browseengine . bobo . facets . RuntimeFacetHandlerFactory ; import com . senseidb . indexing . SenseiIndexPruner ; import com . senseidb . indexing . SenseiIndexPruner . DefaultSenseiIndexPruner ; import com . senseidb . indexing . activity . CompositeActivityManager ; import com . senseidb . jmx . JmxUtil ; import com . senseidb . search . plugin . PluggableSearchEngineManager ; import com . senseidb . search . req . SenseiSystemInfo ; public class SenseiCore { private static final Logger logger = Logger . getLogger ( SenseiServer . class ) ; private SenseiZoieFactory < ? > _zoieFactory ; private SenseiIndexingManager _indexManager ; private SenseiQueryBuilderFactory _queryBuilderFactory ; private final HashSet < Zoie < BoboIndexReader , ? > > zoieSystems = new HashSet < Zoie < BoboIndexReader , ? > > ( ) ; private final int [ ] _partitions ; private final int _id ; private final Map < Integer , Zoie < BoboIndexReader , ? > > _readerFactoryMap ; private SenseiSystemInfo _senseiSystemInfo ; private volatile boolean _started ; private SenseiIndexPruner _pruner ; private PluggableSearchEngineManager pluggableSearchEngineManager ; public SenseiCore ( int id , int [ ] partitions , SenseiZoieFactory < ? > zoieSystemFactory , SenseiIndexingManager indexManager , SenseiQueryBuilderFactory queryBuilderFactory ) { _zoieFactory = zoieSystemFactory ; _indexManager = indexManager ; _queryBuilderFactory = queryBuilderFactory ; _partitions = partitions ; _id = id ; _readerFactoryMap = new HashMap < Integer , Zoie < BoboIndexReader , ? > > ( ) ; _started = false ; _pruner = null ; } public void setIndexPruner ( SenseiIndexPruner pruner ) { _pruner = pruner ; } public SenseiIndexPruner getIndexPruner ( ) { return _pruner == null ? new DefaultSenseiIndexPruner ( ) : _pruner ; } public int getNodeId ( ) { return _id ; } public int [ ] getPartitions ( ) { return _partitions ; } public SenseiSystemInfo getSystemInfo ( ) { if ( _senseiSystemInfo == null ) _senseiSystemInfo = new SenseiSystemInfo ( ) ; if ( _senseiSystemInfo . getFacetInfos ( ) == null ) { Set < SenseiSystemInfo . SenseiFacetInfo > facetInfos = new HashSet < SenseiSystemInfo . SenseiFacetInfo > ( ) ; if ( _zoieFactory . getDecorator ( ) != null && _zoieFactory . getDecorator ( ) . getFacetHandlerList ( ) != null ) { for ( FacetHandler < ? > facetHandler : _zoieFactory . getDecorator ( ) . getFacetHandlerList ( ) ) { facetInfos . add ( new SenseiSystemInfo . SenseiFacetInfo ( facetHandler . getName ( ) ) ) ; } } if ( _zoieFactory . getDecorator ( ) != null && _zoieFactory . getDecorator ( ) . getFacetHandlerFactories ( ) != null ) { for ( RuntimeFacetHandlerFactory < ? , ? > runtimeFacetHandlerFactory : _zoieFactory . getDecorator ( ) . getFacetHandlerFactories ( ) ) { SenseiSystemInfo . SenseiFacetInfo facetInfo = new SenseiSystemInfo . SenseiFacetInfo ( runtimeFacetHandlerFactory . getName ( ) ) ; facetInfo . setRunTime ( true ) ; facetInfos . add ( facetInfo ) ; } } _senseiSystemInfo . setFacetInfos ( facetInfos ) ; } Date lastModified = new Date ( <NUM_LIT> ) ; String version = null ; for ( Zoie < BoboIndexReader , ? > zoieSystem : zoieSystems ) { if ( version == null || _zoieFactory . getVersionComparator ( ) . compare ( version , zoieSystem . getVersion ( ) ) < <NUM_LIT:0> ) version = zoieSystem . getVersion ( ) ; } _senseiSystemInfo . setLastModified ( lastModified . getTime ( ) ) ; if ( version != null ) _senseiSystemInfo . setVersion ( version ) ; return _senseiSystemInfo ; } public void setSystemInfo ( SenseiSystemInfo senseiSystemInfo ) { _senseiSystemInfo = senseiSystemInfo ; } public void start ( ) throws Exception { if ( _started ) return ; for ( int part : _partitions ) { Zoie < BoboIndexReader , ? > zoieSystem = _zoieFactory . getZoieInstance ( _id , part ) ; String [ ] mbeannames = zoieSystem . getStandardMBeanNames ( ) ; for ( String name : mbeannames ) { JmxUtil . registerMBean ( zoieSystem . getStandardMBean ( name ) , "<STR_LIT>" , name + "<STR_LIT:->" + _id + "<STR_LIT:->" + part ) ; } if ( ! zoieSystems . contains ( zoieSystem ) ) { zoieSystem . start ( ) ; zoieSystems . add ( zoieSystem ) ; } _readerFactoryMap . put ( part , zoieSystem ) ; } try { pluggableSearchEngineManager . start ( this ) ; logger . info ( "<STR_LIT>" ) ; if ( _indexManager != null ) { _indexManager . initialize ( _readerFactoryMap ) ; } logger . info ( "<STR_LIT>" ) ; if ( _indexManager != null ) { _indexManager . start ( ) ; } logger . info ( "<STR_LIT>" ) ; } catch ( Exception e ) { logger . error ( "<STR_LIT>" , e ) ; } _started = true ; } public void shutdown ( ) { if ( ! _started ) return ; logger . info ( "<STR_LIT>" ) ; logger . info ( "<STR_LIT>" ) ; if ( _indexManager != null ) { _indexManager . shutdown ( ) ; } logger . info ( "<STR_LIT>" ) ; for ( Zoie < BoboIndexReader , ? > zoieSystem : zoieSystems ) { zoieSystem . shutdown ( ) ; } zoieSystems . clear ( ) ; _started = false ; } public DataProvider getDataProvider ( ) { return _indexManager . getDataProvider ( ) ; } public IndexReaderFactory < ZoieIndexReader < BoboIndexReader > > getIndexReaderFactory ( int partition ) { IndexReaderFactory < ZoieIndexReader < BoboIndexReader > > readerFactory = _readerFactoryMap . get ( partition ) ; if ( readerFactory == null ) { logger . error ( "<STR_LIT>" + partition + "<STR_LIT>" + _readerFactoryMap . keySet ( ) + "<STR_LIT>" ) ; } return readerFactory ; } public SenseiQueryBuilderFactory getQueryBuilderFactory ( ) { return _queryBuilderFactory ; } public void syncWithVersion ( long timeToWait , String version ) throws ZoieException { _indexManager . syncWithVersion ( timeToWait , version ) ; } public void setPluggableSearchEngineManager ( PluggableSearchEngineManager pluggableSearchEngineManager ) { this . pluggableSearchEngineManager = pluggableSearchEngineManager ; } public PluggableSearchEngineManager getPluggableSearchEngineManager ( ) { return pluggableSearchEngineManager ; } } </s>
|
<s> package com . senseidb . search . node ; import com . linkedin . norbert . javacompat . network . RequestBuilder ; import com . linkedin . norbert . network . ResponseIterator ; import com . linkedin . norbert . network . common . ExceptionIterator ; import com . linkedin . norbert . network . common . PartialIterator ; import com . linkedin . norbert . network . common . TimeoutIterator ; import com . senseidb . metrics . MetricFactory ; import com . senseidb . search . req . SenseiRequest ; import it . unimi . dsi . fastutil . ints . IntOpenHashSet ; import it . unimi . dsi . fastutil . ints . IntSet ; import java . util . ArrayList ; import java . util . List ; import java . util . Set ; import java . util . concurrent . Callable ; import java . util . concurrent . ExecutionException ; import java . util . concurrent . TimeUnit ; import org . apache . commons . lang . RandomStringUtils ; import org . apache . log4j . Logger ; import com . linkedin . norbert . NorbertException ; import com . linkedin . norbert . javacompat . cluster . Node ; import com . linkedin . norbert . javacompat . network . PartitionedNetworkClient ; import com . linkedin . norbert . network . Serializer ; import com . senseidb . metrics . MetricsConstants ; import com . senseidb . search . req . AbstractSenseiRequest ; import com . senseidb . search . req . AbstractSenseiResult ; import com . senseidb . search . req . ErrorType ; import com . senseidb . search . req . SenseiError ; import com . senseidb . svc . api . SenseiException ; import com . yammer . metrics . core . Meter ; import com . yammer . metrics . core . MetricName ; import com . yammer . metrics . core . Timer ; public abstract class AbstractConsistentHashBroker < REQUEST extends AbstractSenseiRequest , RESULT extends AbstractSenseiResult > extends AbstractSenseiBroker < REQUEST , RESULT > { private final static Logger logger = Logger . getLogger ( AbstractConsistentHashBroker . class ) ; protected long _timeout = <NUM_LIT> ; protected final Serializer < REQUEST , RESULT > _serializer ; private final Timer _scatterTimer ; private final Timer _gatherTimer ; private final Timer _totalTimer ; private final Meter _searchCounter ; private final Meter _errorMeter ; private final Meter _emptyMeter ; public AbstractConsistentHashBroker ( PartitionedNetworkClient < String > networkClient , Serializer < REQUEST , RESULT > serializer ) throws NorbertException { super ( networkClient ) ; _serializer = serializer ; MetricName scatterMetricName = new MetricName ( MetricsConstants . Domain , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" ) ; _scatterTimer = MetricFactory . newTimer ( scatterMetricName , TimeUnit . MILLISECONDS , TimeUnit . SECONDS ) ; MetricName gatherMetricName = new MetricName ( MetricsConstants . Domain , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" ) ; _gatherTimer = MetricFactory . newTimer ( gatherMetricName , TimeUnit . MILLISECONDS , TimeUnit . SECONDS ) ; MetricName totalMetricName = new MetricName ( MetricsConstants . Domain , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" ) ; _totalTimer = MetricFactory . newTimer ( totalMetricName , TimeUnit . MILLISECONDS , TimeUnit . SECONDS ) ; MetricName searchCounterMetricName = new MetricName ( MetricsConstants . Domain , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" ) ; _searchCounter = MetricFactory . newMeter ( searchCounterMetricName , "<STR_LIT>" , TimeUnit . SECONDS ) ; MetricName errorMetricName = new MetricName ( MetricsConstants . Domain , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" ) ; _errorMeter = MetricFactory . newMeter ( errorMetricName , "<STR_LIT>" , TimeUnit . SECONDS ) ; MetricName emptyMetricName = new MetricName ( MetricsConstants . Domain , "<STR_LIT>" , "<STR_LIT>" , "<STR_LIT>" ) ; _emptyMeter = MetricFactory . newMeter ( emptyMetricName , "<STR_LIT>" , TimeUnit . SECONDS ) ; } public < T > T customizeRequest ( REQUEST request ) { return ( T ) request ; } protected IntSet getPartitions ( Set < Node > nodes ) { IntSet partitionSet = new IntOpenHashSet ( ) ; for ( Node n : nodes ) { partitionSet . addAll ( n . getPartitionIds ( ) ) ; } return partitionSet ; } public abstract RESULT getEmptyResultInstance ( ) ; public RESULT browse ( final REQUEST req ) throws SenseiException { _searchCounter . mark ( ) ; try { return _totalTimer . time ( new Callable < RESULT > ( ) { @ Override public RESULT call ( ) throws Exception { return doBrowse ( _networkClient , req , _partitions ) ; } } ) ; } catch ( Exception e ) { _errorMeter . mark ( ) ; throw new SenseiException ( e . getMessage ( ) , e ) ; } } public abstract RESULT mergeResults ( REQUEST request , List < RESULT > resultList ) ; protected String getRouteParam ( REQUEST req ) { String param = req . getRouteParam ( ) ; if ( param == null ) { return RandomStringUtils . random ( <NUM_LIT:4> ) ; } else { return param ; } } protected RESULT doBrowse ( PartitionedNetworkClient < String > networkClient , final REQUEST req , IntSet partitions ) { final long time = System . currentTimeMillis ( ) ; final List < RESULT > resultList = new ArrayList < RESULT > ( ) ; try { resultList . addAll ( _scatterTimer . time ( new Callable < List < RESULT > > ( ) { @ Override public List < RESULT > call ( ) throws Exception { return doCall ( req ) ; } } ) ) ; } catch ( Exception e ) { _errorMeter . mark ( ) ; RESULT emptyResult = getEmptyResultInstance ( ) ; logger . error ( "<STR_LIT>" , e ) ; emptyResult . addError ( new SenseiError ( "<STR_LIT>" + e . getMessage ( ) , ErrorType . BrokerGatherError ) ) ; return emptyResult ; } if ( resultList . size ( ) == <NUM_LIT:0> ) { logger . error ( "<STR_LIT>" ) ; RESULT emptyResult = getEmptyResultInstance ( ) ; emptyResult . addError ( new SenseiError ( "<STR_LIT>" + "<STR_LIT>" , ErrorType . BrokerGatherError ) ) ; _emptyMeter . mark ( ) ; return emptyResult ; } RESULT result = null ; try { result = _gatherTimer . time ( new Callable < RESULT > ( ) { @ Override public RESULT call ( ) throws Exception { return mergeResults ( req , resultList ) ; } } ) ; } catch ( Exception e ) { result = getEmptyResultInstance ( ) ; logger . error ( "<STR_LIT>" , e ) ; result . addError ( new SenseiError ( "<STR_LIT>" + e . getMessage ( ) , ErrorType . BrokerGatherError ) ) ; _errorMeter . mark ( ) ; } if ( logger . isDebugEnabled ( ) ) { logger . debug ( "<STR_LIT>" + ( System . currentTimeMillis ( ) - time ) + "<STR_LIT>" ) ; } return result ; } protected List < RESULT > doCall ( final REQUEST req ) throws ExecutionException { List < RESULT > resultList = new ArrayList < RESULT > ( ) ; ResponseIterator < RESULT > responseIterator = buildIterator ( _networkClient . sendRequestToOneReplica ( getRouteParam ( req ) , new RequestBuilder < Integer , REQUEST > ( ) { private int count = <NUM_LIT:0> ; @ Override public REQUEST apply ( Node node , Set < Integer > nodePartitions ) { REQUEST clone = ( REQUEST ) ( ( ( SenseiRequest ) req ) . clone ( ) ) ; clone . setPartitions ( nodePartitions ) ; REQUEST customizedRequest = customizeRequest ( clone ) ; return customizedRequest ; } } , _serializer ) ) ; while ( responseIterator . hasNext ( ) ) { resultList . add ( responseIterator . next ( ) ) ; } logger . debug ( String . format ( "<STR_LIT>" , resultList . size ( ) ) ) ; return resultList ; } protected ResponseIterator < RESULT > buildIterator ( ResponseIterator < RESULT > responseIterator ) { TimeoutIterator < RESULT > timeoutIterator = new TimeoutIterator < RESULT > ( responseIterator , _timeout ) ; if ( allowPartialMerge ( ) ) { return new PartialIterator < RESULT > ( new ExceptionIterator < RESULT > ( timeoutIterator ) ) ; } return timeoutIterator ; } public void shutdown ( ) { logger . info ( "<STR_LIT>" ) ; if ( _scatterTimer != null ) { _scatterTimer . stop ( ) ; } if ( _gatherTimer != null ) { _gatherTimer . stop ( ) ; } if ( _totalTimer != null ) { _totalTimer . stop ( ) ; } if ( _searchCounter != null ) { _searchCounter . stop ( ) ; } if ( _errorMeter != null ) { _errorMeter . stop ( ) ; } if ( _emptyMeter != null ) { _emptyMeter . stop ( ) ; } } public long getTimeout ( ) { return _timeout ; } public void setTimeout ( long timeout ) { this . _timeout = timeout ; } public abstract boolean allowPartialMerge ( ) ; } </s>
|
<s> package com . senseidb . search . node . impl ; import java . io . UnsupportedEncodingException ; import org . apache . lucene . queryParser . ParseException ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . Filter ; import org . apache . lucene . search . Query ; import com . senseidb . search . node . SenseiQueryBuilder ; import com . senseidb . search . req . SenseiQuery ; public class SimpleQueryBuilder implements SenseiQueryBuilder { protected Query _query = null ; protected Filter _filter = null ; public SimpleQueryBuilder ( SenseiQuery query , QueryParser parser ) throws Exception { doBuild ( query , parser ) ; } protected void doBuild ( SenseiQuery query , QueryParser parser ) throws Exception { if ( query != null ) { byte [ ] bytes = query . toBytes ( ) ; String qString = null ; try { qString = new String ( bytes , "<STR_LIT:UTF-8>" ) ; } catch ( UnsupportedEncodingException e ) { throw new ParseException ( e . getMessage ( ) ) ; } if ( qString . length ( ) > <NUM_LIT:0> ) { synchronized ( parser ) { _query = parser . parse ( qString ) ; } } } } public Query buildQuery ( ) throws ParseException { return _query ; } public Filter buildFilter ( ) { return _filter ; } } </s>
|
<s> package com . senseidb . search . node . impl ; import java . util . Map ; import proj . zoie . api . DataProvider ; import proj . zoie . api . Zoie ; import proj . zoie . api . ZoieException ; import com . browseengine . bobo . api . BoboIndexReader ; import com . senseidb . search . node . SenseiIndexingManager ; public class NoopIndexingManager < D > implements SenseiIndexingManager < D > { @ Override public void initialize ( Map < Integer , Zoie < BoboIndexReader , D > > zoieSystemMap ) throws Exception { } @ Override public void shutdown ( ) { } @ Override public void start ( ) throws Exception { } @ Override public DataProvider getDataProvider ( ) { return null ; } @ Override public void syncWithVersion ( long timeToWait , String version ) throws ZoieException { } } </s>
|
<s> package com . senseidb . search . node . impl ; import org . apache . lucene . queryParser . QueryParser ; import com . senseidb . search . node . SenseiQueryBuilder ; import com . senseidb . search . node . SenseiQueryBuilderFactory ; import com . senseidb . search . req . SenseiQuery ; public class SimpleQueryBuilderFactory implements SenseiQueryBuilderFactory { private final QueryParser _parser ; public SimpleQueryBuilderFactory ( QueryParser parser ) { _parser = parser ; } public SenseiQueryBuilder getQueryBuilder ( SenseiQuery query ) throws Exception { return new SimpleQueryBuilder ( query , _parser ) ; } } </s>
|
<s> package com . senseidb . search . node . impl ; import java . io . File ; import proj . zoie . api . DirectoryManager . DIRECTORY_MODE ; import proj . zoie . api . indexing . ZoieIndexableInterpreter ; import proj . zoie . impl . indexing . ZoieConfig ; import proj . zoie . impl . indexing . ZoieSystem ; import com . browseengine . bobo . api . BoboIndexReader ; import com . senseidb . conf . ZoieFactoryFactory ; import com . senseidb . search . node . SenseiIndexReaderDecorator ; import com . senseidb . search . node . SenseiZoieFactory ; import com . senseidb . search . node . SenseiZoieSystemFactory ; public class DemoZoieSystemFactory < T > extends SenseiZoieSystemFactory < T > { private ZoieSystem < BoboIndexReader , T > _zoieSystem = null ; public DemoZoieSystemFactory ( File idxDir , ZoieIndexableInterpreter < T > interpreter , SenseiIndexReaderDecorator indexReaderDecorator , ZoieConfig zoieConfig ) { super ( idxDir , DIRECTORY_MODE . SIMPLE , interpreter , indexReaderDecorator , zoieConfig ) ; } public DemoZoieSystemFactory ( File idxDir , ZoieIndexableInterpreter < T > interpreter , ZoieConfig zoieConfig ) { super ( idxDir , DIRECTORY_MODE . SIMPLE , interpreter , new SenseiIndexReaderDecorator ( ) , zoieConfig ) ; } @ Override public ZoieSystem < BoboIndexReader , T > getZoieInstance ( int nodeId , int partitionId ) { if ( _zoieSystem == null ) { _zoieSystem = super . getZoieInstance ( nodeId , partitionId ) ; } return _zoieSystem ; } @ Override public File getPath ( int nodeId , int partitionId ) { return _idxDir ; } public static class DemoZoieFactoryFactory implements ZoieFactoryFactory { @ Override public SenseiZoieFactory < ? > getZoieFactory ( File idxDir , ZoieIndexableInterpreter < ? > interpreter , SenseiIndexReaderDecorator decorator , ZoieConfig config ) { return new DemoZoieSystemFactory ( idxDir , interpreter , decorator , config ) ; } } } </s>
|
<s> package com . senseidb . search . node . impl ; import org . apache . log4j . Logger ; import org . apache . lucene . queryParser . ParseException ; import org . apache . lucene . queryParser . QueryParser ; import org . apache . lucene . search . Filter ; import org . apache . lucene . search . MatchAllDocsQuery ; import org . apache . lucene . search . Query ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . search . node . SenseiQueryBuilder ; import com . senseidb . search . query . QueryConstructor ; import com . senseidb . search . query . filters . FilterConstructor ; public class DefaultJsonQueryBuilderFactory extends AbstractJsonQueryBuilderFactory { private static Logger logger = Logger . getLogger ( DefaultJsonQueryBuilderFactory . class ) ; private final QueryParser _qparser ; public DefaultJsonQueryBuilderFactory ( QueryParser qparser ) { _qparser = qparser ; } @ Override public SenseiQueryBuilder buildQueryBuilder ( JSONObject jsonQuery ) { final JSONObject query ; final JSONObject filter ; if ( jsonQuery != null ) { Object obj = jsonQuery . opt ( "<STR_LIT:query>" ) ; if ( obj == null ) query = null ; else if ( obj instanceof JSONObject ) query = ( JSONObject ) obj ; else if ( obj instanceof String ) { query = new JSONObject ( ) ; JSONObject tmp = new JSONObject ( ) ; try { tmp . put ( "<STR_LIT:query>" , ( String ) obj ) ; query . put ( "<STR_LIT>" , tmp ) ; } catch ( JSONException jse ) { } } else { throw new IllegalArgumentException ( "<STR_LIT>" + jsonQuery ) ; } filter = jsonQuery . optJSONObject ( "<STR_LIT>" ) ; } else { query = null ; filter = null ; } return new SenseiQueryBuilder ( ) { @ Override public Filter buildFilter ( ) throws ParseException { try { return FilterConstructor . constructFilter ( filter , _qparser ) ; } catch ( Exception e ) { logger . error ( e . getMessage ( ) , e ) ; throw new ParseException ( e . getMessage ( ) ) ; } } @ Override public Query buildQuery ( ) throws ParseException { try { Query q = QueryConstructor . constructQuery ( query , _qparser ) ; if ( q == null ) { q = new MatchAllDocsQuery ( ) ; } return q ; } catch ( Exception e ) { logger . error ( e . getMessage ( ) , e ) ; throw new ParseException ( e . getMessage ( ) ) ; } } } ; } } </s>
|
<s> package com . senseidb . search . node . impl ; import java . nio . charset . Charset ; import org . json . JSONException ; import org . json . JSONObject ; import com . senseidb . conf . SenseiSchema ; import proj . zoie . store . ZoieStoreSerializer ; public class JSONDataSerializer implements ZoieStoreSerializer < JSONObject > { private static Charset UTF8 = Charset . forName ( "<STR_LIT:UTF-8>" ) ; private final String _uidField ; private final String _delField ; private final String _skipFieldField ; public JSONDataSerializer ( SenseiSchema schema ) { _uidField = schema . getUidField ( ) ; _delField = schema . getDeleteField ( ) ; _skipFieldField = schema . getSkipField ( ) ; } @ Override public long getUid ( JSONObject data ) { try { return Long . parseLong ( data . optString ( _uidField , "<STR_LIT:-1>" ) ) ; } catch ( Exception e ) { return - <NUM_LIT:1L> ; } } @ Override public byte [ ] toBytes ( JSONObject data ) { return data . toString ( ) . getBytes ( UTF8 ) ; } @ Override public JSONObject fromBytes ( byte [ ] data ) { try { return new JSONObject ( new String ( data , UTF8 ) ) ; } catch ( JSONException e ) { return null ; } } @ Override public boolean isDelete ( JSONObject obj ) { return obj . optBoolean ( _delField ) ; } @ Override public boolean isSkip ( JSONObject obj ) { return obj . optBoolean ( _skipFieldField ) ; } } </s>
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.