Example usage for org.apache.lucene.search Sort rewrite

List of usage examples for org.apache.lucene.search Sort rewrite

Introduction

In this page you can find the example usage for org.apache.lucene.search Sort rewrite.

Prototype

public Sort rewrite(IndexSearcher searcher) throws IOException 

Source Link

Document

Rewrites the SortFields in this Sort, returning a new Sort if any of the fields changes during their rewriting.

Usage

From source file:com.stratio.cassandra.lucene.index.RAMIndex.java

License:Apache License

/**
 * Finds the top {@code count} hits for {@code query} and sorting the hits by {@code sort}.
 *
 * @param query the {@link Query} to search for
 * @param sort the {@link Sort} to be applied
 * @param count the max number of results to be collected
 * @param fields the names of the fields to be loaded
 * @return the found documents/*from   w  w w  .  j  a  v  a 2  s  . c o m*/
 */
public List<Document> search(Query query, Sort sort, Integer count, Set<String> fields) {
    try {
        indexWriter.commit();
        IndexReader reader = DirectoryReader.open(directory);
        IndexSearcher searcher = new IndexSearcher(reader);
        sort = sort.rewrite(searcher);
        TopDocs topDocs = searcher.search(query, count, sort);
        ScoreDoc[] scoreDocs = topDocs.scoreDocs;
        List<Document> documents = new LinkedList<>();
        for (ScoreDoc scoreDoc : scoreDocs) {
            Document document = searcher.doc(scoreDoc.doc, fields);
            documents.add(document);
        }
        searcher.getIndexReader().close();
        return documents;
    } catch (IOException e) {
        throw new IndexException(logger, e, "Error while searching");
    }
}

From source file:io.puntanegra.fhir.index.lucene.LuceneRAMIndex.java

License:Apache License

/**
 * Finds the top {@code count} hits for {@code query} and sorting the hits
 * by {@code sort}./*  w ww .  ja  v a  2  s  . c o m*/
 *
 * @param query
 *            the {@link Query} to search for
 * @param sort
 *            the {@link Sort} to be applied
 * @param count
 *            the max number of results to be collected
 * @param fields
 *            the names of the fields to be loaded
 * @return the found documents
 */
public List<Document> search(Query query, Sort sort, Integer count, Set<String> fields) {
    try {
        indexWriter.commit();
        IndexReader reader = DirectoryReader.open(directory);
        IndexSearcher searcher = new IndexSearcher(reader);
        sort = sort.rewrite(searcher);
        TopDocs topDocs = searcher.search(query, count, sort);
        ScoreDoc[] scoreDocs = topDocs.scoreDocs;
        List<Document> documents = new LinkedList<>();
        for (ScoreDoc scoreDoc : scoreDocs) {
            Document document = searcher.doc(scoreDoc.doc, fields);
            documents.add(document);
        }
        searcher.getIndexReader().close();
        return documents;
    } catch (IOException e) {
        throw new FhirIndexException(e, "Error while searching");
    }
}

From source file:org.alfresco.solr.SolrInformationServer.java

License:Open Source License

@Override
public List<TenantAclIdDbId> getDocsWithUncleanContent(int start, int rows) throws IOException {
    RefCounted<SolrIndexSearcher> refCounted = null;
    try {//from w  ww.j a v  a  2 s .  c  om
        List<TenantAclIdDbId> docIds = new ArrayList<>();
        refCounted = this.core.getSearcher();
        SolrIndexSearcher searcher = refCounted.get();

        /*
        *  Below is the code for purging the cleanContentCache.
        *  The cleanContentCache is an in-memory LRU cache of the transactions that have already
        *  had their content fetched. This is needed because the ContentTracker does not have an up-to-date
        *  snapshot of the index to determine which nodes are marked as dirty/new. The cleanContentCache is used
        *  to filter out nodes that belong to transactions that have already been processed, which stops them from
        *  being re-processed.
        *
        *  The cleanContentCache needs to be purged periodically to support retrying of failed content fetches.
        *  This is because fetches for individual nodes within the transaction may have failed, but the transaction will still be in the
        *  cleanContentCache, which prevents it from being retried.
        *
        *  Once a transaction is purged from the cleanContentCache it will be retried automatically if it is marked dirty/new
        *  in current snapshot of the index.
        *
        *  The code below runs every two minutes and purges transactions from the
        *  cleanContentCache that is more then 20 minutes old.
        *
        */
        long purgeTime = System.currentTimeMillis();
        if (purgeTime - cleanContentLastPurged > 120000) {
            Iterator<Entry<Long, Long>> entries = cleanContentCache.entrySet().iterator();
            while (entries.hasNext()) {
                Entry<Long, Long> entry = entries.next();
                long txnTime = entry.getValue();
                if (purgeTime - txnTime > 1200000) {
                    //Purge the clean content cache of records more then 20 minutes old.
                    entries.remove();
                }
            }
            cleanContentLastPurged = purgeTime;
        }

        long txnFloor;
        Sort sort = new Sort(new SortField(FIELD_INTXID, SortField.Type.LONG));
        sort = sort.rewrite(searcher);
        TopFieldCollector collector = TopFieldCollector.create(sort, 1, null, false, false, false);

        DelegatingCollector delegatingCollector = new TxnCacheFilter(cleanContentCache); //Filter transactions that have already been processed.
        delegatingCollector.setLastDelegate(collector);
        searcher.search(dirtyOrNewContentQuery(), delegatingCollector);

        if (collector.getTotalHits() == 0) {
            return docIds;
        }

        ScoreDoc[] scoreDocs = collector.topDocs().scoreDocs;
        List<LeafReaderContext> leaves = searcher.getTopReaderContext().leaves();
        int index = ReaderUtil.subIndex(scoreDocs[0].doc, leaves);
        LeafReaderContext context = leaves.get(index);
        NumericDocValues longs = context.reader().getNumericDocValues(FIELD_INTXID);
        txnFloor = longs.get(scoreDocs[0].doc - context.docBase);

        //Find the next N transactions
        //The TxnCollector collects the transaction ids from the matching documents
        //The txnIds are limited to a range >= the txnFloor and < an arbitrary transaction ceiling.
        TxnCollector txnCollector = new TxnCollector(txnFloor);
        searcher.search(dirtyOrNewContentQuery(), txnCollector);
        LongHashSet txnSet = txnCollector.getTxnSet();

        if (txnSet.size() == 0) {
            //This should really never be the case, at a minimum the transaction floor should be collected.
            return docIds;
        }

        FieldType fieldType = searcher.getSchema().getField(FIELD_INTXID).getType();
        BooleanQuery.Builder builder = new BooleanQuery.Builder();

        for (LongCursor cursor : txnSet) {
            long txnID = cursor.value;
            //Build up the query for the filter of transactions we need to pull the dirty content for.
            TermQuery txnIDQuery = new TermQuery(
                    new Term(FIELD_INTXID, fieldType.readableToIndexed(Long.toString(txnID))));
            builder.add(new BooleanClause(txnIDQuery, BooleanClause.Occur.SHOULD));
        }

        BooleanQuery txnFilterQuery = builder.build();

        //Get the docs with dirty content for the transactions gathered above.
        DocListCollector docListCollector = new DocListCollector();
        BooleanQuery.Builder builder2 = new BooleanQuery.Builder();

        builder2.add(dirtyOrNewContentQuery(), BooleanClause.Occur.MUST);
        builder2.add(new QueryWrapperFilter(txnFilterQuery), BooleanClause.Occur.MUST);

        searcher.search(builder2.build(), docListCollector);
        IntArrayList docList = docListCollector.getDocs();
        int size = docList.size();

        List<Long> processedTxns = new ArrayList<>();
        for (int i = 0; i < size; ++i) {
            int doc = docList.get(i);
            Document document = searcher.doc(doc, REQUEST_ONLY_ID_FIELD);
            index = ReaderUtil.subIndex(doc, leaves);
            context = leaves.get(index);
            longs = context.reader().getNumericDocValues(FIELD_INTXID);

            long txnId = longs.get(doc - context.docBase);

            if (!cleanContentCache.containsKey(txnId)) {
                processedTxns.add(txnId);
                IndexableField id = document.getField(FIELD_SOLR4_ID);
                String idString = id.stringValue();
                TenantAclIdDbId tenantAndDbId = AlfrescoSolrDataModel.decodeNodeDocumentId(idString);
                docIds.add(tenantAndDbId);
            }
        }

        long txnTime = System.currentTimeMillis();

        for (Long l : processedTxns) {
            //Save the indexVersion so we know when we can clean out this entry
            cleanContentCache.put(l, txnTime);
        }

        return docIds;
    } finally {
        ofNullable(refCounted).ifPresent(RefCounted::decref);
    }
}

From source file:org.apache.solr.handler.component.ExpandComponent.java

License:Apache License

@SuppressWarnings("unchecked")
@Override/*from w  w w .  ja  v  a 2s.  com*/
public void process(ResponseBuilder rb) throws IOException {

    if (!rb.doExpand) {
        return;
    }

    SolrQueryRequest req = rb.req;
    SolrParams params = req.getParams();

    boolean isShard = params.getBool(ShardParams.IS_SHARD, false);
    String ids = params.get(ShardParams.IDS);

    if (ids == null && isShard) {
        return;
    }

    String field = params.get(ExpandParams.EXPAND_FIELD);
    if (field == null) {
        List<Query> filters = rb.getFilters();
        if (filters != null) {
            for (Query q : filters) {
                if (q instanceof CollapsingQParserPlugin.CollapsingPostFilter) {
                    CollapsingQParserPlugin.CollapsingPostFilter cp = (CollapsingQParserPlugin.CollapsingPostFilter) q;
                    field = cp.getField();
                }
            }
        }
    }

    if (field == null) {
        throw new IOException("Expand field is null.");
    }

    String sortParam = params.get(ExpandParams.EXPAND_SORT);
    String[] fqs = params.getParams(ExpandParams.EXPAND_FQ);
    String qs = params.get(ExpandParams.EXPAND_Q);
    int limit = params.getInt(ExpandParams.EXPAND_ROWS, 5);

    Sort sort = null;

    if (sortParam != null) {
        sort = QueryParsing.parseSortSpec(sortParam, rb.req).getSort();
    }

    Query query;
    if (qs == null) {
        query = rb.getQuery();
    } else {
        try {
            QParser parser = QParser.getParser(qs, null, req);
            query = parser.getQuery();
        } catch (Exception e) {
            throw new IOException(e);
        }
    }

    List<Query> newFilters = new ArrayList<>();

    if (fqs == null) {
        List<Query> filters = rb.getFilters();
        if (filters != null) {
            for (Query q : filters) {
                if (!(q instanceof CollapsingQParserPlugin.CollapsingPostFilter)) {
                    newFilters.add(q);
                }
            }
        }
    } else {
        try {
            for (String fq : fqs) {
                if (fq != null && fq.trim().length() != 0 && !fq.equals("*:*")) {
                    QParser fqp = QParser.getParser(fq, null, req);
                    newFilters.add(fqp.getQuery());
                }
            }
        } catch (Exception e) {
            throw new IOException(e);
        }
    }

    SolrIndexSearcher searcher = req.getSearcher();
    AtomicReader reader = searcher.getAtomicReader();
    SortedDocValues values = FieldCache.DEFAULT.getTermsIndex(reader, field);
    FixedBitSet groupBits = new FixedBitSet(values.getValueCount());
    DocList docList = rb.getResults().docList;
    IntOpenHashSet collapsedSet = new IntOpenHashSet(docList.size() * 2);

    DocIterator idit = docList.iterator();

    while (idit.hasNext()) {
        int doc = idit.nextDoc();
        int ord = values.getOrd(doc);
        if (ord > -1) {
            groupBits.set(ord);
            collapsedSet.add(doc);
        }
    }

    Collector collector;
    if (sort != null)
        sort = sort.rewrite(searcher);
    GroupExpandCollector groupExpandCollector = new GroupExpandCollector(values, groupBits, collapsedSet, limit,
            sort);
    SolrIndexSearcher.ProcessedFilter pfilter = searcher.getProcessedFilter(null, newFilters);
    if (pfilter.postFilter != null) {
        pfilter.postFilter.setLastDelegate(groupExpandCollector);
        collector = pfilter.postFilter;
    } else {
        collector = groupExpandCollector;
    }

    searcher.search(query, pfilter.filter, collector);
    IntObjectMap groups = groupExpandCollector.getGroups();
    Map<String, DocSlice> outMap = new HashMap();
    CharsRef charsRef = new CharsRef();
    FieldType fieldType = searcher.getSchema().getField(field).getType();
    for (IntObjectCursor cursor : (Iterable<IntObjectCursor>) groups) {
        int ord = cursor.key;
        TopDocsCollector topDocsCollector = (TopDocsCollector) cursor.value;
        TopDocs topDocs = topDocsCollector.topDocs();
        ScoreDoc[] scoreDocs = topDocs.scoreDocs;
        if (scoreDocs.length > 0) {
            int[] docs = new int[scoreDocs.length];
            float[] scores = new float[scoreDocs.length];
            for (int i = 0; i < docs.length; i++) {
                ScoreDoc scoreDoc = scoreDocs[i];
                docs[i] = scoreDoc.doc;
                scores[i] = scoreDoc.score;
            }
            DocSlice slice = new DocSlice(0, docs.length, docs, scores, topDocs.totalHits,
                    topDocs.getMaxScore());
            final BytesRef bytesRef = values.lookupOrd(ord);
            fieldType.indexedToReadable(bytesRef, charsRef);
            String group = charsRef.toString();
            outMap.put(group, slice);
        }
    }

    rb.rsp.add("expanded", outMap);
}

From source file:org.apache.solr.search.federated.DJoinQParserPlugin.java

License:Apache License

@Override
public QParser createParser(String qstr, SolrParams localParams, SolrParams params, SolrQueryRequest req) {
    return new QParser(qstr, localParams, params, req) {

        @Override// w ww .  j a  v  a 2 s .  c  om
        public Query parse() throws SyntaxError {
            return new RankQuery() {

                private Query mainQuery;

                @Override
                @SuppressWarnings("rawtypes")
                public TopDocsCollector getTopDocsCollector(int len, QueryCommand cmd, IndexSearcher searcher)
                        throws IOException {
                    Sort sort = cmd.getSort();
                    if (sort == null) {
                        return TopScoreDocCollector.create(len, false);
                    } else {
                        return TopFieldCollector.create(sort.rewrite(searcher), len, false, true, true, false);
                    }
                }

                @Override
                public MergeStrategy getMergeStrategy() {
                    return new DJoinMergeStrategy();
                }

                @Override
                public RankQuery wrap(Query mainQuery) {
                    this.mainQuery = mainQuery;
                    return this;
                }

                @Override
                public Query rewrite(IndexReader reader) throws IOException {
                    return mainQuery.rewrite(reader);
                }

                @Override
                public Weight createWeight(IndexSearcher searcher) throws IOException {
                    return mainQuery.createWeight(searcher);
                }

            };
        }

    };
}

From source file:org.apache.solr.search.ReRankCollector.java

License:Apache License

public ReRankCollector(int reRankDocs, int length, Rescorer reRankQueryRescorer, QueryCommand cmd,
        IndexSearcher searcher, Map<BytesRef, Integer> boostedPriority) throws IOException {
    super(null);/*from   w w w.j a v  a 2s. c o  m*/
    this.reRankDocs = reRankDocs;
    this.length = length;
    this.boostedPriority = boostedPriority;
    Sort sort = cmd.getSort();
    if (sort == null) {
        this.mainCollector = TopScoreDocCollector.create(Math.max(this.reRankDocs, length));
    } else {
        sort = sort.rewrite(searcher);
        this.mainCollector = TopFieldCollector.create(sort, Math.max(this.reRankDocs, length), false, true,
                true);
    }
    this.searcher = searcher;
    this.reRankQueryRescorer = reRankQueryRescorer;
}

From source file:org.apache.solr.search.SolrIndexSearcher.java

License:Apache License

/** Returns a weighted sort according to this searcher */
public Sort weightSort(Sort sort) throws IOException {
    return (sort != null) ? sort.rewrite(this) : null;
}