List of usage examples for org.apache.lucene.search Sort rewrite
public Sort rewrite(IndexSearcher searcher) throws IOException
From source file:com.stratio.cassandra.lucene.index.RAMIndex.java
License:Apache License
/** * Finds the top {@code count} hits for {@code query} and sorting the hits by {@code sort}. * * @param query the {@link Query} to search for * @param sort the {@link Sort} to be applied * @param count the max number of results to be collected * @param fields the names of the fields to be loaded * @return the found documents/*from w w w . j a v a 2 s . c o m*/ */ public List<Document> search(Query query, Sort sort, Integer count, Set<String> fields) { try { indexWriter.commit(); IndexReader reader = DirectoryReader.open(directory); IndexSearcher searcher = new IndexSearcher(reader); sort = sort.rewrite(searcher); TopDocs topDocs = searcher.search(query, count, sort); ScoreDoc[] scoreDocs = topDocs.scoreDocs; List<Document> documents = new LinkedList<>(); for (ScoreDoc scoreDoc : scoreDocs) { Document document = searcher.doc(scoreDoc.doc, fields); documents.add(document); } searcher.getIndexReader().close(); return documents; } catch (IOException e) { throw new IndexException(logger, e, "Error while searching"); } }
From source file:io.puntanegra.fhir.index.lucene.LuceneRAMIndex.java
License:Apache License
/** * Finds the top {@code count} hits for {@code query} and sorting the hits * by {@code sort}./* w ww . ja v a 2 s . c o m*/ * * @param query * the {@link Query} to search for * @param sort * the {@link Sort} to be applied * @param count * the max number of results to be collected * @param fields * the names of the fields to be loaded * @return the found documents */ public List<Document> search(Query query, Sort sort, Integer count, Set<String> fields) { try { indexWriter.commit(); IndexReader reader = DirectoryReader.open(directory); IndexSearcher searcher = new IndexSearcher(reader); sort = sort.rewrite(searcher); TopDocs topDocs = searcher.search(query, count, sort); ScoreDoc[] scoreDocs = topDocs.scoreDocs; List<Document> documents = new LinkedList<>(); for (ScoreDoc scoreDoc : scoreDocs) { Document document = searcher.doc(scoreDoc.doc, fields); documents.add(document); } searcher.getIndexReader().close(); return documents; } catch (IOException e) { throw new FhirIndexException(e, "Error while searching"); } }
From source file:org.alfresco.solr.SolrInformationServer.java
License:Open Source License
@Override public List<TenantAclIdDbId> getDocsWithUncleanContent(int start, int rows) throws IOException { RefCounted<SolrIndexSearcher> refCounted = null; try {//from w ww.j a v a 2 s . c om List<TenantAclIdDbId> docIds = new ArrayList<>(); refCounted = this.core.getSearcher(); SolrIndexSearcher searcher = refCounted.get(); /* * Below is the code for purging the cleanContentCache. * The cleanContentCache is an in-memory LRU cache of the transactions that have already * had their content fetched. This is needed because the ContentTracker does not have an up-to-date * snapshot of the index to determine which nodes are marked as dirty/new. The cleanContentCache is used * to filter out nodes that belong to transactions that have already been processed, which stops them from * being re-processed. * * The cleanContentCache needs to be purged periodically to support retrying of failed content fetches. * This is because fetches for individual nodes within the transaction may have failed, but the transaction will still be in the * cleanContentCache, which prevents it from being retried. * * Once a transaction is purged from the cleanContentCache it will be retried automatically if it is marked dirty/new * in current snapshot of the index. * * The code below runs every two minutes and purges transactions from the * cleanContentCache that is more then 20 minutes old. * */ long purgeTime = System.currentTimeMillis(); if (purgeTime - cleanContentLastPurged > 120000) { Iterator<Entry<Long, Long>> entries = cleanContentCache.entrySet().iterator(); while (entries.hasNext()) { Entry<Long, Long> entry = entries.next(); long txnTime = entry.getValue(); if (purgeTime - txnTime > 1200000) { //Purge the clean content cache of records more then 20 minutes old. entries.remove(); } } cleanContentLastPurged = purgeTime; } long txnFloor; Sort sort = new Sort(new SortField(FIELD_INTXID, SortField.Type.LONG)); sort = sort.rewrite(searcher); TopFieldCollector collector = TopFieldCollector.create(sort, 1, null, false, false, false); DelegatingCollector delegatingCollector = new TxnCacheFilter(cleanContentCache); //Filter transactions that have already been processed. delegatingCollector.setLastDelegate(collector); searcher.search(dirtyOrNewContentQuery(), delegatingCollector); if (collector.getTotalHits() == 0) { return docIds; } ScoreDoc[] scoreDocs = collector.topDocs().scoreDocs; List<LeafReaderContext> leaves = searcher.getTopReaderContext().leaves(); int index = ReaderUtil.subIndex(scoreDocs[0].doc, leaves); LeafReaderContext context = leaves.get(index); NumericDocValues longs = context.reader().getNumericDocValues(FIELD_INTXID); txnFloor = longs.get(scoreDocs[0].doc - context.docBase); //Find the next N transactions //The TxnCollector collects the transaction ids from the matching documents //The txnIds are limited to a range >= the txnFloor and < an arbitrary transaction ceiling. TxnCollector txnCollector = new TxnCollector(txnFloor); searcher.search(dirtyOrNewContentQuery(), txnCollector); LongHashSet txnSet = txnCollector.getTxnSet(); if (txnSet.size() == 0) { //This should really never be the case, at a minimum the transaction floor should be collected. return docIds; } FieldType fieldType = searcher.getSchema().getField(FIELD_INTXID).getType(); BooleanQuery.Builder builder = new BooleanQuery.Builder(); for (LongCursor cursor : txnSet) { long txnID = cursor.value; //Build up the query for the filter of transactions we need to pull the dirty content for. TermQuery txnIDQuery = new TermQuery( new Term(FIELD_INTXID, fieldType.readableToIndexed(Long.toString(txnID)))); builder.add(new BooleanClause(txnIDQuery, BooleanClause.Occur.SHOULD)); } BooleanQuery txnFilterQuery = builder.build(); //Get the docs with dirty content for the transactions gathered above. DocListCollector docListCollector = new DocListCollector(); BooleanQuery.Builder builder2 = new BooleanQuery.Builder(); builder2.add(dirtyOrNewContentQuery(), BooleanClause.Occur.MUST); builder2.add(new QueryWrapperFilter(txnFilterQuery), BooleanClause.Occur.MUST); searcher.search(builder2.build(), docListCollector); IntArrayList docList = docListCollector.getDocs(); int size = docList.size(); List<Long> processedTxns = new ArrayList<>(); for (int i = 0; i < size; ++i) { int doc = docList.get(i); Document document = searcher.doc(doc, REQUEST_ONLY_ID_FIELD); index = ReaderUtil.subIndex(doc, leaves); context = leaves.get(index); longs = context.reader().getNumericDocValues(FIELD_INTXID); long txnId = longs.get(doc - context.docBase); if (!cleanContentCache.containsKey(txnId)) { processedTxns.add(txnId); IndexableField id = document.getField(FIELD_SOLR4_ID); String idString = id.stringValue(); TenantAclIdDbId tenantAndDbId = AlfrescoSolrDataModel.decodeNodeDocumentId(idString); docIds.add(tenantAndDbId); } } long txnTime = System.currentTimeMillis(); for (Long l : processedTxns) { //Save the indexVersion so we know when we can clean out this entry cleanContentCache.put(l, txnTime); } return docIds; } finally { ofNullable(refCounted).ifPresent(RefCounted::decref); } }
From source file:org.apache.solr.handler.component.ExpandComponent.java
License:Apache License
@SuppressWarnings("unchecked") @Override/*from w w w . ja v a 2s. com*/ public void process(ResponseBuilder rb) throws IOException { if (!rb.doExpand) { return; } SolrQueryRequest req = rb.req; SolrParams params = req.getParams(); boolean isShard = params.getBool(ShardParams.IS_SHARD, false); String ids = params.get(ShardParams.IDS); if (ids == null && isShard) { return; } String field = params.get(ExpandParams.EXPAND_FIELD); if (field == null) { List<Query> filters = rb.getFilters(); if (filters != null) { for (Query q : filters) { if (q instanceof CollapsingQParserPlugin.CollapsingPostFilter) { CollapsingQParserPlugin.CollapsingPostFilter cp = (CollapsingQParserPlugin.CollapsingPostFilter) q; field = cp.getField(); } } } } if (field == null) { throw new IOException("Expand field is null."); } String sortParam = params.get(ExpandParams.EXPAND_SORT); String[] fqs = params.getParams(ExpandParams.EXPAND_FQ); String qs = params.get(ExpandParams.EXPAND_Q); int limit = params.getInt(ExpandParams.EXPAND_ROWS, 5); Sort sort = null; if (sortParam != null) { sort = QueryParsing.parseSortSpec(sortParam, rb.req).getSort(); } Query query; if (qs == null) { query = rb.getQuery(); } else { try { QParser parser = QParser.getParser(qs, null, req); query = parser.getQuery(); } catch (Exception e) { throw new IOException(e); } } List<Query> newFilters = new ArrayList<>(); if (fqs == null) { List<Query> filters = rb.getFilters(); if (filters != null) { for (Query q : filters) { if (!(q instanceof CollapsingQParserPlugin.CollapsingPostFilter)) { newFilters.add(q); } } } } else { try { for (String fq : fqs) { if (fq != null && fq.trim().length() != 0 && !fq.equals("*:*")) { QParser fqp = QParser.getParser(fq, null, req); newFilters.add(fqp.getQuery()); } } } catch (Exception e) { throw new IOException(e); } } SolrIndexSearcher searcher = req.getSearcher(); AtomicReader reader = searcher.getAtomicReader(); SortedDocValues values = FieldCache.DEFAULT.getTermsIndex(reader, field); FixedBitSet groupBits = new FixedBitSet(values.getValueCount()); DocList docList = rb.getResults().docList; IntOpenHashSet collapsedSet = new IntOpenHashSet(docList.size() * 2); DocIterator idit = docList.iterator(); while (idit.hasNext()) { int doc = idit.nextDoc(); int ord = values.getOrd(doc); if (ord > -1) { groupBits.set(ord); collapsedSet.add(doc); } } Collector collector; if (sort != null) sort = sort.rewrite(searcher); GroupExpandCollector groupExpandCollector = new GroupExpandCollector(values, groupBits, collapsedSet, limit, sort); SolrIndexSearcher.ProcessedFilter pfilter = searcher.getProcessedFilter(null, newFilters); if (pfilter.postFilter != null) { pfilter.postFilter.setLastDelegate(groupExpandCollector); collector = pfilter.postFilter; } else { collector = groupExpandCollector; } searcher.search(query, pfilter.filter, collector); IntObjectMap groups = groupExpandCollector.getGroups(); Map<String, DocSlice> outMap = new HashMap(); CharsRef charsRef = new CharsRef(); FieldType fieldType = searcher.getSchema().getField(field).getType(); for (IntObjectCursor cursor : (Iterable<IntObjectCursor>) groups) { int ord = cursor.key; TopDocsCollector topDocsCollector = (TopDocsCollector) cursor.value; TopDocs topDocs = topDocsCollector.topDocs(); ScoreDoc[] scoreDocs = topDocs.scoreDocs; if (scoreDocs.length > 0) { int[] docs = new int[scoreDocs.length]; float[] scores = new float[scoreDocs.length]; for (int i = 0; i < docs.length; i++) { ScoreDoc scoreDoc = scoreDocs[i]; docs[i] = scoreDoc.doc; scores[i] = scoreDoc.score; } DocSlice slice = new DocSlice(0, docs.length, docs, scores, topDocs.totalHits, topDocs.getMaxScore()); final BytesRef bytesRef = values.lookupOrd(ord); fieldType.indexedToReadable(bytesRef, charsRef); String group = charsRef.toString(); outMap.put(group, slice); } } rb.rsp.add("expanded", outMap); }
From source file:org.apache.solr.search.federated.DJoinQParserPlugin.java
License:Apache License
@Override public QParser createParser(String qstr, SolrParams localParams, SolrParams params, SolrQueryRequest req) { return new QParser(qstr, localParams, params, req) { @Override// w ww . j a v a 2 s . c om public Query parse() throws SyntaxError { return new RankQuery() { private Query mainQuery; @Override @SuppressWarnings("rawtypes") public TopDocsCollector getTopDocsCollector(int len, QueryCommand cmd, IndexSearcher searcher) throws IOException { Sort sort = cmd.getSort(); if (sort == null) { return TopScoreDocCollector.create(len, false); } else { return TopFieldCollector.create(sort.rewrite(searcher), len, false, true, true, false); } } @Override public MergeStrategy getMergeStrategy() { return new DJoinMergeStrategy(); } @Override public RankQuery wrap(Query mainQuery) { this.mainQuery = mainQuery; return this; } @Override public Query rewrite(IndexReader reader) throws IOException { return mainQuery.rewrite(reader); } @Override public Weight createWeight(IndexSearcher searcher) throws IOException { return mainQuery.createWeight(searcher); } }; } }; }
From source file:org.apache.solr.search.ReRankCollector.java
License:Apache License
public ReRankCollector(int reRankDocs, int length, Rescorer reRankQueryRescorer, QueryCommand cmd, IndexSearcher searcher, Map<BytesRef, Integer> boostedPriority) throws IOException { super(null);/*from w w w.j a v a 2s. c o m*/ this.reRankDocs = reRankDocs; this.length = length; this.boostedPriority = boostedPriority; Sort sort = cmd.getSort(); if (sort == null) { this.mainCollector = TopScoreDocCollector.create(Math.max(this.reRankDocs, length)); } else { sort = sort.rewrite(searcher); this.mainCollector = TopFieldCollector.create(sort, Math.max(this.reRankDocs, length), false, true, true); } this.searcher = searcher; this.reRankQueryRescorer = reRankQueryRescorer; }
From source file:org.apache.solr.search.SolrIndexSearcher.java
License:Apache License
/** Returns a weighted sort according to this searcher */ public Sort weightSort(Sort sort) throws IOException { return (sort != null) ? sort.rewrite(this) : null; }