560 lines
22 KiB
Java
560 lines
22 KiB
Java
package it.cavallium.dbengine.lucene;
|
|
|
|
import it.cavallium.dbengine.client.CompositeSnapshot;
|
|
import it.cavallium.dbengine.client.IndicizerAnalyzers;
|
|
import it.cavallium.dbengine.client.IndicizerSimilarities;
|
|
import it.cavallium.dbengine.client.query.QueryParser;
|
|
import it.cavallium.dbengine.client.query.current.data.QueryParams;
|
|
import it.cavallium.dbengine.client.query.current.data.TotalHitsCount;
|
|
import it.cavallium.dbengine.database.EnglishItalianStopFilter;
|
|
import it.cavallium.dbengine.database.LLKeyScore;
|
|
import it.cavallium.dbengine.database.LLUtils;
|
|
import it.cavallium.dbengine.database.collections.DatabaseMapDictionary;
|
|
import it.cavallium.dbengine.database.collections.DatabaseMapDictionaryDeep;
|
|
import it.cavallium.dbengine.database.collections.ValueGetter;
|
|
import it.cavallium.dbengine.database.disk.LLIndexSearchers;
|
|
import it.cavallium.dbengine.lucene.analyzer.NCharGramAnalyzer;
|
|
import it.cavallium.dbengine.lucene.analyzer.NCharGramEdgeAnalyzer;
|
|
import it.cavallium.dbengine.lucene.analyzer.TextFieldsAnalyzer;
|
|
import it.cavallium.dbengine.lucene.analyzer.TextFieldsSimilarity;
|
|
import it.cavallium.dbengine.lucene.analyzer.WordAnalyzer;
|
|
import it.cavallium.dbengine.lucene.mlt.BigCompositeReader;
|
|
import it.cavallium.dbengine.lucene.mlt.MultiMoreLikeThis;
|
|
import it.cavallium.dbengine.lucene.searcher.LocalQueryParams;
|
|
import it.cavallium.dbengine.lucene.similarity.NGramSimilarity;
|
|
import java.io.EOFException;
|
|
import java.io.IOException;
|
|
import java.nio.ByteBuffer;
|
|
import java.nio.channels.FileChannel;
|
|
import java.util.Comparator;
|
|
import java.util.HashMap;
|
|
import java.util.List;
|
|
import java.util.Map;
|
|
import java.util.Map.Entry;
|
|
import java.util.NoSuchElementException;
|
|
import java.util.Set;
|
|
import java.util.stream.Collectors;
|
|
import org.apache.lucene.analysis.Analyzer;
|
|
import org.apache.lucene.analysis.LowerCaseFilter;
|
|
import org.apache.lucene.analysis.TokenStream;
|
|
import org.apache.lucene.analysis.en.EnglishPossessiveFilter;
|
|
import org.apache.lucene.analysis.en.KStemFilter;
|
|
import org.apache.lucene.analysis.miscellaneous.ASCIIFoldingFilter;
|
|
import org.apache.lucene.analysis.miscellaneous.PerFieldAnalyzerWrapper;
|
|
import org.apache.lucene.analysis.standard.StandardAnalyzer;
|
|
import org.apache.lucene.document.Document;
|
|
import org.apache.lucene.index.IndexReader;
|
|
import org.apache.lucene.index.IndexableField;
|
|
import org.apache.lucene.search.BooleanClause.Occur;
|
|
import org.apache.lucene.search.BooleanQuery.Builder;
|
|
import org.apache.lucene.search.ConstantScoreQuery;
|
|
import org.apache.lucene.search.IndexSearcher;
|
|
import org.apache.lucene.search.MatchAllDocsQuery;
|
|
import org.apache.lucene.search.MatchNoDocsQuery;
|
|
import org.apache.lucene.search.Query;
|
|
import org.apache.lucene.search.ScoreDoc;
|
|
import org.apache.lucene.search.Sort;
|
|
import org.apache.lucene.search.TopDocs;
|
|
import org.apache.lucene.search.TopFieldDocs;
|
|
import org.apache.lucene.search.TotalHits;
|
|
import org.apache.lucene.search.similarities.BooleanSimilarity;
|
|
import org.apache.lucene.search.similarities.ClassicSimilarity;
|
|
import org.apache.lucene.search.similarities.PerFieldSimilarityWrapper;
|
|
import org.apache.lucene.search.similarities.Similarity;
|
|
import org.apache.lucene.search.similarities.TFIDFSimilarity;
|
|
import org.jetbrains.annotations.NotNull;
|
|
import org.jetbrains.annotations.Nullable;
|
|
import org.novasearch.lucene.search.similarities.BM25Similarity;
|
|
import org.novasearch.lucene.search.similarities.BM25Similarity.BM25Model;
|
|
import org.novasearch.lucene.search.similarities.LdpSimilarity;
|
|
import org.novasearch.lucene.search.similarities.LtcSimilarity;
|
|
import org.novasearch.lucene.search.similarities.RobertsonSimilarity;
|
|
import org.warp.commonutils.log.Logger;
|
|
import org.warp.commonutils.log.LoggerFactory;
|
|
import reactor.core.publisher.Flux;
|
|
import reactor.core.publisher.Mono;
|
|
import reactor.core.scheduler.Schedulers;
|
|
import reactor.util.concurrent.Queues;
|
|
import reactor.util.function.Tuple2;
|
|
|
|
public class LuceneUtils {
|
|
|
|
private static final Logger logger = LoggerFactory.getLogger(LuceneUtils.class);
|
|
|
|
private static final Analyzer lucene4GramWordsAnalyzerEdgeInstance = new NCharGramEdgeAnalyzer(true, 4, 4);
|
|
private static final Analyzer lucene4GramStringAnalyzerEdgeInstance = new NCharGramEdgeAnalyzer(false, 4, 4);
|
|
private static final Analyzer lucene4GramWordsAnalyzerInstance = new NCharGramAnalyzer(true, 4, 4);
|
|
private static final Analyzer lucene4GramStringAnalyzerInstance = new NCharGramAnalyzer(false, 4, 4);
|
|
private static final Analyzer lucene3To5GramWordsAnalyzerEdgeInstance = new NCharGramEdgeAnalyzer(true, 3, 5);
|
|
private static final Analyzer lucene3To5GramStringAnalyzerEdgeInstance = new NCharGramEdgeAnalyzer(false, 3, 5);
|
|
private static final Analyzer lucene3To5GramWordsAnalyzerInstance = new NCharGramAnalyzer(true, 3, 5);
|
|
private static final Analyzer lucene3To5GramStringAnalyzerInstance = new NCharGramAnalyzer(false, 3, 5);
|
|
private static final Analyzer luceneStandardAnalyzerInstance = new StandardAnalyzer();
|
|
private static final Analyzer luceneWordAnalyzerStopWordsAndStemInstance = new WordAnalyzer(false,true, true);
|
|
private static final Analyzer luceneWordAnalyzerStopWordsInstance = new WordAnalyzer(false, true, false);
|
|
private static final Analyzer luceneWordAnalyzerStemInstance = new WordAnalyzer(false, false, true);
|
|
private static final Analyzer luceneWordAnalyzerSimpleInstance = new WordAnalyzer(false, false, false);
|
|
private static final Analyzer luceneICUCollationKeyInstance = new WordAnalyzer(false, true, true);
|
|
private static final Similarity luceneBM25ClassicSimilarityInstance = new BM25Similarity(BM25Model.CLASSIC);
|
|
private static final Similarity luceneBM25PlusSimilarityInstance = new BM25Similarity(BM25Model.PLUS);
|
|
private static final Similarity luceneBM25LSimilarityInstance = new BM25Similarity(BM25Model.L);
|
|
private static final Similarity luceneBM15PlusSimilarityInstance = new BM25Similarity(1.2f, 0.0f, 0.5f, BM25Model.PLUS);
|
|
private static final Similarity luceneBM11PlusSimilarityInstance = new BM25Similarity(1.2f, 1.0f, 0.5f, BM25Model.PLUS);
|
|
private static final Similarity luceneBM25ClassicNGramSimilarityInstance = NGramSimilarity.bm25(BM25Model.CLASSIC);
|
|
private static final Similarity luceneBM25PlusNGramSimilarityInstance = NGramSimilarity.bm25(BM25Model.PLUS);
|
|
private static final Similarity luceneBM25LNGramSimilarityInstance = NGramSimilarity.bm25(BM25Model.L);
|
|
private static final Similarity luceneBM15PlusNGramSimilarityInstance = NGramSimilarity.bm15(BM25Model.PLUS);
|
|
private static final Similarity luceneBM11PlusNGramSimilarityInstance = NGramSimilarity.bm11(BM25Model.PLUS);
|
|
private static final Similarity luceneClassicSimilarityInstance = new ClassicSimilarity();
|
|
private static final Similarity luceneClassicNGramSimilarityInstance = NGramSimilarity.classic();
|
|
private static final Similarity luceneLTCSimilarityInstance = new LtcSimilarity();
|
|
private static final Similarity luceneLDPSimilarityInstance = new LdpSimilarity();
|
|
private static final Similarity luceneLDPNoLengthSimilarityInstance = new LdpSimilarity(0, 0.5f);
|
|
private static final Similarity luceneBooleanSimilarityInstance = new BooleanSimilarity();
|
|
private static final Similarity luceneRobertsonSimilarityInstance = new RobertsonSimilarity();
|
|
// TODO: remove this default page limits and make the limits configurable into QueryParams
|
|
private static final PageLimits DEFAULT_PAGE_LIMITS = new ExponentialPageLimits();
|
|
|
|
@SuppressWarnings("DuplicatedCode")
|
|
public static Analyzer getAnalyzer(TextFieldsAnalyzer analyzer) {
|
|
return switch (analyzer) {
|
|
case N4GramPartialWords -> lucene4GramWordsAnalyzerInstance;
|
|
case N4GramPartialString -> lucene4GramStringAnalyzerInstance;
|
|
case N4GramPartialWordsEdge -> lucene4GramWordsAnalyzerEdgeInstance;
|
|
case N4GramPartialStringEdge -> lucene4GramStringAnalyzerEdgeInstance;
|
|
case N3To5GramPartialWords -> lucene3To5GramWordsAnalyzerInstance;
|
|
case N3To5GramPartialString -> lucene3To5GramStringAnalyzerInstance;
|
|
case N3To5GramPartialWordsEdge -> lucene3To5GramWordsAnalyzerEdgeInstance;
|
|
case N3To5GramPartialStringEdge -> lucene3To5GramStringAnalyzerEdgeInstance;
|
|
case Standard -> luceneStandardAnalyzerInstance;
|
|
case FullText -> luceneWordAnalyzerStopWordsAndStemInstance;
|
|
case WordWithStopwordsStripping -> luceneWordAnalyzerStopWordsInstance;
|
|
case WordWithStemming -> luceneWordAnalyzerStemInstance;
|
|
case WordSimple -> luceneWordAnalyzerSimpleInstance;
|
|
case ICUCollationKey -> luceneICUCollationKeyInstance;
|
|
//noinspection UnnecessaryDefault
|
|
default -> throw new UnsupportedOperationException("Unknown analyzer: " + analyzer);
|
|
};
|
|
}
|
|
|
|
@SuppressWarnings("DuplicatedCode")
|
|
public static Similarity getSimilarity(TextFieldsSimilarity similarity) {
|
|
return switch (similarity) {
|
|
case BM25Classic -> luceneBM25ClassicSimilarityInstance;
|
|
case NGramBM25Classic -> luceneBM25ClassicNGramSimilarityInstance;
|
|
case BM25L -> luceneBM25LSimilarityInstance;
|
|
case NGramBM25L -> luceneBM25LNGramSimilarityInstance;
|
|
case Classic -> luceneClassicSimilarityInstance;
|
|
case NGramClassic -> luceneClassicNGramSimilarityInstance;
|
|
case BM25Plus -> luceneBM25PlusSimilarityInstance;
|
|
case NGramBM25Plus -> luceneBM25PlusNGramSimilarityInstance;
|
|
case BM15Plus -> luceneBM15PlusSimilarityInstance;
|
|
case NGramBM15Plus -> luceneBM15PlusNGramSimilarityInstance;
|
|
case BM11Plus -> luceneBM11PlusSimilarityInstance;
|
|
case NGramBM11Plus -> luceneBM11PlusNGramSimilarityInstance;
|
|
case LTC -> luceneLTCSimilarityInstance;
|
|
case LDP -> luceneLDPSimilarityInstance;
|
|
case LDPNoLength -> luceneLDPNoLengthSimilarityInstance;
|
|
case Robertson -> luceneRobertsonSimilarityInstance;
|
|
case Boolean -> luceneBooleanSimilarityInstance;
|
|
//noinspection UnnecessaryDefault
|
|
default -> throw new IllegalStateException("Unknown similarity: " + similarity);
|
|
};
|
|
}
|
|
|
|
/**
|
|
*
|
|
* @param stem Enable stem filters on words.
|
|
* Pass false if it will be used with a n-gram filter
|
|
*/
|
|
public static TokenStream newCommonFilter(TokenStream tokenStream, boolean stem) {
|
|
tokenStream = newCommonNormalizer(tokenStream);
|
|
if (stem) {
|
|
tokenStream = new KStemFilter(tokenStream);
|
|
tokenStream = new EnglishPossessiveFilter(tokenStream);
|
|
}
|
|
return tokenStream;
|
|
}
|
|
|
|
public static TokenStream newCommonNormalizer(TokenStream tokenStream) {
|
|
tokenStream = new ASCIIFoldingFilter(tokenStream);
|
|
tokenStream = new LowerCaseFilter(tokenStream);
|
|
return tokenStream;
|
|
}
|
|
|
|
/**
|
|
* @throws NoSuchElementException when the key is not found
|
|
* @throws IOException when an error occurs when reading the document
|
|
*/
|
|
@NotNull
|
|
public static String keyOfTopDoc(int docId, IndexReader indexReader,
|
|
String keyFieldName) throws IOException, NoSuchElementException {
|
|
if (Schedulers.isInNonBlockingThread()) {
|
|
throw new UnsupportedOperationException("Called keyOfTopDoc in a nonblocking thread");
|
|
}
|
|
if (docId > indexReader.maxDoc()) {
|
|
throw new IOException("Document " + docId + " > maxDoc (" +indexReader.maxDoc() + ")");
|
|
}
|
|
DocumentStoredSingleFieldVisitor visitor = new DocumentStoredSingleFieldVisitor(keyFieldName);
|
|
indexReader.document(docId, visitor);
|
|
Document d = visitor.getDocument();
|
|
if (d.getFields().isEmpty()) {
|
|
throw new NoSuchElementException(
|
|
"Can't get key (field \"" + keyFieldName + "\") of document docId: " + docId + ". Available fields: []");
|
|
} else {
|
|
var field = d.getField(keyFieldName);
|
|
if (field == null) {
|
|
throw new NoSuchElementException(
|
|
"Can't get key (field \"" + keyFieldName + "\") of document docId: " + docId + ". Available fields: " + d
|
|
.getFields()
|
|
.stream()
|
|
.map(IndexableField::name)
|
|
.collect(Collectors.joining(",", "[", "]")));
|
|
} else {
|
|
return field.stringValue();
|
|
}
|
|
}
|
|
}
|
|
|
|
public static <T, U, V> ValueGetter<Entry<T, U>, V> getAsyncDbValueGetterDeep(
|
|
CompositeSnapshot snapshot,
|
|
DatabaseMapDictionaryDeep<T, Map<U, V>, DatabaseMapDictionary<U, V>> dictionaryDeep) {
|
|
return entry -> LLUtils.usingResource(dictionaryDeep
|
|
.at(snapshot, entry.getKey()), sub -> sub.getValue(snapshot, entry.getValue()), true);
|
|
}
|
|
|
|
public static PerFieldAnalyzerWrapper toPerFieldAnalyzerWrapper(IndicizerAnalyzers indicizerAnalyzers) {
|
|
HashMap<String, Analyzer> perFieldAnalyzer = new HashMap<>();
|
|
indicizerAnalyzers
|
|
.fieldAnalyzer()
|
|
.forEach((key, value) -> perFieldAnalyzer.put(key, LuceneUtils.getAnalyzer(value)));
|
|
return new PerFieldAnalyzerWrapper(LuceneUtils.getAnalyzer(indicizerAnalyzers.defaultAnalyzer()), perFieldAnalyzer);
|
|
}
|
|
|
|
public static PerFieldSimilarityWrapper toPerFieldSimilarityWrapper(IndicizerSimilarities indicizerSimilarities) {
|
|
HashMap<String, Similarity> perFieldSimilarity = new HashMap<>();
|
|
indicizerSimilarities
|
|
.fieldSimilarity()
|
|
.forEach((key, value) -> perFieldSimilarity.put(key, LuceneUtils.getSimilarity(value)));
|
|
var defaultSimilarity = LuceneUtils.getSimilarity(indicizerSimilarities.defaultSimilarity());
|
|
return new PerFieldSimilarityWrapper() {
|
|
|
|
@Override
|
|
public Similarity get(String name) {
|
|
return perFieldSimilarity.getOrDefault(name, defaultSimilarity);
|
|
}
|
|
};
|
|
}
|
|
|
|
public static int alignUnsigned(int number, boolean expand) {
|
|
if (number % 4096 != 0) {
|
|
if (expand) {
|
|
return number + (4096 - (number % 4096));
|
|
} else {
|
|
return number - (number % 4096);
|
|
}
|
|
} else {
|
|
return number;
|
|
}
|
|
}
|
|
|
|
public static long alignUnsigned(long number, boolean expand) {
|
|
if (number % 4096L != 0) {
|
|
if (expand) {
|
|
return number + (4096L - (number % 4096L));
|
|
} else {
|
|
return number - (number % 4096L);
|
|
}
|
|
} else {
|
|
return number;
|
|
}
|
|
}
|
|
|
|
public static void readInternalAligned(Object ref,
|
|
FileChannel channel,
|
|
long pos,
|
|
ByteBuffer b,
|
|
int readLength,
|
|
int usefulLength,
|
|
long end) throws IOException {
|
|
if (Schedulers.isInNonBlockingThread()) {
|
|
throw new UnsupportedOperationException("Called readInternalAligned in a nonblocking thread");
|
|
}
|
|
int startBufPosition = b.position();
|
|
int readData = 0;
|
|
int i;
|
|
for(; readLength > 0; readLength -= i) {
|
|
int toRead = readLength;
|
|
b.limit(b.position() + toRead);
|
|
|
|
assert b.remaining() == toRead;
|
|
|
|
var beforeReadBufPosition = b.position();
|
|
channel.read(b, pos);
|
|
b.limit(Math.min(startBufPosition + usefulLength, b.position() + toRead));
|
|
var afterReadBufPosition = b.position();
|
|
i = (afterReadBufPosition - beforeReadBufPosition);
|
|
readData += i;
|
|
|
|
if (i < toRead && i > 0) {
|
|
if (readData < usefulLength) {
|
|
throw new EOFException("read past EOF: " + ref + " buffer: " + b + " chunkLen: " + toRead + " end: " + end);
|
|
}
|
|
if (readData == usefulLength) {
|
|
b.limit(b.position());
|
|
// File end reached
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (i < 0) {
|
|
throw new EOFException("read past EOF: " + ref + " buffer: " + b + " chunkLen: " + toRead + " end: " + end);
|
|
}
|
|
|
|
assert i > 0 : "FileChannel.read with non zero-length bb.remaining() must always read at least one byte (FileChannel is in blocking mode, see spec of ReadableByteChannel)";
|
|
|
|
pos += i;
|
|
}
|
|
|
|
assert readLength == 0;
|
|
}
|
|
|
|
public static int safeLongToInt(long l) {
|
|
if (l > 2147483630) {
|
|
return 2147483630;
|
|
} else if (l < -2147483630) {
|
|
return -2147483630;
|
|
} else {
|
|
return (int) l;
|
|
}
|
|
}
|
|
|
|
@Nullable
|
|
public static ScoreDoc getLastScoreDoc(ScoreDoc[] scoreDocs) {
|
|
if (scoreDocs == null) {
|
|
return null;
|
|
}
|
|
if (scoreDocs.length == 0) {
|
|
return null;
|
|
}
|
|
return scoreDocs[scoreDocs.length - 1];
|
|
}
|
|
|
|
public static LocalQueryParams toLocalQueryParams(QueryParams queryParams) {
|
|
return new LocalQueryParams(QueryParser.toQuery(queryParams.query()),
|
|
queryParams.offset(),
|
|
queryParams.limit(),
|
|
DEFAULT_PAGE_LIMITS,
|
|
queryParams.minCompetitiveScore().getNullable(),
|
|
QueryParser.toSort(queryParams.sort()),
|
|
queryParams.complete()
|
|
);
|
|
}
|
|
|
|
public static Flux<LLKeyScore> convertHits(Flux<ScoreDoc> hitsFlux,
|
|
List<IndexSearcher> indexSearchers,
|
|
String keyFieldName,
|
|
boolean preserveOrder) {
|
|
if (preserveOrder) {
|
|
return hitsFlux
|
|
.mapNotNull(hit -> mapHitBlocking(hit, indexSearchers, keyFieldName));
|
|
} else {
|
|
// Compute parallelism
|
|
var availableProcessors = Runtime.getRuntime().availableProcessors();
|
|
var min = Queues.XS_BUFFER_SIZE;
|
|
var maxParallelGroups = Math.max(availableProcessors, min);
|
|
|
|
return hitsFlux
|
|
.groupBy(hit -> hit.shardIndex % maxParallelGroups) // Max n groups
|
|
.flatMap(shardHits -> shardHits
|
|
.mapNotNull(hit -> mapHitBlocking(hit, indexSearchers, keyFieldName)),
|
|
maxParallelGroups // Max n concurrency. Concurrency must be >= total groups count
|
|
);
|
|
}
|
|
}
|
|
|
|
@Nullable
|
|
private static LLKeyScore mapHitBlocking(ScoreDoc hit,
|
|
List<IndexSearcher> indexSearchers,
|
|
String keyFieldName) {
|
|
if (Schedulers.isInNonBlockingThread()) {
|
|
throw new UnsupportedOperationException("Called mapHitBlocking in a nonblocking thread");
|
|
}
|
|
int shardDocId = hit.doc;
|
|
int shardIndex = hit.shardIndex;
|
|
float score = hit.score;
|
|
if (shardIndex == -1 && indexSearchers.size() == 1) {
|
|
shardIndex = 0;
|
|
}
|
|
var indexSearcher = indexSearchers.get(shardIndex);
|
|
try {
|
|
String collectedDoc = keyOfTopDoc(shardDocId, indexSearcher.getIndexReader(), keyFieldName);
|
|
return new LLKeyScore(shardDocId, score, collectedDoc);
|
|
} catch (NoSuchElementException ex) {
|
|
logger.debug("Error: document {} key is not present!", shardDocId);
|
|
return null;
|
|
} catch (Exception ex) {
|
|
logger.error("Failed to read document {}", shardDocId, ex);
|
|
return new LLKeyScore(shardDocId, score, null);
|
|
}
|
|
}
|
|
|
|
public static TopDocs mergeTopDocs(
|
|
@Nullable Sort sort,
|
|
@Nullable Integer startN,
|
|
@Nullable Integer topN,
|
|
TopDocs[] topDocs,
|
|
Comparator<ScoreDoc> tieBreaker) {
|
|
if ((startN == null) != (topN == null)) {
|
|
throw new IllegalArgumentException("You must pass startN and topN together or nothing");
|
|
}
|
|
TopDocs result;
|
|
if (sort != null) {
|
|
if (!(topDocs instanceof TopFieldDocs[])) {
|
|
throw new IllegalStateException("Expected TopFieldDocs[], got TopDocs[]");
|
|
}
|
|
if (startN == null) {
|
|
int defaultTopN = 0;
|
|
for (TopDocs td : topDocs) {
|
|
int length = td.scoreDocs.length;
|
|
defaultTopN += length;
|
|
}
|
|
result = TopDocs.merge(sort, 0, defaultTopN,
|
|
(TopFieldDocs[]) topDocs,
|
|
tieBreaker
|
|
);
|
|
} else {
|
|
result = TopDocs.merge(sort, startN,
|
|
topN,
|
|
(TopFieldDocs[]) topDocs,
|
|
tieBreaker
|
|
);
|
|
}
|
|
} else {
|
|
if (startN == null) {
|
|
int defaultTopN = 0;
|
|
for (TopDocs td : topDocs) {
|
|
int length = td.scoreDocs.length;
|
|
defaultTopN += length;
|
|
}
|
|
result = TopDocs.merge(0,
|
|
defaultTopN,
|
|
topDocs,
|
|
tieBreaker
|
|
);
|
|
} else {
|
|
result = TopDocs.merge(startN,
|
|
topN,
|
|
topDocs,
|
|
tieBreaker
|
|
);
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
|
|
public static int totalHitsThreshold(boolean complete) {
|
|
return complete ? Integer.MAX_VALUE : 1;
|
|
}
|
|
|
|
public static long totalHitsThresholdLong(boolean complete) {
|
|
return complete ? Long.MAX_VALUE : 1;
|
|
}
|
|
|
|
public static TotalHitsCount convertTotalHitsCount(TotalHits totalHits) {
|
|
return switch (totalHits.relation) {
|
|
case EQUAL_TO -> TotalHitsCount.of(totalHits.value, true);
|
|
case GREATER_THAN_OR_EQUAL_TO -> TotalHitsCount.of(totalHits.value, false);
|
|
};
|
|
}
|
|
|
|
public static TotalHitsCount sum(TotalHitsCount totalHitsCount, TotalHitsCount totalHitsCount1) {
|
|
return TotalHitsCount.of(totalHitsCount.value() + totalHitsCount1.value(),
|
|
totalHitsCount.exact() && totalHitsCount1.exact()
|
|
);
|
|
}
|
|
|
|
@SuppressWarnings("unused")
|
|
public static String toHumanReadableString(TotalHitsCount totalHitsCount) {
|
|
if (totalHitsCount.exact()) {
|
|
return Long.toString(totalHitsCount.value());
|
|
} else {
|
|
return totalHitsCount.value() + "+";
|
|
}
|
|
}
|
|
|
|
public static Mono<LocalQueryParams> getMoreLikeThisQuery(
|
|
LLIndexSearchers inputIndexSearchers,
|
|
LocalQueryParams localQueryParams,
|
|
Analyzer analyzer,
|
|
Similarity similarity,
|
|
Flux<Tuple2<String, Set<String>>> mltDocumentFieldsFlux) {
|
|
var indexSearchers = inputIndexSearchers.shards();
|
|
Query luceneAdditionalQuery;
|
|
try {
|
|
luceneAdditionalQuery = localQueryParams.query();
|
|
} catch (Exception e) {
|
|
return Mono.error(e);
|
|
}
|
|
return mltDocumentFieldsFlux
|
|
.collectMap(Tuple2::getT1, Tuple2::getT2, HashMap::new)
|
|
.flatMap(mltDocumentFields -> Mono.fromCallable(() -> {
|
|
mltDocumentFields.entrySet().removeIf(entry -> entry.getValue().isEmpty());
|
|
if (mltDocumentFields.isEmpty()) {
|
|
return new LocalQueryParams(new MatchNoDocsQuery(),
|
|
localQueryParams.offsetLong(),
|
|
localQueryParams.limitLong(),
|
|
DEFAULT_PAGE_LIMITS,
|
|
localQueryParams.minCompetitiveScore(),
|
|
localQueryParams.sort(),
|
|
localQueryParams.complete()
|
|
);
|
|
}
|
|
MultiMoreLikeThis mlt;
|
|
if (indexSearchers.size() == 1) {
|
|
mlt = new MultiMoreLikeThis(new BigCompositeReader<>(indexSearchers.get(0).getIndexReader(), IndexReader[]::new), null);
|
|
} else {
|
|
IndexReader[] indexReaders = new IndexReader[indexSearchers.size()];
|
|
for (int i = 0, size = indexSearchers.size(); i < size; i++) {
|
|
indexReaders[i] = indexSearchers.get(i).getIndexReader();
|
|
}
|
|
mlt = new MultiMoreLikeThis(new BigCompositeReader<>(indexReaders, new ArrayIndexComparator(indexReaders)), null);
|
|
}
|
|
mlt.setAnalyzer(analyzer);
|
|
mlt.setFieldNames(mltDocumentFields.keySet().toArray(String[]::new));
|
|
mlt.setMinTermFreq(1);
|
|
mlt.setMinDocFreq(3);
|
|
mlt.setMaxDocFreqPct(20);
|
|
mlt.setBoost(localQueryParams.needsScores());
|
|
mlt.setStopWords(EnglishItalianStopFilter.getStopWordsString());
|
|
if (similarity instanceof TFIDFSimilarity tfidfSimilarity) {
|
|
mlt.setSimilarity(tfidfSimilarity);
|
|
} else {
|
|
mlt.setSimilarity(new ClassicSimilarity());
|
|
}
|
|
|
|
// Get the reference docId and apply it to MoreLikeThis, to generate the query
|
|
@SuppressWarnings({"unchecked", "rawtypes"})
|
|
var mltQuery = mlt.like((Map) mltDocumentFields);
|
|
Query luceneQuery;
|
|
if (!(luceneAdditionalQuery instanceof MatchAllDocsQuery)) {
|
|
luceneQuery = new Builder()
|
|
.add(mltQuery, Occur.MUST)
|
|
.add(new ConstantScoreQuery(luceneAdditionalQuery), Occur.MUST)
|
|
.build();
|
|
} else {
|
|
luceneQuery = mltQuery;
|
|
}
|
|
|
|
return new LocalQueryParams(luceneQuery,
|
|
localQueryParams.offsetLong(),
|
|
localQueryParams.limitLong(),
|
|
DEFAULT_PAGE_LIMITS,
|
|
localQueryParams.minCompetitiveScore(),
|
|
localQueryParams.sort(),
|
|
localQueryParams.complete()
|
|
);
|
|
}).subscribeOn(Schedulers.boundedElastic()));
|
|
}
|
|
}
|