Refactor iterations
This commit is contained in:
parent
5f3bf768ad
commit
08eb457235
@ -213,13 +213,17 @@ public class DatabaseMapDictionaryDeep<T, U, US extends DatabaseStage<U>> implem
|
||||
@Override
|
||||
public Mono<US> at(@Nullable CompositeSnapshot snapshot, T keySuffix) {
|
||||
byte[] keySuffixData = serializeSuffix(keySuffix);
|
||||
Flux<byte[]> keyFlux;
|
||||
if (this.subStageGetter.needsKeyFlux()) {
|
||||
keyFlux = this.dictionary.getRangeKeys(resolveSnapshot(snapshot), toExtRange(keySuffixData));
|
||||
} else {
|
||||
keyFlux = Flux.empty();
|
||||
}
|
||||
return this.subStageGetter
|
||||
.subStage(dictionary,
|
||||
snapshot,
|
||||
toKeyWithoutExt(keySuffixData),
|
||||
this.subStageGetter.needsKeyFlux()
|
||||
? this.dictionary.getRangeKeys(resolveSnapshot(snapshot), toExtRange(keySuffixData))
|
||||
: Flux.empty()
|
||||
keyFlux
|
||||
);
|
||||
}
|
||||
|
||||
|
@ -84,19 +84,17 @@ public interface DatabaseStageMap<T, U, US extends DatabaseStage<U>> extends Dat
|
||||
}
|
||||
|
||||
default Mono<Void> replaceAllValues(boolean canKeysChange, Function<Entry<T, U>, Mono<Entry<T, U>>> entriesReplacer) {
|
||||
return Mono.defer(() -> {
|
||||
if (canKeysChange) {
|
||||
return this.setAllValues(this.getAllValues(null).flatMap(entriesReplacer)).then();
|
||||
} else {
|
||||
return this
|
||||
.getAllValues(null)
|
||||
.flatMap(entriesReplacer)
|
||||
.flatMap(replacedEntry -> this
|
||||
.at(null, replacedEntry.getKey())
|
||||
.map(entry -> entry.set(replacedEntry.getValue())))
|
||||
.then();
|
||||
}
|
||||
});
|
||||
if (canKeysChange) {
|
||||
return this.setAllValues(this.getAllValues(null).flatMap(entriesReplacer)).then();
|
||||
} else {
|
||||
return this
|
||||
.getAllValues(null)
|
||||
.flatMap(entriesReplacer)
|
||||
.flatMap(replacedEntry -> this
|
||||
.at(null, replacedEntry.getKey())
|
||||
.map(entry -> entry.set(replacedEntry.getValue())))
|
||||
.then();
|
||||
}
|
||||
}
|
||||
|
||||
default Mono<Void> replaceAll(Function<Entry<T, US>, Mono<Void>> entriesReplacer) {
|
||||
|
@ -46,7 +46,7 @@ public class SubStageGetterMap<T, U> implements SubStageGetter<Map<T, U>, Databa
|
||||
|
||||
@Override
|
||||
public boolean needsKeyFlux() {
|
||||
return assertsEnabled;
|
||||
return true;
|
||||
}
|
||||
|
||||
private Mono<Void> checkKeyFluxConsistency(byte[] prefixKey, Flux<byte[]> keyFlux) {
|
||||
|
@ -63,7 +63,7 @@ public class SubStageGetterMapDeep<T, U, US extends DatabaseStage<U>> implements
|
||||
|
||||
@Override
|
||||
public boolean needsKeyFlux() {
|
||||
return assertsEnabled;
|
||||
return true;
|
||||
}
|
||||
|
||||
private Mono<Void> checkKeyFluxConsistency(byte[] prefixKey, Flux<byte[]> keyFlux) {
|
||||
|
@ -168,6 +168,7 @@ public class LLLocalDictionary implements LLDictionary {
|
||||
.fromCallable(() -> {
|
||||
var readOpts = resolveSnapshot(snapshot);
|
||||
readOpts.setVerifyChecksums(false);
|
||||
readOpts.setFillCache(false);
|
||||
if (range.hasMin()) {
|
||||
readOpts.setIterateLowerBound(new Slice(range.getMin()));
|
||||
}
|
||||
@ -355,49 +356,47 @@ public class LLLocalDictionary implements LLDictionary {
|
||||
}
|
||||
|
||||
private Mono<byte[]> getPrevValue(byte[] key, LLDictionaryResultType resultType) {
|
||||
return Mono.defer(() -> {
|
||||
switch (resultType) {
|
||||
case VALUE_CHANGED:
|
||||
return containsKey(null, key).single().map(LLUtils::booleanToResponse);
|
||||
case PREVIOUS_VALUE:
|
||||
return Mono
|
||||
.fromCallable(() -> {
|
||||
StampedLock lock;
|
||||
long stamp;
|
||||
if (updateMode == UpdateMode.ALLOW) {
|
||||
lock = itemsLock.getAt(getLockIndex(key));
|
||||
switch (resultType) {
|
||||
case VALUE_CHANGED:
|
||||
return containsKey(null, key).single().map(LLUtils::booleanToResponse);
|
||||
case PREVIOUS_VALUE:
|
||||
return Mono
|
||||
.fromCallable(() -> {
|
||||
StampedLock lock;
|
||||
long stamp;
|
||||
if (updateMode == UpdateMode.ALLOW) {
|
||||
lock = itemsLock.getAt(getLockIndex(key));
|
||||
|
||||
stamp = lock.readLock();
|
||||
} else {
|
||||
lock = null;
|
||||
stamp = 0;
|
||||
}
|
||||
try {
|
||||
logger.trace("Reading {}", key);
|
||||
var data = new Holder<byte[]>();
|
||||
if (db.keyMayExist(cfh, key, data)) {
|
||||
if (data.getValue() != null) {
|
||||
return data.getValue();
|
||||
} else {
|
||||
return db.get(cfh, key);
|
||||
}
|
||||
stamp = lock.readLock();
|
||||
} else {
|
||||
lock = null;
|
||||
stamp = 0;
|
||||
}
|
||||
try {
|
||||
logger.trace("Reading {}", key);
|
||||
var data = new Holder<byte[]>();
|
||||
if (db.keyMayExist(cfh, key, data)) {
|
||||
if (data.getValue() != null) {
|
||||
return data.getValue();
|
||||
} else {
|
||||
return null;
|
||||
}
|
||||
} finally {
|
||||
if (updateMode == UpdateMode.ALLOW) {
|
||||
lock.unlockRead(stamp);
|
||||
return db.get(cfh, key);
|
||||
}
|
||||
} else {
|
||||
return null;
|
||||
}
|
||||
})
|
||||
.onErrorMap(cause -> new IOException("Failed to read " + Arrays.toString(key), cause))
|
||||
.subscribeOn(dbScheduler);
|
||||
case VOID:
|
||||
return Mono.empty();
|
||||
default:
|
||||
return Mono.error(new IllegalStateException("Unexpected value: " + resultType));
|
||||
}
|
||||
});
|
||||
} finally {
|
||||
if (updateMode == UpdateMode.ALLOW) {
|
||||
lock.unlockRead(stamp);
|
||||
}
|
||||
}
|
||||
})
|
||||
.onErrorMap(cause -> new IOException("Failed to read " + Arrays.toString(key), cause))
|
||||
.subscribeOn(dbScheduler);
|
||||
case VOID:
|
||||
return Mono.empty();
|
||||
default:
|
||||
return Mono.error(new IllegalStateException("Unexpected value: " + resultType));
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
@ -522,26 +521,22 @@ public class LLLocalDictionary implements LLDictionary {
|
||||
|
||||
@Override
|
||||
public Flux<Entry<byte[], byte[]>> getRange(@Nullable LLSnapshot snapshot, LLRange range) {
|
||||
return Flux.defer(() -> {
|
||||
if (range.isSingle()) {
|
||||
return getRangeSingle(snapshot, range.getMin());
|
||||
} else {
|
||||
return getRangeMulti(snapshot, range);
|
||||
}
|
||||
});
|
||||
if (range.isSingle()) {
|
||||
return getRangeSingle(snapshot, range.getMin());
|
||||
} else {
|
||||
return getRangeMulti(snapshot, range);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public Flux<List<Entry<byte[], byte[]>>> getRangeGrouped(@Nullable LLSnapshot snapshot,
|
||||
LLRange range,
|
||||
int prefixLength) {
|
||||
return Flux.defer(() -> {
|
||||
if (range.isSingle()) {
|
||||
return getRangeSingle(snapshot, range.getMin()).map(List::of);
|
||||
} else {
|
||||
return getRangeMultiGrouped(snapshot, range, prefixLength);
|
||||
}
|
||||
});
|
||||
if (range.isSingle()) {
|
||||
return getRangeSingle(snapshot, range.getMin()).map(List::of);
|
||||
} else {
|
||||
return getRangeMultiGrouped(snapshot, range, prefixLength);
|
||||
}
|
||||
}
|
||||
|
||||
private Flux<Entry<byte[],byte[]>> getRangeSingle(LLSnapshot snapshot, byte[] key) {
|
||||
@ -552,22 +547,30 @@ public class LLLocalDictionary implements LLDictionary {
|
||||
}
|
||||
|
||||
private Flux<Entry<byte[],byte[]>> getRangeMulti(LLSnapshot snapshot, LLRange range) {
|
||||
return new LLLocalLuceneEntryReactiveIterator(db, cfh, range, resolveSnapshot(snapshot)).subscribeOn(dbScheduler);
|
||||
return new LLLocalLuceneEntryReactiveIterator(db, cfh, range, resolveSnapshot(snapshot))
|
||||
.flux()
|
||||
.subscribeOn(dbScheduler);
|
||||
}
|
||||
|
||||
private Flux<List<Entry<byte[],byte[]>>> getRangeMultiGrouped(LLSnapshot snapshot, LLRange range, int prefixLength) {
|
||||
return new LLLocalLuceneGroupedEntryReactiveIterator(db, cfh, prefixLength, range, resolveSnapshot(snapshot)).subscribeOn(dbScheduler);
|
||||
return new LLLocalLuceneGroupedEntryReactiveIterator(db,
|
||||
cfh,
|
||||
prefixLength,
|
||||
range,
|
||||
resolveSnapshot(snapshot),
|
||||
"getRangeMultiGrouped"
|
||||
)
|
||||
.flux()
|
||||
.subscribeOn(dbScheduler);
|
||||
}
|
||||
|
||||
@Override
|
||||
public Flux<byte[]> getRangeKeys(@Nullable LLSnapshot snapshot, LLRange range) {
|
||||
return Flux.defer(() -> {
|
||||
if (range.isSingle()) {
|
||||
return getRangeKeysSingle(snapshot, range.getMin());
|
||||
} else {
|
||||
return getRangeKeysMulti(snapshot, range);
|
||||
}
|
||||
});
|
||||
if (range.isSingle()) {
|
||||
return getRangeKeysSingle(snapshot, range.getMin());
|
||||
} else {
|
||||
return getRangeKeysMulti(snapshot, range);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
@ -576,8 +579,9 @@ public class LLLocalDictionary implements LLDictionary {
|
||||
cfh,
|
||||
prefixLength,
|
||||
range,
|
||||
resolveSnapshot(snapshot)
|
||||
).subscribeOn(dbScheduler);
|
||||
resolveSnapshot(snapshot),
|
||||
"getRangeKeysGrouped"
|
||||
).flux().subscribeOn(dbScheduler);
|
||||
}
|
||||
|
||||
private Flux<byte[]> getRangeKeysSingle(LLSnapshot snapshot, byte[] key) {
|
||||
@ -589,14 +593,14 @@ public class LLLocalDictionary implements LLDictionary {
|
||||
}
|
||||
|
||||
private Flux<byte[]> getRangeKeysMulti(LLSnapshot snapshot, LLRange range) {
|
||||
return new LLLocalLuceneKeysReactiveIterator(db, cfh, range, resolveSnapshot(snapshot)).subscribeOn(dbScheduler);
|
||||
return new LLLocalLuceneKeysReactiveIterator(db, cfh, range, resolveSnapshot(snapshot)).flux().subscribeOn(dbScheduler);
|
||||
}
|
||||
|
||||
@Override
|
||||
public Flux<Entry<byte[], byte[]>> setRange(LLRange range,
|
||||
Flux<Entry<byte[], byte[]>> entries,
|
||||
boolean getOldValues) {
|
||||
return Flux.defer(() -> Flux
|
||||
return Flux
|
||||
.usingWhen(
|
||||
Mono
|
||||
.fromCallable(() -> new CappedWriteBatch(db,
|
||||
@ -653,8 +657,7 @@ public class LLLocalDictionary implements LLDictionary {
|
||||
.subscribeOn(dbScheduler)
|
||||
)
|
||||
.subscribeOn(dbScheduler)
|
||||
.onErrorMap(cause -> new IOException("Failed to write range", cause))
|
||||
);
|
||||
.onErrorMap(cause -> new IOException("Failed to write range", cause));
|
||||
}
|
||||
|
||||
private static byte[] incrementLexicographically(byte[] key) {
|
||||
@ -728,44 +731,41 @@ public class LLLocalDictionary implements LLDictionary {
|
||||
|
||||
@Override
|
||||
public Mono<Long> sizeRange(@Nullable LLSnapshot snapshot, LLRange range, boolean fast) {
|
||||
return Mono
|
||||
.defer(() -> {
|
||||
if (range.isAll()) {
|
||||
return Mono
|
||||
.fromCallable(() -> fast ? fastSizeAll(snapshot) : exactSizeAll(snapshot))
|
||||
.onErrorMap(IOException::new)
|
||||
.subscribeOn(dbScheduler);
|
||||
} else {
|
||||
return Mono
|
||||
.fromCallable(() -> {
|
||||
var readOpts = resolveSnapshot(snapshot);
|
||||
readOpts.setFillCache(false);
|
||||
readOpts.setVerifyChecksums(false);
|
||||
if (range.hasMin()) {
|
||||
readOpts.setIterateLowerBound(new Slice(range.getMin()));
|
||||
}
|
||||
if (range.hasMax()) {
|
||||
readOpts.setIterateUpperBound(new Slice(range.getMax()));
|
||||
}
|
||||
if (fast) {
|
||||
readOpts.setIgnoreRangeDeletions(true);
|
||||
if (range.isAll()) {
|
||||
return Mono
|
||||
.fromCallable(() -> fast ? fastSizeAll(snapshot) : exactSizeAll(snapshot))
|
||||
.onErrorMap(IOException::new)
|
||||
.subscribeOn(dbScheduler);
|
||||
} else {
|
||||
return Mono
|
||||
.fromCallable(() -> {
|
||||
var readOpts = resolveSnapshot(snapshot);
|
||||
readOpts.setFillCache(false);
|
||||
readOpts.setVerifyChecksums(false);
|
||||
if (range.hasMin()) {
|
||||
readOpts.setIterateLowerBound(new Slice(range.getMin()));
|
||||
}
|
||||
if (range.hasMax()) {
|
||||
readOpts.setIterateUpperBound(new Slice(range.getMax()));
|
||||
}
|
||||
if (fast) {
|
||||
readOpts.setIgnoreRangeDeletions(true);
|
||||
|
||||
}
|
||||
try (var iter = db.newIterator(cfh, readOpts)) {
|
||||
iter.seekToFirst();
|
||||
long i = 0;
|
||||
while (iter.isValid()) {
|
||||
iter.next();
|
||||
i++;
|
||||
}
|
||||
return i;
|
||||
}
|
||||
})
|
||||
.onErrorMap(cause -> new IOException("Failed to get size of range "
|
||||
+ range.toString(), cause))
|
||||
.subscribeOn(dbScheduler);
|
||||
}
|
||||
});
|
||||
}
|
||||
try (var iter = db.newIterator(cfh, readOpts)) {
|
||||
iter.seekToFirst();
|
||||
long i = 0;
|
||||
while (iter.isValid()) {
|
||||
iter.next();
|
||||
i++;
|
||||
}
|
||||
return i;
|
||||
}
|
||||
})
|
||||
.onErrorMap(cause -> new IOException("Failed to get size of range "
|
||||
+ range.toString(), cause))
|
||||
.subscribeOn(dbScheduler);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -13,8 +13,9 @@ public class LLLocalLuceneGroupedEntryReactiveIterator extends LLLocalLuceneGrou
|
||||
ColumnFamilyHandle cfh,
|
||||
int prefixLength,
|
||||
LLRange range,
|
||||
ReadOptions readOptions) {
|
||||
super(db, cfh, prefixLength, range, readOptions, true);
|
||||
ReadOptions readOptions,
|
||||
String debugName) {
|
||||
super(db, cfh, prefixLength, range, readOptions, true, debugName);
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -11,8 +11,9 @@ public class LLLocalLuceneGroupedKeysReactiveIterator extends LLLocalLuceneGroup
|
||||
ColumnFamilyHandle cfh,
|
||||
int prefixLength,
|
||||
LLRange range,
|
||||
ReadOptions readOptions) {
|
||||
super(db, cfh, prefixLength, range, readOptions, false);
|
||||
ReadOptions readOptions,
|
||||
String debugName) {
|
||||
super(db, cfh, prefixLength, range, readOptions, false, debugName);
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -4,15 +4,13 @@ import it.cavallium.dbengine.database.LLRange;
|
||||
import it.unimi.dsi.fastutil.objects.ObjectArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.List;
|
||||
import org.jetbrains.annotations.NotNull;
|
||||
import org.rocksdb.ColumnFamilyHandle;
|
||||
import org.rocksdb.ReadOptions;
|
||||
import org.rocksdb.RocksDB;
|
||||
import org.rocksdb.Slice;
|
||||
import reactor.core.CoreSubscriber;
|
||||
import reactor.core.publisher.Flux;
|
||||
|
||||
public abstract class LLLocalLuceneGroupedReactiveIterator<T> extends Flux<List<T>> {
|
||||
public abstract class LLLocalLuceneGroupedReactiveIterator<T> {
|
||||
|
||||
private static final byte[] EMPTY = new byte[0];
|
||||
|
||||
@ -22,24 +20,28 @@ public abstract class LLLocalLuceneGroupedReactiveIterator<T> extends Flux<List<
|
||||
private final LLRange range;
|
||||
private final ReadOptions readOptions;
|
||||
private final boolean readValues;
|
||||
private final String debugName;
|
||||
|
||||
public LLLocalLuceneGroupedReactiveIterator(RocksDB db,
|
||||
ColumnFamilyHandle cfh,
|
||||
int prefixLength,
|
||||
LLRange range,
|
||||
ReadOptions readOptions,
|
||||
boolean readValues) {
|
||||
boolean readValues,
|
||||
String debugName) {
|
||||
this.db = db;
|
||||
this.cfh = cfh;
|
||||
this.prefixLength = prefixLength;
|
||||
this.range = range;
|
||||
this.readOptions = readOptions;
|
||||
this.readValues = readValues;
|
||||
this.debugName = debugName;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void subscribe(@NotNull CoreSubscriber<? super List<T>> actual) {
|
||||
Flux<List<T>> flux = Flux
|
||||
|
||||
@SuppressWarnings("Convert2MethodRef")
|
||||
public Flux<List<T>> flux() {
|
||||
return Flux
|
||||
.generate(() -> {
|
||||
var readOptions = new ReadOptions(this.readOptions);
|
||||
readOptions.setFillCache(range.hasMin() && range.hasMax());
|
||||
@ -78,8 +80,7 @@ public abstract class LLLocalLuceneGroupedReactiveIterator<T> extends Flux<List<
|
||||
sink.complete();
|
||||
}
|
||||
return rocksIterator;
|
||||
}, tuple -> {});
|
||||
flux.subscribe(actual);
|
||||
}, rocksIterator1 -> rocksIterator1.close());
|
||||
}
|
||||
|
||||
public abstract T getEntry(byte[] key, byte[] value);
|
||||
|
@ -1,16 +1,13 @@
|
||||
package it.cavallium.dbengine.database.disk;
|
||||
|
||||
import it.cavallium.dbengine.database.LLRange;
|
||||
import org.jetbrains.annotations.NotNull;
|
||||
import org.rocksdb.AbstractImmutableNativeReference;
|
||||
import org.rocksdb.ColumnFamilyHandle;
|
||||
import org.rocksdb.ReadOptions;
|
||||
import org.rocksdb.RocksDB;
|
||||
import org.rocksdb.Slice;
|
||||
import reactor.core.CoreSubscriber;
|
||||
import reactor.core.publisher.Flux;
|
||||
|
||||
public abstract class LLLocalLuceneReactiveIterator<T> extends Flux<T> {
|
||||
public abstract class LLLocalLuceneReactiveIterator<T> {
|
||||
|
||||
private static final byte[] EMPTY = new byte[0];
|
||||
|
||||
@ -32,9 +29,9 @@ public abstract class LLLocalLuceneReactiveIterator<T> extends Flux<T> {
|
||||
this.readValues = readValues;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void subscribe(@NotNull CoreSubscriber<? super T> actual) {
|
||||
Flux<T> flux = Flux
|
||||
@SuppressWarnings("Convert2MethodRef")
|
||||
public Flux<T> flux() {
|
||||
return Flux
|
||||
.generate(() -> {
|
||||
var readOptions = new ReadOptions(this.readOptions);
|
||||
readOptions.setFillCache(range.hasMin() && range.hasMax());
|
||||
@ -61,8 +58,7 @@ public abstract class LLLocalLuceneReactiveIterator<T> extends Flux<T> {
|
||||
sink.complete();
|
||||
}
|
||||
return rocksIterator;
|
||||
}, AbstractImmutableNativeReference::close);
|
||||
flux.subscribe(actual);
|
||||
}, rocksIterator1 -> rocksIterator1.close());
|
||||
}
|
||||
|
||||
public abstract T getEntry(byte[] key, byte[] value);
|
||||
|
Loading…
Reference in New Issue
Block a user