2020-12-07 22:15:18 +01:00
|
|
|
package it.cavallium.dbengine.database.disk;
|
|
|
|
|
2021-01-17 18:31:25 +01:00
|
|
|
import it.cavallium.dbengine.database.LLDictionary;
|
|
|
|
import it.cavallium.dbengine.database.LLDictionaryResultType;
|
2021-01-30 00:24:55 +01:00
|
|
|
import it.cavallium.dbengine.database.LLRange;
|
2021-01-17 18:31:25 +01:00
|
|
|
import it.cavallium.dbengine.database.LLSnapshot;
|
|
|
|
import it.cavallium.dbengine.database.LLUtils;
|
2021-02-01 02:21:53 +01:00
|
|
|
import it.unimi.dsi.fastutil.objects.ObjectArrayList;
|
2020-12-07 22:15:18 +01:00
|
|
|
import java.io.IOException;
|
2021-02-01 02:21:53 +01:00
|
|
|
import java.util.ArrayList;
|
2020-12-07 22:15:18 +01:00
|
|
|
import java.util.Arrays;
|
2021-02-01 02:21:53 +01:00
|
|
|
import java.util.List;
|
2020-12-07 22:15:18 +01:00
|
|
|
import java.util.Map;
|
|
|
|
import java.util.Map.Entry;
|
|
|
|
import java.util.Objects;
|
|
|
|
import java.util.function.Function;
|
|
|
|
import org.jetbrains.annotations.NotNull;
|
|
|
|
import org.jetbrains.annotations.Nullable;
|
|
|
|
import org.rocksdb.ColumnFamilyHandle;
|
|
|
|
import org.rocksdb.FlushOptions;
|
|
|
|
import org.rocksdb.Holder;
|
|
|
|
import org.rocksdb.ReadOptions;
|
|
|
|
import org.rocksdb.RocksDB;
|
|
|
|
import org.rocksdb.RocksDBException;
|
|
|
|
import org.rocksdb.RocksIterator;
|
|
|
|
import org.rocksdb.Snapshot;
|
|
|
|
import org.rocksdb.WriteOptions;
|
|
|
|
import org.warp.commonutils.concurrency.atomicity.NotAtomic;
|
2021-01-30 00:24:55 +01:00
|
|
|
import reactor.core.publisher.Flux;
|
|
|
|
import reactor.core.publisher.Mono;
|
2021-02-01 02:21:53 +01:00
|
|
|
import reactor.core.scheduler.Scheduler;
|
2020-12-07 22:15:18 +01:00
|
|
|
|
|
|
|
@NotAtomic
|
|
|
|
public class LLLocalDictionary implements LLDictionary {
|
|
|
|
|
|
|
|
private static final boolean USE_CURRENT_FASTSIZE_FOR_OLD_SNAPSHOTS = true;
|
|
|
|
static final int RESERVED_WRITE_BATCH_SIZE = 2 * 1024 * 1024; // 2MiB
|
|
|
|
static final long MAX_WRITE_BATCH_SIZE = 1024L * 1024L * 1024L; // 1GiB
|
|
|
|
static final int CAPPED_WRITE_BATCH_CAP = 50000; // 50K operations
|
2021-02-01 02:21:53 +01:00
|
|
|
static final int MULTI_GET_WINDOW = 500;
|
2020-12-07 22:15:18 +01:00
|
|
|
static final WriteOptions BATCH_WRITE_OPTIONS = new WriteOptions().setLowPri(true);
|
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
private static final byte[] FIRST_KEY = new byte[]{};
|
2020-12-07 22:15:18 +01:00
|
|
|
private static final byte[] NO_DATA = new byte[0];
|
|
|
|
private static final ReadOptions EMPTY_READ_OPTIONS = new ReadOptions();
|
|
|
|
private final RocksDB db;
|
|
|
|
private final ColumnFamilyHandle cfh;
|
|
|
|
private final String databaseName;
|
2021-02-01 02:21:53 +01:00
|
|
|
private final Scheduler dbScheduler;
|
2020-12-07 22:15:18 +01:00
|
|
|
private final Function<LLSnapshot, Snapshot> snapshotResolver;
|
|
|
|
|
|
|
|
public LLLocalDictionary(@NotNull RocksDB db,
|
|
|
|
@NotNull ColumnFamilyHandle columnFamilyHandle,
|
|
|
|
String databaseName,
|
2021-02-01 02:21:53 +01:00
|
|
|
Scheduler dbScheduler,
|
2020-12-07 22:15:18 +01:00
|
|
|
Function<LLSnapshot, Snapshot> snapshotResolver) {
|
|
|
|
Objects.requireNonNull(db);
|
|
|
|
this.db = db;
|
|
|
|
Objects.requireNonNull(columnFamilyHandle);
|
|
|
|
this.cfh = columnFamilyHandle;
|
|
|
|
this.databaseName = databaseName;
|
2021-02-01 02:21:53 +01:00
|
|
|
this.dbScheduler = dbScheduler;
|
2020-12-07 22:15:18 +01:00
|
|
|
this.snapshotResolver = snapshotResolver;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public String getDatabaseName() {
|
|
|
|
return databaseName;
|
|
|
|
}
|
|
|
|
|
|
|
|
private ReadOptions resolveSnapshot(LLSnapshot snapshot) {
|
|
|
|
if (snapshot != null) {
|
|
|
|
return getReadOptions(snapshotResolver.apply(snapshot));
|
|
|
|
} else {
|
|
|
|
return EMPTY_READ_OPTIONS;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private ReadOptions getReadOptions(Snapshot snapshot) {
|
|
|
|
if (snapshot != null) {
|
|
|
|
return new ReadOptions().setSnapshot(snapshot);
|
|
|
|
} else {
|
|
|
|
return EMPTY_READ_OPTIONS;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
2021-01-30 00:24:55 +01:00
|
|
|
public Mono<byte[]> get(@Nullable LLSnapshot snapshot, byte[] key) {
|
|
|
|
return Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
Holder<byte[]> data = new Holder<>();
|
|
|
|
if (db.keyMayExist(cfh, resolveSnapshot(snapshot), key, data)) {
|
|
|
|
if (data.getValue() != null) {
|
|
|
|
return data.getValue();
|
|
|
|
} else {
|
|
|
|
return db.get(cfh, resolveSnapshot(snapshot), key);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
return null;
|
|
|
|
}
|
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
2021-01-30 00:26:58 +01:00
|
|
|
public Mono<Boolean> isRangeEmpty(@Nullable LLSnapshot snapshot, LLRange range) {
|
2021-01-30 00:24:55 +01:00
|
|
|
if (range.isSingle()) {
|
|
|
|
return containsKey(snapshot, range.getSingle()).map(contains -> !contains);
|
|
|
|
} else {
|
|
|
|
return containsRange(snapshot, range).map(contains -> !contains);
|
|
|
|
}
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
public Mono<Boolean> containsRange(@Nullable LLSnapshot snapshot, LLRange range) {
|
|
|
|
return Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
try (RocksIterator iter = db.newIterator(cfh, resolveSnapshot(snapshot))) {
|
|
|
|
if (range.hasMin()) {
|
|
|
|
iter.seek(range.getMin());
|
|
|
|
} else {
|
|
|
|
iter.seekToFirst();
|
|
|
|
}
|
|
|
|
if (!iter.isValid()) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (range.hasMax()) {
|
|
|
|
byte[] key1 = iter.key();
|
|
|
|
return Arrays.compareUnsigned(key1, range.getMax()) <= 0;
|
|
|
|
} else {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
private Mono<Boolean> containsKey(@Nullable LLSnapshot snapshot, byte[] key) {
|
|
|
|
return Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
int size = RocksDB.NOT_FOUND;
|
|
|
|
Holder<byte[]> data = new Holder<>();
|
|
|
|
if (db.keyMayExist(cfh, resolveSnapshot(snapshot), key, data)) {
|
2020-12-07 22:15:18 +01:00
|
|
|
if (data.getValue() != null) {
|
2021-01-30 00:24:55 +01:00
|
|
|
size = data.getValue().length;
|
2020-12-07 22:15:18 +01:00
|
|
|
} else {
|
2021-01-30 00:24:55 +01:00
|
|
|
size = db.get(cfh, resolveSnapshot(snapshot), key, NO_DATA);
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
}
|
2021-01-30 00:24:55 +01:00
|
|
|
return size != RocksDB.NOT_FOUND;
|
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public Mono<byte[]> put(byte[] key, byte[] value, LLDictionaryResultType resultType) {
|
2021-02-02 00:09:46 +01:00
|
|
|
return getPrevValue(key, resultType)
|
|
|
|
.concatWith(Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
db.put(cfh, key, value);
|
|
|
|
return null;
|
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
|
|
|
.subscribeOn(dbScheduler)
|
|
|
|
.then(Mono.empty())
|
|
|
|
).singleOrEmpty();
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
2021-01-30 00:24:55 +01:00
|
|
|
public Mono<byte[]> remove(byte[] key, LLDictionaryResultType resultType) {
|
2021-02-02 00:09:46 +01:00
|
|
|
return getPrevValue(key, resultType)
|
|
|
|
.concatWith(Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
db.delete(cfh, key);
|
|
|
|
return null;
|
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
|
|
|
.subscribeOn(dbScheduler)
|
|
|
|
.then(Mono.empty())
|
|
|
|
).singleOrEmpty();
|
2021-01-30 00:35:03 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
private Mono<byte[]> getPrevValue(byte[] key, LLDictionaryResultType resultType) {
|
2021-02-02 00:09:46 +01:00
|
|
|
return Mono.defer(() -> {
|
|
|
|
switch (resultType) {
|
|
|
|
case VALUE_CHANGED:
|
|
|
|
return containsKey(null, key).single().map(LLUtils::booleanToResponse);
|
|
|
|
case PREVIOUS_VALUE:
|
|
|
|
return Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
var data = new Holder<byte[]>();
|
|
|
|
if (db.keyMayExist(cfh, key, data)) {
|
|
|
|
if (data.getValue() != null) {
|
|
|
|
return data.getValue();
|
|
|
|
} else {
|
|
|
|
return db.get(cfh, key);
|
|
|
|
}
|
2021-01-30 00:24:55 +01:00
|
|
|
} else {
|
2021-02-02 00:09:46 +01:00
|
|
|
return null;
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
2021-02-02 00:09:46 +01:00
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
|
|
|
.subscribeOn(dbScheduler);
|
|
|
|
case VOID:
|
|
|
|
return Mono.empty();
|
|
|
|
default:
|
|
|
|
return Mono.error(new IllegalStateException("Unexpected value: " + resultType));
|
|
|
|
}
|
|
|
|
});
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
2020-12-07 22:15:18 +01:00
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
@Override
|
|
|
|
public Flux<Entry<byte[], byte[]>> getMulti(@Nullable LLSnapshot snapshot, Flux<byte[]> keys) {
|
2021-02-01 02:21:53 +01:00
|
|
|
return keys
|
|
|
|
.window(MULTI_GET_WINDOW)
|
|
|
|
.flatMap(keysWindowFlux -> keysWindowFlux.collectList()
|
|
|
|
.flatMapMany(keysWindow -> Mono
|
|
|
|
.<ArrayList<Entry<byte[], byte[]>>>fromCallable(() -> {
|
|
|
|
var handlesArray = new ColumnFamilyHandle[keysWindow.size()];
|
|
|
|
Arrays.fill(handlesArray, cfh);
|
|
|
|
var handles = ObjectArrayList.wrap(handlesArray, handlesArray.length);
|
|
|
|
var results = db.multiGetAsList(resolveSnapshot(snapshot), handles, keysWindow);
|
|
|
|
var mappedResults = new ArrayList<Entry<byte[], byte[]>>(results.size());
|
|
|
|
for (int i = 0; i < results.size(); i++) {
|
|
|
|
var val = results.get(i);
|
|
|
|
if (val != null) {
|
|
|
|
results.set(i, null);
|
|
|
|
mappedResults.add(Map.entry(keysWindow.get(i), val));
|
|
|
|
}
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
2021-02-01 02:21:53 +01:00
|
|
|
return mappedResults;
|
2021-01-30 00:24:55 +01:00
|
|
|
})
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler)
|
|
|
|
.<Entry<byte[], byte[]>>flatMapMany(Flux::fromIterable)
|
2021-01-30 00:24:55 +01:00
|
|
|
)
|
|
|
|
)
|
|
|
|
.onErrorMap(IOException::new);
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
2021-02-01 02:21:53 +01:00
|
|
|
@Override
|
|
|
|
public Flux<Entry<byte[], byte[]>> putMulti(Flux<Entry<byte[], byte[]>> entries, boolean getOldValues) {
|
|
|
|
return entries
|
|
|
|
.window(Math.min(MULTI_GET_WINDOW, CAPPED_WRITE_BATCH_CAP))
|
|
|
|
.flatMap(Flux::collectList)
|
|
|
|
.flatMap(entriesWindow -> this
|
|
|
|
.getMulti(null, Flux.fromIterable(entriesWindow).map(Entry::getKey))
|
2021-02-02 00:09:46 +01:00
|
|
|
.publishOn(dbScheduler)
|
2021-02-01 02:21:53 +01:00
|
|
|
.concatWith(Mono.fromCallable(() -> {
|
2021-02-02 00:09:46 +01:00
|
|
|
//System.out.println(Thread.currentThread()+"\tTest");
|
2021-02-01 02:21:53 +01:00
|
|
|
var batch = new CappedWriteBatch(db,
|
|
|
|
CAPPED_WRITE_BATCH_CAP,
|
|
|
|
RESERVED_WRITE_BATCH_SIZE,
|
|
|
|
MAX_WRITE_BATCH_SIZE,
|
|
|
|
BATCH_WRITE_OPTIONS
|
|
|
|
);
|
|
|
|
for (Entry<byte[], byte[]> entry : entriesWindow) {
|
|
|
|
batch.put(entry.getKey(), entry.getValue());
|
|
|
|
}
|
|
|
|
batch.writeToDbAndClose();
|
|
|
|
batch.close();
|
|
|
|
return null;
|
|
|
|
})));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2021-01-30 00:33:36 +01:00
|
|
|
@NotNull
|
|
|
|
private Mono<Entry<byte[], byte[]>> putEntryToWriteBatch(Entry<byte[], byte[]> newEntry, boolean getOldValues,
|
|
|
|
CappedWriteBatch writeBatch) {
|
|
|
|
return Mono.from(Mono
|
2021-01-30 00:35:03 +01:00
|
|
|
.defer(() -> {
|
2021-01-30 00:33:36 +01:00
|
|
|
if (getOldValues) {
|
|
|
|
return get(null, newEntry.getKey());
|
|
|
|
} else {
|
|
|
|
return Mono.empty();
|
|
|
|
}
|
|
|
|
})
|
|
|
|
.concatWith(Mono.<byte[]>fromCallable(() -> {
|
|
|
|
synchronized (writeBatch) {
|
|
|
|
writeBatch.put(cfh, newEntry.getKey(), newEntry.getValue());
|
|
|
|
}
|
|
|
|
return null;
|
|
|
|
})
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler))
|
2021-01-30 00:33:36 +01:00
|
|
|
.map(oldValue -> Map.entry(newEntry.getKey(), oldValue)));
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
2021-02-01 02:21:53 +01:00
|
|
|
@NotNull
|
|
|
|
private Flux<Entry<byte[], byte[]>> putEntryToWriteBatch(List<Entry<byte[], byte[]>> newEntries, boolean getOldValues,
|
|
|
|
CappedWriteBatch writeBatch) {
|
|
|
|
return Flux
|
|
|
|
.from(Flux
|
|
|
|
.defer(() -> {
|
|
|
|
if (getOldValues) {
|
|
|
|
return getMulti(null, Flux.fromIterable(newEntries).map(Entry::getKey));
|
|
|
|
} else {
|
|
|
|
return Flux.empty();
|
|
|
|
}
|
|
|
|
})
|
|
|
|
.concatWith(Mono
|
|
|
|
.<Entry<byte[], byte[]>>fromCallable(() -> {
|
|
|
|
synchronized (writeBatch) {
|
|
|
|
for (Entry<byte[], byte[]> newEntry : newEntries) {
|
|
|
|
writeBatch.put(cfh, newEntry.getKey(), newEntry.getValue());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return null;
|
|
|
|
}).subscribeOn(dbScheduler)
|
|
|
|
)
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
2020-12-07 22:15:18 +01:00
|
|
|
@Override
|
2021-01-30 00:24:55 +01:00
|
|
|
public Flux<Entry<byte[], byte[]>> getRange(@Nullable LLSnapshot snapshot, LLRange range) {
|
2021-02-02 00:09:46 +01:00
|
|
|
return Flux.defer(() -> {
|
|
|
|
if (range.isSingle()) {
|
|
|
|
return getRangeSingle(snapshot, range.getMin());
|
|
|
|
} else {
|
|
|
|
return getRangeMulti(snapshot, range);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public Flux<List<Entry<byte[], byte[]>>> getRangeGrouped(@Nullable LLSnapshot snapshot,
|
|
|
|
LLRange range,
|
|
|
|
int prefixLength) {
|
|
|
|
return Flux.defer(() -> {
|
|
|
|
if (range.isSingle()) {
|
|
|
|
return getRangeSingle(snapshot, range.getMin()).map(List::of);
|
|
|
|
} else {
|
|
|
|
return getRangeMultiGrouped(snapshot, range, prefixLength);
|
|
|
|
}
|
|
|
|
});
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
|
|
|
|
2021-01-30 20:16:14 +01:00
|
|
|
private Flux<Entry<byte[],byte[]>> getRangeSingle(LLSnapshot snapshot, byte[] key) {
|
|
|
|
return this
|
|
|
|
.get(snapshot, key)
|
|
|
|
.map(value -> Map.entry(key, value))
|
|
|
|
.flux();
|
|
|
|
}
|
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
private Flux<Entry<byte[],byte[]>> getRangeMulti(LLSnapshot snapshot, LLRange range) {
|
2021-02-02 00:09:46 +01:00
|
|
|
return Flux
|
|
|
|
.<Entry<byte[], byte[]>>push(sink -> {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tPreparing Read rande item");
|
|
|
|
try (var rocksIterator = db.newIterator(cfh, resolveSnapshot(snapshot))) {
|
|
|
|
if (range.hasMin()) {
|
|
|
|
rocksIterator.seek(range.getMin());
|
|
|
|
} else {
|
|
|
|
rocksIterator.seekToFirst();
|
|
|
|
}
|
|
|
|
byte[] key;
|
|
|
|
while (rocksIterator.isValid()) {
|
|
|
|
key = rocksIterator.key();
|
|
|
|
if (range.hasMax() && Arrays.compareUnsigned(key, range.getMax()) > 0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
//System.out.println(Thread.currentThread() + "\tRead rande item");
|
|
|
|
sink.next(Map.entry(key, rocksIterator.value()));
|
|
|
|
rocksIterator.next();
|
|
|
|
}
|
|
|
|
} finally {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tFinish Read rande item");
|
|
|
|
sink.complete();
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
2021-01-30 00:24:55 +01:00
|
|
|
})
|
2021-02-02 00:09:46 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
|
|
|
}
|
|
|
|
|
|
|
|
private Flux<List<Entry<byte[],byte[]>>> getRangeMultiGrouped(LLSnapshot snapshot, LLRange range, int prefixLength) {
|
|
|
|
return Flux
|
|
|
|
.<List<Entry<byte[], byte[]>>>push(sink -> {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tPreparing Read rande item");
|
|
|
|
try (var rocksIterator = db.newIterator(cfh, resolveSnapshot(snapshot))) {
|
|
|
|
if (range.hasMin()) {
|
|
|
|
rocksIterator.seek(range.getMin());
|
|
|
|
} else {
|
|
|
|
rocksIterator.seekToFirst();
|
|
|
|
}
|
|
|
|
byte[] firstGroupKey = null;
|
|
|
|
List<Entry<byte[], byte[]>> currentGroupValues = new ArrayList<>();
|
2021-01-30 00:24:55 +01:00
|
|
|
|
2021-02-02 00:09:46 +01:00
|
|
|
byte[] key;
|
|
|
|
while (rocksIterator.isValid()) {
|
|
|
|
key = rocksIterator.key();
|
|
|
|
if (firstGroupKey == null) { // Fix first value
|
|
|
|
firstGroupKey = key;
|
|
|
|
}
|
|
|
|
if (range.hasMax() && Arrays.compareUnsigned(key, range.getMax()) > 0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (Arrays.equals(firstGroupKey, 0, prefixLength, key, 0, prefixLength)) {
|
|
|
|
currentGroupValues.add(Map.entry(key, rocksIterator.value()));
|
|
|
|
} else {
|
|
|
|
if (!currentGroupValues.isEmpty()) {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tRead rande item");
|
|
|
|
sink.next(currentGroupValues);
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
2021-02-02 00:09:46 +01:00
|
|
|
firstGroupKey = key;
|
|
|
|
currentGroupValues = new ArrayList<>();
|
|
|
|
}
|
|
|
|
rocksIterator.next();
|
|
|
|
}
|
|
|
|
if (!currentGroupValues.isEmpty()) {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tRead rande item");
|
|
|
|
sink.next(currentGroupValues);
|
|
|
|
}
|
|
|
|
} finally {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tFinish Read rande item");
|
|
|
|
sink.complete();
|
|
|
|
}
|
|
|
|
})
|
|
|
|
.subscribeOn(dbScheduler);
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
|
|
|
|
2021-01-30 20:16:14 +01:00
|
|
|
@Override
|
|
|
|
public Flux<byte[]> getRangeKeys(@Nullable LLSnapshot snapshot, LLRange range) {
|
2021-02-02 00:09:46 +01:00
|
|
|
return Flux.defer(() -> {
|
|
|
|
if (range.isSingle()) {
|
|
|
|
//System.out.println(Thread.currentThread() + "getRangeKeys single");
|
|
|
|
return getRangeKeysSingle(snapshot, range.getMin()).doOnTerminate(() -> {}/*System.out.println(Thread.currentThread() + "getRangeKeys single end")*/);
|
|
|
|
} else {
|
|
|
|
//System.out.println(Thread.currentThread() + "getRangeKeys multi");
|
|
|
|
return getRangeKeysMulti(snapshot, range);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public Flux<List<byte[]>> getRangeKeysGrouped(@Nullable LLSnapshot snapshot, LLRange range, int prefixLength) {
|
|
|
|
return Flux
|
|
|
|
.<List<byte[]>>push(sink -> {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tPreparing Read rande item");
|
|
|
|
try (var rocksIterator = db.newIterator(cfh, resolveSnapshot(snapshot))) {
|
|
|
|
if (range.hasMin()) {
|
|
|
|
rocksIterator.seek(range.getMin());
|
|
|
|
} else {
|
|
|
|
rocksIterator.seekToFirst();
|
|
|
|
}
|
|
|
|
byte[] firstGroupKey = null;
|
|
|
|
List<byte[]> currentGroupValues = new ArrayList<>();
|
|
|
|
|
|
|
|
byte[] key;
|
|
|
|
while (rocksIterator.isValid()) {
|
|
|
|
key = rocksIterator.key();
|
|
|
|
if (firstGroupKey == null) { // Fix first value
|
|
|
|
firstGroupKey = key;
|
|
|
|
}
|
|
|
|
if (range.hasMax() && Arrays.compareUnsigned(key, range.getMax()) > 0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (Arrays.equals(firstGroupKey, 0, prefixLength, key, 0, prefixLength)) {
|
|
|
|
currentGroupValues.add(key);
|
|
|
|
} else {
|
|
|
|
if (!currentGroupValues.isEmpty()) {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tRead rande item");
|
|
|
|
sink.next(currentGroupValues);
|
|
|
|
}
|
|
|
|
firstGroupKey = key;
|
|
|
|
currentGroupValues = new ArrayList<>();
|
|
|
|
currentGroupValues.add(key);
|
|
|
|
}
|
|
|
|
rocksIterator.next();
|
|
|
|
}
|
|
|
|
if (!currentGroupValues.isEmpty()) {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tRead rande item");
|
|
|
|
sink.next(currentGroupValues);
|
|
|
|
}
|
|
|
|
} finally {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tFinish Read rande item");
|
|
|
|
sink.complete();
|
|
|
|
}
|
|
|
|
})
|
|
|
|
.subscribeOn(dbScheduler);
|
2021-01-30 20:16:14 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
private Flux<byte[]> getRangeKeysSingle(LLSnapshot snapshot, byte[] key) {
|
2021-01-30 00:24:55 +01:00
|
|
|
return this
|
2021-01-30 20:16:14 +01:00
|
|
|
.containsKey(snapshot, key)
|
|
|
|
.filter(contains -> contains)
|
|
|
|
.map(contains -> key)
|
2021-01-30 00:24:55 +01:00
|
|
|
.flux();
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
2021-01-30 20:16:14 +01:00
|
|
|
private Flux<byte[]> getRangeKeysMulti(LLSnapshot snapshot, LLRange range) {
|
2021-02-02 00:09:46 +01:00
|
|
|
return Flux
|
|
|
|
.<byte[]>push(sink -> {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tkPreparing Read rande item");
|
|
|
|
try (var rocksIterator = db.newIterator(cfh, resolveSnapshot(snapshot))) {
|
|
|
|
if (range.hasMin()) {
|
|
|
|
rocksIterator.seek(range.getMin());
|
|
|
|
} else {
|
|
|
|
rocksIterator.seekToFirst();
|
|
|
|
}
|
|
|
|
byte[] key;
|
|
|
|
sink.onRequest(l -> {}/*System.out.println(Thread.currentThread() + "\tkRequested " + l)*/);
|
|
|
|
while (rocksIterator.isValid()) {
|
|
|
|
key = rocksIterator.key();
|
|
|
|
if (range.hasMax() && Arrays.compareUnsigned(key, range.getMax()) > 0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
//System.out.println(Thread.currentThread() + "\tkRead rande item");
|
|
|
|
sink.next(key);
|
|
|
|
rocksIterator.next();
|
|
|
|
}
|
|
|
|
} finally {
|
|
|
|
//System.out.println(Thread.currentThread() + "\tkFinish Read rande item");
|
|
|
|
sink.complete();
|
2021-01-30 20:16:14 +01:00
|
|
|
}
|
2021-02-02 00:09:46 +01:00
|
|
|
//System.out.println(Thread.currentThread() + "\tkFinish end Read rande item");
|
2021-01-30 20:16:14 +01:00
|
|
|
})
|
2021-02-02 00:09:46 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
2021-01-30 20:16:14 +01:00
|
|
|
}
|
|
|
|
|
2020-12-07 22:15:18 +01:00
|
|
|
@Override
|
2021-01-30 00:24:55 +01:00
|
|
|
public Flux<Entry<byte[], byte[]>> setRange(LLRange range,
|
|
|
|
Flux<Entry<byte[], byte[]>> entries,
|
|
|
|
boolean getOldValues) {
|
2021-02-02 00:09:46 +01:00
|
|
|
return Flux.defer(() -> {
|
|
|
|
if (range.isAll()) {
|
|
|
|
return clear().thenMany(Flux.empty());
|
|
|
|
} else {
|
|
|
|
return Mono
|
|
|
|
.fromCallable(() -> new CappedWriteBatch(db,
|
|
|
|
CAPPED_WRITE_BATCH_CAP,
|
|
|
|
RESERVED_WRITE_BATCH_SIZE,
|
|
|
|
MAX_WRITE_BATCH_SIZE,
|
|
|
|
BATCH_WRITE_OPTIONS
|
|
|
|
))
|
|
|
|
.subscribeOn(dbScheduler)
|
|
|
|
.flatMapMany(writeBatch -> Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
synchronized (writeBatch) {
|
|
|
|
if (range.hasMin() && range.hasMax()) {
|
|
|
|
writeBatch.deleteRange(cfh, range.getMin(), range.getMax());
|
|
|
|
writeBatch.delete(cfh, range.getMax());
|
|
|
|
} else if (range.hasMax()) {
|
|
|
|
writeBatch.deleteRange(cfh, FIRST_KEY, range.getMax());
|
|
|
|
writeBatch.delete(cfh, range.getMax());
|
|
|
|
} else {
|
|
|
|
try (var it = db.newIterator(cfh, getReadOptions(null))) {
|
|
|
|
it.seekToLast();
|
|
|
|
if (it.isValid()) {
|
|
|
|
writeBatch.deleteRange(cfh, range.getMin(), it.key());
|
|
|
|
writeBatch.delete(cfh, it.key());
|
|
|
|
}
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2021-02-02 00:09:46 +01:00
|
|
|
return null;
|
|
|
|
})
|
|
|
|
.subscribeOn(dbScheduler)
|
|
|
|
.thenMany(entries)
|
|
|
|
.flatMap(newEntry -> putEntryToWriteBatch(newEntry, getOldValues, writeBatch))
|
|
|
|
.concatWith(Mono.<Entry<byte[], byte[]>>fromCallable(() -> {
|
|
|
|
synchronized (writeBatch) {
|
|
|
|
writeBatch.writeToDbAndClose();
|
|
|
|
writeBatch.close();
|
|
|
|
}
|
|
|
|
return null;
|
|
|
|
}).subscribeOn(dbScheduler))
|
|
|
|
.doFinally(signalType -> {
|
|
|
|
synchronized (writeBatch) {
|
|
|
|
writeBatch.close();
|
|
|
|
}
|
|
|
|
}))
|
|
|
|
.onErrorMap(IOException::new);
|
|
|
|
}
|
|
|
|
});
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
public Mono<Void> clear() {
|
|
|
|
return Mono
|
|
|
|
.<Void>fromCallable(() -> {
|
|
|
|
try (RocksIterator iter = db.newIterator(cfh); CappedWriteBatch writeBatch = new CappedWriteBatch(db,
|
|
|
|
CAPPED_WRITE_BATCH_CAP,
|
|
|
|
RESERVED_WRITE_BATCH_SIZE,
|
|
|
|
MAX_WRITE_BATCH_SIZE,
|
|
|
|
BATCH_WRITE_OPTIONS
|
|
|
|
)) {
|
2020-12-07 22:15:18 +01:00
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
iter.seekToFirst();
|
2020-12-07 22:15:18 +01:00
|
|
|
|
|
|
|
while (iter.isValid()) {
|
|
|
|
writeBatch.delete(cfh, iter.key());
|
|
|
|
|
|
|
|
iter.next();
|
|
|
|
}
|
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
writeBatch.writeToDbAndClose();
|
2020-12-07 22:15:18 +01:00
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
// Compact range
|
|
|
|
db.compactRange(cfh);
|
2020-12-07 22:15:18 +01:00
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
db.flush(new FlushOptions().setWaitForFlush(true).setAllowWriteStall(true), cfh);
|
|
|
|
db.flushWal(true);
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
2021-01-30 00:24:55 +01:00
|
|
|
return null;
|
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
2020-12-07 22:15:18 +01:00
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
2021-01-30 00:24:55 +01:00
|
|
|
public Mono<Long> sizeRange(@Nullable LLSnapshot snapshot, LLRange range, boolean fast) {
|
|
|
|
return Mono
|
|
|
|
.defer(() -> {
|
|
|
|
if (range.isAll()) {
|
|
|
|
return Mono
|
|
|
|
.fromCallable(() -> fast ? fastSizeAll(snapshot) : exactSizeAll(snapshot))
|
|
|
|
.onErrorMap(IOException::new)
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
2021-01-30 00:24:55 +01:00
|
|
|
} else {
|
|
|
|
return Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
try (var iter = db.newIterator(cfh, resolveSnapshot(snapshot))) {
|
|
|
|
if (range.hasMin()) {
|
|
|
|
iter.seek(range.getMin());
|
|
|
|
} else {
|
|
|
|
iter.seekToFirst();
|
|
|
|
}
|
|
|
|
long i = 0;
|
|
|
|
while (iter.isValid()) {
|
|
|
|
if (range.hasMax()) {
|
|
|
|
byte[] key1 = iter.key();
|
|
|
|
if (Arrays.compareUnsigned(key1, range.getMax()) > 0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
iter.next();
|
|
|
|
i++;
|
|
|
|
}
|
|
|
|
return i;
|
|
|
|
}
|
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
2021-01-30 00:24:55 +01:00
|
|
|
}
|
|
|
|
});
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
private long fastSizeAll(@Nullable LLSnapshot snapshot) {
|
2020-12-07 22:15:18 +01:00
|
|
|
var rocksdbSnapshot = resolveSnapshot(snapshot);
|
|
|
|
if (USE_CURRENT_FASTSIZE_FOR_OLD_SNAPSHOTS || rocksdbSnapshot.snapshot() == null) {
|
|
|
|
try {
|
|
|
|
return db.getLongProperty(cfh, "rocksdb.estimate-num-keys");
|
|
|
|
} catch (RocksDBException e) {
|
|
|
|
e.printStackTrace();
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
long count = 0;
|
|
|
|
try (RocksIterator iter = db.newIterator(cfh, rocksdbSnapshot)) {
|
|
|
|
iter.seekToFirst();
|
|
|
|
// If it's a fast size of a snapshot, count only up to 1000 elements
|
|
|
|
while (iter.isValid() && count < 1000) {
|
|
|
|
count++;
|
|
|
|
iter.next();
|
|
|
|
}
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-30 00:24:55 +01:00
|
|
|
private long exactSizeAll(@Nullable LLSnapshot snapshot) {
|
2020-12-07 22:15:18 +01:00
|
|
|
long count = 0;
|
|
|
|
try (RocksIterator iter = db.newIterator(cfh, resolveSnapshot(snapshot))) {
|
|
|
|
iter.seekToFirst();
|
|
|
|
while (iter.isValid()) {
|
|
|
|
count++;
|
|
|
|
iter.next();
|
|
|
|
}
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
2021-01-30 00:24:55 +01:00
|
|
|
public Mono<Entry<byte[], byte[]>> removeOne(LLRange range) {
|
|
|
|
return Mono
|
|
|
|
.fromCallable(() -> {
|
|
|
|
try (RocksIterator iter = db.newIterator(cfh)) {
|
|
|
|
if (range.hasMin()) {
|
|
|
|
iter.seek(range.getMin());
|
|
|
|
} else {
|
|
|
|
iter.seekToFirst();
|
|
|
|
}
|
|
|
|
if (!iter.isValid()) {
|
|
|
|
return null;
|
|
|
|
}
|
|
|
|
if (range.hasMax() && Arrays.compareUnsigned(iter.key(), range.getMax()) > 0) {
|
|
|
|
return null;
|
|
|
|
}
|
|
|
|
byte[] key = iter.key();
|
|
|
|
byte[] value = iter.value();
|
|
|
|
db.delete(cfh, key);
|
|
|
|
return Map.entry(key, value);
|
|
|
|
}
|
|
|
|
})
|
|
|
|
.onErrorMap(IOException::new)
|
2021-02-01 02:21:53 +01:00
|
|
|
.subscribeOn(dbScheduler);
|
2020-12-07 22:15:18 +01:00
|
|
|
}
|
|
|
|
}
|