Move CompactionFilter tests in db_test.cc to db_compaction_filter_test.cc
Summary: Move CompactionFilter tests in db_test.cc to db_compaction_filter_test.cc Test Plan: db_test db_compaction_filter_test Reviewers: igor, sdong, IslamAbdelRahman, anthony Reviewed By: anthony Subscribers: dhruba, leveldb Differential Revision: https://reviews.facebook.net/D42207
This commit is contained in:
parent
0936362a70
commit
c3f98bb89b
@ -255,6 +255,8 @@ set(TESTS
|
|||||||
db/cuckoo_table_db_test.cc
|
db/cuckoo_table_db_test.cc
|
||||||
db/db_iter_test.cc
|
db/db_iter_test.cc
|
||||||
db/db_test.cc
|
db/db_test.cc
|
||||||
|
db/db_compaction_filter_test.cc
|
||||||
|
db/db_dynamic_level_test.cc
|
||||||
db/dbformat_test.cc
|
db/dbformat_test.cc
|
||||||
db/deletefile_test.cc
|
db/deletefile_test.cc
|
||||||
db/fault_injection_test.cc
|
db/fault_injection_test.cc
|
||||||
|
3
Makefile
3
Makefile
@ -676,6 +676,9 @@ slice_transform_test: util/slice_transform_test.o $(LIBOBJECTS) $(TESTHARNESS)
|
|||||||
db_test: db/db_test.o util/db_test_util.o $(LIBOBJECTS) $(TESTHARNESS)
|
db_test: db/db_test.o util/db_test_util.o $(LIBOBJECTS) $(TESTHARNESS)
|
||||||
$(AM_LINK)
|
$(AM_LINK)
|
||||||
|
|
||||||
|
db_compaction_filter_test: db/db_compaction_filter_test.o util/db_test_util.o $(LIBOBJECTS) $(TESTHARNESS)
|
||||||
|
$(AM_LINK)
|
||||||
|
|
||||||
db_dynamic_level_test: db/db_dynamic_level_test.o util/db_test_util.o $(LIBOBJECTS) $(TESTHARNESS)
|
db_dynamic_level_test: db/db_dynamic_level_test.o util/db_test_util.o $(LIBOBJECTS) $(TESTHARNESS)
|
||||||
$(AM_LINK)
|
$(AM_LINK)
|
||||||
|
|
||||||
|
853
db/db_compaction_filter_test.cc
Normal file
853
db/db_compaction_filter_test.cc
Normal file
@ -0,0 +1,853 @@
|
|||||||
|
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
||||||
|
// This source code is licensed under the BSD-style license found in the
|
||||||
|
// LICENSE file in the root directory of this source tree. An additional grant
|
||||||
|
// of patent rights can be found in the PATENTS file in the same directory.
|
||||||
|
//
|
||||||
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style license that can be
|
||||||
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
||||||
|
|
||||||
|
#include "port/stack_trace.h"
|
||||||
|
#include "util/db_test_util.h"
|
||||||
|
|
||||||
|
namespace rocksdb {
|
||||||
|
|
||||||
|
static int cfilter_count = 0;
|
||||||
|
|
||||||
|
// This is a static filter used for filtering
|
||||||
|
// kvs during the compaction process.
|
||||||
|
static std::string NEW_VALUE = "NewValue";
|
||||||
|
|
||||||
|
class DBTestCompactionFilter : public DBTestBase {
|
||||||
|
public:
|
||||||
|
DBTestCompactionFilter() : DBTestBase("/db_compaction_filter_test") {}
|
||||||
|
};
|
||||||
|
|
||||||
|
class KeepFilter : public CompactionFilter {
|
||||||
|
public:
|
||||||
|
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
||||||
|
std::string* new_value, bool* value_changed) const
|
||||||
|
override {
|
||||||
|
cfilter_count++;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "KeepFilter"; }
|
||||||
|
};
|
||||||
|
|
||||||
|
class DeleteFilter : public CompactionFilter {
|
||||||
|
public:
|
||||||
|
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
||||||
|
std::string* new_value, bool* value_changed) const
|
||||||
|
override {
|
||||||
|
cfilter_count++;
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "DeleteFilter"; }
|
||||||
|
};
|
||||||
|
|
||||||
|
class DelayFilter : public CompactionFilter {
|
||||||
|
public:
|
||||||
|
explicit DelayFilter(DBTestBase* d) : db_test(d) {}
|
||||||
|
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
||||||
|
std::string* new_value,
|
||||||
|
bool* value_changed) const override {
|
||||||
|
db_test->env_->addon_time_.fetch_add(1000);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "DelayFilter"; }
|
||||||
|
|
||||||
|
private:
|
||||||
|
DBTestBase* db_test;
|
||||||
|
};
|
||||||
|
|
||||||
|
class ConditionalFilter : public CompactionFilter {
|
||||||
|
public:
|
||||||
|
explicit ConditionalFilter(const std::string* filtered_value)
|
||||||
|
: filtered_value_(filtered_value) {}
|
||||||
|
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
||||||
|
std::string* new_value,
|
||||||
|
bool* value_changed) const override {
|
||||||
|
return value.ToString() == *filtered_value_;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "ConditionalFilter"; }
|
||||||
|
|
||||||
|
private:
|
||||||
|
const std::string* filtered_value_;
|
||||||
|
};
|
||||||
|
|
||||||
|
class ChangeFilter : public CompactionFilter {
|
||||||
|
public:
|
||||||
|
explicit ChangeFilter() {}
|
||||||
|
|
||||||
|
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
||||||
|
std::string* new_value, bool* value_changed) const
|
||||||
|
override {
|
||||||
|
assert(new_value != nullptr);
|
||||||
|
*new_value = NEW_VALUE;
|
||||||
|
*value_changed = true;
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "ChangeFilter"; }
|
||||||
|
};
|
||||||
|
|
||||||
|
class KeepFilterFactory : public CompactionFilterFactory {
|
||||||
|
public:
|
||||||
|
explicit KeepFilterFactory(bool check_context = false)
|
||||||
|
: check_context_(check_context) {}
|
||||||
|
|
||||||
|
virtual std::unique_ptr<CompactionFilter> CreateCompactionFilter(
|
||||||
|
const CompactionFilter::Context& context) override {
|
||||||
|
if (check_context_) {
|
||||||
|
EXPECT_EQ(expect_full_compaction_.load(), context.is_full_compaction);
|
||||||
|
EXPECT_EQ(expect_manual_compaction_.load(), context.is_manual_compaction);
|
||||||
|
}
|
||||||
|
return std::unique_ptr<CompactionFilter>(new KeepFilter());
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "KeepFilterFactory"; }
|
||||||
|
bool check_context_;
|
||||||
|
std::atomic_bool expect_full_compaction_;
|
||||||
|
std::atomic_bool expect_manual_compaction_;
|
||||||
|
};
|
||||||
|
|
||||||
|
class DeleteFilterFactory : public CompactionFilterFactory {
|
||||||
|
public:
|
||||||
|
virtual std::unique_ptr<CompactionFilter> CreateCompactionFilter(
|
||||||
|
const CompactionFilter::Context& context) override {
|
||||||
|
if (context.is_manual_compaction) {
|
||||||
|
return std::unique_ptr<CompactionFilter>(new DeleteFilter());
|
||||||
|
} else {
|
||||||
|
return std::unique_ptr<CompactionFilter>(nullptr);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "DeleteFilterFactory"; }
|
||||||
|
};
|
||||||
|
|
||||||
|
class DelayFilterFactory : public CompactionFilterFactory {
|
||||||
|
public:
|
||||||
|
explicit DelayFilterFactory(DBTestBase* d) : db_test(d) {}
|
||||||
|
virtual std::unique_ptr<CompactionFilter> CreateCompactionFilter(
|
||||||
|
const CompactionFilter::Context& context) override {
|
||||||
|
return std::unique_ptr<CompactionFilter>(new DelayFilter(db_test));
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "DelayFilterFactory"; }
|
||||||
|
|
||||||
|
private:
|
||||||
|
DBTestBase* db_test;
|
||||||
|
};
|
||||||
|
|
||||||
|
class ConditionalFilterFactory : public CompactionFilterFactory {
|
||||||
|
public:
|
||||||
|
explicit ConditionalFilterFactory(const Slice& filtered_value)
|
||||||
|
: filtered_value_(filtered_value.ToString()) {}
|
||||||
|
|
||||||
|
virtual std::unique_ptr<CompactionFilter> CreateCompactionFilter(
|
||||||
|
const CompactionFilter::Context& context) override {
|
||||||
|
return std::unique_ptr<CompactionFilter>(
|
||||||
|
new ConditionalFilter(&filtered_value_));
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override {
|
||||||
|
return "ConditionalFilterFactory";
|
||||||
|
}
|
||||||
|
|
||||||
|
private:
|
||||||
|
std::string filtered_value_;
|
||||||
|
};
|
||||||
|
|
||||||
|
class ChangeFilterFactory : public CompactionFilterFactory {
|
||||||
|
public:
|
||||||
|
explicit ChangeFilterFactory() {}
|
||||||
|
|
||||||
|
virtual std::unique_ptr<CompactionFilter> CreateCompactionFilter(
|
||||||
|
const CompactionFilter::Context& context) override {
|
||||||
|
return std::unique_ptr<CompactionFilter>(new ChangeFilter());
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override { return "ChangeFilterFactory"; }
|
||||||
|
};
|
||||||
|
|
||||||
|
TEST_F(DBTestCompactionFilter, CompactionFilter) {
|
||||||
|
Options options = CurrentOptions();
|
||||||
|
options.max_open_files = -1;
|
||||||
|
options.num_levels = 3;
|
||||||
|
options.max_mem_compaction_level = 0;
|
||||||
|
options.compaction_filter_factory = std::make_shared<KeepFilterFactory>();
|
||||||
|
options = CurrentOptions(options);
|
||||||
|
CreateAndReopenWithCF({"pikachu"}, options);
|
||||||
|
|
||||||
|
// Write 100K keys, these are written to a few files in L0.
|
||||||
|
const std::string value(10, 'x');
|
||||||
|
for (int i = 0; i < 100000; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%010d", i);
|
||||||
|
Put(1, key, value);
|
||||||
|
}
|
||||||
|
ASSERT_OK(Flush(1));
|
||||||
|
|
||||||
|
// Push all files to the highest level L2. Verify that
|
||||||
|
// the compaction is each level invokes the filter for
|
||||||
|
// all the keys in that level.
|
||||||
|
cfilter_count = 0;
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
||||||
|
ASSERT_EQ(cfilter_count, 100000);
|
||||||
|
cfilter_count = 0;
|
||||||
|
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
||||||
|
ASSERT_EQ(cfilter_count, 100000);
|
||||||
|
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(0, 1), 0);
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(1, 1), 0);
|
||||||
|
ASSERT_NE(NumTableFilesAtLevel(2, 1), 0);
|
||||||
|
cfilter_count = 0;
|
||||||
|
|
||||||
|
// All the files are in the lowest level.
|
||||||
|
// Verify that all but the 100001st record
|
||||||
|
// has sequence number zero. The 100001st record
|
||||||
|
// is at the tip of this snapshot and cannot
|
||||||
|
// be zeroed out.
|
||||||
|
int count = 0;
|
||||||
|
int total = 0;
|
||||||
|
Arena arena;
|
||||||
|
{
|
||||||
|
ScopedArenaIterator iter(
|
||||||
|
dbfull()->TEST_NewInternalIterator(&arena, handles_[1]));
|
||||||
|
iter->SeekToFirst();
|
||||||
|
ASSERT_OK(iter->status());
|
||||||
|
while (iter->Valid()) {
|
||||||
|
ParsedInternalKey ikey(Slice(), 0, kTypeValue);
|
||||||
|
ikey.sequence = -1;
|
||||||
|
ASSERT_EQ(ParseInternalKey(iter->key(), &ikey), true);
|
||||||
|
total++;
|
||||||
|
if (ikey.sequence != 0) {
|
||||||
|
count++;
|
||||||
|
}
|
||||||
|
iter->Next();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ASSERT_EQ(total, 100000);
|
||||||
|
ASSERT_EQ(count, 1);
|
||||||
|
|
||||||
|
// overwrite all the 100K keys once again.
|
||||||
|
for (int i = 0; i < 100000; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%010d", i);
|
||||||
|
ASSERT_OK(Put(1, key, value));
|
||||||
|
}
|
||||||
|
ASSERT_OK(Flush(1));
|
||||||
|
|
||||||
|
// push all files to the highest level L2. This
|
||||||
|
// means that all keys should pass at least once
|
||||||
|
// via the compaction filter
|
||||||
|
cfilter_count = 0;
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
||||||
|
ASSERT_EQ(cfilter_count, 100000);
|
||||||
|
cfilter_count = 0;
|
||||||
|
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
||||||
|
ASSERT_EQ(cfilter_count, 100000);
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(0, 1), 0);
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(1, 1), 0);
|
||||||
|
ASSERT_NE(NumTableFilesAtLevel(2, 1), 0);
|
||||||
|
|
||||||
|
// create a new database with the compaction
|
||||||
|
// filter in such a way that it deletes all keys
|
||||||
|
options.compaction_filter_factory = std::make_shared<DeleteFilterFactory>();
|
||||||
|
options.create_if_missing = true;
|
||||||
|
DestroyAndReopen(options);
|
||||||
|
CreateAndReopenWithCF({"pikachu"}, options);
|
||||||
|
|
||||||
|
// write all the keys once again.
|
||||||
|
for (int i = 0; i < 100000; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%010d", i);
|
||||||
|
ASSERT_OK(Put(1, key, value));
|
||||||
|
}
|
||||||
|
ASSERT_OK(Flush(1));
|
||||||
|
ASSERT_NE(NumTableFilesAtLevel(0, 1), 0);
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(1, 1), 0);
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(2, 1), 0);
|
||||||
|
|
||||||
|
// Push all files to the highest level L2. This
|
||||||
|
// triggers the compaction filter to delete all keys,
|
||||||
|
// verify that at the end of the compaction process,
|
||||||
|
// nothing is left.
|
||||||
|
cfilter_count = 0;
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
||||||
|
ASSERT_EQ(cfilter_count, 100000);
|
||||||
|
cfilter_count = 0;
|
||||||
|
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
||||||
|
ASSERT_EQ(cfilter_count, 0);
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(0, 1), 0);
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(1, 1), 0);
|
||||||
|
|
||||||
|
{
|
||||||
|
// Scan the entire database to ensure that nothing is left
|
||||||
|
std::unique_ptr<Iterator> iter(
|
||||||
|
db_->NewIterator(ReadOptions(), handles_[1]));
|
||||||
|
iter->SeekToFirst();
|
||||||
|
count = 0;
|
||||||
|
while (iter->Valid()) {
|
||||||
|
count++;
|
||||||
|
iter->Next();
|
||||||
|
}
|
||||||
|
ASSERT_EQ(count, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
// The sequence number of the remaining record
|
||||||
|
// is not zeroed out even though it is at the
|
||||||
|
// level Lmax because this record is at the tip
|
||||||
|
count = 0;
|
||||||
|
{
|
||||||
|
ScopedArenaIterator iter(
|
||||||
|
dbfull()->TEST_NewInternalIterator(&arena, handles_[1]));
|
||||||
|
iter->SeekToFirst();
|
||||||
|
ASSERT_OK(iter->status());
|
||||||
|
while (iter->Valid()) {
|
||||||
|
ParsedInternalKey ikey(Slice(), 0, kTypeValue);
|
||||||
|
ASSERT_EQ(ParseInternalKey(iter->key(), &ikey), true);
|
||||||
|
ASSERT_NE(ikey.sequence, (unsigned)0);
|
||||||
|
count++;
|
||||||
|
iter->Next();
|
||||||
|
}
|
||||||
|
ASSERT_EQ(count, 0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Tests the edge case where compaction does not produce any output -- all
|
||||||
|
// entries are deleted. The compaction should create bunch of 'DeleteFile'
|
||||||
|
// entries in VersionEdit, but none of the 'AddFile's.
|
||||||
|
TEST_F(DBTestCompactionFilter, CompactionFilterDeletesAll) {
|
||||||
|
Options options;
|
||||||
|
options.compaction_filter_factory = std::make_shared<DeleteFilterFactory>();
|
||||||
|
options.disable_auto_compactions = true;
|
||||||
|
options.create_if_missing = true;
|
||||||
|
options = CurrentOptions(options);
|
||||||
|
DestroyAndReopen(options);
|
||||||
|
|
||||||
|
// put some data
|
||||||
|
for (int table = 0; table < 4; ++table) {
|
||||||
|
for (int i = 0; i < 10 + table; ++i) {
|
||||||
|
Put(ToString(table * 100 + i), "val");
|
||||||
|
}
|
||||||
|
Flush();
|
||||||
|
}
|
||||||
|
|
||||||
|
// this will produce empty file (delete compaction filter)
|
||||||
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
||||||
|
ASSERT_EQ(0U, CountLiveFiles());
|
||||||
|
|
||||||
|
Reopen(options);
|
||||||
|
|
||||||
|
Iterator* itr = db_->NewIterator(ReadOptions());
|
||||||
|
itr->SeekToFirst();
|
||||||
|
// empty db
|
||||||
|
ASSERT_TRUE(!itr->Valid());
|
||||||
|
|
||||||
|
delete itr;
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST_F(DBTestCompactionFilter, CompactionFilterWithValueChange) {
|
||||||
|
do {
|
||||||
|
Options options;
|
||||||
|
options.num_levels = 3;
|
||||||
|
options.max_mem_compaction_level = 0;
|
||||||
|
options.compaction_filter_factory =
|
||||||
|
std::make_shared<ChangeFilterFactory>();
|
||||||
|
options = CurrentOptions(options);
|
||||||
|
CreateAndReopenWithCF({"pikachu"}, options);
|
||||||
|
|
||||||
|
// Write 100K+1 keys, these are written to a few files
|
||||||
|
// in L0. We do this so that the current snapshot points
|
||||||
|
// to the 100001 key.The compaction filter is not invoked
|
||||||
|
// on keys that are visible via a snapshot because we
|
||||||
|
// anyways cannot delete it.
|
||||||
|
const std::string value(10, 'x');
|
||||||
|
for (int i = 0; i < 100001; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%010d", i);
|
||||||
|
Put(1, key, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
// push all files to lower levels
|
||||||
|
ASSERT_OK(Flush(1));
|
||||||
|
if (option_config_ != kUniversalCompactionMultiLevel) {
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
||||||
|
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
||||||
|
} else {
|
||||||
|
dbfull()->CompactRange(CompactRangeOptions(), handles_[1], nullptr,
|
||||||
|
nullptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
// re-write all data again
|
||||||
|
for (int i = 0; i < 100001; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%010d", i);
|
||||||
|
Put(1, key, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
// push all files to lower levels. This should
|
||||||
|
// invoke the compaction filter for all 100000 keys.
|
||||||
|
ASSERT_OK(Flush(1));
|
||||||
|
if (option_config_ != kUniversalCompactionMultiLevel) {
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
||||||
|
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
||||||
|
} else {
|
||||||
|
dbfull()->CompactRange(CompactRangeOptions(), handles_[1], nullptr,
|
||||||
|
nullptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
// verify that all keys now have the new value that
|
||||||
|
// was set by the compaction process.
|
||||||
|
for (int i = 0; i < 100001; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%010d", i);
|
||||||
|
std::string newvalue = Get(1, key);
|
||||||
|
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
||||||
|
}
|
||||||
|
} while (ChangeCompactOptions());
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST_F(DBTestCompactionFilter, CompactionFilterWithMergeOperator) {
|
||||||
|
std::string one, two, three, four;
|
||||||
|
PutFixed64(&one, 1);
|
||||||
|
PutFixed64(&two, 2);
|
||||||
|
PutFixed64(&three, 3);
|
||||||
|
PutFixed64(&four, 4);
|
||||||
|
|
||||||
|
Options options;
|
||||||
|
options = CurrentOptions(options);
|
||||||
|
options.create_if_missing = true;
|
||||||
|
options.merge_operator = MergeOperators::CreateUInt64AddOperator();
|
||||||
|
options.num_levels = 3;
|
||||||
|
options.max_mem_compaction_level = 0;
|
||||||
|
// Filter out keys with value is 2.
|
||||||
|
options.compaction_filter_factory =
|
||||||
|
std::make_shared<ConditionalFilterFactory>(two);
|
||||||
|
DestroyAndReopen(options);
|
||||||
|
|
||||||
|
// In the same compaction, a value type needs to be deleted based on
|
||||||
|
// compaction filter, and there is a merge type for the key. compaction
|
||||||
|
// filter result is ignored.
|
||||||
|
ASSERT_OK(db_->Put(WriteOptions(), "foo", two));
|
||||||
|
ASSERT_OK(Flush());
|
||||||
|
ASSERT_OK(db_->Merge(WriteOptions(), "foo", one));
|
||||||
|
ASSERT_OK(Flush());
|
||||||
|
std::string newvalue = Get("foo");
|
||||||
|
ASSERT_EQ(newvalue, three);
|
||||||
|
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
||||||
|
newvalue = Get("foo");
|
||||||
|
ASSERT_EQ(newvalue, three);
|
||||||
|
|
||||||
|
// value key can be deleted based on compaction filter, leaving only
|
||||||
|
// merge keys.
|
||||||
|
ASSERT_OK(db_->Put(WriteOptions(), "bar", two));
|
||||||
|
ASSERT_OK(Flush());
|
||||||
|
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
||||||
|
newvalue = Get("bar");
|
||||||
|
ASSERT_EQ("NOT_FOUND", newvalue);
|
||||||
|
ASSERT_OK(db_->Merge(WriteOptions(), "bar", two));
|
||||||
|
ASSERT_OK(Flush());
|
||||||
|
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
||||||
|
newvalue = Get("bar");
|
||||||
|
ASSERT_EQ(two, two);
|
||||||
|
|
||||||
|
// Compaction filter never applies to merge keys.
|
||||||
|
ASSERT_OK(db_->Put(WriteOptions(), "foobar", one));
|
||||||
|
ASSERT_OK(Flush());
|
||||||
|
ASSERT_OK(db_->Merge(WriteOptions(), "foobar", two));
|
||||||
|
ASSERT_OK(Flush());
|
||||||
|
newvalue = Get("foobar");
|
||||||
|
ASSERT_EQ(newvalue, three);
|
||||||
|
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
||||||
|
newvalue = Get("foobar");
|
||||||
|
ASSERT_EQ(newvalue, three);
|
||||||
|
|
||||||
|
// In the same compaction, both of value type and merge type keys need to be
|
||||||
|
// deleted based on compaction filter, and there is a merge type for the key.
|
||||||
|
// For both keys, compaction filter results are ignored.
|
||||||
|
ASSERT_OK(db_->Put(WriteOptions(), "barfoo", two));
|
||||||
|
ASSERT_OK(Flush());
|
||||||
|
ASSERT_OK(db_->Merge(WriteOptions(), "barfoo", two));
|
||||||
|
ASSERT_OK(Flush());
|
||||||
|
newvalue = Get("barfoo");
|
||||||
|
ASSERT_EQ(newvalue, four);
|
||||||
|
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
||||||
|
newvalue = Get("barfoo");
|
||||||
|
ASSERT_EQ(newvalue, four);
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST_F(DBTestCompactionFilter, CompactionFilterContextManual) {
|
||||||
|
KeepFilterFactory* filter = new KeepFilterFactory();
|
||||||
|
|
||||||
|
Options options = CurrentOptions();
|
||||||
|
options.compaction_style = kCompactionStyleUniversal;
|
||||||
|
options.compaction_filter_factory.reset(filter);
|
||||||
|
options.compression = kNoCompression;
|
||||||
|
options.level0_file_num_compaction_trigger = 8;
|
||||||
|
Reopen(options);
|
||||||
|
int num_keys_per_file = 400;
|
||||||
|
for (int j = 0; j < 3; j++) {
|
||||||
|
// Write several keys.
|
||||||
|
const std::string value(10, 'x');
|
||||||
|
for (int i = 0; i < num_keys_per_file; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%08d%02d", i, j);
|
||||||
|
Put(key, value);
|
||||||
|
}
|
||||||
|
dbfull()->TEST_FlushMemTable();
|
||||||
|
// Make sure next file is much smaller so automatic compaction will not
|
||||||
|
// be triggered.
|
||||||
|
num_keys_per_file /= 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Force a manual compaction
|
||||||
|
cfilter_count = 0;
|
||||||
|
filter->expect_manual_compaction_.store(true);
|
||||||
|
filter->expect_full_compaction_.store(false); // Manual compaction always
|
||||||
|
// set this flag.
|
||||||
|
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
||||||
|
ASSERT_EQ(cfilter_count, 700);
|
||||||
|
ASSERT_EQ(NumSortedRuns(0), 1);
|
||||||
|
|
||||||
|
// Verify total number of keys is correct after manual compaction.
|
||||||
|
{
|
||||||
|
int count = 0;
|
||||||
|
int total = 0;
|
||||||
|
Arena arena;
|
||||||
|
ScopedArenaIterator iter(dbfull()->TEST_NewInternalIterator(&arena));
|
||||||
|
iter->SeekToFirst();
|
||||||
|
ASSERT_OK(iter->status());
|
||||||
|
while (iter->Valid()) {
|
||||||
|
ParsedInternalKey ikey(Slice(), 0, kTypeValue);
|
||||||
|
ikey.sequence = -1;
|
||||||
|
ASSERT_EQ(ParseInternalKey(iter->key(), &ikey), true);
|
||||||
|
total++;
|
||||||
|
if (ikey.sequence != 0) {
|
||||||
|
count++;
|
||||||
|
}
|
||||||
|
iter->Next();
|
||||||
|
}
|
||||||
|
ASSERT_EQ(total, 700);
|
||||||
|
ASSERT_EQ(count, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
class KeepFilterV2 : public CompactionFilterV2 {
|
||||||
|
public:
|
||||||
|
virtual std::vector<bool> Filter(int level,
|
||||||
|
const SliceVector& keys,
|
||||||
|
const SliceVector& existing_values,
|
||||||
|
std::vector<std::string>* new_values,
|
||||||
|
std::vector<bool>* values_changed)
|
||||||
|
const override {
|
||||||
|
cfilter_count++;
|
||||||
|
std::vector<bool> ret;
|
||||||
|
new_values->clear();
|
||||||
|
values_changed->clear();
|
||||||
|
for (unsigned int i = 0; i < keys.size(); ++i) {
|
||||||
|
values_changed->push_back(false);
|
||||||
|
ret.push_back(false);
|
||||||
|
}
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override {
|
||||||
|
return "KeepFilterV2";
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
class DeleteFilterV2 : public CompactionFilterV2 {
|
||||||
|
public:
|
||||||
|
virtual std::vector<bool> Filter(int level,
|
||||||
|
const SliceVector& keys,
|
||||||
|
const SliceVector& existing_values,
|
||||||
|
std::vector<std::string>* new_values,
|
||||||
|
std::vector<bool>* values_changed)
|
||||||
|
const override {
|
||||||
|
cfilter_count++;
|
||||||
|
new_values->clear();
|
||||||
|
values_changed->clear();
|
||||||
|
std::vector<bool> ret;
|
||||||
|
for (unsigned int i = 0; i < keys.size(); ++i) {
|
||||||
|
values_changed->push_back(false);
|
||||||
|
ret.push_back(true);
|
||||||
|
}
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override {
|
||||||
|
return "DeleteFilterV2";
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
class ChangeFilterV2 : public CompactionFilterV2 {
|
||||||
|
public:
|
||||||
|
virtual std::vector<bool> Filter(int level,
|
||||||
|
const SliceVector& keys,
|
||||||
|
const SliceVector& existing_values,
|
||||||
|
std::vector<std::string>* new_values,
|
||||||
|
std::vector<bool>* values_changed)
|
||||||
|
const override {
|
||||||
|
std::vector<bool> ret;
|
||||||
|
new_values->clear();
|
||||||
|
values_changed->clear();
|
||||||
|
for (unsigned int i = 0; i < keys.size(); ++i) {
|
||||||
|
values_changed->push_back(true);
|
||||||
|
new_values->push_back(NEW_VALUE);
|
||||||
|
ret.push_back(false);
|
||||||
|
}
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override {
|
||||||
|
return "ChangeFilterV2";
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
class KeepFilterFactoryV2 : public CompactionFilterFactoryV2 {
|
||||||
|
public:
|
||||||
|
explicit KeepFilterFactoryV2(const SliceTransform* prefix_extractor)
|
||||||
|
: CompactionFilterFactoryV2(prefix_extractor) { }
|
||||||
|
|
||||||
|
virtual std::unique_ptr<CompactionFilterV2>
|
||||||
|
CreateCompactionFilterV2(
|
||||||
|
const CompactionFilterContext& context) override {
|
||||||
|
return std::unique_ptr<CompactionFilterV2>(new KeepFilterV2());
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override {
|
||||||
|
return "KeepFilterFactoryV2";
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
class DeleteFilterFactoryV2 : public CompactionFilterFactoryV2 {
|
||||||
|
public:
|
||||||
|
explicit DeleteFilterFactoryV2(const SliceTransform* prefix_extractor)
|
||||||
|
: CompactionFilterFactoryV2(prefix_extractor) { }
|
||||||
|
|
||||||
|
virtual std::unique_ptr<CompactionFilterV2>
|
||||||
|
CreateCompactionFilterV2(
|
||||||
|
const CompactionFilterContext& context) override {
|
||||||
|
return std::unique_ptr<CompactionFilterV2>(new DeleteFilterV2());
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override {
|
||||||
|
return "DeleteFilterFactoryV2";
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
class ChangeFilterFactoryV2 : public CompactionFilterFactoryV2 {
|
||||||
|
public:
|
||||||
|
explicit ChangeFilterFactoryV2(const SliceTransform* prefix_extractor)
|
||||||
|
: CompactionFilterFactoryV2(prefix_extractor) { }
|
||||||
|
|
||||||
|
virtual std::unique_ptr<CompactionFilterV2>
|
||||||
|
CreateCompactionFilterV2(
|
||||||
|
const CompactionFilterContext& context) override {
|
||||||
|
return std::unique_ptr<CompactionFilterV2>(new ChangeFilterV2());
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual const char* Name() const override {
|
||||||
|
return "ChangeFilterFactoryV2";
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
TEST_F(DBTestCompactionFilter, CompactionFilterV2) {
|
||||||
|
Options options = CurrentOptions();
|
||||||
|
options.num_levels = 3;
|
||||||
|
options.max_mem_compaction_level = 0;
|
||||||
|
// extract prefix
|
||||||
|
std::unique_ptr<const SliceTransform> prefix_extractor;
|
||||||
|
prefix_extractor.reset(NewFixedPrefixTransform(8));
|
||||||
|
|
||||||
|
options.compaction_filter_factory_v2
|
||||||
|
= std::make_shared<KeepFilterFactoryV2>(prefix_extractor.get());
|
||||||
|
// In a testing environment, we can only flush the application
|
||||||
|
// compaction filter buffer using universal compaction
|
||||||
|
option_config_ = kUniversalCompaction;
|
||||||
|
options.compaction_style = (rocksdb::CompactionStyle)1;
|
||||||
|
Reopen(options);
|
||||||
|
|
||||||
|
// Write 100K keys, these are written to a few files in L0.
|
||||||
|
const std::string value(10, 'x');
|
||||||
|
for (int i = 0; i < 100000; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%08d%010d", i , i);
|
||||||
|
Put(key, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
dbfull()->TEST_FlushMemTable();
|
||||||
|
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
||||||
|
dbfull()->TEST_CompactRange(1, nullptr, nullptr);
|
||||||
|
|
||||||
|
ASSERT_EQ(NumSortedRuns(0), 1);
|
||||||
|
|
||||||
|
// All the files are in the lowest level.
|
||||||
|
int count = 0;
|
||||||
|
int total = 0;
|
||||||
|
{
|
||||||
|
Arena arena;
|
||||||
|
ScopedArenaIterator iter(dbfull()->TEST_NewInternalIterator(&arena));
|
||||||
|
iter->SeekToFirst();
|
||||||
|
ASSERT_OK(iter->status());
|
||||||
|
while (iter->Valid()) {
|
||||||
|
ParsedInternalKey ikey(Slice(), 0, kTypeValue);
|
||||||
|
ikey.sequence = -1;
|
||||||
|
ASSERT_EQ(ParseInternalKey(iter->key(), &ikey), true);
|
||||||
|
total++;
|
||||||
|
if (ikey.sequence != 0) {
|
||||||
|
count++;
|
||||||
|
}
|
||||||
|
iter->Next();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
ASSERT_EQ(total, 100000);
|
||||||
|
// 1 snapshot only. Since we are using universal compacton,
|
||||||
|
// the sequence no is cleared for better compression
|
||||||
|
ASSERT_EQ(count, 1);
|
||||||
|
|
||||||
|
// create a new database with the compaction
|
||||||
|
// filter in such a way that it deletes all keys
|
||||||
|
options.compaction_filter_factory_v2 =
|
||||||
|
std::make_shared<DeleteFilterFactoryV2>(prefix_extractor.get());
|
||||||
|
options.create_if_missing = true;
|
||||||
|
DestroyAndReopen(options);
|
||||||
|
|
||||||
|
// write all the keys once again.
|
||||||
|
for (int i = 0; i < 100000; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%08d%010d", i, i);
|
||||||
|
Put(key, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
dbfull()->TEST_FlushMemTable();
|
||||||
|
ASSERT_NE(NumTableFilesAtLevel(0), 0);
|
||||||
|
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
||||||
|
dbfull()->TEST_CompactRange(1, nullptr, nullptr);
|
||||||
|
ASSERT_EQ(NumTableFilesAtLevel(1), 0);
|
||||||
|
|
||||||
|
// Scan the entire database to ensure that nothing is left
|
||||||
|
Iterator* iter = db_->NewIterator(ReadOptions());
|
||||||
|
iter->SeekToFirst();
|
||||||
|
count = 0;
|
||||||
|
while (iter->Valid()) {
|
||||||
|
count++;
|
||||||
|
iter->Next();
|
||||||
|
}
|
||||||
|
|
||||||
|
ASSERT_EQ(count, 0);
|
||||||
|
delete iter;
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST_F(DBTestCompactionFilter, CompactionFilterV2WithValueChange) {
|
||||||
|
Options options = CurrentOptions();
|
||||||
|
options.num_levels = 3;
|
||||||
|
options.max_mem_compaction_level = 0;
|
||||||
|
std::unique_ptr<const SliceTransform> prefix_extractor;
|
||||||
|
prefix_extractor.reset(NewFixedPrefixTransform(8));
|
||||||
|
options.compaction_filter_factory_v2 =
|
||||||
|
std::make_shared<ChangeFilterFactoryV2>(prefix_extractor.get());
|
||||||
|
// In a testing environment, we can only flush the application
|
||||||
|
// compaction filter buffer using universal compaction
|
||||||
|
option_config_ = kUniversalCompaction;
|
||||||
|
options.compaction_style = (rocksdb::CompactionStyle)1;
|
||||||
|
options = CurrentOptions(options);
|
||||||
|
Reopen(options);
|
||||||
|
|
||||||
|
// Write 100K+1 keys, these are written to a few files
|
||||||
|
// in L0. We do this so that the current snapshot points
|
||||||
|
// to the 100001 key.The compaction filter is not invoked
|
||||||
|
// on keys that are visible via a snapshot because we
|
||||||
|
// anyways cannot delete it.
|
||||||
|
const std::string value(10, 'x');
|
||||||
|
for (int i = 0; i < 100001; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%08d%010d", i, i);
|
||||||
|
Put(key, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
// push all files to lower levels
|
||||||
|
dbfull()->TEST_FlushMemTable();
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
||||||
|
dbfull()->TEST_CompactRange(1, nullptr, nullptr);
|
||||||
|
|
||||||
|
// verify that all keys now have the new value that
|
||||||
|
// was set by the compaction process.
|
||||||
|
for (int i = 0; i < 100001; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "B%08d%010d", i, i);
|
||||||
|
std::string newvalue = Get(key);
|
||||||
|
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST_F(DBTestCompactionFilter, CompactionFilterV2NULLPrefix) {
|
||||||
|
Options options = CurrentOptions();
|
||||||
|
options.num_levels = 3;
|
||||||
|
options.max_mem_compaction_level = 0;
|
||||||
|
std::unique_ptr<const SliceTransform> prefix_extractor;
|
||||||
|
prefix_extractor.reset(NewFixedPrefixTransform(8));
|
||||||
|
options.compaction_filter_factory_v2 =
|
||||||
|
std::make_shared<ChangeFilterFactoryV2>(prefix_extractor.get());
|
||||||
|
// In a testing environment, we can only flush the application
|
||||||
|
// compaction filter buffer using universal compaction
|
||||||
|
option_config_ = kUniversalCompaction;
|
||||||
|
options.compaction_style = (rocksdb::CompactionStyle)1;
|
||||||
|
Reopen(options);
|
||||||
|
|
||||||
|
// Write 100K+1 keys, these are written to a few files
|
||||||
|
// in L0. We do this so that the current snapshot points
|
||||||
|
// to the 100001 key.The compaction filter is not invoked
|
||||||
|
// on keys that are visible via a snapshot because we
|
||||||
|
// anyways cannot delete it.
|
||||||
|
const std::string value(10, 'x');
|
||||||
|
char first_key[100];
|
||||||
|
snprintf(first_key, sizeof(first_key), "%s0000%010d", "NULL", 1);
|
||||||
|
Put(first_key, value);
|
||||||
|
for (int i = 1; i < 100000; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "%08d%010d", i, i);
|
||||||
|
Put(key, value);
|
||||||
|
}
|
||||||
|
|
||||||
|
char last_key[100];
|
||||||
|
snprintf(last_key, sizeof(last_key), "%s0000%010d", "NULL", 2);
|
||||||
|
Put(last_key, value);
|
||||||
|
|
||||||
|
// push all files to lower levels
|
||||||
|
dbfull()->TEST_FlushMemTable();
|
||||||
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
||||||
|
|
||||||
|
// verify that all keys now have the new value that
|
||||||
|
// was set by the compaction process.
|
||||||
|
std::string newvalue = Get(first_key);
|
||||||
|
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
||||||
|
newvalue = Get(last_key);
|
||||||
|
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
||||||
|
for (int i = 1; i < 100000; i++) {
|
||||||
|
char key[100];
|
||||||
|
snprintf(key, sizeof(key), "%08d%010d", i, i);
|
||||||
|
newvalue = Get(key);
|
||||||
|
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} // namespace rocksdb
|
||||||
|
|
||||||
|
int main(int argc, char** argv) {
|
||||||
|
#if !(defined NDEBUG) || !defined(OS_WIN)
|
||||||
|
rocksdb::port::InstallStackTraceHandler();
|
||||||
|
::testing::InitGoogleTest(&argc, argv);
|
||||||
|
return RUN_ALL_TESTS();
|
||||||
|
#else
|
||||||
|
return 0;
|
||||||
|
#endif
|
||||||
|
}
|
806
db/db_test.cc
806
db/db_test.cc
@ -2880,83 +2880,28 @@ TEST_F(DBTest, CompactionDeletionTriggerReopen) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// This is a static filter used for filtering
|
class DBTestUniversalCompactionBase
|
||||||
// kvs during the compaction process.
|
: public DBTest,
|
||||||
static int cfilter_count;
|
public ::testing::WithParamInterface<int> {
|
||||||
static std::string NEW_VALUE = "NewValue";
|
public:
|
||||||
|
virtual void SetUp() override { num_levels_ = GetParam(); }
|
||||||
|
int num_levels_;
|
||||||
|
};
|
||||||
|
|
||||||
|
class DBTestUniversalCompaction : public DBTestUniversalCompactionBase {};
|
||||||
|
|
||||||
|
namespace {
|
||||||
class KeepFilter : public CompactionFilter {
|
class KeepFilter : public CompactionFilter {
|
||||||
public:
|
public:
|
||||||
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
||||||
std::string* new_value, bool* value_changed) const
|
std::string* new_value, bool* value_changed) const
|
||||||
override {
|
override {
|
||||||
cfilter_count++;
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
virtual const char* Name() const override { return "KeepFilter"; }
|
virtual const char* Name() const override { return "KeepFilter"; }
|
||||||
};
|
};
|
||||||
|
|
||||||
class DeleteFilter : public CompactionFilter {
|
|
||||||
public:
|
|
||||||
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
|
||||||
std::string* new_value, bool* value_changed) const
|
|
||||||
override {
|
|
||||||
cfilter_count++;
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override { return "DeleteFilter"; }
|
|
||||||
};
|
|
||||||
|
|
||||||
class DelayFilter : public CompactionFilter {
|
|
||||||
public:
|
|
||||||
explicit DelayFilter(DBTest* d) : db_test(d) {}
|
|
||||||
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
|
||||||
std::string* new_value,
|
|
||||||
bool* value_changed) const override {
|
|
||||||
db_test->env_->addon_time_.fetch_add(1000);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override { return "DelayFilter"; }
|
|
||||||
|
|
||||||
private:
|
|
||||||
DBTest* db_test;
|
|
||||||
};
|
|
||||||
|
|
||||||
class ConditionalFilter : public CompactionFilter {
|
|
||||||
public:
|
|
||||||
explicit ConditionalFilter(const std::string* filtered_value)
|
|
||||||
: filtered_value_(filtered_value) {}
|
|
||||||
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
|
||||||
std::string* new_value,
|
|
||||||
bool* value_changed) const override {
|
|
||||||
return value.ToString() == *filtered_value_;
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override { return "ConditionalFilter"; }
|
|
||||||
|
|
||||||
private:
|
|
||||||
const std::string* filtered_value_;
|
|
||||||
};
|
|
||||||
|
|
||||||
class ChangeFilter : public CompactionFilter {
|
|
||||||
public:
|
|
||||||
explicit ChangeFilter() {}
|
|
||||||
|
|
||||||
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
|
||||||
std::string* new_value, bool* value_changed) const
|
|
||||||
override {
|
|
||||||
assert(new_value != nullptr);
|
|
||||||
*new_value = NEW_VALUE;
|
|
||||||
*value_changed = true;
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override { return "ChangeFilter"; }
|
|
||||||
};
|
|
||||||
|
|
||||||
class KeepFilterFactory : public CompactionFilterFactory {
|
class KeepFilterFactory : public CompactionFilterFactory {
|
||||||
public:
|
public:
|
||||||
explicit KeepFilterFactory(bool check_context = false)
|
explicit KeepFilterFactory(bool check_context = false)
|
||||||
@ -2977,18 +2922,20 @@ class KeepFilterFactory : public CompactionFilterFactory {
|
|||||||
std::atomic_bool expect_manual_compaction_;
|
std::atomic_bool expect_manual_compaction_;
|
||||||
};
|
};
|
||||||
|
|
||||||
class DeleteFilterFactory : public CompactionFilterFactory {
|
class DelayFilter : public CompactionFilter {
|
||||||
public:
|
public:
|
||||||
virtual std::unique_ptr<CompactionFilter> CreateCompactionFilter(
|
explicit DelayFilter(DBTest* d) : db_test(d) {}
|
||||||
const CompactionFilter::Context& context) override {
|
virtual bool Filter(int level, const Slice& key, const Slice& value,
|
||||||
if (context.is_manual_compaction) {
|
std::string* new_value,
|
||||||
return std::unique_ptr<CompactionFilter>(new DeleteFilter());
|
bool* value_changed) const override {
|
||||||
} else {
|
db_test->env_->addon_time_.fetch_add(1000);
|
||||||
return std::unique_ptr<CompactionFilter>(nullptr);
|
return true;
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
virtual const char* Name() const override { return "DeleteFilterFactory"; }
|
virtual const char* Name() const override { return "DelayFilter"; }
|
||||||
|
|
||||||
|
private:
|
||||||
|
DBTest* db_test;
|
||||||
};
|
};
|
||||||
|
|
||||||
class DelayFilterFactory : public CompactionFilterFactory {
|
class DelayFilterFactory : public CompactionFilterFactory {
|
||||||
@ -3004,47 +2951,7 @@ class DelayFilterFactory : public CompactionFilterFactory {
|
|||||||
private:
|
private:
|
||||||
DBTest* db_test;
|
DBTest* db_test;
|
||||||
};
|
};
|
||||||
|
} // namespace
|
||||||
class ConditionalFilterFactory : public CompactionFilterFactory {
|
|
||||||
public:
|
|
||||||
explicit ConditionalFilterFactory(const Slice& filtered_value)
|
|
||||||
: filtered_value_(filtered_value.ToString()) {}
|
|
||||||
|
|
||||||
virtual std::unique_ptr<CompactionFilter> CreateCompactionFilter(
|
|
||||||
const CompactionFilter::Context& context) override {
|
|
||||||
return std::unique_ptr<CompactionFilter>(
|
|
||||||
new ConditionalFilter(&filtered_value_));
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override {
|
|
||||||
return "ConditionalFilterFactory";
|
|
||||||
}
|
|
||||||
|
|
||||||
private:
|
|
||||||
std::string filtered_value_;
|
|
||||||
};
|
|
||||||
|
|
||||||
class ChangeFilterFactory : public CompactionFilterFactory {
|
|
||||||
public:
|
|
||||||
explicit ChangeFilterFactory() {}
|
|
||||||
|
|
||||||
virtual std::unique_ptr<CompactionFilter> CreateCompactionFilter(
|
|
||||||
const CompactionFilter::Context& context) override {
|
|
||||||
return std::unique_ptr<CompactionFilter>(new ChangeFilter());
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override { return "ChangeFilterFactory"; }
|
|
||||||
};
|
|
||||||
|
|
||||||
class DBTestUniversalCompactionBase
|
|
||||||
: public DBTest,
|
|
||||||
public ::testing::WithParamInterface<int> {
|
|
||||||
public:
|
|
||||||
virtual void SetUp() override { num_levels_ = GetParam(); }
|
|
||||||
int num_levels_;
|
|
||||||
};
|
|
||||||
|
|
||||||
class DBTestUniversalCompaction : public DBTestUniversalCompactionBase {};
|
|
||||||
|
|
||||||
// TODO(kailiu) The tests on UniversalCompaction has some issues:
|
// TODO(kailiu) The tests on UniversalCompaction has some issues:
|
||||||
// 1. A lot of magic numbers ("11" or "12").
|
// 1. A lot of magic numbers ("11" or "12").
|
||||||
@ -4667,675 +4574,6 @@ TEST_F(DBTest, InPlaceUpdateCallbackNoAction) {
|
|||||||
} while (ChangeCompactOptions());
|
} while (ChangeCompactOptions());
|
||||||
}
|
}
|
||||||
|
|
||||||
TEST_F(DBTest, CompactionFilter) {
|
|
||||||
Options options = CurrentOptions();
|
|
||||||
options.max_open_files = -1;
|
|
||||||
options.num_levels = 3;
|
|
||||||
options.max_mem_compaction_level = 0;
|
|
||||||
options.compaction_filter_factory = std::make_shared<KeepFilterFactory>();
|
|
||||||
options = CurrentOptions(options);
|
|
||||||
CreateAndReopenWithCF({"pikachu"}, options);
|
|
||||||
|
|
||||||
// Write 100K keys, these are written to a few files in L0.
|
|
||||||
const std::string value(10, 'x');
|
|
||||||
for (int i = 0; i < 100000; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%010d", i);
|
|
||||||
Put(1, key, value);
|
|
||||||
}
|
|
||||||
ASSERT_OK(Flush(1));
|
|
||||||
|
|
||||||
// Push all files to the highest level L2. Verify that
|
|
||||||
// the compaction is each level invokes the filter for
|
|
||||||
// all the keys in that level.
|
|
||||||
cfilter_count = 0;
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
|
||||||
ASSERT_EQ(cfilter_count, 100000);
|
|
||||||
cfilter_count = 0;
|
|
||||||
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
|
||||||
ASSERT_EQ(cfilter_count, 100000);
|
|
||||||
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(0, 1), 0);
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(1, 1), 0);
|
|
||||||
ASSERT_NE(NumTableFilesAtLevel(2, 1), 0);
|
|
||||||
cfilter_count = 0;
|
|
||||||
|
|
||||||
// All the files are in the lowest level.
|
|
||||||
// Verify that all but the 100001st record
|
|
||||||
// has sequence number zero. The 100001st record
|
|
||||||
// is at the tip of this snapshot and cannot
|
|
||||||
// be zeroed out.
|
|
||||||
// TODO: figure out sequence number squashtoo
|
|
||||||
int count = 0;
|
|
||||||
int total = 0;
|
|
||||||
Arena arena;
|
|
||||||
{
|
|
||||||
ScopedArenaIterator iter(
|
|
||||||
dbfull()->TEST_NewInternalIterator(&arena, handles_[1]));
|
|
||||||
iter->SeekToFirst();
|
|
||||||
ASSERT_OK(iter->status());
|
|
||||||
while (iter->Valid()) {
|
|
||||||
ParsedInternalKey ikey(Slice(), 0, kTypeValue);
|
|
||||||
ikey.sequence = -1;
|
|
||||||
ASSERT_EQ(ParseInternalKey(iter->key(), &ikey), true);
|
|
||||||
total++;
|
|
||||||
if (ikey.sequence != 0) {
|
|
||||||
count++;
|
|
||||||
}
|
|
||||||
iter->Next();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
ASSERT_EQ(total, 100000);
|
|
||||||
ASSERT_EQ(count, 1);
|
|
||||||
|
|
||||||
// overwrite all the 100K keys once again.
|
|
||||||
for (int i = 0; i < 100000; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%010d", i);
|
|
||||||
ASSERT_OK(Put(1, key, value));
|
|
||||||
}
|
|
||||||
ASSERT_OK(Flush(1));
|
|
||||||
|
|
||||||
// push all files to the highest level L2. This
|
|
||||||
// means that all keys should pass at least once
|
|
||||||
// via the compaction filter
|
|
||||||
cfilter_count = 0;
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
|
||||||
ASSERT_EQ(cfilter_count, 100000);
|
|
||||||
cfilter_count = 0;
|
|
||||||
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
|
||||||
ASSERT_EQ(cfilter_count, 100000);
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(0, 1), 0);
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(1, 1), 0);
|
|
||||||
ASSERT_NE(NumTableFilesAtLevel(2, 1), 0);
|
|
||||||
|
|
||||||
// create a new database with the compaction
|
|
||||||
// filter in such a way that it deletes all keys
|
|
||||||
options.compaction_filter_factory = std::make_shared<DeleteFilterFactory>();
|
|
||||||
options.create_if_missing = true;
|
|
||||||
DestroyAndReopen(options);
|
|
||||||
CreateAndReopenWithCF({"pikachu"}, options);
|
|
||||||
|
|
||||||
// write all the keys once again.
|
|
||||||
for (int i = 0; i < 100000; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%010d", i);
|
|
||||||
ASSERT_OK(Put(1, key, value));
|
|
||||||
}
|
|
||||||
ASSERT_OK(Flush(1));
|
|
||||||
ASSERT_NE(NumTableFilesAtLevel(0, 1), 0);
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(1, 1), 0);
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(2, 1), 0);
|
|
||||||
|
|
||||||
// Push all files to the highest level L2. This
|
|
||||||
// triggers the compaction filter to delete all keys,
|
|
||||||
// verify that at the end of the compaction process,
|
|
||||||
// nothing is left.
|
|
||||||
cfilter_count = 0;
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
|
||||||
ASSERT_EQ(cfilter_count, 100000);
|
|
||||||
cfilter_count = 0;
|
|
||||||
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
|
||||||
ASSERT_EQ(cfilter_count, 0);
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(0, 1), 0);
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(1, 1), 0);
|
|
||||||
|
|
||||||
{
|
|
||||||
// Scan the entire database to ensure that nothing is left
|
|
||||||
std::unique_ptr<Iterator> iter(
|
|
||||||
db_->NewIterator(ReadOptions(), handles_[1]));
|
|
||||||
iter->SeekToFirst();
|
|
||||||
count = 0;
|
|
||||||
while (iter->Valid()) {
|
|
||||||
count++;
|
|
||||||
iter->Next();
|
|
||||||
}
|
|
||||||
ASSERT_EQ(count, 0);
|
|
||||||
}
|
|
||||||
|
|
||||||
// The sequence number of the remaining record
|
|
||||||
// is not zeroed out even though it is at the
|
|
||||||
// level Lmax because this record is at the tip
|
|
||||||
// TODO: remove the following or design a different
|
|
||||||
// test
|
|
||||||
count = 0;
|
|
||||||
{
|
|
||||||
ScopedArenaIterator iter(
|
|
||||||
dbfull()->TEST_NewInternalIterator(&arena, handles_[1]));
|
|
||||||
iter->SeekToFirst();
|
|
||||||
ASSERT_OK(iter->status());
|
|
||||||
while (iter->Valid()) {
|
|
||||||
ParsedInternalKey ikey(Slice(), 0, kTypeValue);
|
|
||||||
ASSERT_EQ(ParseInternalKey(iter->key(), &ikey), true);
|
|
||||||
ASSERT_NE(ikey.sequence, (unsigned)0);
|
|
||||||
count++;
|
|
||||||
iter->Next();
|
|
||||||
}
|
|
||||||
ASSERT_EQ(count, 0);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Tests the edge case where compaction does not produce any output -- all
|
|
||||||
// entries are deleted. The compaction should create bunch of 'DeleteFile'
|
|
||||||
// entries in VersionEdit, but none of the 'AddFile's.
|
|
||||||
TEST_F(DBTest, CompactionFilterDeletesAll) {
|
|
||||||
Options options;
|
|
||||||
options.compaction_filter_factory = std::make_shared<DeleteFilterFactory>();
|
|
||||||
options.disable_auto_compactions = true;
|
|
||||||
options.create_if_missing = true;
|
|
||||||
options = CurrentOptions(options);
|
|
||||||
DestroyAndReopen(options);
|
|
||||||
|
|
||||||
// put some data
|
|
||||||
for (int table = 0; table < 4; ++table) {
|
|
||||||
for (int i = 0; i < 10 + table; ++i) {
|
|
||||||
Put(ToString(table * 100 + i), "val");
|
|
||||||
}
|
|
||||||
Flush();
|
|
||||||
}
|
|
||||||
|
|
||||||
// this will produce empty file (delete compaction filter)
|
|
||||||
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
||||||
ASSERT_EQ(0U, CountLiveFiles());
|
|
||||||
|
|
||||||
Reopen(options);
|
|
||||||
|
|
||||||
Iterator* itr = db_->NewIterator(ReadOptions());
|
|
||||||
itr->SeekToFirst();
|
|
||||||
// empty db
|
|
||||||
ASSERT_TRUE(!itr->Valid());
|
|
||||||
|
|
||||||
delete itr;
|
|
||||||
}
|
|
||||||
|
|
||||||
TEST_F(DBTest, CompactionFilterWithValueChange) {
|
|
||||||
do {
|
|
||||||
Options options;
|
|
||||||
options.num_levels = 3;
|
|
||||||
options.max_mem_compaction_level = 0;
|
|
||||||
options.compaction_filter_factory =
|
|
||||||
std::make_shared<ChangeFilterFactory>();
|
|
||||||
options = CurrentOptions(options);
|
|
||||||
CreateAndReopenWithCF({"pikachu"}, options);
|
|
||||||
|
|
||||||
// Write 100K+1 keys, these are written to a few files
|
|
||||||
// in L0. We do this so that the current snapshot points
|
|
||||||
// to the 100001 key.The compaction filter is not invoked
|
|
||||||
// on keys that are visible via a snapshot because we
|
|
||||||
// anyways cannot delete it.
|
|
||||||
const std::string value(10, 'x');
|
|
||||||
for (int i = 0; i < 100001; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%010d", i);
|
|
||||||
Put(1, key, value);
|
|
||||||
}
|
|
||||||
|
|
||||||
// push all files to lower levels
|
|
||||||
ASSERT_OK(Flush(1));
|
|
||||||
if (option_config_ != kUniversalCompactionMultiLevel) {
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
|
||||||
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
|
||||||
} else {
|
|
||||||
dbfull()->CompactRange(CompactRangeOptions(), handles_[1], nullptr,
|
|
||||||
nullptr);
|
|
||||||
}
|
|
||||||
|
|
||||||
// re-write all data again
|
|
||||||
for (int i = 0; i < 100001; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%010d", i);
|
|
||||||
Put(1, key, value);
|
|
||||||
}
|
|
||||||
|
|
||||||
// push all files to lower levels. This should
|
|
||||||
// invoke the compaction filter for all 100000 keys.
|
|
||||||
ASSERT_OK(Flush(1));
|
|
||||||
if (option_config_ != kUniversalCompactionMultiLevel) {
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
|
||||||
dbfull()->TEST_CompactRange(1, nullptr, nullptr, handles_[1]);
|
|
||||||
} else {
|
|
||||||
dbfull()->CompactRange(CompactRangeOptions(), handles_[1], nullptr,
|
|
||||||
nullptr);
|
|
||||||
}
|
|
||||||
|
|
||||||
// verify that all keys now have the new value that
|
|
||||||
// was set by the compaction process.
|
|
||||||
for (int i = 0; i < 100001; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%010d", i);
|
|
||||||
std::string newvalue = Get(1, key);
|
|
||||||
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
|
||||||
}
|
|
||||||
} while (ChangeCompactOptions());
|
|
||||||
}
|
|
||||||
|
|
||||||
TEST_F(DBTest, CompactionFilterWithMergeOperator) {
|
|
||||||
std::string one, two, three, four;
|
|
||||||
PutFixed64(&one, 1);
|
|
||||||
PutFixed64(&two, 2);
|
|
||||||
PutFixed64(&three, 3);
|
|
||||||
PutFixed64(&four, 4);
|
|
||||||
|
|
||||||
Options options;
|
|
||||||
options = CurrentOptions(options);
|
|
||||||
options.create_if_missing = true;
|
|
||||||
options.merge_operator = MergeOperators::CreateUInt64AddOperator();
|
|
||||||
options.num_levels = 3;
|
|
||||||
options.max_mem_compaction_level = 0;
|
|
||||||
// Filter out keys with value is 2.
|
|
||||||
options.compaction_filter_factory =
|
|
||||||
std::make_shared<ConditionalFilterFactory>(two);
|
|
||||||
DestroyAndReopen(options);
|
|
||||||
|
|
||||||
// In the same compaction, a value type needs to be deleted based on
|
|
||||||
// compaction filter, and there is a merge type for the key. compaction
|
|
||||||
// filter result is ignored.
|
|
||||||
ASSERT_OK(db_->Put(WriteOptions(), "foo", two));
|
|
||||||
ASSERT_OK(Flush());
|
|
||||||
ASSERT_OK(db_->Merge(WriteOptions(), "foo", one));
|
|
||||||
ASSERT_OK(Flush());
|
|
||||||
std::string newvalue = Get("foo");
|
|
||||||
ASSERT_EQ(newvalue, three);
|
|
||||||
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
||||||
newvalue = Get("foo");
|
|
||||||
ASSERT_EQ(newvalue, three);
|
|
||||||
|
|
||||||
// value key can be deleted based on compaction filter, leaving only
|
|
||||||
// merge keys.
|
|
||||||
ASSERT_OK(db_->Put(WriteOptions(), "bar", two));
|
|
||||||
ASSERT_OK(Flush());
|
|
||||||
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
||||||
newvalue = Get("bar");
|
|
||||||
ASSERT_EQ("NOT_FOUND", newvalue);
|
|
||||||
ASSERT_OK(db_->Merge(WriteOptions(), "bar", two));
|
|
||||||
ASSERT_OK(Flush());
|
|
||||||
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
||||||
newvalue = Get("bar");
|
|
||||||
ASSERT_EQ(two, two);
|
|
||||||
|
|
||||||
// Compaction filter never applies to merge keys.
|
|
||||||
ASSERT_OK(db_->Put(WriteOptions(), "foobar", one));
|
|
||||||
ASSERT_OK(Flush());
|
|
||||||
ASSERT_OK(db_->Merge(WriteOptions(), "foobar", two));
|
|
||||||
ASSERT_OK(Flush());
|
|
||||||
newvalue = Get("foobar");
|
|
||||||
ASSERT_EQ(newvalue, three);
|
|
||||||
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
||||||
newvalue = Get("foobar");
|
|
||||||
ASSERT_EQ(newvalue, three);
|
|
||||||
|
|
||||||
// In the same compaction, both of value type and merge type keys need to be
|
|
||||||
// deleted based on compaction filter, and there is a merge type for the key.
|
|
||||||
// For both keys, compaction filter results are ignored.
|
|
||||||
ASSERT_OK(db_->Put(WriteOptions(), "barfoo", two));
|
|
||||||
ASSERT_OK(Flush());
|
|
||||||
ASSERT_OK(db_->Merge(WriteOptions(), "barfoo", two));
|
|
||||||
ASSERT_OK(Flush());
|
|
||||||
newvalue = Get("barfoo");
|
|
||||||
ASSERT_EQ(newvalue, four);
|
|
||||||
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
||||||
newvalue = Get("barfoo");
|
|
||||||
ASSERT_EQ(newvalue, four);
|
|
||||||
}
|
|
||||||
|
|
||||||
TEST_F(DBTest, CompactionFilterContextManual) {
|
|
||||||
KeepFilterFactory* filter = new KeepFilterFactory();
|
|
||||||
|
|
||||||
Options options = CurrentOptions();
|
|
||||||
options.compaction_style = kCompactionStyleUniversal;
|
|
||||||
options.compaction_filter_factory.reset(filter);
|
|
||||||
options.compression = kNoCompression;
|
|
||||||
options.level0_file_num_compaction_trigger = 8;
|
|
||||||
Reopen(options);
|
|
||||||
int num_keys_per_file = 400;
|
|
||||||
for (int j = 0; j < 3; j++) {
|
|
||||||
// Write several keys.
|
|
||||||
const std::string value(10, 'x');
|
|
||||||
for (int i = 0; i < num_keys_per_file; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%08d%02d", i, j);
|
|
||||||
Put(key, value);
|
|
||||||
}
|
|
||||||
dbfull()->TEST_FlushMemTable();
|
|
||||||
// Make sure next file is much smaller so automatic compaction will not
|
|
||||||
// be triggered.
|
|
||||||
num_keys_per_file /= 2;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Force a manual compaction
|
|
||||||
cfilter_count = 0;
|
|
||||||
filter->expect_manual_compaction_.store(true);
|
|
||||||
filter->expect_full_compaction_.store(false); // Manual compaction always
|
|
||||||
// set this flag.
|
|
||||||
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
||||||
ASSERT_EQ(cfilter_count, 700);
|
|
||||||
ASSERT_EQ(NumSortedRuns(0), 1);
|
|
||||||
|
|
||||||
// Verify total number of keys is correct after manual compaction.
|
|
||||||
{
|
|
||||||
int count = 0;
|
|
||||||
int total = 0;
|
|
||||||
Arena arena;
|
|
||||||
ScopedArenaIterator iter(dbfull()->TEST_NewInternalIterator(&arena));
|
|
||||||
iter->SeekToFirst();
|
|
||||||
ASSERT_OK(iter->status());
|
|
||||||
while (iter->Valid()) {
|
|
||||||
ParsedInternalKey ikey(Slice(), 0, kTypeValue);
|
|
||||||
ikey.sequence = -1;
|
|
||||||
ASSERT_EQ(ParseInternalKey(iter->key(), &ikey), true);
|
|
||||||
total++;
|
|
||||||
if (ikey.sequence != 0) {
|
|
||||||
count++;
|
|
||||||
}
|
|
||||||
iter->Next();
|
|
||||||
}
|
|
||||||
ASSERT_EQ(total, 700);
|
|
||||||
ASSERT_EQ(count, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
class KeepFilterV2 : public CompactionFilterV2 {
|
|
||||||
public:
|
|
||||||
virtual std::vector<bool> Filter(int level,
|
|
||||||
const SliceVector& keys,
|
|
||||||
const SliceVector& existing_values,
|
|
||||||
std::vector<std::string>* new_values,
|
|
||||||
std::vector<bool>* values_changed)
|
|
||||||
const override {
|
|
||||||
cfilter_count++;
|
|
||||||
std::vector<bool> ret;
|
|
||||||
new_values->clear();
|
|
||||||
values_changed->clear();
|
|
||||||
for (unsigned int i = 0; i < keys.size(); ++i) {
|
|
||||||
values_changed->push_back(false);
|
|
||||||
ret.push_back(false);
|
|
||||||
}
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override {
|
|
||||||
return "KeepFilterV2";
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
class DeleteFilterV2 : public CompactionFilterV2 {
|
|
||||||
public:
|
|
||||||
virtual std::vector<bool> Filter(int level,
|
|
||||||
const SliceVector& keys,
|
|
||||||
const SliceVector& existing_values,
|
|
||||||
std::vector<std::string>* new_values,
|
|
||||||
std::vector<bool>* values_changed)
|
|
||||||
const override {
|
|
||||||
cfilter_count++;
|
|
||||||
new_values->clear();
|
|
||||||
values_changed->clear();
|
|
||||||
std::vector<bool> ret;
|
|
||||||
for (unsigned int i = 0; i < keys.size(); ++i) {
|
|
||||||
values_changed->push_back(false);
|
|
||||||
ret.push_back(true);
|
|
||||||
}
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override {
|
|
||||||
return "DeleteFilterV2";
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
class ChangeFilterV2 : public CompactionFilterV2 {
|
|
||||||
public:
|
|
||||||
virtual std::vector<bool> Filter(int level,
|
|
||||||
const SliceVector& keys,
|
|
||||||
const SliceVector& existing_values,
|
|
||||||
std::vector<std::string>* new_values,
|
|
||||||
std::vector<bool>* values_changed)
|
|
||||||
const override {
|
|
||||||
std::vector<bool> ret;
|
|
||||||
new_values->clear();
|
|
||||||
values_changed->clear();
|
|
||||||
for (unsigned int i = 0; i < keys.size(); ++i) {
|
|
||||||
values_changed->push_back(true);
|
|
||||||
new_values->push_back(NEW_VALUE);
|
|
||||||
ret.push_back(false);
|
|
||||||
}
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override {
|
|
||||||
return "ChangeFilterV2";
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
class KeepFilterFactoryV2 : public CompactionFilterFactoryV2 {
|
|
||||||
public:
|
|
||||||
explicit KeepFilterFactoryV2(const SliceTransform* prefix_extractor)
|
|
||||||
: CompactionFilterFactoryV2(prefix_extractor) { }
|
|
||||||
|
|
||||||
virtual std::unique_ptr<CompactionFilterV2>
|
|
||||||
CreateCompactionFilterV2(
|
|
||||||
const CompactionFilterContext& context) override {
|
|
||||||
return std::unique_ptr<CompactionFilterV2>(new KeepFilterV2());
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override {
|
|
||||||
return "KeepFilterFactoryV2";
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
class DeleteFilterFactoryV2 : public CompactionFilterFactoryV2 {
|
|
||||||
public:
|
|
||||||
explicit DeleteFilterFactoryV2(const SliceTransform* prefix_extractor)
|
|
||||||
: CompactionFilterFactoryV2(prefix_extractor) { }
|
|
||||||
|
|
||||||
virtual std::unique_ptr<CompactionFilterV2>
|
|
||||||
CreateCompactionFilterV2(
|
|
||||||
const CompactionFilterContext& context) override {
|
|
||||||
return std::unique_ptr<CompactionFilterV2>(new DeleteFilterV2());
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override {
|
|
||||||
return "DeleteFilterFactoryV2";
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
class ChangeFilterFactoryV2 : public CompactionFilterFactoryV2 {
|
|
||||||
public:
|
|
||||||
explicit ChangeFilterFactoryV2(const SliceTransform* prefix_extractor)
|
|
||||||
: CompactionFilterFactoryV2(prefix_extractor) { }
|
|
||||||
|
|
||||||
virtual std::unique_ptr<CompactionFilterV2>
|
|
||||||
CreateCompactionFilterV2(
|
|
||||||
const CompactionFilterContext& context) override {
|
|
||||||
return std::unique_ptr<CompactionFilterV2>(new ChangeFilterV2());
|
|
||||||
}
|
|
||||||
|
|
||||||
virtual const char* Name() const override {
|
|
||||||
return "ChangeFilterFactoryV2";
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
TEST_F(DBTest, CompactionFilterV2) {
|
|
||||||
Options options = CurrentOptions();
|
|
||||||
options.num_levels = 3;
|
|
||||||
options.max_mem_compaction_level = 0;
|
|
||||||
// extract prefix
|
|
||||||
std::unique_ptr<const SliceTransform> prefix_extractor;
|
|
||||||
prefix_extractor.reset(NewFixedPrefixTransform(8));
|
|
||||||
|
|
||||||
options.compaction_filter_factory_v2
|
|
||||||
= std::make_shared<KeepFilterFactoryV2>(prefix_extractor.get());
|
|
||||||
// In a testing environment, we can only flush the application
|
|
||||||
// compaction filter buffer using universal compaction
|
|
||||||
option_config_ = kUniversalCompaction;
|
|
||||||
options.compaction_style = (rocksdb::CompactionStyle)1;
|
|
||||||
Reopen(options);
|
|
||||||
|
|
||||||
// Write 100K keys, these are written to a few files in L0.
|
|
||||||
const std::string value(10, 'x');
|
|
||||||
for (int i = 0; i < 100000; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%08d%010d", i , i);
|
|
||||||
Put(key, value);
|
|
||||||
}
|
|
||||||
|
|
||||||
dbfull()->TEST_FlushMemTable();
|
|
||||||
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
|
||||||
dbfull()->TEST_CompactRange(1, nullptr, nullptr);
|
|
||||||
|
|
||||||
ASSERT_EQ(NumSortedRuns(0), 1);
|
|
||||||
|
|
||||||
// All the files are in the lowest level.
|
|
||||||
int count = 0;
|
|
||||||
int total = 0;
|
|
||||||
{
|
|
||||||
Arena arena;
|
|
||||||
ScopedArenaIterator iter(dbfull()->TEST_NewInternalIterator(&arena));
|
|
||||||
iter->SeekToFirst();
|
|
||||||
ASSERT_OK(iter->status());
|
|
||||||
while (iter->Valid()) {
|
|
||||||
ParsedInternalKey ikey(Slice(), 0, kTypeValue);
|
|
||||||
ikey.sequence = -1;
|
|
||||||
ASSERT_EQ(ParseInternalKey(iter->key(), &ikey), true);
|
|
||||||
total++;
|
|
||||||
if (ikey.sequence != 0) {
|
|
||||||
count++;
|
|
||||||
}
|
|
||||||
iter->Next();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
ASSERT_EQ(total, 100000);
|
|
||||||
// 1 snapshot only. Since we are using universal compacton,
|
|
||||||
// the sequence no is cleared for better compression
|
|
||||||
ASSERT_EQ(count, 1);
|
|
||||||
|
|
||||||
// create a new database with the compaction
|
|
||||||
// filter in such a way that it deletes all keys
|
|
||||||
options.compaction_filter_factory_v2 =
|
|
||||||
std::make_shared<DeleteFilterFactoryV2>(prefix_extractor.get());
|
|
||||||
options.create_if_missing = true;
|
|
||||||
DestroyAndReopen(options);
|
|
||||||
|
|
||||||
// write all the keys once again.
|
|
||||||
for (int i = 0; i < 100000; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%08d%010d", i, i);
|
|
||||||
Put(key, value);
|
|
||||||
}
|
|
||||||
|
|
||||||
dbfull()->TEST_FlushMemTable();
|
|
||||||
ASSERT_NE(NumTableFilesAtLevel(0), 0);
|
|
||||||
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
|
||||||
dbfull()->TEST_CompactRange(1, nullptr, nullptr);
|
|
||||||
ASSERT_EQ(NumTableFilesAtLevel(1), 0);
|
|
||||||
|
|
||||||
// Scan the entire database to ensure that nothing is left
|
|
||||||
Iterator* iter = db_->NewIterator(ReadOptions());
|
|
||||||
iter->SeekToFirst();
|
|
||||||
count = 0;
|
|
||||||
while (iter->Valid()) {
|
|
||||||
count++;
|
|
||||||
iter->Next();
|
|
||||||
}
|
|
||||||
|
|
||||||
ASSERT_EQ(count, 0);
|
|
||||||
delete iter;
|
|
||||||
}
|
|
||||||
|
|
||||||
TEST_F(DBTest, CompactionFilterV2WithValueChange) {
|
|
||||||
Options options = CurrentOptions();
|
|
||||||
options.num_levels = 3;
|
|
||||||
options.max_mem_compaction_level = 0;
|
|
||||||
std::unique_ptr<const SliceTransform> prefix_extractor;
|
|
||||||
prefix_extractor.reset(NewFixedPrefixTransform(8));
|
|
||||||
options.compaction_filter_factory_v2 =
|
|
||||||
std::make_shared<ChangeFilterFactoryV2>(prefix_extractor.get());
|
|
||||||
// In a testing environment, we can only flush the application
|
|
||||||
// compaction filter buffer using universal compaction
|
|
||||||
option_config_ = kUniversalCompaction;
|
|
||||||
options.compaction_style = (rocksdb::CompactionStyle)1;
|
|
||||||
options = CurrentOptions(options);
|
|
||||||
Reopen(options);
|
|
||||||
|
|
||||||
// Write 100K+1 keys, these are written to a few files
|
|
||||||
// in L0. We do this so that the current snapshot points
|
|
||||||
// to the 100001 key.The compaction filter is not invoked
|
|
||||||
// on keys that are visible via a snapshot because we
|
|
||||||
// anyways cannot delete it.
|
|
||||||
const std::string value(10, 'x');
|
|
||||||
for (int i = 0; i < 100001; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%08d%010d", i, i);
|
|
||||||
Put(key, value);
|
|
||||||
}
|
|
||||||
|
|
||||||
// push all files to lower levels
|
|
||||||
dbfull()->TEST_FlushMemTable();
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
|
||||||
dbfull()->TEST_CompactRange(1, nullptr, nullptr);
|
|
||||||
|
|
||||||
// verify that all keys now have the new value that
|
|
||||||
// was set by the compaction process.
|
|
||||||
for (int i = 0; i < 100001; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "B%08d%010d", i, i);
|
|
||||||
std::string newvalue = Get(key);
|
|
||||||
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
TEST_F(DBTest, CompactionFilterV2NULLPrefix) {
|
|
||||||
Options options = CurrentOptions();
|
|
||||||
options.num_levels = 3;
|
|
||||||
options.max_mem_compaction_level = 0;
|
|
||||||
std::unique_ptr<const SliceTransform> prefix_extractor;
|
|
||||||
prefix_extractor.reset(NewFixedPrefixTransform(8));
|
|
||||||
options.compaction_filter_factory_v2 =
|
|
||||||
std::make_shared<ChangeFilterFactoryV2>(prefix_extractor.get());
|
|
||||||
// In a testing environment, we can only flush the application
|
|
||||||
// compaction filter buffer using universal compaction
|
|
||||||
option_config_ = kUniversalCompaction;
|
|
||||||
options.compaction_style = (rocksdb::CompactionStyle)1;
|
|
||||||
Reopen(options);
|
|
||||||
|
|
||||||
// Write 100K+1 keys, these are written to a few files
|
|
||||||
// in L0. We do this so that the current snapshot points
|
|
||||||
// to the 100001 key.The compaction filter is not invoked
|
|
||||||
// on keys that are visible via a snapshot because we
|
|
||||||
// anyways cannot delete it.
|
|
||||||
const std::string value(10, 'x');
|
|
||||||
char first_key[100];
|
|
||||||
snprintf(first_key, sizeof(first_key), "%s0000%010d", "NULL", 1);
|
|
||||||
Put(first_key, value);
|
|
||||||
for (int i = 1; i < 100000; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "%08d%010d", i, i);
|
|
||||||
Put(key, value);
|
|
||||||
}
|
|
||||||
|
|
||||||
char last_key[100];
|
|
||||||
snprintf(last_key, sizeof(last_key), "%s0000%010d", "NULL", 2);
|
|
||||||
Put(last_key, value);
|
|
||||||
|
|
||||||
// push all files to lower levels
|
|
||||||
dbfull()->TEST_FlushMemTable();
|
|
||||||
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
|
||||||
|
|
||||||
// verify that all keys now have the new value that
|
|
||||||
// was set by the compaction process.
|
|
||||||
std::string newvalue = Get(first_key);
|
|
||||||
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
|
||||||
newvalue = Get(last_key);
|
|
||||||
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
|
||||||
for (int i = 1; i < 100000; i++) {
|
|
||||||
char key[100];
|
|
||||||
snprintf(key, sizeof(key), "%08d%010d", i, i);
|
|
||||||
newvalue = Get(key);
|
|
||||||
ASSERT_EQ(newvalue.compare(NEW_VALUE), 0);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
TEST_F(DBTest, SparseMerge) {
|
TEST_F(DBTest, SparseMerge) {
|
||||||
do {
|
do {
|
||||||
|
1
src.mk
1
src.mk
@ -163,6 +163,7 @@ TEST_BENCH_SOURCES = \
|
|||||||
db/dbformat_test.cc \
|
db/dbformat_test.cc \
|
||||||
db/db_iter_test.cc \
|
db/db_iter_test.cc \
|
||||||
db/db_test.cc \
|
db/db_test.cc \
|
||||||
|
db/db_compaction_filter_test.cc \
|
||||||
db/db_dynamic_level_test.cc \
|
db/db_dynamic_level_test.cc \
|
||||||
db/deletefile_test.cc \
|
db/deletefile_test.cc \
|
||||||
db/fault_injection_test.cc \
|
db/fault_injection_test.cc \
|
||||||
|
Loading…
Reference in New Issue
Block a user