2016-09-08 00:41:54 +02:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2017-07-16 01:03:42 +02:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2016-09-08 00:41:54 +02:00
|
|
|
|
2017-02-28 20:05:08 +01:00
|
|
|
#ifndef ROCKSDB_LITE
|
2016-12-16 20:17:26 +01:00
|
|
|
|
2017-02-28 20:05:08 +01:00
|
|
|
#include <functional>
|
2016-09-08 00:41:54 +02:00
|
|
|
#include "db/db_test_util.h"
|
2019-05-30 05:44:08 +02:00
|
|
|
#include "file/filename.h"
|
2016-09-08 00:41:54 +02:00
|
|
|
#include "port/port.h"
|
|
|
|
#include "port/stack_trace.h"
|
|
|
|
#include "rocksdb/sst_file_writer.h"
|
2019-05-30 20:21:38 +02:00
|
|
|
#include "test_util/fault_injection_test_env.h"
|
|
|
|
#include "test_util/testutil.h"
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
namespace rocksdb {
|
|
|
|
|
2019-05-24 06:54:23 +02:00
|
|
|
// A test environment that can be configured to fail the Link operation.
|
|
|
|
class ExternalSSTTestEnv : public EnvWrapper {
|
|
|
|
public:
|
|
|
|
ExternalSSTTestEnv(Env* t, bool fail_link)
|
|
|
|
: EnvWrapper(t), fail_link_(fail_link) {}
|
|
|
|
|
|
|
|
Status LinkFile(const std::string& s, const std::string& t) override {
|
|
|
|
if (fail_link_) {
|
|
|
|
return Status::NotSupported("Link failed");
|
|
|
|
}
|
|
|
|
return target()->LinkFile(s, t);
|
|
|
|
}
|
|
|
|
|
|
|
|
void set_fail_link(bool fail_link) { fail_link_ = fail_link; }
|
|
|
|
|
|
|
|
private:
|
|
|
|
bool fail_link_;
|
|
|
|
};
|
|
|
|
|
|
|
|
class ExternSSTFileLinkFailFallbackTest
|
|
|
|
: public DBTestBase,
|
|
|
|
public ::testing::WithParamInterface<std::tuple<bool, bool>> {
|
|
|
|
public:
|
|
|
|
ExternSSTFileLinkFailFallbackTest()
|
|
|
|
: DBTestBase("/external_sst_file_test"),
|
|
|
|
test_env_(new ExternalSSTTestEnv(env_, true)) {
|
|
|
|
sst_files_dir_ = dbname_ + "/sst_files/";
|
|
|
|
test::DestroyDir(env_, sst_files_dir_);
|
|
|
|
env_->CreateDir(sst_files_dir_);
|
|
|
|
options_ = CurrentOptions();
|
|
|
|
options_.disable_auto_compactions = true;
|
|
|
|
options_.env = test_env_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void TearDown() override {
|
|
|
|
delete db_;
|
|
|
|
db_ = nullptr;
|
|
|
|
ASSERT_OK(DestroyDB(dbname_, options_));
|
|
|
|
delete test_env_;
|
|
|
|
test_env_ = nullptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
protected:
|
|
|
|
std::string sst_files_dir_;
|
|
|
|
Options options_;
|
|
|
|
ExternalSSTTestEnv* test_env_;
|
|
|
|
};
|
|
|
|
|
2019-01-30 01:16:53 +01:00
|
|
|
class ExternalSSTFileTest
|
|
|
|
: public DBTestBase,
|
|
|
|
public ::testing::WithParamInterface<std::tuple<bool, bool>> {
|
2016-09-08 00:41:54 +02:00
|
|
|
public:
|
|
|
|
ExternalSSTFileTest() : DBTestBase("/external_sst_file_test") {
|
2016-12-08 22:30:09 +01:00
|
|
|
sst_files_dir_ = dbname_ + "/sst_files/";
|
2016-09-08 00:41:54 +02:00
|
|
|
DestroyAndRecreateExternalSSTFilesDir();
|
|
|
|
}
|
|
|
|
|
|
|
|
void DestroyAndRecreateExternalSSTFilesDir() {
|
|
|
|
test::DestroyDir(env_, sst_files_dir_);
|
|
|
|
env_->CreateDir(sst_files_dir_);
|
|
|
|
}
|
|
|
|
|
2019-02-13 04:07:25 +01:00
|
|
|
Status GenerateOneExternalFile(
|
|
|
|
const Options& options, ColumnFamilyHandle* cfh,
|
|
|
|
std::vector<std::pair<std::string, std::string>>& data, int file_id,
|
|
|
|
bool sort_data, std::string* external_file_path,
|
|
|
|
std::map<std::string, std::string>* true_data) {
|
|
|
|
// Generate a file id if not provided
|
|
|
|
if (-1 == file_id) {
|
|
|
|
file_id = (++last_file_id_);
|
|
|
|
}
|
|
|
|
// Sort data if asked to do so
|
|
|
|
if (sort_data) {
|
|
|
|
std::sort(data.begin(), data.end(),
|
|
|
|
[&](const std::pair<std::string, std::string>& e1,
|
|
|
|
const std::pair<std::string, std::string>& e2) {
|
|
|
|
return options.comparator->Compare(e1.first, e2.first) < 0;
|
|
|
|
});
|
|
|
|
auto uniq_iter = std::unique(
|
|
|
|
data.begin(), data.end(),
|
|
|
|
[&](const std::pair<std::string, std::string>& e1,
|
|
|
|
const std::pair<std::string, std::string>& e2) {
|
|
|
|
return options.comparator->Compare(e1.first, e2.first) == 0;
|
|
|
|
});
|
|
|
|
data.resize(uniq_iter - data.begin());
|
|
|
|
}
|
|
|
|
std::string file_path = sst_files_dir_ + ToString(file_id);
|
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options, cfh);
|
|
|
|
Status s = sst_file_writer.Open(file_path);
|
|
|
|
if (!s.ok()) {
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
for (const auto& entry : data) {
|
|
|
|
s = sst_file_writer.Put(entry.first, entry.second);
|
|
|
|
if (!s.ok()) {
|
|
|
|
sst_file_writer.Finish();
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
s = sst_file_writer.Finish();
|
|
|
|
if (s.ok() && external_file_path != nullptr) {
|
|
|
|
*external_file_path = file_path;
|
|
|
|
}
|
|
|
|
if (s.ok() && nullptr != true_data) {
|
|
|
|
for (const auto& entry : data) {
|
|
|
|
true_data->insert({entry.first, entry.second});
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
Status GenerateAndAddExternalFile(
|
|
|
|
const Options options,
|
|
|
|
std::vector<std::pair<std::string, std::string>> data, int file_id = -1,
|
2018-11-02 00:21:30 +01:00
|
|
|
bool allow_global_seqno = false, bool write_global_seqno = false,
|
2019-04-08 20:12:25 +02:00
|
|
|
bool verify_checksums_before_ingest = true, bool ingest_behind = false,
|
|
|
|
bool sort_data = false,
|
2016-12-06 22:56:17 +01:00
|
|
|
std::map<std::string, std::string>* true_data = nullptr,
|
|
|
|
ColumnFamilyHandle* cfh = nullptr) {
|
2016-10-21 02:05:32 +02:00
|
|
|
// Generate a file id if not provided
|
|
|
|
if (file_id == -1) {
|
|
|
|
file_id = last_file_id_ + 1;
|
|
|
|
last_file_id_++;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Sort data if asked to do so
|
|
|
|
if (sort_data) {
|
|
|
|
std::sort(data.begin(), data.end(),
|
|
|
|
[&](const std::pair<std::string, std::string>& e1,
|
|
|
|
const std::pair<std::string, std::string>& e2) {
|
|
|
|
return options.comparator->Compare(e1.first, e2.first) < 0;
|
|
|
|
});
|
|
|
|
auto uniq_iter = std::unique(
|
|
|
|
data.begin(), data.end(),
|
|
|
|
[&](const std::pair<std::string, std::string>& e1,
|
|
|
|
const std::pair<std::string, std::string>& e2) {
|
|
|
|
return options.comparator->Compare(e1.first, e2.first) == 0;
|
|
|
|
});
|
|
|
|
data.resize(uniq_iter - data.begin());
|
|
|
|
}
|
2016-10-19 01:59:37 +02:00
|
|
|
std::string file_path = sst_files_dir_ + ToString(file_id);
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options, cfh);
|
2016-10-19 01:59:37 +02:00
|
|
|
|
|
|
|
Status s = sst_file_writer.Open(file_path);
|
|
|
|
if (!s.ok()) {
|
|
|
|
return s;
|
|
|
|
}
|
2016-10-21 02:05:32 +02:00
|
|
|
for (auto& entry : data) {
|
2017-05-26 21:05:19 +02:00
|
|
|
s = sst_file_writer.Put(entry.first, entry.second);
|
2016-10-19 01:59:37 +02:00
|
|
|
if (!s.ok()) {
|
2016-10-21 02:05:32 +02:00
|
|
|
sst_file_writer.Finish();
|
2016-10-19 01:59:37 +02:00
|
|
|
return s;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
s = sst_file_writer.Finish();
|
|
|
|
|
|
|
|
if (s.ok()) {
|
2016-10-21 02:05:32 +02:00
|
|
|
IngestExternalFileOptions ifo;
|
|
|
|
ifo.allow_global_seqno = allow_global_seqno;
|
2018-11-02 00:21:30 +01:00
|
|
|
ifo.write_global_seqno = allow_global_seqno ? write_global_seqno : false;
|
2019-01-30 01:16:53 +01:00
|
|
|
ifo.verify_checksums_before_ingest = verify_checksums_before_ingest;
|
2019-04-08 20:12:25 +02:00
|
|
|
ifo.ingest_behind = ingest_behind;
|
2017-05-17 20:32:26 +02:00
|
|
|
if (cfh) {
|
|
|
|
s = db_->IngestExternalFile(cfh, {file_path}, ifo);
|
|
|
|
} else {
|
|
|
|
s = db_->IngestExternalFile({file_path}, ifo);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s.ok() && true_data) {
|
|
|
|
for (auto& entry : data) {
|
|
|
|
(*true_data)[entry.first] = entry.second;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2019-02-13 04:07:25 +01:00
|
|
|
Status GenerateAndAddExternalFiles(
|
|
|
|
const Options& options,
|
|
|
|
const std::vector<ColumnFamilyHandle*>& column_families,
|
|
|
|
const std::vector<IngestExternalFileOptions>& ifos,
|
|
|
|
std::vector<std::vector<std::pair<std::string, std::string>>>& data,
|
|
|
|
int file_id, bool sort_data,
|
|
|
|
std::vector<std::map<std::string, std::string>>& true_data) {
|
|
|
|
if (-1 == file_id) {
|
|
|
|
file_id = (++last_file_id_);
|
|
|
|
}
|
|
|
|
// Generate external SST files, one for each column family
|
|
|
|
size_t num_cfs = column_families.size();
|
|
|
|
assert(ifos.size() == num_cfs);
|
|
|
|
assert(data.size() == num_cfs);
|
|
|
|
Status s;
|
|
|
|
std::vector<IngestExternalFileArg> args(num_cfs);
|
|
|
|
for (size_t i = 0; i != num_cfs; ++i) {
|
|
|
|
std::string external_file_path;
|
|
|
|
s = GenerateOneExternalFile(
|
|
|
|
options, column_families[i], data[i], file_id, sort_data,
|
|
|
|
&external_file_path,
|
|
|
|
true_data.size() == num_cfs ? &true_data[i] : nullptr);
|
|
|
|
if (!s.ok()) {
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
++file_id;
|
|
|
|
|
|
|
|
args[i].column_family = column_families[i];
|
|
|
|
args[i].external_files.push_back(external_file_path);
|
|
|
|
args[i].options = ifos[i];
|
|
|
|
}
|
|
|
|
s = db_->IngestExternalFiles(args);
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
Status GenerateAndAddExternalFile(
|
|
|
|
const Options options, std::vector<std::pair<int, std::string>> data,
|
2018-11-02 00:21:30 +01:00
|
|
|
int file_id = -1, bool allow_global_seqno = false,
|
2019-01-30 01:16:53 +01:00
|
|
|
bool write_global_seqno = false,
|
2019-04-08 20:12:25 +02:00
|
|
|
bool verify_checksums_before_ingest = true, bool ingest_behind = false,
|
|
|
|
bool sort_data = false,
|
2016-12-06 22:56:17 +01:00
|
|
|
std::map<std::string, std::string>* true_data = nullptr,
|
|
|
|
ColumnFamilyHandle* cfh = nullptr) {
|
2016-10-21 02:05:32 +02:00
|
|
|
std::vector<std::pair<std::string, std::string>> file_data;
|
|
|
|
for (auto& entry : data) {
|
|
|
|
file_data.emplace_back(Key(entry.first), entry.second);
|
|
|
|
}
|
2019-04-08 20:12:25 +02:00
|
|
|
return GenerateAndAddExternalFile(options, file_data, file_id,
|
|
|
|
allow_global_seqno, write_global_seqno,
|
|
|
|
verify_checksums_before_ingest,
|
|
|
|
ingest_behind, sort_data, true_data, cfh);
|
2016-10-21 02:05:32 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
Status GenerateAndAddExternalFile(
|
|
|
|
const Options options, std::vector<int> keys, int file_id = -1,
|
2018-11-02 00:21:30 +01:00
|
|
|
bool allow_global_seqno = false, bool write_global_seqno = false,
|
2019-04-08 20:12:25 +02:00
|
|
|
bool verify_checksums_before_ingest = true, bool ingest_behind = false,
|
|
|
|
bool sort_data = false,
|
2016-12-06 22:56:17 +01:00
|
|
|
std::map<std::string, std::string>* true_data = nullptr,
|
|
|
|
ColumnFamilyHandle* cfh = nullptr) {
|
2016-10-21 02:05:32 +02:00
|
|
|
std::vector<std::pair<std::string, std::string>> file_data;
|
|
|
|
for (auto& k : keys) {
|
|
|
|
file_data.emplace_back(Key(k), Key(k) + ToString(file_id));
|
|
|
|
}
|
2019-04-08 20:12:25 +02:00
|
|
|
return GenerateAndAddExternalFile(options, file_data, file_id,
|
|
|
|
allow_global_seqno, write_global_seqno,
|
|
|
|
verify_checksums_before_ingest,
|
|
|
|
ingest_behind, sort_data, true_data, cfh);
|
2016-10-21 02:05:32 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
Status DeprecatedAddFile(const std::vector<std::string>& files,
|
|
|
|
bool move_files = false,
|
2018-11-02 00:21:30 +01:00
|
|
|
bool skip_snapshot_check = false,
|
|
|
|
bool skip_write_global_seqno = false) {
|
2016-10-21 02:05:32 +02:00
|
|
|
IngestExternalFileOptions opts;
|
|
|
|
opts.move_files = move_files;
|
|
|
|
opts.snapshot_consistency = !skip_snapshot_check;
|
|
|
|
opts.allow_global_seqno = false;
|
|
|
|
opts.allow_blocking_flush = false;
|
2018-11-02 00:21:30 +01:00
|
|
|
opts.write_global_seqno = !skip_write_global_seqno;
|
2016-10-21 02:05:32 +02:00
|
|
|
return db_->IngestExternalFile(files, opts);
|
|
|
|
}
|
|
|
|
|
2019-02-14 22:52:47 +01:00
|
|
|
~ExternalSSTFileTest() override { test::DestroyDir(env_, sst_files_dir_); }
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
protected:
|
2016-10-21 02:05:32 +02:00
|
|
|
int last_file_id_ = 0;
|
2016-09-08 00:41:54 +02:00
|
|
|
std::string sst_files_dir_;
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, Basic) {
|
|
|
|
do {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
2016-12-17 03:26:56 +01:00
|
|
|
// Current file size should be 0 after sst_file_writer init and before open a file.
|
|
|
|
ASSERT_EQ(sst_file_writer.FileSize(), 0);
|
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
// file1.sst (0 => 99)
|
|
|
|
std::string file1 = sst_files_dir_ + "file1.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file1));
|
|
|
|
for (int k = 0; k < 100; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file1_info;
|
|
|
|
Status s = sst_file_writer.Finish(&file1_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
2016-12-17 03:26:56 +01:00
|
|
|
|
|
|
|
// Current file size should be non-zero after success write.
|
|
|
|
ASSERT_GT(sst_file_writer.FileSize(), 0);
|
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(file1_info.file_path, file1);
|
|
|
|
ASSERT_EQ(file1_info.num_entries, 100);
|
|
|
|
ASSERT_EQ(file1_info.smallest_key, Key(0));
|
|
|
|
ASSERT_EQ(file1_info.largest_key, Key(99));
|
2018-07-14 07:40:23 +02:00
|
|
|
ASSERT_EQ(file1_info.num_range_del_entries, 0);
|
|
|
|
ASSERT_EQ(file1_info.smallest_range_del_key, "");
|
|
|
|
ASSERT_EQ(file1_info.largest_range_del_key, "");
|
2016-09-08 00:41:54 +02:00
|
|
|
// sst_file_writer already finished, cannot add this value
|
2017-05-26 21:05:19 +02:00
|
|
|
s = sst_file_writer.Put(Key(100), "bad_val");
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
|
|
|
|
// file2.sst (100 => 199)
|
|
|
|
std::string file2 = sst_files_dir_ + "file2.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file2));
|
|
|
|
for (int k = 100; k < 200; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
// Cannot add this key because it's not after last added key
|
2017-05-26 21:05:19 +02:00
|
|
|
s = sst_file_writer.Put(Key(99), "bad_val");
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
ExternalSstFileInfo file2_info;
|
|
|
|
s = sst_file_writer.Finish(&file2_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file2_info.file_path, file2);
|
|
|
|
ASSERT_EQ(file2_info.num_entries, 100);
|
|
|
|
ASSERT_EQ(file2_info.smallest_key, Key(100));
|
|
|
|
ASSERT_EQ(file2_info.largest_key, Key(199));
|
|
|
|
|
|
|
|
// file3.sst (195 => 299)
|
|
|
|
// This file values overlap with file2 values
|
|
|
|
std::string file3 = sst_files_dir_ + "file3.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file3));
|
|
|
|
for (int k = 195; k < 300; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val_overlap"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file3_info;
|
|
|
|
s = sst_file_writer.Finish(&file3_info);
|
2016-12-17 03:26:56 +01:00
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
2016-12-17 03:26:56 +01:00
|
|
|
// Current file size should be non-zero after success finish.
|
|
|
|
ASSERT_GT(sst_file_writer.FileSize(), 0);
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(file3_info.file_path, file3);
|
|
|
|
ASSERT_EQ(file3_info.num_entries, 105);
|
|
|
|
ASSERT_EQ(file3_info.smallest_key, Key(195));
|
|
|
|
ASSERT_EQ(file3_info.largest_key, Key(299));
|
|
|
|
|
|
|
|
// file4.sst (30 => 39)
|
|
|
|
// This file values overlap with file1 values
|
|
|
|
std::string file4 = sst_files_dir_ + "file4.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file4));
|
|
|
|
for (int k = 30; k < 40; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val_overlap"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file4_info;
|
|
|
|
s = sst_file_writer.Finish(&file4_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file4_info.file_path, file4);
|
|
|
|
ASSERT_EQ(file4_info.num_entries, 10);
|
|
|
|
ASSERT_EQ(file4_info.smallest_key, Key(30));
|
|
|
|
ASSERT_EQ(file4_info.largest_key, Key(39));
|
|
|
|
|
|
|
|
// file5.sst (400 => 499)
|
|
|
|
std::string file5 = sst_files_dir_ + "file5.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file5));
|
|
|
|
for (int k = 400; k < 500; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file5_info;
|
|
|
|
s = sst_file_writer.Finish(&file5_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file5_info.file_path, file5);
|
|
|
|
ASSERT_EQ(file5_info.num_entries, 100);
|
|
|
|
ASSERT_EQ(file5_info.smallest_key, Key(400));
|
|
|
|
ASSERT_EQ(file5_info.largest_key, Key(499));
|
|
|
|
|
2018-07-14 07:40:23 +02:00
|
|
|
// file6.sst (delete 400 => 500)
|
|
|
|
std::string file6 = sst_files_dir_ + "file6.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file6));
|
|
|
|
sst_file_writer.DeleteRange(Key(400), Key(500));
|
|
|
|
ExternalSstFileInfo file6_info;
|
|
|
|
s = sst_file_writer.Finish(&file6_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file6_info.file_path, file6);
|
|
|
|
ASSERT_EQ(file6_info.num_entries, 0);
|
|
|
|
ASSERT_EQ(file6_info.smallest_key, "");
|
|
|
|
ASSERT_EQ(file6_info.largest_key, "");
|
|
|
|
ASSERT_EQ(file6_info.num_range_del_entries, 1);
|
|
|
|
ASSERT_EQ(file6_info.smallest_range_del_key, Key(400));
|
|
|
|
ASSERT_EQ(file6_info.largest_range_del_key, Key(500));
|
|
|
|
|
|
|
|
// file7.sst (delete 500 => 570, put 520 => 599 divisible by 2)
|
|
|
|
std::string file7 = sst_files_dir_ + "file7.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file7));
|
|
|
|
sst_file_writer.DeleteRange(Key(500), Key(550));
|
|
|
|
for (int k = 520; k < 560; k += 2) {
|
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
|
|
|
}
|
|
|
|
sst_file_writer.DeleteRange(Key(525), Key(575));
|
|
|
|
for (int k = 560; k < 600; k += 2) {
|
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
|
|
|
}
|
|
|
|
ExternalSstFileInfo file7_info;
|
|
|
|
s = sst_file_writer.Finish(&file7_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file7_info.file_path, file7);
|
|
|
|
ASSERT_EQ(file7_info.num_entries, 40);
|
|
|
|
ASSERT_EQ(file7_info.smallest_key, Key(520));
|
|
|
|
ASSERT_EQ(file7_info.largest_key, Key(598));
|
|
|
|
ASSERT_EQ(file7_info.num_range_del_entries, 2);
|
|
|
|
ASSERT_EQ(file7_info.smallest_range_del_key, Key(500));
|
|
|
|
ASSERT_EQ(file7_info.largest_range_del_key, Key(575));
|
|
|
|
|
|
|
|
// file8.sst (delete 600 => 700)
|
|
|
|
std::string file8 = sst_files_dir_ + "file8.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file8));
|
|
|
|
sst_file_writer.DeleteRange(Key(600), Key(700));
|
|
|
|
ExternalSstFileInfo file8_info;
|
|
|
|
s = sst_file_writer.Finish(&file8_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file8_info.file_path, file8);
|
|
|
|
ASSERT_EQ(file8_info.num_entries, 0);
|
|
|
|
ASSERT_EQ(file8_info.smallest_key, "");
|
|
|
|
ASSERT_EQ(file8_info.largest_key, "");
|
|
|
|
ASSERT_EQ(file8_info.num_range_del_entries, 1);
|
|
|
|
ASSERT_EQ(file8_info.smallest_range_del_key, Key(600));
|
|
|
|
ASSERT_EQ(file8_info.largest_range_del_key, Key(700));
|
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
// Cannot create an empty sst file
|
|
|
|
std::string file_empty = sst_files_dir_ + "file_empty.sst";
|
|
|
|
ExternalSstFileInfo file_empty_info;
|
|
|
|
s = sst_file_writer.Finish(&file_empty_info);
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
// Add file using file path
|
2016-10-21 02:05:32 +02:00
|
|
|
s = DeprecatedAddFile({file1});
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(db_->GetLatestSequenceNumber(), 0U);
|
|
|
|
for (int k = 0; k < 100; k++) {
|
|
|
|
ASSERT_EQ(Get(Key(k)), Key(k) + "_val");
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add file while holding a snapshot will fail
|
|
|
|
const Snapshot* s1 = db_->GetSnapshot();
|
|
|
|
if (s1 != nullptr) {
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_NOK(DeprecatedAddFile({file2}));
|
2016-09-08 00:41:54 +02:00
|
|
|
db_->ReleaseSnapshot(s1);
|
|
|
|
}
|
|
|
|
// We can add the file after releaseing the snapshot
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile({file2}));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
ASSERT_EQ(db_->GetLatestSequenceNumber(), 0U);
|
|
|
|
for (int k = 0; k < 200; k++) {
|
|
|
|
ASSERT_EQ(Get(Key(k)), Key(k) + "_val");
|
|
|
|
}
|
|
|
|
|
2017-01-23 21:50:40 +01:00
|
|
|
// This file has overlapping values with the existing data
|
2016-10-21 02:05:32 +02:00
|
|
|
s = DeprecatedAddFile({file3});
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
|
2017-01-23 21:50:40 +01:00
|
|
|
// This file has overlapping values with the existing data
|
2016-10-21 02:05:32 +02:00
|
|
|
s = DeprecatedAddFile({file4});
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
|
|
|
|
// Overwrite values of keys divisible by 5
|
|
|
|
for (int k = 0; k < 200; k += 5) {
|
|
|
|
ASSERT_OK(Put(Key(k), Key(k) + "_val_new"));
|
|
|
|
}
|
|
|
|
ASSERT_NE(db_->GetLatestSequenceNumber(), 0U);
|
|
|
|
|
|
|
|
// Key range of file5 (400 => 499) dont overlap with any keys in DB
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile({file5}));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
2018-07-14 07:40:23 +02:00
|
|
|
// This file has overlapping values with the existing data
|
|
|
|
s = DeprecatedAddFile({file6});
|
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
|
|
|
|
// Key range of file7 (500 => 598) dont overlap with any keys in DB
|
|
|
|
ASSERT_OK(DeprecatedAddFile({file7}));
|
|
|
|
|
|
|
|
// Key range of file7 (600 => 700) dont overlap with any keys in DB
|
|
|
|
ASSERT_OK(DeprecatedAddFile({file8}));
|
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
// Make sure values are correct before and after flush/compaction
|
|
|
|
for (int i = 0; i < 2; i++) {
|
|
|
|
for (int k = 0; k < 200; k++) {
|
|
|
|
std::string value = Key(k) + "_val";
|
|
|
|
if (k % 5 == 0) {
|
|
|
|
value += "_new";
|
|
|
|
}
|
|
|
|
ASSERT_EQ(Get(Key(k)), value);
|
|
|
|
}
|
|
|
|
for (int k = 400; k < 500; k++) {
|
|
|
|
std::string value = Key(k) + "_val";
|
|
|
|
ASSERT_EQ(Get(Key(k)), value);
|
|
|
|
}
|
2018-07-14 07:40:23 +02:00
|
|
|
for (int k = 500; k < 600; k++) {
|
|
|
|
std::string value = Key(k) + "_val";
|
|
|
|
if (k < 520 || k % 2 == 1) {
|
|
|
|
value = "NOT_FOUND";
|
|
|
|
}
|
|
|
|
ASSERT_EQ(Get(Key(k)), value);
|
|
|
|
}
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
}
|
|
|
|
|
|
|
|
Close();
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
|
|
|
// Delete keys in range (400 => 499)
|
|
|
|
for (int k = 400; k < 500; k++) {
|
|
|
|
ASSERT_OK(Delete(Key(k)));
|
|
|
|
}
|
|
|
|
// We deleted range (400 => 499) but cannot add file5 because
|
|
|
|
// of the range tombstones
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_NOK(DeprecatedAddFile({file5}));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Compacting the DB will remove the tombstones
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
|
|
|
|
// Now we can add the file
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile({file5}));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Verify values of file5 in DB
|
|
|
|
for (int k = 400; k < 500; k++) {
|
|
|
|
std::string value = Key(k) + "_val";
|
|
|
|
ASSERT_EQ(Get(Key(k)), value);
|
|
|
|
}
|
|
|
|
DestroyAndRecreateExternalSSTFilesDir();
|
2018-07-14 07:40:23 +02:00
|
|
|
} while (ChangeOptions(kSkipPlainTable | kSkipFIFOCompaction |
|
|
|
|
kRangeDelSkipConfigs));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
2018-11-02 00:21:30 +01:00
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
class SstFileWriterCollector : public TablePropertiesCollector {
|
|
|
|
public:
|
|
|
|
explicit SstFileWriterCollector(const std::string prefix) : prefix_(prefix) {
|
|
|
|
name_ = prefix_ + "_SstFileWriterCollector";
|
|
|
|
}
|
|
|
|
|
|
|
|
const char* Name() const override { return name_.c_str(); }
|
|
|
|
|
|
|
|
Status Finish(UserCollectedProperties* properties) override {
|
2018-02-01 22:54:50 +01:00
|
|
|
std::string count = std::to_string(count_);
|
2016-09-08 00:41:54 +02:00
|
|
|
*properties = UserCollectedProperties{
|
|
|
|
{prefix_ + "_SstFileWriterCollector", "YES"},
|
2018-02-01 22:54:50 +01:00
|
|
|
{prefix_ + "_Count", count},
|
2016-09-08 00:41:54 +02:00
|
|
|
};
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
2018-03-05 22:08:17 +01:00
|
|
|
Status AddUserKey(const Slice& /*user_key*/, const Slice& /*value*/,
|
|
|
|
EntryType /*type*/, SequenceNumber /*seq*/,
|
|
|
|
uint64_t /*file_size*/) override {
|
2016-09-08 00:41:54 +02:00
|
|
|
++count_;
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
2019-02-14 22:52:47 +01:00
|
|
|
UserCollectedProperties GetReadableProperties() const override {
|
2016-09-08 00:41:54 +02:00
|
|
|
return UserCollectedProperties{};
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
uint32_t count_ = 0;
|
|
|
|
std::string prefix_;
|
|
|
|
std::string name_;
|
|
|
|
};
|
|
|
|
|
|
|
|
class SstFileWriterCollectorFactory : public TablePropertiesCollectorFactory {
|
|
|
|
public:
|
|
|
|
explicit SstFileWriterCollectorFactory(std::string prefix)
|
|
|
|
: prefix_(prefix), num_created_(0) {}
|
2019-02-14 22:52:47 +01:00
|
|
|
TablePropertiesCollector* CreateTablePropertiesCollector(
|
2018-03-05 22:08:17 +01:00
|
|
|
TablePropertiesCollectorFactory::Context /*context*/) override {
|
2016-09-08 00:41:54 +02:00
|
|
|
num_created_++;
|
|
|
|
return new SstFileWriterCollector(prefix_);
|
|
|
|
}
|
|
|
|
const char* Name() const override { return "SstFileWriterCollectorFactory"; }
|
|
|
|
|
|
|
|
std::string prefix_;
|
|
|
|
uint32_t num_created_;
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, AddList) {
|
|
|
|
do {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
|
|
|
|
auto abc_collector = std::make_shared<SstFileWriterCollectorFactory>("abc");
|
|
|
|
auto xyz_collector = std::make_shared<SstFileWriterCollectorFactory>("xyz");
|
|
|
|
|
|
|
|
options.table_properties_collector_factories.emplace_back(abc_collector);
|
|
|
|
options.table_properties_collector_factories.emplace_back(xyz_collector);
|
|
|
|
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// file1.sst (0 => 99)
|
|
|
|
std::string file1 = sst_files_dir_ + "file1.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file1));
|
|
|
|
for (int k = 0; k < 100; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file1_info;
|
|
|
|
Status s = sst_file_writer.Finish(&file1_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file1_info.file_path, file1);
|
|
|
|
ASSERT_EQ(file1_info.num_entries, 100);
|
|
|
|
ASSERT_EQ(file1_info.smallest_key, Key(0));
|
|
|
|
ASSERT_EQ(file1_info.largest_key, Key(99));
|
|
|
|
// sst_file_writer already finished, cannot add this value
|
2017-05-26 21:05:19 +02:00
|
|
|
s = sst_file_writer.Put(Key(100), "bad_val");
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
|
|
|
|
// file2.sst (100 => 199)
|
|
|
|
std::string file2 = sst_files_dir_ + "file2.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file2));
|
|
|
|
for (int k = 100; k < 200; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
// Cannot add this key because it's not after last added key
|
2017-05-26 21:05:19 +02:00
|
|
|
s = sst_file_writer.Put(Key(99), "bad_val");
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
ExternalSstFileInfo file2_info;
|
|
|
|
s = sst_file_writer.Finish(&file2_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file2_info.file_path, file2);
|
|
|
|
ASSERT_EQ(file2_info.num_entries, 100);
|
|
|
|
ASSERT_EQ(file2_info.smallest_key, Key(100));
|
|
|
|
ASSERT_EQ(file2_info.largest_key, Key(199));
|
|
|
|
|
|
|
|
// file3.sst (195 => 199)
|
|
|
|
// This file values overlap with file2 values
|
|
|
|
std::string file3 = sst_files_dir_ + "file3.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file3));
|
|
|
|
for (int k = 195; k < 200; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val_overlap"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file3_info;
|
|
|
|
s = sst_file_writer.Finish(&file3_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file3_info.file_path, file3);
|
|
|
|
ASSERT_EQ(file3_info.num_entries, 5);
|
|
|
|
ASSERT_EQ(file3_info.smallest_key, Key(195));
|
|
|
|
ASSERT_EQ(file3_info.largest_key, Key(199));
|
|
|
|
|
|
|
|
// file4.sst (30 => 39)
|
|
|
|
// This file values overlap with file1 values
|
|
|
|
std::string file4 = sst_files_dir_ + "file4.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file4));
|
|
|
|
for (int k = 30; k < 40; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val_overlap"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file4_info;
|
|
|
|
s = sst_file_writer.Finish(&file4_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file4_info.file_path, file4);
|
|
|
|
ASSERT_EQ(file4_info.num_entries, 10);
|
|
|
|
ASSERT_EQ(file4_info.smallest_key, Key(30));
|
|
|
|
ASSERT_EQ(file4_info.largest_key, Key(39));
|
|
|
|
|
|
|
|
// file5.sst (200 => 299)
|
|
|
|
std::string file5 = sst_files_dir_ + "file5.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file5));
|
|
|
|
for (int k = 200; k < 300; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file5_info;
|
|
|
|
s = sst_file_writer.Finish(&file5_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file5_info.file_path, file5);
|
|
|
|
ASSERT_EQ(file5_info.num_entries, 100);
|
|
|
|
ASSERT_EQ(file5_info.smallest_key, Key(200));
|
|
|
|
ASSERT_EQ(file5_info.largest_key, Key(299));
|
|
|
|
|
2018-07-14 07:40:23 +02:00
|
|
|
// file6.sst (delete 0 => 100)
|
|
|
|
std::string file6 = sst_files_dir_ + "file6.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file6));
|
|
|
|
ASSERT_OK(sst_file_writer.DeleteRange(Key(0), Key(75)));
|
|
|
|
ASSERT_OK(sst_file_writer.DeleteRange(Key(25), Key(100)));
|
|
|
|
ExternalSstFileInfo file6_info;
|
|
|
|
s = sst_file_writer.Finish(&file6_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file6_info.file_path, file6);
|
|
|
|
ASSERT_EQ(file6_info.num_entries, 0);
|
|
|
|
ASSERT_EQ(file6_info.smallest_key, "");
|
|
|
|
ASSERT_EQ(file6_info.largest_key, "");
|
|
|
|
ASSERT_EQ(file6_info.num_range_del_entries, 2);
|
|
|
|
ASSERT_EQ(file6_info.smallest_range_del_key, Key(0));
|
|
|
|
ASSERT_EQ(file6_info.largest_range_del_key, Key(100));
|
|
|
|
|
2019-08-15 05:58:59 +02:00
|
|
|
// file7.sst (delete 99 => 201)
|
2018-07-14 07:40:23 +02:00
|
|
|
std::string file7 = sst_files_dir_ + "file7.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file7));
|
2019-08-15 05:58:59 +02:00
|
|
|
ASSERT_OK(sst_file_writer.DeleteRange(Key(99), Key(201)));
|
2018-07-14 07:40:23 +02:00
|
|
|
ExternalSstFileInfo file7_info;
|
|
|
|
s = sst_file_writer.Finish(&file7_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file7_info.file_path, file7);
|
|
|
|
ASSERT_EQ(file7_info.num_entries, 0);
|
|
|
|
ASSERT_EQ(file7_info.smallest_key, "");
|
|
|
|
ASSERT_EQ(file7_info.largest_key, "");
|
|
|
|
ASSERT_EQ(file7_info.num_range_del_entries, 1);
|
2019-08-15 05:58:59 +02:00
|
|
|
ASSERT_EQ(file7_info.smallest_range_del_key, Key(99));
|
|
|
|
ASSERT_EQ(file7_info.largest_range_del_key, Key(201));
|
2018-07-14 07:40:23 +02:00
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
// list 1 has internal key range conflict
|
|
|
|
std::vector<std::string> file_list0({file1, file2});
|
|
|
|
std::vector<std::string> file_list1({file3, file2, file1});
|
|
|
|
std::vector<std::string> file_list2({file5});
|
|
|
|
std::vector<std::string> file_list3({file3, file4});
|
2018-07-14 07:40:23 +02:00
|
|
|
std::vector<std::string> file_list4({file5, file7});
|
|
|
|
std::vector<std::string> file_list5({file6, file7});
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
2018-07-14 07:40:23 +02:00
|
|
|
// These lists of files have key ranges that overlap with each other
|
2016-10-21 02:05:32 +02:00
|
|
|
s = DeprecatedAddFile(file_list1);
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
2019-08-15 05:58:59 +02:00
|
|
|
// Both of the following overlap on the range deletion tombstone.
|
2018-07-14 07:40:23 +02:00
|
|
|
s = DeprecatedAddFile(file_list4);
|
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
s = DeprecatedAddFile(file_list5);
|
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Add files using file path list
|
2016-10-21 02:05:32 +02:00
|
|
|
s = DeprecatedAddFile(file_list0);
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(db_->GetLatestSequenceNumber(), 0U);
|
|
|
|
for (int k = 0; k < 200; k++) {
|
|
|
|
ASSERT_EQ(Get(Key(k)), Key(k) + "_val");
|
|
|
|
}
|
|
|
|
|
|
|
|
TablePropertiesCollection props;
|
|
|
|
ASSERT_OK(db_->GetPropertiesOfAllTables(&props));
|
|
|
|
ASSERT_EQ(props.size(), 2);
|
|
|
|
for (auto file_props : props) {
|
|
|
|
auto user_props = file_props.second->user_collected_properties;
|
|
|
|
ASSERT_EQ(user_props["abc_SstFileWriterCollector"], "YES");
|
|
|
|
ASSERT_EQ(user_props["xyz_SstFileWriterCollector"], "YES");
|
|
|
|
ASSERT_EQ(user_props["abc_Count"], "100");
|
|
|
|
ASSERT_EQ(user_props["xyz_Count"], "100");
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add file while holding a snapshot will fail
|
|
|
|
const Snapshot* s1 = db_->GetSnapshot();
|
|
|
|
if (s1 != nullptr) {
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_NOK(DeprecatedAddFile(file_list2));
|
2016-09-08 00:41:54 +02:00
|
|
|
db_->ReleaseSnapshot(s1);
|
|
|
|
}
|
|
|
|
// We can add the file after releaseing the snapshot
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile(file_list2));
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(db_->GetLatestSequenceNumber(), 0U);
|
|
|
|
for (int k = 0; k < 300; k++) {
|
|
|
|
ASSERT_EQ(Get(Key(k)), Key(k) + "_val");
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT_OK(db_->GetPropertiesOfAllTables(&props));
|
|
|
|
ASSERT_EQ(props.size(), 3);
|
|
|
|
for (auto file_props : props) {
|
|
|
|
auto user_props = file_props.second->user_collected_properties;
|
|
|
|
ASSERT_EQ(user_props["abc_SstFileWriterCollector"], "YES");
|
|
|
|
ASSERT_EQ(user_props["xyz_SstFileWriterCollector"], "YES");
|
|
|
|
ASSERT_EQ(user_props["abc_Count"], "100");
|
|
|
|
ASSERT_EQ(user_props["xyz_Count"], "100");
|
|
|
|
}
|
|
|
|
|
2017-01-23 21:50:40 +01:00
|
|
|
// This file list has overlapping values with the existing data
|
2016-10-21 02:05:32 +02:00
|
|
|
s = DeprecatedAddFile(file_list3);
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_FALSE(s.ok()) << s.ToString();
|
|
|
|
|
|
|
|
// Overwrite values of keys divisible by 5
|
|
|
|
for (int k = 0; k < 200; k += 5) {
|
|
|
|
ASSERT_OK(Put(Key(k), Key(k) + "_val_new"));
|
|
|
|
}
|
|
|
|
ASSERT_NE(db_->GetLatestSequenceNumber(), 0U);
|
|
|
|
|
|
|
|
// Make sure values are correct before and after flush/compaction
|
|
|
|
for (int i = 0; i < 2; i++) {
|
|
|
|
for (int k = 0; k < 200; k++) {
|
|
|
|
std::string value = Key(k) + "_val";
|
|
|
|
if (k % 5 == 0) {
|
|
|
|
value += "_new";
|
|
|
|
}
|
|
|
|
ASSERT_EQ(Get(Key(k)), value);
|
|
|
|
}
|
|
|
|
for (int k = 200; k < 300; k++) {
|
|
|
|
std::string value = Key(k) + "_val";
|
|
|
|
ASSERT_EQ(Get(Key(k)), value);
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete keys in range (200 => 299)
|
|
|
|
for (int k = 200; k < 300; k++) {
|
|
|
|
ASSERT_OK(Delete(Key(k)));
|
|
|
|
}
|
|
|
|
// We deleted range (200 => 299) but cannot add file5 because
|
|
|
|
// of the range tombstones
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_NOK(DeprecatedAddFile(file_list2));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Compacting the DB will remove the tombstones
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
|
|
|
|
// Now we can add the file
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile(file_list2));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Verify values of file5 in DB
|
|
|
|
for (int k = 200; k < 300; k++) {
|
|
|
|
std::string value = Key(k) + "_val";
|
|
|
|
ASSERT_EQ(Get(Key(k)), value);
|
|
|
|
}
|
|
|
|
DestroyAndRecreateExternalSSTFilesDir();
|
2018-07-14 07:40:23 +02:00
|
|
|
} while (ChangeOptions(kSkipPlainTable | kSkipFIFOCompaction |
|
|
|
|
kRangeDelSkipConfigs));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, AddListAtomicity) {
|
|
|
|
do {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// files[0].sst (0 => 99)
|
|
|
|
// files[1].sst (100 => 199)
|
|
|
|
// ...
|
|
|
|
// file[8].sst (800 => 899)
|
|
|
|
int n = 9;
|
|
|
|
std::vector<std::string> files(n);
|
|
|
|
std::vector<ExternalSstFileInfo> files_info(n);
|
|
|
|
for (int i = 0; i < n; i++) {
|
|
|
|
files[i] = sst_files_dir_ + "file" + std::to_string(i) + ".sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(files[i]));
|
|
|
|
for (int k = i * 100; k < (i + 1) * 100; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
Status s = sst_file_writer.Finish(&files_info[i]);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(files_info[i].file_path, files[i]);
|
|
|
|
ASSERT_EQ(files_info[i].num_entries, 100);
|
|
|
|
ASSERT_EQ(files_info[i].smallest_key, Key(i * 100));
|
|
|
|
ASSERT_EQ(files_info[i].largest_key, Key((i + 1) * 100 - 1));
|
|
|
|
}
|
|
|
|
files.push_back(sst_files_dir_ + "file" + std::to_string(n) + ".sst");
|
2016-10-21 02:05:32 +02:00
|
|
|
auto s = DeprecatedAddFile(files);
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_NOK(s) << s.ToString();
|
|
|
|
for (int k = 0; k < n * 100; k++) {
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get(Key(k)));
|
|
|
|
}
|
2016-10-21 02:05:32 +02:00
|
|
|
files.pop_back();
|
|
|
|
ASSERT_OK(DeprecatedAddFile(files));
|
2016-09-08 00:41:54 +02:00
|
|
|
for (int k = 0; k < n * 100; k++) {
|
|
|
|
std::string value = Key(k) + "_val";
|
|
|
|
ASSERT_EQ(Get(Key(k)), value);
|
|
|
|
}
|
|
|
|
DestroyAndRecreateExternalSSTFilesDir();
|
2017-04-26 22:28:39 +02:00
|
|
|
} while (ChangeOptions(kSkipPlainTable | kSkipFIFOCompaction));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
// This test reporduce a bug that can happen in some cases if the DB started
|
|
|
|
// purging obsolete files when we are adding an external sst file.
|
|
|
|
// This situation may result in deleting the file while it's being added.
|
|
|
|
TEST_F(ExternalSSTFileTest, PurgeObsoleteFilesBug) {
|
|
|
|
Options options = CurrentOptions();
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// file1.sst (0 => 500)
|
|
|
|
std::string sst_file_path = sst_files_dir_ + "file1.sst";
|
|
|
|
Status s = sst_file_writer.Open(sst_file_path);
|
|
|
|
ASSERT_OK(s);
|
|
|
|
for (int i = 0; i < 500; i++) {
|
|
|
|
std::string k = Key(i);
|
2017-05-26 21:05:19 +02:00
|
|
|
s = sst_file_writer.Put(k, k + "_val");
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_OK(s);
|
|
|
|
}
|
|
|
|
|
|
|
|
ExternalSstFileInfo sst_file_info;
|
|
|
|
s = sst_file_writer.Finish(&sst_file_info);
|
|
|
|
ASSERT_OK(s);
|
|
|
|
|
|
|
|
options.delete_obsolete_files_period_micros = 0;
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
2018-04-13 19:47:54 +02:00
|
|
|
"ExternalSstFileIngestionJob::Prepare:FileAdded", [&](void* /* arg */) {
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_OK(Put("aaa", "bbb"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_OK(Put("aaa", "xxx"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
db_->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
|
|
});
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
s = DeprecatedAddFile({sst_file_path});
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_OK(s);
|
|
|
|
|
|
|
|
for (int i = 0; i < 500; i++) {
|
|
|
|
std::string k = Key(i);
|
|
|
|
std::string v = k + "_val";
|
|
|
|
ASSERT_EQ(Get(k), v);
|
|
|
|
}
|
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, SkipSnapshot) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// file1.sst (0 => 99)
|
|
|
|
std::string file1 = sst_files_dir_ + "file1.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file1));
|
|
|
|
for (int k = 0; k < 100; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file1_info;
|
|
|
|
Status s = sst_file_writer.Finish(&file1_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file1_info.file_path, file1);
|
|
|
|
ASSERT_EQ(file1_info.num_entries, 100);
|
|
|
|
ASSERT_EQ(file1_info.smallest_key, Key(0));
|
|
|
|
ASSERT_EQ(file1_info.largest_key, Key(99));
|
|
|
|
|
|
|
|
// file2.sst (100 => 299)
|
|
|
|
std::string file2 = sst_files_dir_ + "file2.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file2));
|
|
|
|
for (int k = 100; k < 300; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file2_info;
|
|
|
|
s = sst_file_writer.Finish(&file2_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file2_info.file_path, file2);
|
|
|
|
ASSERT_EQ(file2_info.num_entries, 200);
|
|
|
|
ASSERT_EQ(file2_info.smallest_key, Key(100));
|
|
|
|
ASSERT_EQ(file2_info.largest_key, Key(299));
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile({file1}));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Add file will fail when holding snapshot and use the default
|
|
|
|
// skip_snapshot_check to false
|
|
|
|
const Snapshot* s1 = db_->GetSnapshot();
|
|
|
|
if (s1 != nullptr) {
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_NOK(DeprecatedAddFile({file2}));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Add file will success when set skip_snapshot_check to true even db holding
|
|
|
|
// snapshot
|
|
|
|
if (s1 != nullptr) {
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile({file2}, false, true));
|
2016-09-08 00:41:54 +02:00
|
|
|
db_->ReleaseSnapshot(s1);
|
|
|
|
}
|
|
|
|
|
|
|
|
// file3.sst (300 => 399)
|
|
|
|
std::string file3 = sst_files_dir_ + "file3.sst";
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file3));
|
|
|
|
for (int k = 300; k < 400; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k) + "_val"));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file3_info;
|
|
|
|
s = sst_file_writer.Finish(&file3_info);
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
ASSERT_EQ(file3_info.file_path, file3);
|
|
|
|
ASSERT_EQ(file3_info.num_entries, 100);
|
|
|
|
ASSERT_EQ(file3_info.smallest_key, Key(300));
|
|
|
|
ASSERT_EQ(file3_info.largest_key, Key(399));
|
|
|
|
|
|
|
|
// check that we have change the old key
|
|
|
|
ASSERT_EQ(Get(Key(300)), "NOT_FOUND");
|
|
|
|
const Snapshot* s2 = db_->GetSnapshot();
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile({file3}, false, true));
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(Get(Key(300)), Key(300) + ("_val"));
|
|
|
|
ASSERT_EQ(Get(Key(300), s2), Key(300) + ("_val"));
|
|
|
|
|
|
|
|
db_->ReleaseSnapshot(s2);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, MultiThreaded) {
|
|
|
|
// Bulk load 10 files every file contain 1000 keys
|
|
|
|
int num_files = 10;
|
|
|
|
int keys_per_file = 1000;
|
|
|
|
|
|
|
|
// Generate file names
|
|
|
|
std::vector<std::string> file_names;
|
|
|
|
for (int i = 0; i < num_files; i++) {
|
|
|
|
std::string file_name = "file_" + ToString(i) + ".sst";
|
|
|
|
file_names.push_back(sst_files_dir_ + file_name);
|
|
|
|
}
|
|
|
|
|
|
|
|
do {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
|
|
|
|
std::atomic<int> thread_num(0);
|
|
|
|
std::function<void()> write_file_func = [&]() {
|
|
|
|
int file_idx = thread_num.fetch_add(1);
|
|
|
|
int range_start = file_idx * keys_per_file;
|
|
|
|
int range_end = range_start + keys_per_file;
|
|
|
|
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file_names[file_idx]));
|
|
|
|
|
|
|
|
for (int k = range_start; k < range_end; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k)));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
Status s = sst_file_writer.Finish();
|
|
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
|
|
};
|
|
|
|
// Write num_files files in parallel
|
2017-02-06 23:43:55 +01:00
|
|
|
std::vector<port::Thread> sst_writer_threads;
|
2016-09-08 00:41:54 +02:00
|
|
|
for (int i = 0; i < num_files; ++i) {
|
|
|
|
sst_writer_threads.emplace_back(write_file_func);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (auto& t : sst_writer_threads) {
|
|
|
|
t.join();
|
|
|
|
}
|
|
|
|
|
|
|
|
fprintf(stderr, "Wrote %d files (%d keys)\n", num_files,
|
|
|
|
num_files * keys_per_file);
|
|
|
|
|
|
|
|
thread_num.store(0);
|
|
|
|
std::atomic<int> files_added(0);
|
|
|
|
// Thread 0 -> Load {f0,f1}
|
|
|
|
// Thread 1 -> Load {f0,f1}
|
|
|
|
// Thread 2 -> Load {f2,f3}
|
|
|
|
// Thread 3 -> Load {f2,f3}
|
|
|
|
// Thread 4 -> Load {f4,f5}
|
|
|
|
// Thread 5 -> Load {f4,f5}
|
|
|
|
// ...
|
|
|
|
std::function<void()> load_file_func = [&]() {
|
|
|
|
// We intentionally add every file twice, and assert that it was added
|
|
|
|
// only once and the other add failed
|
|
|
|
int thread_id = thread_num.fetch_add(1);
|
|
|
|
int file_idx = (thread_id / 2) * 2;
|
|
|
|
// sometimes we use copy, sometimes link .. the result should be the same
|
|
|
|
bool move_file = (thread_id % 3 == 0);
|
|
|
|
|
|
|
|
std::vector<std::string> files_to_add;
|
|
|
|
|
|
|
|
files_to_add = {file_names[file_idx]};
|
|
|
|
if (static_cast<size_t>(file_idx + 1) < file_names.size()) {
|
|
|
|
files_to_add.push_back(file_names[file_idx + 1]);
|
|
|
|
}
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
Status s = DeprecatedAddFile(files_to_add, move_file);
|
2016-09-08 00:41:54 +02:00
|
|
|
if (s.ok()) {
|
|
|
|
files_added += static_cast<int>(files_to_add.size());
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// Bulk load num_files files in parallel
|
2017-02-06 23:43:55 +01:00
|
|
|
std::vector<port::Thread> add_file_threads;
|
2016-09-08 00:41:54 +02:00
|
|
|
DestroyAndReopen(options);
|
|
|
|
for (int i = 0; i < num_files; ++i) {
|
|
|
|
add_file_threads.emplace_back(load_file_func);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (auto& t : add_file_threads) {
|
|
|
|
t.join();
|
|
|
|
}
|
|
|
|
ASSERT_EQ(files_added.load(), num_files);
|
|
|
|
fprintf(stderr, "Loaded %d files (%d keys)\n", num_files,
|
|
|
|
num_files * keys_per_file);
|
|
|
|
|
|
|
|
// Overwrite values of keys divisible by 100
|
|
|
|
for (int k = 0; k < num_files * keys_per_file; k += 100) {
|
|
|
|
std::string key = Key(k);
|
|
|
|
Status s = Put(key, key + "_new");
|
|
|
|
ASSERT_TRUE(s.ok());
|
|
|
|
}
|
|
|
|
|
|
|
|
for (int i = 0; i < 2; i++) {
|
|
|
|
// Make sure the values are correct before and after flush/compaction
|
|
|
|
for (int k = 0; k < num_files * keys_per_file; ++k) {
|
|
|
|
std::string key = Key(k);
|
|
|
|
std::string value = (k % 100 == 0) ? (key + "_new") : key;
|
|
|
|
ASSERT_EQ(Get(key), value);
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
}
|
|
|
|
|
|
|
|
fprintf(stderr, "Verified %d values\n", num_files * keys_per_file);
|
|
|
|
DestroyAndRecreateExternalSSTFilesDir();
|
2017-04-26 22:28:39 +02:00
|
|
|
} while (ChangeOptions(kSkipPlainTable | kSkipFIFOCompaction));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, OverlappingRanges) {
|
|
|
|
Random rnd(301);
|
2018-06-18 23:38:50 +02:00
|
|
|
SequenceNumber assigned_seqno = 0;
|
2017-04-26 22:28:39 +02:00
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
2018-06-18 23:38:50 +02:00
|
|
|
"ExternalSstFileIngestionJob::Run", [&assigned_seqno](void* arg) {
|
2017-04-26 22:28:39 +02:00
|
|
|
ASSERT_TRUE(arg != nullptr);
|
2018-06-18 23:38:50 +02:00
|
|
|
assigned_seqno = *(static_cast<SequenceNumber*>(arg));
|
2017-04-26 22:28:39 +02:00
|
|
|
});
|
|
|
|
bool need_flush = false;
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"DBImpl::IngestExternalFile:NeedFlush", [&need_flush](void* arg) {
|
|
|
|
ASSERT_TRUE(arg != nullptr);
|
|
|
|
need_flush = *(static_cast<bool*>(arg));
|
|
|
|
});
|
|
|
|
bool overlap_with_db = false;
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"ExternalSstFileIngestionJob::AssignLevelAndSeqnoForIngestedFile",
|
|
|
|
[&overlap_with_db](void* arg) {
|
|
|
|
ASSERT_TRUE(arg != nullptr);
|
|
|
|
overlap_with_db = *(static_cast<bool*>(arg));
|
|
|
|
});
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
2016-09-08 00:41:54 +02:00
|
|
|
do {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
printf("Option config = %d\n", option_config_);
|
|
|
|
std::vector<std::pair<int, int>> key_ranges;
|
2018-07-13 02:40:31 +02:00
|
|
|
for (int i = 0; i < 100; i++) {
|
2016-09-08 00:41:54 +02:00
|
|
|
int range_start = rnd.Uniform(20000);
|
|
|
|
int keys_per_range = 10 + rnd.Uniform(41);
|
|
|
|
|
|
|
|
key_ranges.emplace_back(range_start, range_start + keys_per_range);
|
|
|
|
}
|
|
|
|
|
|
|
|
int memtable_add = 0;
|
|
|
|
int success_add_file = 0;
|
|
|
|
int failed_add_file = 0;
|
|
|
|
std::map<std::string, std::string> true_data;
|
|
|
|
for (size_t i = 0; i < key_ranges.size(); i++) {
|
|
|
|
int range_start = key_ranges[i].first;
|
|
|
|
int range_end = key_ranges[i].second;
|
|
|
|
|
|
|
|
Status s;
|
|
|
|
std::string range_val = "range_" + ToString(i);
|
|
|
|
|
|
|
|
// For 20% of ranges we use DB::Put, for 80% we use DB::AddFile
|
|
|
|
if (i && i % 5 == 0) {
|
|
|
|
// Use DB::Put to insert range (insert into memtable)
|
|
|
|
range_val += "_put";
|
|
|
|
for (int k = range_start; k <= range_end; k++) {
|
|
|
|
s = Put(Key(k), range_val);
|
|
|
|
ASSERT_OK(s);
|
|
|
|
}
|
|
|
|
memtable_add++;
|
|
|
|
} else {
|
|
|
|
// Use DB::AddFile to insert range
|
|
|
|
range_val += "_add_file";
|
|
|
|
|
|
|
|
// Generate the file containing the range
|
|
|
|
std::string file_name = sst_files_dir_ + env_->GenerateUniqueId();
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file_name));
|
|
|
|
for (int k = range_start; k <= range_end; k++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
s = sst_file_writer.Put(Key(k), range_val);
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_OK(s);
|
|
|
|
}
|
|
|
|
ExternalSstFileInfo file_info;
|
|
|
|
s = sst_file_writer.Finish(&file_info);
|
|
|
|
ASSERT_OK(s);
|
|
|
|
|
|
|
|
// Insert the generated file
|
2016-10-21 02:05:32 +02:00
|
|
|
s = DeprecatedAddFile({file_name});
|
2016-09-08 00:41:54 +02:00
|
|
|
auto it = true_data.lower_bound(Key(range_start));
|
2017-04-26 22:28:39 +02:00
|
|
|
if (option_config_ != kUniversalCompaction &&
|
2018-07-13 20:08:57 +02:00
|
|
|
option_config_ != kUniversalCompactionMultiLevel &&
|
|
|
|
option_config_ != kUniversalSubcompactions) {
|
2017-04-26 22:28:39 +02:00
|
|
|
if (it != true_data.end() && it->first <= Key(range_end)) {
|
|
|
|
// This range overlap with data already exist in DB
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
failed_add_file++;
|
|
|
|
} else {
|
|
|
|
ASSERT_OK(s);
|
|
|
|
success_add_file++;
|
|
|
|
}
|
2016-09-08 00:41:54 +02:00
|
|
|
} else {
|
2017-04-26 22:28:39 +02:00
|
|
|
if ((it != true_data.end() && it->first <= Key(range_end)) ||
|
2018-06-18 23:38:50 +02:00
|
|
|
need_flush || assigned_seqno > 0 || overlap_with_db) {
|
2017-04-26 22:28:39 +02:00
|
|
|
// This range overlap with data already exist in DB
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
failed_add_file++;
|
|
|
|
} else {
|
|
|
|
ASSERT_OK(s);
|
|
|
|
success_add_file++;
|
|
|
|
}
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s.ok()) {
|
|
|
|
// Update true_data map to include the new inserted data
|
|
|
|
for (int k = range_start; k <= range_end; k++) {
|
|
|
|
true_data[Key(k)] = range_val;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Flush / Compact the DB
|
|
|
|
if (i && i % 50 == 0) {
|
|
|
|
Flush();
|
|
|
|
}
|
|
|
|
if (i && i % 75 == 0) {
|
|
|
|
db_->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
printf("Total: %" ROCKSDB_PRIszt
|
|
|
|
" ranges\n"
|
|
|
|
"AddFile()|Success: %d ranges\n"
|
|
|
|
"AddFile()|RangeConflict: %d ranges\n"
|
|
|
|
"Put(): %d ranges\n",
|
|
|
|
key_ranges.size(), success_add_file, failed_add_file, memtable_add);
|
|
|
|
|
|
|
|
// Verify the correctness of the data
|
|
|
|
for (const auto& kv : true_data) {
|
|
|
|
ASSERT_EQ(Get(kv.first), kv.second);
|
|
|
|
}
|
|
|
|
printf("keys/values verified\n");
|
|
|
|
DestroyAndRecreateExternalSSTFilesDir();
|
2017-04-26 22:28:39 +02:00
|
|
|
} while (ChangeOptions(kSkipPlainTable | kSkipFIFOCompaction));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
|
2018-11-02 00:21:30 +01:00
|
|
|
TEST_P(ExternalSSTFileTest, PickedLevel) {
|
2016-09-08 00:41:54 +02:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.disable_auto_compactions = false;
|
|
|
|
options.level0_file_num_compaction_trigger = 4;
|
|
|
|
options.num_levels = 4;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
std::map<std::string, std::string> true_data;
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// File 0 will go to last level (L3)
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {1, 10}, -1, false, false, true,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
EXPECT_EQ(FilesPerLevel(), "0,0,0,1");
|
|
|
|
|
|
|
|
// File 1 will go to level L2 (since it overlap with file 0 in L3)
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {2, 9}, -1, false, false, true,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
EXPECT_EQ(FilesPerLevel(), "0,0,1,1");
|
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->LoadDependency({
|
|
|
|
{"ExternalSSTFileTest::PickedLevel:0", "BackgroundCallCompaction:0"},
|
|
|
|
{"DBImpl::BackgroundCompaction:Start",
|
|
|
|
"ExternalSSTFileTest::PickedLevel:1"},
|
|
|
|
{"ExternalSSTFileTest::PickedLevel:2",
|
|
|
|
"DBImpl::BackgroundCompaction:NonTrivial:AfterRun"},
|
|
|
|
});
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
// Flush 4 files containing the same keys
|
|
|
|
for (int i = 0; i < 4; i++) {
|
|
|
|
ASSERT_OK(Put(Key(3), Key(3) + "put"));
|
|
|
|
ASSERT_OK(Put(Key(8), Key(8) + "put"));
|
2016-10-21 02:05:32 +02:00
|
|
|
true_data[Key(3)] = Key(3) + "put";
|
|
|
|
true_data[Key(8)] = Key(8) + "put";
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_OK(Flush());
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for BackgroundCompaction() to be called
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevel:0");
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevel:1");
|
|
|
|
|
|
|
|
EXPECT_EQ(FilesPerLevel(), "4,0,1,1");
|
|
|
|
|
|
|
|
// This file overlaps with file 0 (L3), file 1 (L2) and the
|
|
|
|
// output of compaction going to L1
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {4, 7}, -1, false, false, true,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
EXPECT_EQ(FilesPerLevel(), "5,0,1,1");
|
|
|
|
|
|
|
|
// This file does not overlap with any file or with the running compaction
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {9000, 9001}, -1, false, false,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
EXPECT_EQ(FilesPerLevel(), "5,0,1,2");
|
|
|
|
|
|
|
|
// Hold compaction from finishing
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevel:2");
|
|
|
|
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
EXPECT_EQ(FilesPerLevel(), "1,1,1,2");
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
size_t kcnt = 0;
|
|
|
|
VerifyDBFromMap(true_data, &kcnt, false);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
}
|
|
|
|
|
2016-09-26 23:44:48 +02:00
|
|
|
TEST_F(ExternalSSTFileTest, PickedLevelBug) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.disable_auto_compactions = false;
|
|
|
|
options.level0_file_num_compaction_trigger = 3;
|
|
|
|
options.num_levels = 2;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
std::vector<int> file_keys;
|
|
|
|
|
|
|
|
// file #1 in L0
|
|
|
|
file_keys = {0, 5, 7};
|
|
|
|
for (int k : file_keys) {
|
|
|
|
ASSERT_OK(Put(Key(k), Key(k)));
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
// file #2 in L0
|
|
|
|
file_keys = {4, 6, 8, 9};
|
|
|
|
for (int k : file_keys) {
|
|
|
|
ASSERT_OK(Put(Key(k), Key(k)));
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
// We have 2 overlapping files in L0
|
|
|
|
EXPECT_EQ(FilesPerLevel(), "2");
|
|
|
|
|
2016-09-29 00:42:06 +02:00
|
|
|
rocksdb::SyncPoint::GetInstance()->LoadDependency(
|
|
|
|
{{"DBImpl::AddFile:MutexLock", "ExternalSSTFileTest::PickedLevelBug:0"},
|
|
|
|
{"ExternalSSTFileTest::PickedLevelBug:1", "DBImpl::AddFile:MutexUnlock"},
|
|
|
|
{"ExternalSSTFileTest::PickedLevelBug:2",
|
|
|
|
"DBImpl::RunManualCompaction:0"},
|
|
|
|
{"ExternalSSTFileTest::PickedLevelBug:3",
|
|
|
|
"DBImpl::RunManualCompaction:1"}});
|
2016-09-26 23:44:48 +02:00
|
|
|
|
|
|
|
std::atomic<bool> bg_compact_started(false);
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"DBImpl::BackgroundCompaction:Start",
|
2018-04-13 02:55:14 +02:00
|
|
|
[&](void* /*arg*/) { bg_compact_started.store(true); });
|
2016-09-26 23:44:48 +02:00
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
2016-09-29 00:42:06 +02:00
|
|
|
// While writing the MANIFEST start a thread that will ask for compaction
|
2017-02-06 23:43:55 +01:00
|
|
|
rocksdb::port::Thread bg_compact([&]() {
|
2016-09-29 00:42:06 +02:00
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
});
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevelBug:2");
|
|
|
|
|
2016-09-26 23:44:48 +02:00
|
|
|
// Start a thread that will ingest a new file
|
2017-02-06 23:43:55 +01:00
|
|
|
rocksdb::port::Thread bg_addfile([&]() {
|
2016-09-26 23:44:48 +02:00
|
|
|
file_keys = {1, 2, 3};
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, file_keys, 1));
|
|
|
|
});
|
|
|
|
|
|
|
|
// Wait for AddFile to start picking levels and writing MANIFEST
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevelBug:0");
|
|
|
|
|
2016-09-29 00:42:06 +02:00
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevelBug:3");
|
2016-09-26 23:44:48 +02:00
|
|
|
|
|
|
|
// We need to verify that no compactions can run while AddFile is
|
|
|
|
// ingesting the files into the levels it find suitable. So we will
|
|
|
|
// wait for 2 seconds to give a chance for compactions to run during
|
|
|
|
// this period, and then make sure that no compactions where able to run
|
|
|
|
env_->SleepForMicroseconds(1000000 * 2);
|
|
|
|
ASSERT_FALSE(bg_compact_started.load());
|
|
|
|
|
|
|
|
// Hold AddFile from finishing writing the MANIFEST
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevelBug:1");
|
|
|
|
|
|
|
|
bg_addfile.join();
|
|
|
|
bg_compact.join();
|
|
|
|
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
|
|
|
|
int total_keys = 0;
|
|
|
|
Iterator* iter = db_->NewIterator(ReadOptions());
|
|
|
|
for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
|
|
|
|
ASSERT_OK(iter->status());
|
|
|
|
total_keys++;
|
|
|
|
}
|
|
|
|
ASSERT_EQ(total_keys, 10);
|
|
|
|
|
|
|
|
delete iter;
|
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
}
|
|
|
|
|
2018-07-17 23:04:18 +02:00
|
|
|
TEST_F(ExternalSSTFileTest, IngestNonExistingFile) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
Status s = db_->IngestExternalFile({"non_existing_file"},
|
|
|
|
IngestExternalFileOptions());
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
|
|
|
|
// Verify file deletion is not impacted (verify a bug fix)
|
|
|
|
ASSERT_OK(Put(Key(1), Key(1)));
|
|
|
|
ASSERT_OK(Put(Key(9), Key(9)));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
ASSERT_OK(Put(Key(1), Key(1)));
|
|
|
|
ASSERT_OK(Put(Key(9), Key(9)));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
2018-11-03 01:22:37 +01:00
|
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact(true));
|
2019-01-30 01:16:53 +01:00
|
|
|
|
2018-07-17 23:04:18 +02:00
|
|
|
// After full compaction, there should be only 1 file.
|
|
|
|
std::vector<std::string> files;
|
|
|
|
env_->GetChildren(dbname_, &files);
|
|
|
|
int num_sst_files = 0;
|
|
|
|
for (auto& f : files) {
|
|
|
|
uint64_t number;
|
|
|
|
FileType type;
|
|
|
|
if (ParseFileName(f, &number, &type) && type == kTableFile) {
|
|
|
|
num_sst_files++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ASSERT_EQ(1, num_sst_files);
|
|
|
|
}
|
|
|
|
|
2016-09-29 00:42:06 +02:00
|
|
|
TEST_F(ExternalSSTFileTest, CompactDuringAddFileRandom) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.disable_auto_compactions = false;
|
|
|
|
options.level0_file_num_compaction_trigger = 2;
|
|
|
|
options.num_levels = 2;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
std::function<void()> bg_compact = [&]() {
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
};
|
|
|
|
|
|
|
|
int range_id = 0;
|
|
|
|
std::vector<int> file_keys;
|
|
|
|
std::function<void()> bg_addfile = [&]() {
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, file_keys, range_id));
|
|
|
|
};
|
|
|
|
|
2018-07-12 23:39:07 +02:00
|
|
|
const int num_of_ranges = 1000;
|
2017-02-06 23:43:55 +01:00
|
|
|
std::vector<port::Thread> threads;
|
2018-07-12 23:39:07 +02:00
|
|
|
while (range_id < num_of_ranges) {
|
2016-10-19 01:59:37 +02:00
|
|
|
int range_start = range_id * 10;
|
2016-09-29 00:42:06 +02:00
|
|
|
int range_end = range_start + 10;
|
|
|
|
|
|
|
|
file_keys.clear();
|
|
|
|
for (int k = range_start + 1; k < range_end; k++) {
|
|
|
|
file_keys.push_back(k);
|
|
|
|
}
|
|
|
|
ASSERT_OK(Put(Key(range_start), Key(range_start)));
|
|
|
|
ASSERT_OK(Put(Key(range_end), Key(range_end)));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
if (range_id % 10 == 0) {
|
|
|
|
threads.emplace_back(bg_compact);
|
|
|
|
}
|
|
|
|
threads.emplace_back(bg_addfile);
|
|
|
|
|
|
|
|
for (auto& t : threads) {
|
|
|
|
t.join();
|
|
|
|
}
|
|
|
|
threads.clear();
|
|
|
|
|
|
|
|
range_id++;
|
|
|
|
}
|
2016-10-19 01:59:37 +02:00
|
|
|
|
2018-07-12 23:39:07 +02:00
|
|
|
for (int rid = 0; rid < num_of_ranges; rid++) {
|
2016-10-19 01:59:37 +02:00
|
|
|
int range_start = rid * 10;
|
|
|
|
int range_end = range_start + 10;
|
|
|
|
|
|
|
|
ASSERT_EQ(Get(Key(range_start)), Key(range_start)) << rid;
|
|
|
|
ASSERT_EQ(Get(Key(range_end)), Key(range_end)) << rid;
|
|
|
|
for (int k = range_start + 1; k < range_end; k++) {
|
|
|
|
std::string v = Key(k) + ToString(rid);
|
|
|
|
ASSERT_EQ(Get(Key(k)), v) << rid;
|
|
|
|
}
|
|
|
|
}
|
2016-09-29 00:42:06 +02:00
|
|
|
}
|
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
TEST_F(ExternalSSTFileTest, PickedLevelDynamic) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.disable_auto_compactions = false;
|
|
|
|
options.level0_file_num_compaction_trigger = 4;
|
|
|
|
options.level_compaction_dynamic_level_bytes = true;
|
|
|
|
options.num_levels = 4;
|
|
|
|
DestroyAndReopen(options);
|
2016-10-21 02:05:32 +02:00
|
|
|
std::map<std::string, std::string> true_data;
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->LoadDependency({
|
|
|
|
{"ExternalSSTFileTest::PickedLevelDynamic:0",
|
|
|
|
"BackgroundCallCompaction:0"},
|
|
|
|
{"DBImpl::BackgroundCompaction:Start",
|
|
|
|
"ExternalSSTFileTest::PickedLevelDynamic:1"},
|
|
|
|
{"ExternalSSTFileTest::PickedLevelDynamic:2",
|
|
|
|
"DBImpl::BackgroundCompaction:NonTrivial:AfterRun"},
|
|
|
|
});
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
// Flush 4 files containing the same keys
|
|
|
|
for (int i = 0; i < 4; i++) {
|
|
|
|
for (int k = 20; k <= 30; k++) {
|
|
|
|
ASSERT_OK(Put(Key(k), Key(k) + "put"));
|
2016-10-21 02:05:32 +02:00
|
|
|
true_data[Key(k)] = Key(k) + "put";
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
for (int k = 50; k <= 60; k++) {
|
|
|
|
ASSERT_OK(Put(Key(k), Key(k) + "put"));
|
2016-10-21 02:05:32 +02:00
|
|
|
true_data[Key(k)] = Key(k) + "put";
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for BackgroundCompaction() to be called
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevelDynamic:0");
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevelDynamic:1");
|
|
|
|
|
|
|
|
// This file overlaps with the output of the compaction (going to L3)
|
|
|
|
// so the file will be added to L0 since L3 is the base level
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {31, 32, 33, 34}, -1, false,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, true, false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
EXPECT_EQ(FilesPerLevel(), "5");
|
|
|
|
|
|
|
|
// This file does not overlap with the current running compactiong
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {9000, 9001}, -1, false, false,
|
2019-04-08 20:12:25 +02:00
|
|
|
true, false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
EXPECT_EQ(FilesPerLevel(), "5,0,0,1");
|
|
|
|
|
|
|
|
// Hold compaction from finishing
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::PickedLevelDynamic:2");
|
|
|
|
|
|
|
|
// Output of the compaction will go to L3
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
EXPECT_EQ(FilesPerLevel(), "1,0,0,2");
|
|
|
|
|
|
|
|
Close();
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {1, 15, 19}, -1, false, false,
|
2019-04-08 20:12:25 +02:00
|
|
|
true, false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(FilesPerLevel(), "1,0,0,3");
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {1000, 1001, 1002}, -1, false,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, true, false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(FilesPerLevel(), "1,0,0,4");
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {500, 600, 700}, -1, false,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, true, false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(FilesPerLevel(), "1,0,0,5");
|
|
|
|
|
|
|
|
// File 5 overlaps with file 2 (L3 / base level)
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {2, 10}, -1, false, false, true,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(FilesPerLevel(), "2,0,0,5");
|
|
|
|
|
|
|
|
// File 6 overlaps with file 2 (L3 / base level) and file 5 (L0)
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {3, 9}, -1, false, false, true,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(FilesPerLevel(), "3,0,0,5");
|
|
|
|
|
|
|
|
// Verify data in files
|
2016-10-21 02:05:32 +02:00
|
|
|
size_t kcnt = 0;
|
|
|
|
VerifyDBFromMap(true_data, &kcnt, false);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Write range [5 => 10] to L0
|
|
|
|
for (int i = 5; i <= 10; i++) {
|
|
|
|
std::string k = Key(i);
|
|
|
|
std::string v = k + "put";
|
|
|
|
ASSERT_OK(Put(k, v));
|
2016-10-21 02:05:32 +02:00
|
|
|
true_data[k] = v;
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_EQ(FilesPerLevel(), "4,0,0,5");
|
|
|
|
|
|
|
|
// File 7 overlaps with file 4 (L3)
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {650, 651, 652}, -1, false,
|
2019-04-08 20:12:25 +02:00
|
|
|
false, true, false, false, &true_data));
|
2016-09-08 00:41:54 +02:00
|
|
|
ASSERT_EQ(FilesPerLevel(), "5,0,0,5");
|
|
|
|
|
2016-10-21 02:05:32 +02:00
|
|
|
VerifyDBFromMap(true_data, &kcnt, false);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, AddExternalSstFileWithCustomCompartor) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.comparator = ReverseBytewiseComparator();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Generate files with these key ranges
|
|
|
|
// {14 -> 0}
|
|
|
|
// {24 -> 10}
|
|
|
|
// {34 -> 20}
|
|
|
|
// {44 -> 30}
|
|
|
|
// ..
|
|
|
|
std::vector<std::string> generated_files;
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
std::string file_name = sst_files_dir_ + env_->GenerateUniqueId();
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file_name));
|
|
|
|
|
|
|
|
int range_end = i * 10;
|
|
|
|
int range_start = range_end + 15;
|
|
|
|
for (int k = (range_start - 1); k >= range_end; k--) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(k), Key(k)));
|
2016-09-08 00:41:54 +02:00
|
|
|
}
|
|
|
|
ExternalSstFileInfo file_info;
|
|
|
|
ASSERT_OK(sst_file_writer.Finish(&file_info));
|
|
|
|
generated_files.push_back(file_name);
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<std::string> in_files;
|
|
|
|
|
|
|
|
// These 2nd and 3rd files overlap with each other
|
|
|
|
in_files = {generated_files[0], generated_files[4], generated_files[5],
|
|
|
|
generated_files[7]};
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_NOK(DeprecatedAddFile(in_files));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// These 2 files dont overlap with each other
|
|
|
|
in_files = {generated_files[0], generated_files[2]};
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile(in_files));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// These 2 files dont overlap with each other but overlap with keys in DB
|
|
|
|
in_files = {generated_files[3], generated_files[7]};
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_NOK(DeprecatedAddFile(in_files));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
// Files dont overlap and dont overlap with DB key range
|
|
|
|
in_files = {generated_files[4], generated_files[6], generated_files[8]};
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile(in_files));
|
2016-09-08 00:41:54 +02:00
|
|
|
|
|
|
|
for (int i = 0; i < 100; i++) {
|
|
|
|
if (i % 20 <= 14) {
|
|
|
|
ASSERT_EQ(Get(Key(i)), Key(i));
|
|
|
|
} else {
|
|
|
|
ASSERT_EQ(Get(Key(i)), "NOT_FOUND");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-10-13 19:49:06 +02:00
|
|
|
TEST_F(ExternalSSTFileTest, AddFileTrivialMoveBug) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.num_levels = 3;
|
|
|
|
options.IncreaseParallelism(20);
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {1, 4}, 1)); // L3
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {2, 3}, 2)); // L2
|
|
|
|
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {10, 14}, 3)); // L3
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {12, 13}, 4)); // L2
|
|
|
|
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {20, 24}, 5)); // L3
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {22, 23}, 6)); // L2
|
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
2018-04-13 02:55:14 +02:00
|
|
|
"CompactionJob::Run():Start", [&](void* /*arg*/) {
|
2016-10-13 19:49:06 +02:00
|
|
|
// fit in L3 but will overlap with compaction so will be added
|
|
|
|
// to L2 but a compaction will trivially move it to L3
|
|
|
|
// and break LSM consistency
|
Skip deleted WALs during recovery
Summary:
This patch record min log number to keep to the manifest while flushing SST files to ignore them and any WAL older than them during recovery. This is to avoid scenarios when we have a gap between the WAL files are fed to the recovery procedure. The gap could happen by for example out-of-order WAL deletion. Such gap could cause problems in 2PC recovery where the prepared and commit entry are placed into two separate WAL and gap in the WALs could result into not processing the WAL with the commit entry and hence breaking the 2PC recovery logic.
Before the commit, for 2PC case, we determined which log number to keep in FindObsoleteFiles(). We looked at the earliest logs with outstanding prepare entries, or prepare entries whose respective commit or abort are in memtable. With the commit, the same calculation is done while we apply the SST flush. Just before installing the flush file, we precompute the earliest log file to keep after the flush finishes using the same logic (but skipping the memtables just flushed), record this information to the manifest entry for this new flushed SST file. This pre-computed value is also remembered in memory, and will later be used to determine whether a log file can be deleted. This value is unlikely to change until next flush because the commit entry will stay in memtable. (In WritePrepared, we could have removed the older log files as soon as all prepared entries are committed. It's not yet done anyway. Even if we do it, the only thing we loss with this new approach is earlier log deletion between two flushes, which does not guarantee to happen anyway because the obsolete file clean-up function is only executed after flush or compaction)
This min log number to keep is stored in the manifest using the safely-ignore customized field of AddFile entry, in order to guarantee that the DB generated using newer release can be opened by previous releases no older than 4.2.
Closes https://github.com/facebook/rocksdb/pull/3765
Differential Revision: D7747618
Pulled By: siying
fbshipit-source-id: d00c92105b4f83852e9754a1b70d6b64cb590729
2018-05-04 00:35:11 +02:00
|
|
|
static std::atomic<bool> called = {false};
|
|
|
|
if (!called) {
|
|
|
|
called = true;
|
|
|
|
ASSERT_OK(dbfull()->SetOptions({{"max_bytes_for_level_base", "1"}}));
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {15, 16}, 7));
|
|
|
|
}
|
2016-10-13 19:49:06 +02:00
|
|
|
});
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
CompactRangeOptions cro;
|
|
|
|
cro.exclusive_manual_compaction = false;
|
|
|
|
ASSERT_OK(db_->CompactRange(cro, nullptr, nullptr));
|
|
|
|
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, CompactAddedFiles) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.num_levels = 3;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {1, 10}, 1)); // L3
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {2, 9}, 2)); // L2
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {3, 8}, 3)); // L1
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {4, 7}, 4)); // L0
|
|
|
|
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
}
|
2016-10-19 01:59:37 +02:00
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, SstFileWriterNonSharedKeys) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
std::string file_path = sst_files_dir_ + "/not_shared";
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-10-19 01:59:37 +02:00
|
|
|
|
|
|
|
std::string suffix(100, 'X');
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file_path));
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put("A" + suffix, "VAL"));
|
|
|
|
ASSERT_OK(sst_file_writer.Put("BB" + suffix, "VAL"));
|
|
|
|
ASSERT_OK(sst_file_writer.Put("CC" + suffix, "VAL"));
|
|
|
|
ASSERT_OK(sst_file_writer.Put("CXD" + suffix, "VAL"));
|
|
|
|
ASSERT_OK(sst_file_writer.Put("CZZZ" + suffix, "VAL"));
|
|
|
|
ASSERT_OK(sst_file_writer.Put("ZAAAX" + suffix, "VAL"));
|
2016-10-19 01:59:37 +02:00
|
|
|
|
|
|
|
ASSERT_OK(sst_file_writer.Finish());
|
2016-10-21 02:05:32 +02:00
|
|
|
ASSERT_OK(DeprecatedAddFile({file_path}));
|
|
|
|
}
|
|
|
|
|
2018-11-02 00:21:30 +01:00
|
|
|
TEST_P(ExternalSSTFileTest, IngestFileWithGlobalSeqnoRandomized) {
|
2016-10-21 02:05:32 +02:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.IncreaseParallelism(20);
|
|
|
|
options.level0_slowdown_writes_trigger = 256;
|
|
|
|
options.level0_stop_writes_trigger = 256;
|
|
|
|
|
2019-01-30 01:16:53 +01:00
|
|
|
bool write_global_seqno = std::get<0>(GetParam());
|
|
|
|
bool verify_checksums_before_ingest = std::get<1>(GetParam());
|
2016-10-21 02:05:32 +02:00
|
|
|
for (int iter = 0; iter < 2; iter++) {
|
|
|
|
bool write_to_memtable = (iter == 0);
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
Random rnd(301);
|
|
|
|
std::map<std::string, std::string> true_data;
|
2018-07-14 02:25:34 +02:00
|
|
|
for (int i = 0; i < 500; i++) {
|
2016-10-21 02:05:32 +02:00
|
|
|
std::vector<std::pair<std::string, std::string>> random_data;
|
|
|
|
for (int j = 0; j < 100; j++) {
|
|
|
|
std::string k;
|
|
|
|
std::string v;
|
|
|
|
test::RandomString(&rnd, rnd.Next() % 20, &k);
|
|
|
|
test::RandomString(&rnd, rnd.Next() % 50, &v);
|
|
|
|
random_data.emplace_back(k, v);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (write_to_memtable && rnd.OneIn(4)) {
|
|
|
|
// 25% of writes go through memtable
|
|
|
|
for (auto& entry : random_data) {
|
|
|
|
ASSERT_OK(Put(entry.first, entry.second));
|
|
|
|
true_data[entry.first] = entry.second;
|
|
|
|
}
|
|
|
|
} else {
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, random_data, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, true, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
size_t kcnt = 0;
|
|
|
|
VerifyDBFromMap(true_data, &kcnt, false);
|
|
|
|
db_->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
|
|
VerifyDBFromMap(true_data, &kcnt, false);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-11-02 00:21:30 +01:00
|
|
|
TEST_P(ExternalSSTFileTest, IngestFileWithGlobalSeqnoAssignedLevel) {
|
2016-10-21 02:05:32 +02:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.num_levels = 5;
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
std::vector<std::pair<std::string, std::string>> file_data;
|
|
|
|
std::map<std::string, std::string> true_data;
|
|
|
|
|
|
|
|
// Insert 100 -> 200 into the memtable
|
|
|
|
for (int i = 100; i <= 200; i++) {
|
|
|
|
ASSERT_OK(Put(Key(i), "memtable"));
|
|
|
|
true_data[Key(i)] = "memtable";
|
|
|
|
}
|
|
|
|
|
|
|
|
// Insert 0 -> 20 using AddFile
|
|
|
|
file_data.clear();
|
|
|
|
for (int i = 0; i <= 20; i++) {
|
|
|
|
file_data.emplace_back(Key(i), "L4");
|
|
|
|
}
|
2019-01-30 01:16:53 +01:00
|
|
|
bool write_global_seqno = std::get<0>(GetParam());
|
|
|
|
bool verify_checksums_before_ingest = std::get<1>(GetParam());
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, file_data, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, false, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
|
|
|
|
// This file dont overlap with anything in the DB, will go to L4
|
|
|
|
ASSERT_EQ("0,0,0,0,1", FilesPerLevel());
|
|
|
|
|
|
|
|
// Insert 80 -> 130 using AddFile
|
|
|
|
file_data.clear();
|
|
|
|
for (int i = 80; i <= 130; i++) {
|
|
|
|
file_data.emplace_back(Key(i), "L0");
|
|
|
|
}
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, file_data, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, false, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
|
|
|
|
// This file overlap with the memtable, so it will flush it and add
|
|
|
|
// it self to L0
|
|
|
|
ASSERT_EQ("2,0,0,0,1", FilesPerLevel());
|
|
|
|
|
|
|
|
// Insert 30 -> 50 using AddFile
|
|
|
|
file_data.clear();
|
|
|
|
for (int i = 30; i <= 50; i++) {
|
|
|
|
file_data.emplace_back(Key(i), "L4");
|
|
|
|
}
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, file_data, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, false, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
|
|
|
|
// This file dont overlap with anything in the DB and fit in L4 as well
|
|
|
|
ASSERT_EQ("2,0,0,0,2", FilesPerLevel());
|
|
|
|
|
|
|
|
// Insert 10 -> 40 using AddFile
|
|
|
|
file_data.clear();
|
|
|
|
for (int i = 10; i <= 40; i++) {
|
|
|
|
file_data.emplace_back(Key(i), "L3");
|
|
|
|
}
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, file_data, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, false, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
|
|
|
|
// This file overlap with files in L4, we will ingest it in L3
|
|
|
|
ASSERT_EQ("2,0,0,1,2", FilesPerLevel());
|
|
|
|
|
|
|
|
size_t kcnt = 0;
|
|
|
|
VerifyDBFromMap(true_data, &kcnt, false);
|
2016-10-19 01:59:37 +02:00
|
|
|
}
|
|
|
|
|
2018-11-02 00:21:30 +01:00
|
|
|
TEST_P(ExternalSSTFileTest, IngestFileWithGlobalSeqnoMemtableFlush) {
|
2016-10-21 02:05:32 +02:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
uint64_t entries_in_memtable;
|
|
|
|
std::map<std::string, std::string> true_data;
|
|
|
|
|
|
|
|
for (int k : {10, 20, 40, 80}) {
|
|
|
|
ASSERT_OK(Put(Key(k), "memtable"));
|
|
|
|
true_data[Key(k)] = "memtable";
|
|
|
|
}
|
|
|
|
db_->GetIntProperty(DB::Properties::kNumEntriesActiveMemTable,
|
|
|
|
&entries_in_memtable);
|
|
|
|
ASSERT_GE(entries_in_memtable, 1);
|
|
|
|
|
2019-01-30 01:16:53 +01:00
|
|
|
bool write_global_seqno = std::get<0>(GetParam());
|
|
|
|
bool verify_checksums_before_ingest = std::get<1>(GetParam());
|
2016-10-21 02:05:32 +02:00
|
|
|
// No need for flush
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, {90, 100, 110}, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, false, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
db_->GetIntProperty(DB::Properties::kNumEntriesActiveMemTable,
|
|
|
|
&entries_in_memtable);
|
|
|
|
ASSERT_GE(entries_in_memtable, 1);
|
|
|
|
|
|
|
|
// This file will flush the memtable
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, {19, 20, 21}, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, false, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
db_->GetIntProperty(DB::Properties::kNumEntriesActiveMemTable,
|
|
|
|
&entries_in_memtable);
|
|
|
|
ASSERT_EQ(entries_in_memtable, 0);
|
|
|
|
|
|
|
|
for (int k : {200, 201, 205, 206}) {
|
|
|
|
ASSERT_OK(Put(Key(k), "memtable"));
|
|
|
|
true_data[Key(k)] = "memtable";
|
|
|
|
}
|
|
|
|
db_->GetIntProperty(DB::Properties::kNumEntriesActiveMemTable,
|
|
|
|
&entries_in_memtable);
|
|
|
|
ASSERT_GE(entries_in_memtable, 1);
|
|
|
|
|
|
|
|
// No need for flush, this file keys fit between the memtable keys
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, {202, 203, 204}, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, false, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
db_->GetIntProperty(DB::Properties::kNumEntriesActiveMemTable,
|
|
|
|
&entries_in_memtable);
|
|
|
|
ASSERT_GE(entries_in_memtable, 1);
|
|
|
|
|
|
|
|
// This file will flush the memtable
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
2019-04-08 20:12:25 +02:00
|
|
|
options, {206, 207}, -1, true, write_global_seqno,
|
|
|
|
verify_checksums_before_ingest, false, false, &true_data));
|
2016-10-21 02:05:32 +02:00
|
|
|
db_->GetIntProperty(DB::Properties::kNumEntriesActiveMemTable,
|
|
|
|
&entries_in_memtable);
|
|
|
|
ASSERT_EQ(entries_in_memtable, 0);
|
|
|
|
|
|
|
|
size_t kcnt = 0;
|
|
|
|
VerifyDBFromMap(true_data, &kcnt, false);
|
|
|
|
}
|
|
|
|
|
2018-11-02 00:21:30 +01:00
|
|
|
TEST_P(ExternalSSTFileTest, L0SortingIssue) {
|
2016-10-21 02:05:32 +02:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.num_levels = 2;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
std::map<std::string, std::string> true_data;
|
|
|
|
|
|
|
|
ASSERT_OK(Put(Key(1), "memtable"));
|
|
|
|
ASSERT_OK(Put(Key(10), "memtable"));
|
|
|
|
|
2019-01-30 01:16:53 +01:00
|
|
|
bool write_global_seqno = std::get<0>(GetParam());
|
|
|
|
bool verify_checksums_before_ingest = std::get<1>(GetParam());
|
2016-10-21 02:05:32 +02:00
|
|
|
// No Flush needed, No global seqno needed, Ingest in L1
|
2019-04-08 20:12:25 +02:00
|
|
|
ASSERT_OK(
|
|
|
|
GenerateAndAddExternalFile(options, {7, 8}, -1, true, write_global_seqno,
|
|
|
|
verify_checksums_before_ingest, false, false));
|
2016-10-21 02:05:32 +02:00
|
|
|
// No Flush needed, but need a global seqno, Ingest in L0
|
2019-04-08 20:12:25 +02:00
|
|
|
ASSERT_OK(
|
|
|
|
GenerateAndAddExternalFile(options, {7, 8}, -1, true, write_global_seqno,
|
|
|
|
verify_checksums_before_ingest, false, false));
|
2016-10-21 02:05:32 +02:00
|
|
|
printf("%s\n", FilesPerLevel().c_str());
|
|
|
|
|
|
|
|
// Overwrite what we added using external files
|
|
|
|
ASSERT_OK(Put(Key(7), "memtable"));
|
|
|
|
ASSERT_OK(Put(Key(8), "memtable"));
|
|
|
|
|
|
|
|
// Read values from memtable
|
|
|
|
ASSERT_EQ(Get(Key(7)), "memtable");
|
|
|
|
ASSERT_EQ(Get(Key(8)), "memtable");
|
|
|
|
|
|
|
|
// Flush and read from L0
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
printf("%s\n", FilesPerLevel().c_str());
|
|
|
|
ASSERT_EQ(Get(Key(7)), "memtable");
|
|
|
|
ASSERT_EQ(Get(Key(8)), "memtable");
|
|
|
|
}
|
2016-11-09 19:37:51 +01:00
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, CompactionDeadlock) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.num_levels = 2;
|
|
|
|
options.level0_file_num_compaction_trigger = 4;
|
|
|
|
options.level0_slowdown_writes_trigger = 4;
|
|
|
|
options.level0_stop_writes_trigger = 4;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
// atomic conter of currently running bg threads
|
|
|
|
std::atomic<int> running_threads(0);
|
|
|
|
|
|
|
|
rocksdb::SyncPoint::GetInstance()->LoadDependency({
|
|
|
|
{"DBImpl::DelayWrite:Wait", "ExternalSSTFileTest::DeadLock:0"},
|
|
|
|
{"ExternalSSTFileTest::DeadLock:1", "DBImpl::AddFile:Start"},
|
|
|
|
{"DBImpl::AddFile:MutexLock", "ExternalSSTFileTest::DeadLock:2"},
|
|
|
|
{"ExternalSSTFileTest::DeadLock:3", "BackgroundCallCompaction:0"},
|
|
|
|
});
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
// Start ingesting and extrnal file in the background
|
2017-02-06 23:43:55 +01:00
|
|
|
rocksdb::port::Thread bg_ingest_file([&]() {
|
2016-11-09 19:37:51 +01:00
|
|
|
running_threads += 1;
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, {5, 6}));
|
|
|
|
running_threads -= 1;
|
|
|
|
});
|
|
|
|
|
|
|
|
ASSERT_OK(Put(Key(1), "memtable"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
ASSERT_OK(Put(Key(2), "memtable"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
ASSERT_OK(Put(Key(3), "memtable"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
ASSERT_OK(Put(Key(4), "memtable"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
// This thread will try to insert into the memtable but since we have 4 L0
|
|
|
|
// files this thread will be blocked and hold the writer thread
|
2017-02-06 23:43:55 +01:00
|
|
|
rocksdb::port::Thread bg_block_put([&]() {
|
2016-11-09 19:37:51 +01:00
|
|
|
running_threads += 1;
|
|
|
|
ASSERT_OK(Put(Key(10), "memtable"));
|
|
|
|
running_threads -= 1;
|
|
|
|
});
|
|
|
|
|
|
|
|
// Make sure DelayWrite is called first
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::DeadLock:0");
|
|
|
|
|
|
|
|
// `DBImpl::AddFile:Start` will wait until we be here
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::DeadLock:1");
|
|
|
|
|
|
|
|
// Wait for IngestExternalFile() to start and aquire mutex
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::DeadLock:2");
|
|
|
|
|
|
|
|
// Now let compaction start
|
|
|
|
TEST_SYNC_POINT("ExternalSSTFileTest::DeadLock:3");
|
|
|
|
|
|
|
|
// Wait for max 5 seconds, if we did not finish all bg threads
|
|
|
|
// then we hit the deadlock bug
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
if (running_threads.load() == 0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
env_->SleepForMicroseconds(500000);
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT_EQ(running_threads.load(), 0);
|
|
|
|
|
|
|
|
bg_ingest_file.join();
|
|
|
|
bg_block_put.join();
|
|
|
|
}
|
|
|
|
|
2016-11-13 05:03:39 +01:00
|
|
|
TEST_F(ExternalSSTFileTest, DirtyExit) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
std::string file_path = sst_files_dir_ + "/dirty_exit";
|
|
|
|
std::unique_ptr<SstFileWriter> sst_file_writer;
|
|
|
|
|
|
|
|
// Destruct SstFileWriter without calling Finish()
|
2017-03-13 19:17:19 +01:00
|
|
|
sst_file_writer.reset(new SstFileWriter(EnvOptions(), options));
|
2016-11-13 05:03:39 +01:00
|
|
|
ASSERT_OK(sst_file_writer->Open(file_path));
|
|
|
|
sst_file_writer.reset();
|
|
|
|
|
|
|
|
// Destruct SstFileWriter with a failing Finish
|
2017-03-13 19:17:19 +01:00
|
|
|
sst_file_writer.reset(new SstFileWriter(EnvOptions(), options));
|
2016-11-13 05:03:39 +01:00
|
|
|
ASSERT_OK(sst_file_writer->Open(file_path));
|
|
|
|
ASSERT_NOK(sst_file_writer->Finish());
|
|
|
|
}
|
|
|
|
|
2016-12-05 23:16:23 +01:00
|
|
|
TEST_F(ExternalSSTFileTest, FileWithCFInfo) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
CreateAndReopenWithCF({"koko", "toto"}, options);
|
|
|
|
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sfw_default(EnvOptions(), options, handles_[0]);
|
|
|
|
SstFileWriter sfw_cf1(EnvOptions(), options, handles_[1]);
|
|
|
|
SstFileWriter sfw_cf2(EnvOptions(), options, handles_[2]);
|
|
|
|
SstFileWriter sfw_unknown(EnvOptions(), options);
|
2016-12-05 23:16:23 +01:00
|
|
|
|
|
|
|
// default_cf.sst
|
|
|
|
const std::string cf_default_sst = sst_files_dir_ + "/default_cf.sst";
|
|
|
|
ASSERT_OK(sfw_default.Open(cf_default_sst));
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sfw_default.Put("K1", "V1"));
|
|
|
|
ASSERT_OK(sfw_default.Put("K2", "V2"));
|
2016-12-05 23:16:23 +01:00
|
|
|
ASSERT_OK(sfw_default.Finish());
|
|
|
|
|
|
|
|
// cf1.sst
|
|
|
|
const std::string cf1_sst = sst_files_dir_ + "/cf1.sst";
|
|
|
|
ASSERT_OK(sfw_cf1.Open(cf1_sst));
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sfw_cf1.Put("K3", "V1"));
|
|
|
|
ASSERT_OK(sfw_cf1.Put("K4", "V2"));
|
2016-12-05 23:16:23 +01:00
|
|
|
ASSERT_OK(sfw_cf1.Finish());
|
|
|
|
|
|
|
|
// cf_unknown.sst
|
|
|
|
const std::string unknown_sst = sst_files_dir_ + "/cf_unknown.sst";
|
|
|
|
ASSERT_OK(sfw_unknown.Open(unknown_sst));
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sfw_unknown.Put("K5", "V1"));
|
|
|
|
ASSERT_OK(sfw_unknown.Put("K6", "V2"));
|
2016-12-05 23:16:23 +01:00
|
|
|
ASSERT_OK(sfw_unknown.Finish());
|
|
|
|
|
|
|
|
IngestExternalFileOptions ifo;
|
|
|
|
|
|
|
|
// SST CF dont match
|
|
|
|
ASSERT_NOK(db_->IngestExternalFile(handles_[0], {cf1_sst}, ifo));
|
|
|
|
// SST CF dont match
|
|
|
|
ASSERT_NOK(db_->IngestExternalFile(handles_[2], {cf1_sst}, ifo));
|
|
|
|
// SST CF match
|
|
|
|
ASSERT_OK(db_->IngestExternalFile(handles_[1], {cf1_sst}, ifo));
|
|
|
|
|
|
|
|
// SST CF dont match
|
|
|
|
ASSERT_NOK(db_->IngestExternalFile(handles_[1], {cf_default_sst}, ifo));
|
|
|
|
// SST CF dont match
|
|
|
|
ASSERT_NOK(db_->IngestExternalFile(handles_[2], {cf_default_sst}, ifo));
|
|
|
|
// SST CF match
|
|
|
|
ASSERT_OK(db_->IngestExternalFile(handles_[0], {cf_default_sst}, ifo));
|
|
|
|
|
|
|
|
// SST CF unknown
|
|
|
|
ASSERT_OK(db_->IngestExternalFile(handles_[1], {unknown_sst}, ifo));
|
|
|
|
// SST CF unknown
|
|
|
|
ASSERT_OK(db_->IngestExternalFile(handles_[2], {unknown_sst}, ifo));
|
|
|
|
// SST CF unknown
|
|
|
|
ASSERT_OK(db_->IngestExternalFile(handles_[0], {unknown_sst}, ifo));
|
2016-12-13 09:47:52 +01:00
|
|
|
|
|
|
|
// Cannot ingest a file into a dropped CF
|
|
|
|
ASSERT_OK(db_->DropColumnFamily(handles_[1]));
|
|
|
|
ASSERT_NOK(db_->IngestExternalFile(handles_[1], {unknown_sst}, ifo));
|
|
|
|
|
|
|
|
// CF was not dropped, ok to Ingest
|
|
|
|
ASSERT_OK(db_->IngestExternalFile(handles_[2], {unknown_sst}, ifo));
|
2016-12-05 23:16:23 +01:00
|
|
|
}
|
|
|
|
|
2018-04-13 19:47:54 +02:00
|
|
|
/*
|
2019-05-24 06:54:23 +02:00
|
|
|
* Test and verify the functionality of ingestion_options.move_files and
|
|
|
|
* ingestion_options.failed_move_fall_back_to_copy
|
2018-04-13 19:47:54 +02:00
|
|
|
*/
|
2019-05-24 06:54:23 +02:00
|
|
|
TEST_P(ExternSSTFileLinkFailFallbackTest, LinkFailFallBackExternalSst) {
|
|
|
|
const bool fail_link = std::get<0>(GetParam());
|
|
|
|
const bool failed_move_fall_back_to_copy = std::get<1>(GetParam());
|
|
|
|
test_env_->set_fail_link(fail_link);
|
|
|
|
const EnvOptions env_options;
|
|
|
|
DestroyAndReopen(options_);
|
2018-04-13 19:47:54 +02:00
|
|
|
const int kNumKeys = 10000;
|
2019-05-24 06:54:23 +02:00
|
|
|
IngestExternalFileOptions ifo;
|
|
|
|
ifo.move_files = true;
|
|
|
|
ifo.failed_move_fall_back_to_copy = failed_move_fall_back_to_copy;
|
2018-04-13 19:47:54 +02:00
|
|
|
|
|
|
|
std::string file_path = sst_files_dir_ + "file1.sst";
|
|
|
|
// Create SstFileWriter for default column family
|
2019-05-24 06:54:23 +02:00
|
|
|
SstFileWriter sst_file_writer(env_options, options_);
|
2018-04-13 19:47:54 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Open(file_path));
|
|
|
|
for (int i = 0; i < kNumKeys; i++) {
|
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(i), Key(i) + "_value"));
|
|
|
|
}
|
|
|
|
ASSERT_OK(sst_file_writer.Finish());
|
|
|
|
uint64_t file_size = 0;
|
|
|
|
ASSERT_OK(env_->GetFileSize(file_path, &file_size));
|
|
|
|
|
2019-05-24 06:54:23 +02:00
|
|
|
bool copyfile = false;
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"ExternalSstFileIngestionJob::Prepare:CopyFile",
|
|
|
|
[&](void* /* arg */) { copyfile = true; });
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
const Status s = db_->IngestExternalFile({file_path}, ifo);
|
2018-04-13 19:47:54 +02:00
|
|
|
|
|
|
|
ColumnFamilyHandleImpl* cfh =
|
|
|
|
static_cast<ColumnFamilyHandleImpl*>(dbfull()->DefaultColumnFamily());
|
|
|
|
ColumnFamilyData* cfd = cfh->cfd();
|
|
|
|
const InternalStats* internal_stats_ptr = cfd->internal_stats();
|
|
|
|
const std::vector<InternalStats::CompactionStats>& comp_stats =
|
|
|
|
internal_stats_ptr->TEST_GetCompactionStats();
|
|
|
|
uint64_t bytes_copied = 0;
|
|
|
|
uint64_t bytes_moved = 0;
|
|
|
|
for (const auto& stats : comp_stats) {
|
|
|
|
bytes_copied += stats.bytes_written;
|
|
|
|
bytes_moved += stats.bytes_moved;
|
|
|
|
}
|
2019-05-24 06:54:23 +02:00
|
|
|
|
|
|
|
if (!fail_link) {
|
|
|
|
// Link operation succeeds. External SST should be moved.
|
|
|
|
ASSERT_OK(s);
|
2018-04-13 19:47:54 +02:00
|
|
|
ASSERT_EQ(0, bytes_copied);
|
|
|
|
ASSERT_EQ(file_size, bytes_moved);
|
2019-05-24 06:54:23 +02:00
|
|
|
ASSERT_FALSE(copyfile);
|
2018-04-13 19:47:54 +02:00
|
|
|
} else {
|
2019-05-24 06:54:23 +02:00
|
|
|
// Link operation fails.
|
|
|
|
ASSERT_EQ(0, bytes_moved);
|
|
|
|
if (failed_move_fall_back_to_copy) {
|
|
|
|
ASSERT_OK(s);
|
|
|
|
// Copy file is true since a failed link falls back to copy file.
|
|
|
|
ASSERT_TRUE(copyfile);
|
|
|
|
ASSERT_EQ(file_size, bytes_copied);
|
|
|
|
} else {
|
|
|
|
ASSERT_TRUE(s.IsNotSupported());
|
|
|
|
// Copy file is false since a failed link does not fall back to copy file.
|
|
|
|
ASSERT_FALSE(copyfile);
|
|
|
|
ASSERT_EQ(0, bytes_copied);
|
|
|
|
}
|
2018-04-13 19:47:54 +02:00
|
|
|
}
|
2019-05-24 06:54:23 +02:00
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
2018-04-13 19:47:54 +02:00
|
|
|
}
|
|
|
|
|
2016-12-06 22:56:17 +01:00
|
|
|
class TestIngestExternalFileListener : public EventListener {
|
|
|
|
public:
|
2018-03-05 22:08:17 +01:00
|
|
|
void OnExternalFileIngested(DB* /*db*/,
|
2016-12-06 22:56:17 +01:00
|
|
|
const ExternalFileIngestionInfo& info) override {
|
|
|
|
ingested_files.push_back(info);
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<ExternalFileIngestionInfo> ingested_files;
|
|
|
|
};
|
|
|
|
|
2018-11-02 00:21:30 +01:00
|
|
|
TEST_P(ExternalSSTFileTest, IngestionListener) {
|
2016-12-06 22:56:17 +01:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
TestIngestExternalFileListener* listener =
|
|
|
|
new TestIngestExternalFileListener();
|
|
|
|
options.listeners.emplace_back(listener);
|
|
|
|
CreateAndReopenWithCF({"koko", "toto"}, options);
|
|
|
|
|
2019-01-30 01:16:53 +01:00
|
|
|
bool write_global_seqno = std::get<0>(GetParam());
|
|
|
|
bool verify_checksums_before_ingest = std::get<1>(GetParam());
|
2016-12-06 22:56:17 +01:00
|
|
|
// Ingest into default cf
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, {1, 2}, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, true, nullptr, handles_[0]));
|
2016-12-06 22:56:17 +01:00
|
|
|
ASSERT_EQ(listener->ingested_files.size(), 1);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().cf_name, "default");
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().global_seqno, 0);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().table_properties.column_family_id,
|
|
|
|
0);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().table_properties.column_family_name,
|
|
|
|
"default");
|
|
|
|
|
|
|
|
// Ingest into cf1
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, {1, 2}, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, true, nullptr, handles_[1]));
|
2016-12-06 22:56:17 +01:00
|
|
|
ASSERT_EQ(listener->ingested_files.size(), 2);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().cf_name, "koko");
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().global_seqno, 0);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().table_properties.column_family_id,
|
|
|
|
1);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().table_properties.column_family_name,
|
|
|
|
"koko");
|
|
|
|
|
|
|
|
// Ingest into cf2
|
2019-01-30 01:16:53 +01:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, {1, 2}, -1, true, write_global_seqno,
|
2019-04-08 20:12:25 +02:00
|
|
|
verify_checksums_before_ingest, false, true, nullptr, handles_[2]));
|
2016-12-06 22:56:17 +01:00
|
|
|
ASSERT_EQ(listener->ingested_files.size(), 3);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().cf_name, "toto");
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().global_seqno, 0);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().table_properties.column_family_id,
|
|
|
|
2);
|
|
|
|
ASSERT_EQ(listener->ingested_files.back().table_properties.column_family_name,
|
|
|
|
"toto");
|
|
|
|
}
|
2016-12-08 22:30:09 +01:00
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, SnapshotInconsistencyBug) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
const int kNumKeys = 10000;
|
|
|
|
|
|
|
|
// Insert keys using normal path and take a snapshot
|
|
|
|
for (int i = 0; i < kNumKeys; i++) {
|
|
|
|
ASSERT_OK(Put(Key(i), Key(i) + "_V1"));
|
|
|
|
}
|
|
|
|
const Snapshot* snap = db_->GetSnapshot();
|
|
|
|
|
|
|
|
// Overwrite all keys using IngestExternalFile
|
|
|
|
std::string sst_file_path = sst_files_dir_ + "file1.sst";
|
2017-03-13 19:17:19 +01:00
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
2016-12-08 22:30:09 +01:00
|
|
|
ASSERT_OK(sst_file_writer.Open(sst_file_path));
|
|
|
|
for (int i = 0; i < kNumKeys; i++) {
|
2017-05-26 21:05:19 +02:00
|
|
|
ASSERT_OK(sst_file_writer.Put(Key(i), Key(i) + "_V2"));
|
2016-12-08 22:30:09 +01:00
|
|
|
}
|
|
|
|
ASSERT_OK(sst_file_writer.Finish());
|
|
|
|
|
|
|
|
IngestExternalFileOptions ifo;
|
|
|
|
ifo.move_files = true;
|
|
|
|
ASSERT_OK(db_->IngestExternalFile({sst_file_path}, ifo));
|
|
|
|
|
|
|
|
for (int i = 0; i < kNumKeys; i++) {
|
|
|
|
ASSERT_EQ(Get(Key(i), snap), Key(i) + "_V1");
|
|
|
|
ASSERT_EQ(Get(Key(i)), Key(i) + "_V2");
|
|
|
|
}
|
|
|
|
|
|
|
|
db_->ReleaseSnapshot(snap);
|
|
|
|
}
|
2017-05-17 20:32:26 +02:00
|
|
|
|
2018-11-02 00:21:30 +01:00
|
|
|
TEST_P(ExternalSSTFileTest, IngestBehind) {
|
2017-05-17 20:32:26 +02:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.compaction_style = kCompactionStyleUniversal;
|
|
|
|
options.num_levels = 3;
|
|
|
|
options.disable_auto_compactions = false;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
std::vector<std::pair<std::string, std::string>> file_data;
|
|
|
|
std::map<std::string, std::string> true_data;
|
|
|
|
|
|
|
|
// Insert 100 -> 200 into the memtable
|
|
|
|
for (int i = 100; i <= 200; i++) {
|
|
|
|
ASSERT_OK(Put(Key(i), "memtable"));
|
|
|
|
true_data[Key(i)] = "memtable";
|
|
|
|
}
|
|
|
|
|
|
|
|
// Insert 100 -> 200 using IngestExternalFile
|
|
|
|
file_data.clear();
|
|
|
|
for (int i = 0; i <= 20; i++) {
|
|
|
|
file_data.emplace_back(Key(i), "ingest_behind");
|
|
|
|
}
|
|
|
|
|
2019-04-08 20:12:25 +02:00
|
|
|
bool allow_global_seqno = true;
|
|
|
|
bool ingest_behind = true;
|
|
|
|
bool write_global_seqno = std::get<0>(GetParam());
|
|
|
|
bool verify_checksums_before_ingest = std::get<1>(GetParam());
|
2017-05-17 20:32:26 +02:00
|
|
|
|
|
|
|
// Can't ingest behind since allow_ingest_behind isn't set to true
|
2019-04-08 20:12:25 +02:00
|
|
|
ASSERT_NOK(GenerateAndAddExternalFile(
|
|
|
|
options, file_data, -1, allow_global_seqno, write_global_seqno,
|
|
|
|
verify_checksums_before_ingest, ingest_behind, false /*sort_data*/,
|
|
|
|
&true_data));
|
2017-05-17 20:32:26 +02:00
|
|
|
|
|
|
|
options.allow_ingest_behind = true;
|
|
|
|
// check that we still can open the DB, as num_levels should be
|
|
|
|
// sanitized to 3
|
|
|
|
options.num_levels = 2;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
options.num_levels = 3;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
// Insert 100 -> 200 into the memtable
|
|
|
|
for (int i = 100; i <= 200; i++) {
|
|
|
|
ASSERT_OK(Put(Key(i), "memtable"));
|
|
|
|
true_data[Key(i)] = "memtable";
|
|
|
|
}
|
|
|
|
db_->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
|
|
// Universal picker should go at second from the bottom level
|
|
|
|
ASSERT_EQ("0,1", FilesPerLevel());
|
2019-04-08 20:12:25 +02:00
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(
|
|
|
|
options, file_data, -1, allow_global_seqno, write_global_seqno,
|
|
|
|
verify_checksums_before_ingest, true /*ingest_behind*/,
|
|
|
|
false /*sort_data*/, &true_data));
|
2017-05-17 20:32:26 +02:00
|
|
|
ASSERT_EQ("0,1,1", FilesPerLevel());
|
|
|
|
// this time ingest should fail as the file doesn't fit to the bottom level
|
2019-04-08 20:12:25 +02:00
|
|
|
ASSERT_NOK(GenerateAndAddExternalFile(
|
|
|
|
options, file_data, -1, allow_global_seqno, write_global_seqno,
|
|
|
|
verify_checksums_before_ingest, true /*ingest_behind*/,
|
|
|
|
false /*sort_data*/, &true_data));
|
2017-05-17 20:32:26 +02:00
|
|
|
ASSERT_EQ("0,1,1", FilesPerLevel());
|
|
|
|
db_->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
|
|
// bottom level should be empty
|
|
|
|
ASSERT_EQ("0,1", FilesPerLevel());
|
|
|
|
|
|
|
|
size_t kcnt = 0;
|
|
|
|
VerifyDBFromMap(true_data, &kcnt, false);
|
|
|
|
}
|
2018-01-22 23:37:37 +01:00
|
|
|
|
|
|
|
TEST_F(ExternalSSTFileTest, SkipBloomFilter) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.filter_policy.reset(NewBloomFilterPolicy(10));
|
|
|
|
table_options.cache_index_and_filter_blocks = true;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
|
|
|
|
|
|
|
|
// Create external SST file and include bloom filters
|
|
|
|
options.statistics = rocksdb::CreateDBStatistics();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
{
|
|
|
|
std::string file_path = sst_files_dir_ + "sst_with_bloom.sst";
|
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options);
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file_path));
|
|
|
|
ASSERT_OK(sst_file_writer.Put("Key1", "Value1"));
|
|
|
|
ASSERT_OK(sst_file_writer.Finish());
|
|
|
|
|
|
|
|
ASSERT_OK(
|
|
|
|
db_->IngestExternalFile({file_path}, IngestExternalFileOptions()));
|
|
|
|
|
|
|
|
ASSERT_EQ(Get("Key1"), "Value1");
|
|
|
|
ASSERT_GE(
|
|
|
|
options.statistics->getTickerCount(Tickers::BLOCK_CACHE_FILTER_ADD), 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create external SST file but skip bloom filters
|
|
|
|
options.statistics = rocksdb::CreateDBStatistics();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
{
|
|
|
|
std::string file_path = sst_files_dir_ + "sst_with_no_bloom.sst";
|
|
|
|
SstFileWriter sst_file_writer(EnvOptions(), options, nullptr, true,
|
|
|
|
Env::IOPriority::IO_TOTAL,
|
|
|
|
true /* skip_filters */);
|
|
|
|
ASSERT_OK(sst_file_writer.Open(file_path));
|
|
|
|
ASSERT_OK(sst_file_writer.Put("Key1", "Value1"));
|
|
|
|
ASSERT_OK(sst_file_writer.Finish());
|
|
|
|
|
|
|
|
ASSERT_OK(
|
|
|
|
db_->IngestExternalFile({file_path}, IngestExternalFileOptions()));
|
|
|
|
|
|
|
|
ASSERT_EQ(Get("Key1"), "Value1");
|
|
|
|
ASSERT_EQ(
|
|
|
|
options.statistics->getTickerCount(Tickers::BLOCK_CACHE_FILTER_ADD), 0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-02-14 19:16:12 +01:00
|
|
|
TEST_F(ExternalSSTFileTest, IngestFileWrittenWithCompressionDictionary) {
|
|
|
|
if (!ZSTD_Supported()) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
const int kNumEntries = 1 << 10;
|
|
|
|
const int kNumBytesPerEntry = 1 << 10;
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.compression = kZSTD;
|
|
|
|
options.compression_opts.max_dict_bytes = 1 << 14; // 16KB
|
|
|
|
options.compression_opts.zstd_max_train_bytes = 1 << 18; // 256KB
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
std::atomic<int> num_compression_dicts(0);
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"BlockBasedTableBuilder::WriteCompressionDictBlock:RawDict",
|
2019-02-19 21:12:25 +01:00
|
|
|
[&](void* /* arg */) { ++num_compression_dicts; });
|
2019-02-14 19:16:12 +01:00
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
Random rnd(301);
|
|
|
|
std::vector<std::pair<std::string, std::string>> random_data;
|
|
|
|
for (int i = 0; i < kNumEntries; i++) {
|
|
|
|
std::string val;
|
|
|
|
test::RandomString(&rnd, kNumBytesPerEntry, &val);
|
|
|
|
random_data.emplace_back(Key(i), std::move(val));
|
|
|
|
}
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFile(options, std::move(random_data)));
|
|
|
|
ASSERT_EQ(1, num_compression_dicts);
|
|
|
|
}
|
|
|
|
|
2019-02-13 04:07:25 +01:00
|
|
|
TEST_P(ExternalSSTFileTest, IngestFilesIntoMultipleColumnFamilies_Success) {
|
|
|
|
std::unique_ptr<FaultInjectionTestEnv> fault_injection_env(
|
|
|
|
new FaultInjectionTestEnv(env_));
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.env = fault_injection_env.get();
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
std::vector<ColumnFamilyHandle*> column_families;
|
|
|
|
column_families.push_back(handles_[0]);
|
|
|
|
column_families.push_back(handles_[1]);
|
|
|
|
std::vector<IngestExternalFileOptions> ifos(column_families.size());
|
|
|
|
for (auto& ifo : ifos) {
|
|
|
|
ifo.allow_global_seqno = true; // Always allow global_seqno
|
|
|
|
// May or may not write global_seqno
|
|
|
|
ifo.write_global_seqno = std::get<0>(GetParam());
|
|
|
|
// Whether to verify checksums before ingestion
|
|
|
|
ifo.verify_checksums_before_ingest = std::get<1>(GetParam());
|
|
|
|
}
|
|
|
|
std::vector<std::vector<std::pair<std::string, std::string>>> data;
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("foo1", "fv1"), std::make_pair("foo2", "fv2")});
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("bar1", "bv1"), std::make_pair("bar2", "bv2")});
|
|
|
|
// Resize the true_data vector upon construction to avoid re-alloc
|
|
|
|
std::vector<std::map<std::string, std::string>> true_data(
|
|
|
|
column_families.size());
|
|
|
|
Status s = GenerateAndAddExternalFiles(options, column_families, ifos, data,
|
|
|
|
-1, true, true_data);
|
|
|
|
ASSERT_OK(s);
|
|
|
|
Close();
|
|
|
|
ReopenWithColumnFamilies({kDefaultColumnFamilyName, "pikachu"}, options);
|
|
|
|
ASSERT_EQ(2, handles_.size());
|
|
|
|
int cf = 0;
|
|
|
|
for (const auto& verify_map : true_data) {
|
|
|
|
for (const auto& elem : verify_map) {
|
|
|
|
const std::string& key = elem.first;
|
|
|
|
const std::string& value = elem.second;
|
|
|
|
ASSERT_EQ(value, Get(cf, key));
|
|
|
|
}
|
|
|
|
++cf;
|
|
|
|
}
|
|
|
|
Close();
|
|
|
|
Destroy(options, true /* delete_cf_paths */);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ExternalSSTFileTest,
|
|
|
|
IngestFilesIntoMultipleColumnFamilies_NoMixedStateWithSnapshot) {
|
|
|
|
std::unique_ptr<FaultInjectionTestEnv> fault_injection_env(
|
|
|
|
new FaultInjectionTestEnv(env_));
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
SyncPoint::GetInstance()->LoadDependency({
|
|
|
|
{"DBImpl::IngestExternalFiles:InstallSVForFirstCF:0",
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_MixedState:"
|
|
|
|
"BeforeRead"},
|
|
|
|
{"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_MixedState:"
|
|
|
|
"AfterRead",
|
|
|
|
"DBImpl::IngestExternalFiles:InstallSVForFirstCF:1"},
|
|
|
|
});
|
|
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.env = fault_injection_env.get();
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
const std::vector<std::map<std::string, std::string>> data_before_ingestion =
|
|
|
|
{{{"foo1", "fv1_0"}, {"foo2", "fv2_0"}, {"foo3", "fv3_0"}},
|
|
|
|
{{"bar1", "bv1_0"}, {"bar2", "bv2_0"}, {"bar3", "bv3_0"}}};
|
|
|
|
for (size_t i = 0; i != handles_.size(); ++i) {
|
|
|
|
int cf = static_cast<int>(i);
|
|
|
|
const auto& orig_data = data_before_ingestion[i];
|
|
|
|
for (const auto& kv : orig_data) {
|
|
|
|
ASSERT_OK(Put(cf, kv.first, kv.second));
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush(cf));
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<ColumnFamilyHandle*> column_families;
|
|
|
|
column_families.push_back(handles_[0]);
|
|
|
|
column_families.push_back(handles_[1]);
|
|
|
|
std::vector<IngestExternalFileOptions> ifos(column_families.size());
|
|
|
|
for (auto& ifo : ifos) {
|
|
|
|
ifo.allow_global_seqno = true; // Always allow global_seqno
|
|
|
|
// May or may not write global_seqno
|
|
|
|
ifo.write_global_seqno = std::get<0>(GetParam());
|
|
|
|
// Whether to verify checksums before ingestion
|
|
|
|
ifo.verify_checksums_before_ingest = std::get<1>(GetParam());
|
|
|
|
}
|
|
|
|
std::vector<std::vector<std::pair<std::string, std::string>>> data;
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("foo1", "fv1"), std::make_pair("foo2", "fv2")});
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("bar1", "bv1"), std::make_pair("bar2", "bv2")});
|
|
|
|
// Resize the true_data vector upon construction to avoid re-alloc
|
|
|
|
std::vector<std::map<std::string, std::string>> true_data(
|
|
|
|
column_families.size());
|
|
|
|
// Take snapshot before ingestion starts
|
|
|
|
ReadOptions read_opts;
|
|
|
|
read_opts.total_order_seek = true;
|
|
|
|
read_opts.snapshot = dbfull()->GetSnapshot();
|
|
|
|
std::vector<Iterator*> iters(handles_.size());
|
|
|
|
|
|
|
|
// Range scan checks first kv of each CF before ingestion starts.
|
|
|
|
for (size_t i = 0; i != handles_.size(); ++i) {
|
|
|
|
iters[i] = dbfull()->NewIterator(read_opts, handles_[i]);
|
|
|
|
iters[i]->SeekToFirst();
|
|
|
|
ASSERT_TRUE(iters[i]->Valid());
|
|
|
|
const std::string& key = iters[i]->key().ToString();
|
|
|
|
const std::string& value = iters[i]->value().ToString();
|
|
|
|
const std::map<std::string, std::string>& orig_data =
|
|
|
|
data_before_ingestion[i];
|
|
|
|
std::map<std::string, std::string>::const_iterator it = orig_data.find(key);
|
|
|
|
ASSERT_NE(orig_data.end(), it);
|
|
|
|
ASSERT_EQ(it->second, value);
|
|
|
|
iters[i]->Next();
|
|
|
|
}
|
|
|
|
port::Thread ingest_thread([&]() {
|
|
|
|
ASSERT_OK(GenerateAndAddExternalFiles(options, column_families, ifos, data,
|
|
|
|
-1, true, true_data));
|
|
|
|
});
|
|
|
|
TEST_SYNC_POINT(
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_MixedState:"
|
|
|
|
"BeforeRead");
|
|
|
|
// Should see only data before ingestion
|
|
|
|
for (size_t i = 0; i != handles_.size(); ++i) {
|
|
|
|
const auto& orig_data = data_before_ingestion[i];
|
|
|
|
for (; iters[i]->Valid(); iters[i]->Next()) {
|
|
|
|
const std::string& key = iters[i]->key().ToString();
|
|
|
|
const std::string& value = iters[i]->value().ToString();
|
|
|
|
std::map<std::string, std::string>::const_iterator it =
|
|
|
|
orig_data.find(key);
|
|
|
|
ASSERT_NE(orig_data.end(), it);
|
|
|
|
ASSERT_EQ(it->second, value);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
TEST_SYNC_POINT(
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_MixedState:"
|
|
|
|
"AfterRead");
|
|
|
|
ingest_thread.join();
|
|
|
|
for (auto* iter : iters) {
|
|
|
|
delete iter;
|
|
|
|
}
|
|
|
|
iters.clear();
|
|
|
|
dbfull()->ReleaseSnapshot(read_opts.snapshot);
|
|
|
|
|
|
|
|
Close();
|
|
|
|
ReopenWithColumnFamilies({kDefaultColumnFamilyName, "pikachu"}, options);
|
|
|
|
// Should see consistent state after ingestion for all column families even
|
|
|
|
// without snapshot.
|
|
|
|
ASSERT_EQ(2, handles_.size());
|
|
|
|
int cf = 0;
|
|
|
|
for (const auto& verify_map : true_data) {
|
|
|
|
for (const auto& elem : verify_map) {
|
|
|
|
const std::string& key = elem.first;
|
|
|
|
const std::string& value = elem.second;
|
|
|
|
ASSERT_EQ(value, Get(cf, key));
|
|
|
|
}
|
|
|
|
++cf;
|
|
|
|
}
|
|
|
|
Close();
|
|
|
|
Destroy(options, true /* delete_cf_paths */);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ExternalSSTFileTest, IngestFilesIntoMultipleColumnFamilies_PrepareFail) {
|
|
|
|
std::unique_ptr<FaultInjectionTestEnv> fault_injection_env(
|
|
|
|
new FaultInjectionTestEnv(env_));
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.env = fault_injection_env.get();
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
SyncPoint::GetInstance()->LoadDependency({
|
|
|
|
{"DBImpl::IngestExternalFiles:BeforeLastJobPrepare:0",
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_PrepareFail:"
|
|
|
|
"0"},
|
|
|
|
{"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies:PrepareFail:"
|
|
|
|
"1",
|
|
|
|
"DBImpl::IngestExternalFiles:BeforeLastJobPrepare:1"},
|
|
|
|
});
|
|
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
std::vector<ColumnFamilyHandle*> column_families;
|
|
|
|
column_families.push_back(handles_[0]);
|
|
|
|
column_families.push_back(handles_[1]);
|
|
|
|
std::vector<IngestExternalFileOptions> ifos(column_families.size());
|
|
|
|
for (auto& ifo : ifos) {
|
|
|
|
ifo.allow_global_seqno = true; // Always allow global_seqno
|
|
|
|
// May or may not write global_seqno
|
|
|
|
ifo.write_global_seqno = std::get<0>(GetParam());
|
|
|
|
// Whether to verify block checksums before ingest
|
|
|
|
ifo.verify_checksums_before_ingest = std::get<1>(GetParam());
|
|
|
|
}
|
|
|
|
std::vector<std::vector<std::pair<std::string, std::string>>> data;
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("foo1", "fv1"), std::make_pair("foo2", "fv2")});
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("bar1", "bv1"), std::make_pair("bar2", "bv2")});
|
|
|
|
// Resize the true_data vector upon construction to avoid re-alloc
|
|
|
|
std::vector<std::map<std::string, std::string>> true_data(
|
|
|
|
column_families.size());
|
|
|
|
port::Thread ingest_thread([&]() {
|
|
|
|
Status s = GenerateAndAddExternalFiles(options, column_families, ifos, data,
|
|
|
|
-1, true, true_data);
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
});
|
|
|
|
TEST_SYNC_POINT(
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_PrepareFail:"
|
|
|
|
"0");
|
|
|
|
fault_injection_env->SetFilesystemActive(false);
|
|
|
|
TEST_SYNC_POINT(
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies:PrepareFail:"
|
|
|
|
"1");
|
|
|
|
ingest_thread.join();
|
|
|
|
|
|
|
|
fault_injection_env->SetFilesystemActive(true);
|
|
|
|
Close();
|
|
|
|
ReopenWithColumnFamilies({kDefaultColumnFamilyName, "pikachu"}, options);
|
|
|
|
ASSERT_EQ(2, handles_.size());
|
|
|
|
int cf = 0;
|
|
|
|
for (const auto& verify_map : true_data) {
|
|
|
|
for (const auto& elem : verify_map) {
|
|
|
|
const std::string& key = elem.first;
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get(cf, key));
|
|
|
|
}
|
|
|
|
++cf;
|
|
|
|
}
|
|
|
|
Close();
|
|
|
|
Destroy(options, true /* delete_cf_paths */);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ExternalSSTFileTest, IngestFilesIntoMultipleColumnFamilies_CommitFail) {
|
|
|
|
std::unique_ptr<FaultInjectionTestEnv> fault_injection_env(
|
|
|
|
new FaultInjectionTestEnv(env_));
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.env = fault_injection_env.get();
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
SyncPoint::GetInstance()->LoadDependency({
|
|
|
|
{"DBImpl::IngestExternalFiles:BeforeJobsRun:0",
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_CommitFail:"
|
|
|
|
"0"},
|
|
|
|
{"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_CommitFail:"
|
|
|
|
"1",
|
|
|
|
"DBImpl::IngestExternalFiles:BeforeJobsRun:1"},
|
|
|
|
});
|
|
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
std::vector<ColumnFamilyHandle*> column_families;
|
|
|
|
column_families.push_back(handles_[0]);
|
|
|
|
column_families.push_back(handles_[1]);
|
|
|
|
std::vector<IngestExternalFileOptions> ifos(column_families.size());
|
|
|
|
for (auto& ifo : ifos) {
|
|
|
|
ifo.allow_global_seqno = true; // Always allow global_seqno
|
|
|
|
// May or may not write global_seqno
|
|
|
|
ifo.write_global_seqno = std::get<0>(GetParam());
|
|
|
|
// Whether to verify block checksums before ingestion
|
|
|
|
ifo.verify_checksums_before_ingest = std::get<1>(GetParam());
|
|
|
|
}
|
|
|
|
std::vector<std::vector<std::pair<std::string, std::string>>> data;
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("foo1", "fv1"), std::make_pair("foo2", "fv2")});
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("bar1", "bv1"), std::make_pair("bar2", "bv2")});
|
|
|
|
// Resize the true_data vector upon construction to avoid re-alloc
|
|
|
|
std::vector<std::map<std::string, std::string>> true_data(
|
|
|
|
column_families.size());
|
|
|
|
port::Thread ingest_thread([&]() {
|
|
|
|
Status s = GenerateAndAddExternalFiles(options, column_families, ifos, data,
|
|
|
|
-1, true, true_data);
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
});
|
|
|
|
TEST_SYNC_POINT(
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_CommitFail:"
|
|
|
|
"0");
|
|
|
|
fault_injection_env->SetFilesystemActive(false);
|
|
|
|
TEST_SYNC_POINT(
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_CommitFail:"
|
|
|
|
"1");
|
|
|
|
ingest_thread.join();
|
|
|
|
|
|
|
|
fault_injection_env->SetFilesystemActive(true);
|
|
|
|
Close();
|
|
|
|
ReopenWithColumnFamilies({kDefaultColumnFamilyName, "pikachu"}, options);
|
|
|
|
ASSERT_EQ(2, handles_.size());
|
|
|
|
int cf = 0;
|
|
|
|
for (const auto& verify_map : true_data) {
|
|
|
|
for (const auto& elem : verify_map) {
|
|
|
|
const std::string& key = elem.first;
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get(cf, key));
|
|
|
|
}
|
|
|
|
++cf;
|
|
|
|
}
|
|
|
|
Close();
|
|
|
|
Destroy(options, true /* delete_cf_paths */);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ExternalSSTFileTest,
|
|
|
|
IngestFilesIntoMultipleColumnFamilies_PartialManifestWriteFail) {
|
|
|
|
std::unique_ptr<FaultInjectionTestEnv> fault_injection_env(
|
|
|
|
new FaultInjectionTestEnv(env_));
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.env = fault_injection_env.get();
|
|
|
|
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
|
|
|
|
SyncPoint::GetInstance()->ClearTrace();
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
SyncPoint::GetInstance()->LoadDependency({
|
|
|
|
{"VersionSet::ProcessManifestWrites:BeforeWriteLastVersionEdit:0",
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_"
|
|
|
|
"PartialManifestWriteFail:0"},
|
|
|
|
{"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_"
|
|
|
|
"PartialManifestWriteFail:1",
|
|
|
|
"VersionSet::ProcessManifestWrites:BeforeWriteLastVersionEdit:1"},
|
|
|
|
});
|
|
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
std::vector<ColumnFamilyHandle*> column_families;
|
|
|
|
column_families.push_back(handles_[0]);
|
|
|
|
column_families.push_back(handles_[1]);
|
|
|
|
std::vector<IngestExternalFileOptions> ifos(column_families.size());
|
|
|
|
for (auto& ifo : ifos) {
|
|
|
|
ifo.allow_global_seqno = true; // Always allow global_seqno
|
|
|
|
// May or may not write global_seqno
|
|
|
|
ifo.write_global_seqno = std::get<0>(GetParam());
|
|
|
|
// Whether to verify block checksums before ingestion
|
|
|
|
ifo.verify_checksums_before_ingest = std::get<1>(GetParam());
|
|
|
|
}
|
|
|
|
std::vector<std::vector<std::pair<std::string, std::string>>> data;
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("foo1", "fv1"), std::make_pair("foo2", "fv2")});
|
|
|
|
data.push_back(
|
|
|
|
{std::make_pair("bar1", "bv1"), std::make_pair("bar2", "bv2")});
|
|
|
|
// Resize the true_data vector upon construction to avoid re-alloc
|
|
|
|
std::vector<std::map<std::string, std::string>> true_data(
|
|
|
|
column_families.size());
|
|
|
|
port::Thread ingest_thread([&]() {
|
|
|
|
Status s = GenerateAndAddExternalFiles(options, column_families, ifos, data,
|
|
|
|
-1, true, true_data);
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
});
|
|
|
|
TEST_SYNC_POINT(
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_"
|
|
|
|
"PartialManifestWriteFail:0");
|
|
|
|
fault_injection_env->SetFilesystemActive(false);
|
|
|
|
TEST_SYNC_POINT(
|
|
|
|
"ExternalSSTFileTest::IngestFilesIntoMultipleColumnFamilies_"
|
|
|
|
"PartialManifestWriteFail:1");
|
|
|
|
ingest_thread.join();
|
|
|
|
|
|
|
|
fault_injection_env->DropUnsyncedFileData();
|
|
|
|
fault_injection_env->SetFilesystemActive(true);
|
|
|
|
Close();
|
|
|
|
ReopenWithColumnFamilies({kDefaultColumnFamilyName, "pikachu"}, options);
|
|
|
|
ASSERT_EQ(2, handles_.size());
|
|
|
|
int cf = 0;
|
|
|
|
for (const auto& verify_map : true_data) {
|
|
|
|
for (const auto& elem : verify_map) {
|
|
|
|
const std::string& key = elem.first;
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get(cf, key));
|
|
|
|
}
|
|
|
|
++cf;
|
|
|
|
}
|
|
|
|
Close();
|
|
|
|
Destroy(options, true /* delete_cf_paths */);
|
|
|
|
}
|
|
|
|
|
2018-11-02 00:21:30 +01:00
|
|
|
INSTANTIATE_TEST_CASE_P(ExternalSSTFileTest, ExternalSSTFileTest,
|
2019-01-30 01:16:53 +01:00
|
|
|
testing::Values(std::make_tuple(false, false),
|
|
|
|
std::make_tuple(false, true),
|
|
|
|
std::make_tuple(true, false),
|
|
|
|
std::make_tuple(true, true)));
|
2018-11-02 00:21:30 +01:00
|
|
|
|
2019-05-24 06:54:23 +02:00
|
|
|
INSTANTIATE_TEST_CASE_P(ExternSSTFileLinkFailFallbackTest,
|
|
|
|
ExternSSTFileLinkFailFallbackTest,
|
|
|
|
testing::Values(std::make_tuple(true, false),
|
|
|
|
std::make_tuple(true, true),
|
|
|
|
std::make_tuple(false, false)));
|
|
|
|
|
2016-09-08 00:41:54 +02:00
|
|
|
} // namespace rocksdb
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
|
|
rocksdb::port::InstallStackTraceHandler();
|
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
|
|
return RUN_ALL_TESTS();
|
|
|
|
}
|
2017-02-28 20:05:08 +01:00
|
|
|
|
|
|
|
#else
|
|
|
|
#include <stdio.h>
|
|
|
|
|
2018-04-16 02:19:57 +02:00
|
|
|
int main(int /*argc*/, char** /*argv*/) {
|
2017-02-28 20:05:08 +01:00
|
|
|
fprintf(stderr,
|
|
|
|
"SKIPPED as External SST File Writer and Ingestion are not supported "
|
|
|
|
"in ROCKSDB_LITE\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif // !ROCKSDB_LITE
|