// Copyright (c) 2011-present, Facebook, Inc. All rights reserved. // This source code is licensed under the BSD-style license found in the // LICENSE file in the root directory of this source tree. An additional grant // of patent rights can be found in the PATENTS file in the same directory. // // Copyright (c) 2011 The LevelDB Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. See the AUTHORS file for names of contributors. #ifndef ROCKSDB_LITE #include #include #include #include #include "db/db_impl.h" #include "db/filename.h" #include "db/version_set.h" #include "db/write_batch_internal.h" #include "rocksdb/db.h" #include "rocksdb/env.h" #include "rocksdb/transaction_log.h" #include "util/string_util.h" #include "util/sync_point.h" #include "util/testharness.h" #include "util/testutil.h" namespace rocksdb { class DeleteFileTest : public testing::Test { public: std::string dbname_; Options options_; DB* db_; Env* env_; int numlevels_; DeleteFileTest() { db_ = nullptr; env_ = Env::Default(); options_.delete_obsolete_files_period_micros = 0; // always do full purge options_.enable_thread_tracking = true; options_.write_buffer_size = 1024*1024*1000; options_.target_file_size_base = 1024*1024*1000; options_.max_bytes_for_level_base = 1024*1024*1000; options_.WAL_ttl_seconds = 300; // Used to test log files options_.WAL_size_limit_MB = 1024; // Used to test log files dbname_ = test::TmpDir() + "/deletefile_test"; options_.wal_dir = dbname_ + "/wal_files"; // clean up all the files that might have been there before std::vector old_files; env_->GetChildren(dbname_, &old_files); for (auto file : old_files) { env_->DeleteFile(dbname_ + "/" + file); } env_->GetChildren(options_.wal_dir, &old_files); for (auto file : old_files) { env_->DeleteFile(options_.wal_dir + "/" + file); } DestroyDB(dbname_, options_); numlevels_ = 7; EXPECT_OK(ReopenDB(true)); } Status ReopenDB(bool create) { delete db_; if (create) { DestroyDB(dbname_, options_); } db_ = nullptr; options_.create_if_missing = create; return DB::Open(options_, dbname_, &db_); } void CloseDB() { delete db_; db_ = nullptr; } void AddKeys(int numkeys, int startkey = 0) { WriteOptions options; options.sync = false; ReadOptions roptions; for (int i = startkey; i < (numkeys + startkey) ; i++) { std::string temp = ToString(i); Slice key(temp); Slice value(temp); ASSERT_OK(db_->Put(options, key, value)); } } int numKeysInLevels( std::vector &metadata, std::vector *keysperlevel = nullptr) { if (keysperlevel != nullptr) { keysperlevel->resize(numlevels_); } int numKeys = 0; for (size_t i = 0; i < metadata.size(); i++) { int startkey = atoi(metadata[i].smallestkey.c_str()); int endkey = atoi(metadata[i].largestkey.c_str()); int numkeysinfile = (endkey - startkey + 1); numKeys += numkeysinfile; if (keysperlevel != nullptr) { (*keysperlevel)[(int)metadata[i].level] += numkeysinfile; } fprintf(stderr, "level %d name %s smallest %s largest %s\n", metadata[i].level, metadata[i].name.c_str(), metadata[i].smallestkey.c_str(), metadata[i].largestkey.c_str()); } return numKeys; } void CreateTwoLevels() { AddKeys(50000, 10000); DBImpl* dbi = reinterpret_cast(db_); ASSERT_OK(dbi->TEST_FlushMemTable()); ASSERT_OK(dbi->TEST_WaitForFlushMemTable()); for (int i = 0; i < 2; ++i) { ASSERT_OK(dbi->TEST_CompactRange(i, nullptr, nullptr)); } AddKeys(50000, 10000); ASSERT_OK(dbi->TEST_FlushMemTable()); ASSERT_OK(dbi->TEST_WaitForFlushMemTable()); ASSERT_OK(dbi->TEST_CompactRange(0, nullptr, nullptr)); } void CheckFileTypeCounts(std::string& dir, int required_log, int required_sst, int required_manifest) { std::vector filenames; env_->GetChildren(dir, &filenames); int log_cnt = 0, sst_cnt = 0, manifest_cnt = 0; for (auto file : filenames) { uint64_t number; FileType type; if (ParseFileName(file, &number, &type)) { log_cnt += (type == kLogFile); sst_cnt += (type == kTableFile); manifest_cnt += (type == kDescriptorFile); } } ASSERT_EQ(required_log, log_cnt); ASSERT_EQ(required_sst, sst_cnt); ASSERT_EQ(required_manifest, manifest_cnt); } static void DoSleep(void* arg) { auto test = reinterpret_cast(arg); test->env_->SleepForMicroseconds(2 * 1000 * 1000); } // An empty job to guard all jobs are processed static void GuardFinish(void* arg) { TEST_SYNC_POINT("DeleteFileTest::GuardFinish"); } }; TEST_F(DeleteFileTest, AddKeysAndQueryLevels) { CreateTwoLevels(); std::vector metadata; db_->GetLiveFilesMetaData(&metadata); std::string level1file = ""; int level1keycount = 0; std::string level2file = ""; int level2keycount = 0; int level1index = 0; int level2index = 1; ASSERT_EQ((int)metadata.size(), 2); if (metadata[0].level == 2) { level1index = 1; level2index = 0; } level1file = metadata[level1index].name; int startkey = atoi(metadata[level1index].smallestkey.c_str()); int endkey = atoi(metadata[level1index].largestkey.c_str()); level1keycount = (endkey - startkey + 1); level2file = metadata[level2index].name; startkey = atoi(metadata[level2index].smallestkey.c_str()); endkey = atoi(metadata[level2index].largestkey.c_str()); level2keycount = (endkey - startkey + 1); // COntrolled setup. Levels 1 and 2 should both have 50K files. // This is a little fragile as it depends on the current // compaction heuristics. ASSERT_EQ(level1keycount, 50000); ASSERT_EQ(level2keycount, 50000); Status status = db_->DeleteFile("0.sst"); ASSERT_TRUE(status.IsInvalidArgument()); // intermediate level files cannot be deleted. status = db_->DeleteFile(level1file); ASSERT_TRUE(status.IsInvalidArgument()); // Lowest level file deletion should succeed. ASSERT_OK(db_->DeleteFile(level2file)); CloseDB(); } TEST_F(DeleteFileTest, PurgeObsoleteFilesTest) { CreateTwoLevels(); // there should be only one (empty) log file because CreateTwoLevels() // flushes the memtables to disk CheckFileTypeCounts(options_.wal_dir, 1, 0, 0); // 2 ssts, 1 manifest CheckFileTypeCounts(dbname_, 0, 2, 1); std::string first("0"), last("999999"); CompactRangeOptions compact_options; compact_options.change_level = true; compact_options.target_level = 2; Slice first_slice(first), last_slice(last); db_->CompactRange(compact_options, &first_slice, &last_slice); // 1 sst after compaction CheckFileTypeCounts(dbname_, 0, 1, 1); // this time, we keep an iterator alive ReopenDB(true); Iterator *itr = 0; CreateTwoLevels(); itr = db_->NewIterator(ReadOptions()); db_->CompactRange(compact_options, &first_slice, &last_slice); // 3 sst after compaction with live iterator CheckFileTypeCounts(dbname_, 0, 3, 1); delete itr; // 1 sst after iterator deletion CheckFileTypeCounts(dbname_, 0, 1, 1); CloseDB(); } TEST_F(DeleteFileTest, BackgroundPurgeTest) { std::string first("0"), last("999999"); CompactRangeOptions compact_options; compact_options.change_level = true; compact_options.target_level = 2; Slice first_slice(first), last_slice(last); // We keep an iterator alive Iterator* itr = 0; CreateTwoLevels(); ReadOptions options; options.background_purge_on_iterator_cleanup = true; itr = db_->NewIterator(options); db_->CompactRange(compact_options, &first_slice, &last_slice); // 3 sst after compaction with live iterator CheckFileTypeCounts(dbname_, 0, 3, 1); test::SleepingBackgroundTask sleeping_task_before; env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask, &sleeping_task_before, Env::Priority::HIGH); delete itr; test::SleepingBackgroundTask sleeping_task_after; env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask, &sleeping_task_after, Env::Priority::HIGH); // Make sure no purges are executed foreground CheckFileTypeCounts(dbname_, 0, 3, 1); sleeping_task_before.WakeUp(); sleeping_task_before.WaitUntilDone(); // Make sure all background purges are executed sleeping_task_after.WakeUp(); sleeping_task_after.WaitUntilDone(); // 1 sst after iterator deletion CheckFileTypeCounts(dbname_, 0, 1, 1); CloseDB(); } TEST_F(DeleteFileTest, BackgroundPurgeTestMultipleJobs) { std::string first("0"), last("999999"); CompactRangeOptions compact_options; compact_options.change_level = true; compact_options.target_level = 2; Slice first_slice(first), last_slice(last); // We keep an iterator alive CreateTwoLevels(); ReadOptions options; options.background_purge_on_iterator_cleanup = true; Iterator* itr1 = db_->NewIterator(options); CreateTwoLevels(); Iterator* itr2 = db_->NewIterator(options); db_->CompactRange(compact_options, &first_slice, &last_slice); // 5 sst files after 2 compactions with 2 live iterators CheckFileTypeCounts(dbname_, 0, 5, 1); // ~DBImpl should wait until all BGWorkPurge are finished rocksdb::SyncPoint::GetInstance()->LoadDependency( {{"DBImpl::~DBImpl:WaitJob", "DBImpl::BGWorkPurge"}, {"DeleteFileTest::GuardFinish", "DeleteFileTest::BackgroundPurgeTestMultipleJobs:DBClose"}}); rocksdb::SyncPoint::GetInstance()->EnableProcessing(); delete itr1; env_->Schedule(&DeleteFileTest::DoSleep, this, Env::Priority::HIGH); delete itr2; env_->Schedule(&DeleteFileTest::GuardFinish, nullptr, Env::Priority::HIGH); CloseDB(); TEST_SYNC_POINT("DeleteFileTest::BackgroundPurgeTestMultipleJobs:DBClose"); // 1 sst after iterator deletion CheckFileTypeCounts(dbname_, 0, 1, 1); rocksdb::SyncPoint::GetInstance()->DisableProcessing(); } TEST_F(DeleteFileTest, DeleteFileWithIterator) { CreateTwoLevels(); ReadOptions options; Iterator* it = db_->NewIterator(options); std::vector metadata; db_->GetLiveFilesMetaData(&metadata); std::string level2file = ""; ASSERT_EQ((int)metadata.size(), 2); if (metadata[0].level == 1) { level2file = metadata[1].name; } else { level2file = metadata[0].name; } Status status = db_->DeleteFile(level2file); fprintf(stdout, "Deletion status %s: %s\n", level2file.c_str(), status.ToString().c_str()); ASSERT_TRUE(status.ok()); it->SeekToFirst(); int numKeysIterated = 0; while(it->Valid()) { numKeysIterated++; it->Next(); } ASSERT_EQ(numKeysIterated, 50000); delete it; CloseDB(); } TEST_F(DeleteFileTest, DeleteLogFiles) { AddKeys(10, 0); VectorLogPtr logfiles; db_->GetSortedWalFiles(logfiles); ASSERT_GT(logfiles.size(), 0UL); // Take the last log file which is expected to be alive and try to delete it // Should not succeed because live logs are not allowed to be deleted std::unique_ptr alive_log = std::move(logfiles.back()); ASSERT_EQ(alive_log->Type(), kAliveLogFile); ASSERT_OK(env_->FileExists(options_.wal_dir + "/" + alive_log->PathName())); fprintf(stdout, "Deleting alive log file %s\n", alive_log->PathName().c_str()); ASSERT_TRUE(!db_->DeleteFile(alive_log->PathName()).ok()); ASSERT_OK(env_->FileExists(options_.wal_dir + "/" + alive_log->PathName())); logfiles.clear(); // Call Flush to bring about a new working log file and add more keys // Call Flush again to flush out memtable and move alive log to archived log // and try to delete the archived log file FlushOptions fopts; db_->Flush(fopts); AddKeys(10, 0); db_->Flush(fopts); db_->GetSortedWalFiles(logfiles); ASSERT_GT(logfiles.size(), 0UL); std::unique_ptr archived_log = std::move(logfiles.front()); ASSERT_EQ(archived_log->Type(), kArchivedLogFile); ASSERT_OK( env_->FileExists(options_.wal_dir + "/" + archived_log->PathName())); fprintf(stdout, "Deleting archived log file %s\n", archived_log->PathName().c_str()); ASSERT_OK(db_->DeleteFile(archived_log->PathName())); ASSERT_EQ(Status::NotFound(), env_->FileExists(options_.wal_dir + "/" + archived_log->PathName())); CloseDB(); } TEST_F(DeleteFileTest, DeleteNonDefaultColumnFamily) { CloseDB(); DBOptions db_options; db_options.create_if_missing = true; db_options.create_missing_column_families = true; std::vector column_families; column_families.emplace_back(); column_families.emplace_back("new_cf", ColumnFamilyOptions()); std::vector handles; rocksdb::DB* db; ASSERT_OK(DB::Open(db_options, dbname_, column_families, &handles, &db)); Random rnd(5); for (int i = 0; i < 1000; ++i) { ASSERT_OK(db->Put(WriteOptions(), handles[1], test::RandomKey(&rnd, 10), test::RandomKey(&rnd, 10))); } ASSERT_OK(db->Flush(FlushOptions(), handles[1])); for (int i = 0; i < 1000; ++i) { ASSERT_OK(db->Put(WriteOptions(), handles[1], test::RandomKey(&rnd, 10), test::RandomKey(&rnd, 10))); } ASSERT_OK(db->Flush(FlushOptions(), handles[1])); std::vector metadata; db->GetLiveFilesMetaData(&metadata); ASSERT_EQ(2U, metadata.size()); ASSERT_EQ("new_cf", metadata[0].column_family_name); ASSERT_EQ("new_cf", metadata[1].column_family_name); auto old_file = metadata[0].smallest_seqno < metadata[1].smallest_seqno ? metadata[0].name : metadata[1].name; auto new_file = metadata[0].smallest_seqno > metadata[1].smallest_seqno ? metadata[0].name : metadata[1].name; ASSERT_TRUE(db->DeleteFile(new_file).IsInvalidArgument()); ASSERT_OK(db->DeleteFile(old_file)); { std::unique_ptr itr(db->NewIterator(ReadOptions(), handles[1])); int count = 0; for (itr->SeekToFirst(); itr->Valid(); itr->Next()) { ASSERT_OK(itr->status()); ++count; } ASSERT_EQ(count, 1000); } delete handles[0]; delete handles[1]; delete db; ASSERT_OK(DB::Open(db_options, dbname_, column_families, &handles, &db)); { std::unique_ptr itr(db->NewIterator(ReadOptions(), handles[1])); int count = 0; for (itr->SeekToFirst(); itr->Valid(); itr->Next()) { ASSERT_OK(itr->status()); ++count; } ASSERT_EQ(count, 1000); } delete handles[0]; delete handles[1]; delete db; } } //namespace rocksdb int main(int argc, char** argv) { ::testing::InitGoogleTest(&argc, argv); return RUN_ALL_TESTS(); } #else #include int main(int argc, char** argv) { fprintf(stderr, "SKIPPED as DBImpl::DeleteFile is not supported in ROCKSDB_LITE\n"); return 0; } #endif // !ROCKSDB_LITE