// Copyright (c) 2011-present, Facebook, Inc. All rights reserved. // This source code is licensed under both the GPLv2 (found in the // COPYING file in the root directory) and Apache 2.0 License // (found in the LICENSE.Apache file in the root directory). #pragma once #ifndef __STDC_FORMAT_MACROS #define __STDC_FORMAT_MACROS #endif #include #include #include #include #include #include "db/db_impl.h" #include "rocksdb/db.h" #include "rocksdb/options.h" #include "rocksdb/utilities/transaction.h" #include "rocksdb/utilities/transaction_db.h" #include "table/mock_table.h" #include "util/fault_injection_test_env.h" #include "util/random.h" #include "util/string_util.h" #include "util/sync_point.h" #include "util/testharness.h" #include "util/testutil.h" #include "util/transaction_test_util.h" #include "utilities/merge_operators.h" #include "utilities/merge_operators/string_append/stringappend.h" #include "utilities/transactions/pessimistic_transaction_db.h" #include "port/port.h" namespace rocksdb { // Return true if the ith bit is set in combination represented by comb bool IsInCombination(size_t i, size_t comb) { return comb & (size_t(1) << i); } enum WriteOrdering : bool { kOrderedWrite, kUnorderedWrite }; class TransactionTestBase : public ::testing::Test { public: TransactionDB* db; FaultInjectionTestEnv* env; std::string dbname; Options options; TransactionDBOptions txn_db_options; bool use_stackable_db_; TransactionTestBase(bool use_stackable_db, bool two_write_queue, TxnDBWritePolicy write_policy, WriteOrdering write_ordering) : db(nullptr), env(nullptr), use_stackable_db_(use_stackable_db) { options.create_if_missing = true; options.max_write_buffer_number = 2; options.write_buffer_size = 4 * 1024; options.unordered_write = write_ordering == kUnorderedWrite; options.level0_file_num_compaction_trigger = 2; options.merge_operator = MergeOperators::CreateFromStringId("stringappend"); env = new FaultInjectionTestEnv(Env::Default()); options.env = env; options.two_write_queues = two_write_queue; dbname = test::PerThreadDBPath("transaction_testdb"); DestroyDB(dbname, options); txn_db_options.transaction_lock_timeout = 0; txn_db_options.default_lock_timeout = 0; txn_db_options.write_policy = write_policy; txn_db_options.rollback_merge_operands = true; Status s; if (use_stackable_db == false) { s = TransactionDB::Open(options, txn_db_options, dbname, &db); } else { s = OpenWithStackableDB(); } assert(s.ok()); } ~TransactionTestBase() { delete db; db = nullptr; // This is to skip the assert statement in FaultInjectionTestEnv. There // seems to be a bug in btrfs that the makes readdir return recently // unlink-ed files. By using the default fs we simply ignore errors resulted // from attempting to delete such files in DestroyDB. options.env = Env::Default(); DestroyDB(dbname, options); delete env; } Status ReOpenNoDelete() { delete db; db = nullptr; env->AssertNoOpenFile(); env->DropUnsyncedFileData(); env->ResetState(); Status s; if (use_stackable_db_ == false) { s = TransactionDB::Open(options, txn_db_options, dbname, &db); } else { s = OpenWithStackableDB(); } assert(!s.ok() || db != nullptr); return s; } Status ReOpenNoDelete(std::vector& cfs, std::vector* handles) { for (auto h : *handles) { delete h; } handles->clear(); delete db; db = nullptr; env->AssertNoOpenFile(); env->DropUnsyncedFileData(); env->ResetState(); Status s; if (use_stackable_db_ == false) { s = TransactionDB::Open(options, txn_db_options, dbname, cfs, handles, &db); } else { s = OpenWithStackableDB(cfs, handles); } assert(db != nullptr); return s; } Status ReOpen() { delete db; db = nullptr; DestroyDB(dbname, options); Status s; if (use_stackable_db_ == false) { s = TransactionDB::Open(options, txn_db_options, dbname, &db); } else { s = OpenWithStackableDB(); } assert(db != nullptr); return s; } Status OpenWithStackableDB(std::vector& cfs, std::vector* handles) { std::vector compaction_enabled_cf_indices; TransactionDB::PrepareWrap(&options, &cfs, &compaction_enabled_cf_indices); DB* root_db = nullptr; Options options_copy(options); const bool use_seq_per_batch = txn_db_options.write_policy == WRITE_PREPARED || txn_db_options.write_policy == WRITE_UNPREPARED; const bool use_batch_per_txn = txn_db_options.write_policy == WRITE_COMMITTED || txn_db_options.write_policy == WRITE_PREPARED; Status s = DBImpl::Open(options_copy, dbname, cfs, handles, &root_db, use_seq_per_batch, use_batch_per_txn); StackableDB* stackable_db = new StackableDB(root_db); if (s.ok()) { assert(root_db != nullptr); s = TransactionDB::WrapStackableDB(stackable_db, txn_db_options, compaction_enabled_cf_indices, *handles, &db); } if (!s.ok()) { delete stackable_db; // just in case it was not deleted (and not set to nullptr). delete root_db; } return s; } Status OpenWithStackableDB() { std::vector compaction_enabled_cf_indices; std::vector column_families{ColumnFamilyDescriptor( kDefaultColumnFamilyName, ColumnFamilyOptions(options))}; TransactionDB::PrepareWrap(&options, &column_families, &compaction_enabled_cf_indices); std::vector handles; DB* root_db = nullptr; Options options_copy(options); const bool use_seq_per_batch = txn_db_options.write_policy == WRITE_PREPARED || txn_db_options.write_policy == WRITE_UNPREPARED; const bool use_batch_per_txn = txn_db_options.write_policy == WRITE_COMMITTED || txn_db_options.write_policy == WRITE_PREPARED; Status s = DBImpl::Open(options_copy, dbname, column_families, &handles, &root_db, use_seq_per_batch, use_batch_per_txn); if (!s.ok()) { delete root_db; return s; } StackableDB* stackable_db = new StackableDB(root_db); assert(root_db != nullptr); assert(handles.size() == 1); s = TransactionDB::WrapStackableDB(stackable_db, txn_db_options, compaction_enabled_cf_indices, handles, &db); delete handles[0]; if (!s.ok()) { delete stackable_db; // just in case it was not deleted (and not set to nullptr). delete root_db; } return s; } std::atomic linked = {0}; std::atomic exp_seq = {0}; std::atomic commit_writes = {0}; std::atomic expected_commits = {0}; // Without Prepare, the commit does not write to WAL std::atomic with_empty_commits = {0}; std::function txn_t0_with_status = [&](size_t index, Status exp_s) { // Test DB's internal txn. It involves no prepare phase nor a commit marker. WriteOptions wopts; auto s = db->Put(wopts, "key" + std::to_string(index), "value"); ASSERT_EQ(exp_s, s); if (txn_db_options.write_policy == TxnDBWritePolicy::WRITE_COMMITTED) { // Consume one seq per key exp_seq++; } else { // Consume one seq per batch exp_seq++; if (options.two_write_queues) { // Consume one seq for commit exp_seq++; } } with_empty_commits++; }; std::function txn_t0 = [&](size_t index) { return txn_t0_with_status(index, Status::OK()); }; std::function txn_t1 = [&](size_t index) { // Testing directly writing a write batch. Functionality-wise it is // equivalent to commit without prepare. WriteBatch wb; auto istr = std::to_string(index); ASSERT_OK(wb.Put("k1" + istr, "v1")); ASSERT_OK(wb.Put("k2" + istr, "v2")); ASSERT_OK(wb.Put("k3" + istr, "v3")); WriteOptions wopts; auto s = db->Write(wopts, &wb); if (txn_db_options.write_policy == TxnDBWritePolicy::WRITE_COMMITTED) { // Consume one seq per key exp_seq += 3; } else { // Consume one seq per batch exp_seq++; if (options.two_write_queues) { // Consume one seq for commit exp_seq++; } } ASSERT_OK(s); with_empty_commits++; }; std::function txn_t2 = [&](size_t index) { // Commit without prepare. It should write to DB without a commit marker. TransactionOptions txn_options; WriteOptions write_options; Transaction* txn = db->BeginTransaction(write_options, txn_options); auto istr = std::to_string(index); ASSERT_OK(txn->SetName("xid" + istr)); ASSERT_OK(txn->Put(Slice("foo" + istr), Slice("bar"))); ASSERT_OK(txn->Put(Slice("foo2" + istr), Slice("bar2"))); ASSERT_OK(txn->Put(Slice("foo3" + istr), Slice("bar3"))); ASSERT_OK(txn->Put(Slice("foo4" + istr), Slice("bar4"))); ASSERT_OK(txn->Commit()); if (txn_db_options.write_policy == TxnDBWritePolicy::WRITE_COMMITTED) { // Consume one seq per key exp_seq += 4; } else { // Consume one seq per batch exp_seq++; if (options.two_write_queues) { // Consume one seq for commit exp_seq++; } } delete txn; with_empty_commits++; }; std::function txn_t3 = [&](size_t index) { // A full 2pc txn that also involves a commit marker. TransactionOptions txn_options; WriteOptions write_options; Transaction* txn = db->BeginTransaction(write_options, txn_options); auto istr = std::to_string(index); ASSERT_OK(txn->SetName("xid" + istr)); ASSERT_OK(txn->Put(Slice("foo" + istr), Slice("bar"))); ASSERT_OK(txn->Put(Slice("foo2" + istr), Slice("bar2"))); ASSERT_OK(txn->Put(Slice("foo3" + istr), Slice("bar3"))); ASSERT_OK(txn->Put(Slice("foo4" + istr), Slice("bar4"))); ASSERT_OK(txn->Put(Slice("foo5" + istr), Slice("bar5"))); expected_commits++; ASSERT_OK(txn->Prepare()); commit_writes++; ASSERT_OK(txn->Commit()); if (txn_db_options.write_policy == TxnDBWritePolicy::WRITE_COMMITTED) { // Consume one seq per key exp_seq += 5; } else { // Consume one seq per batch exp_seq++; // Consume one seq per commit marker exp_seq++; } delete txn; }; std::function txn_t4 = [&](size_t index) { // A full 2pc txn that also involves a commit marker. TransactionOptions txn_options; WriteOptions write_options; Transaction* txn = db->BeginTransaction(write_options, txn_options); auto istr = std::to_string(index); ASSERT_OK(txn->SetName("xid" + istr)); ASSERT_OK(txn->Put(Slice("foo" + istr), Slice("bar"))); ASSERT_OK(txn->Put(Slice("foo2" + istr), Slice("bar2"))); ASSERT_OK(txn->Put(Slice("foo3" + istr), Slice("bar3"))); ASSERT_OK(txn->Put(Slice("foo4" + istr), Slice("bar4"))); ASSERT_OK(txn->Put(Slice("foo5" + istr), Slice("bar5"))); expected_commits++; ASSERT_OK(txn->Prepare()); commit_writes++; ASSERT_OK(txn->Rollback()); if (txn_db_options.write_policy == TxnDBWritePolicy::WRITE_COMMITTED) { // No seq is consumed for deleting the txn buffer exp_seq += 0; } else { // Consume one seq per batch exp_seq++; // Consume one seq per rollback batch exp_seq++; if (options.two_write_queues) { // Consume one seq for rollback commit exp_seq++; } } delete txn; }; // Test that we can change write policy after a clean shutdown (which would // empty the WAL) void CrossCompatibilityTest(TxnDBWritePolicy from_policy, TxnDBWritePolicy to_policy, bool empty_wal) { TransactionOptions txn_options; ReadOptions read_options; WriteOptions write_options; uint32_t index = 0; Random rnd(1103); options.write_buffer_size = 1024; // To create more sst files std::unordered_map committed_kvs; Transaction* txn; txn_db_options.write_policy = from_policy; if (txn_db_options.write_policy == WRITE_COMMITTED) { options.unordered_write = false; } ReOpen(); for (int i = 0; i < 1024; i++) { auto istr = std::to_string(index); auto k = Slice("foo-" + istr).ToString(); auto v = Slice("bar-" + istr).ToString(); // For test the duplicate keys auto v2 = Slice("bar2-" + istr).ToString(); auto type = rnd.Uniform(4); switch (type) { case 0: committed_kvs[k] = v; ASSERT_OK(db->Put(write_options, k, v)); committed_kvs[k] = v2; ASSERT_OK(db->Put(write_options, k, v2)); break; case 1: { WriteBatch wb; committed_kvs[k] = v; wb.Put(k, v); committed_kvs[k] = v2; wb.Put(k, v2); ASSERT_OK(db->Write(write_options, &wb)); } break; case 2: case 3: txn = db->BeginTransaction(write_options, txn_options); ASSERT_OK(txn->SetName("xid" + istr)); committed_kvs[k] = v; ASSERT_OK(txn->Put(k, v)); committed_kvs[k] = v2; ASSERT_OK(txn->Put(k, v2)); if (type == 3) { ASSERT_OK(txn->Prepare()); } ASSERT_OK(txn->Commit()); delete txn; break; default: assert(0); } index++; } // for i txn_db_options.write_policy = to_policy; if (txn_db_options.write_policy == WRITE_COMMITTED) { options.unordered_write = false; } auto db_impl = reinterpret_cast(db->GetRootDB()); // Before upgrade/downgrade the WAL must be emptied if (empty_wal) { db_impl->TEST_FlushMemTable(); } else { db_impl->FlushWAL(true); } auto s = ReOpenNoDelete(); if (empty_wal) { ASSERT_OK(s); } else { // Test that we can detect the WAL that is produced by an incompatible // WritePolicy and fail fast before mis-interpreting the WAL. ASSERT_TRUE(s.IsNotSupported()); return; } db_impl = reinterpret_cast(db->GetRootDB()); // Check that WAL is empty VectorLogPtr log_files; db_impl->GetSortedWalFiles(log_files); ASSERT_EQ(0, log_files.size()); for (auto& kv : committed_kvs) { std::string value; s = db->Get(read_options, kv.first, &value); if (s.IsNotFound()) { printf("key = %s\n", kv.first.c_str()); } ASSERT_OK(s); if (kv.second != value) { printf("key = %s\n", kv.first.c_str()); } ASSERT_EQ(kv.second, value); } } }; class TransactionTest : public TransactionTestBase, virtual public ::testing::WithParamInterface< std::tuple> { public: TransactionTest() : TransactionTestBase(std::get<0>(GetParam()), std::get<1>(GetParam()), std::get<2>(GetParam()), std::get<3>(GetParam())){}; }; class TransactionStressTest : public TransactionTest {}; class MySQLStyleTransactionTest : public TransactionTestBase, virtual public ::testing::WithParamInterface< std::tuple> { public: MySQLStyleTransactionTest() : TransactionTestBase(std::get<0>(GetParam()), std::get<1>(GetParam()), std::get<2>(GetParam()), std::get<3>(GetParam())), with_slow_threads_(std::get<4>(GetParam())) { if (with_slow_threads_ && (txn_db_options.write_policy == WRITE_PREPARED || txn_db_options.write_policy == WRITE_UNPREPARED)) { // The corner case with slow threads involves the caches filling // over which would not happen even with artifial delays. To help // such cases to show up we lower the size of the cache-related data // structures. txn_db_options.wp_snapshot_cache_bits = 1; txn_db_options.wp_commit_cache_bits = 10; options.write_buffer_size = 1024; EXPECT_OK(ReOpen()); } }; protected: // Also emulate slow threads by addin artiftial delays const bool with_slow_threads_; }; } // namespace rocksdb