920386f2b7
Summary: Note: rebase on and merge after https://github.com/facebook/rocksdb/pull/9349, https://github.com/facebook/rocksdb/pull/9345, (optional) https://github.com/facebook/rocksdb/pull/9393 **Context:** (Quoted from pdillinger) Layers of information during new Bloom/Ribbon Filter construction in building block-based tables includes the following: a) set of keys to add to filter b) set of hashes to add to filter (64-bit hash applied to each key) c) set of Bloom indices to set in filter, with duplicates d) set of Bloom indices to set in filter, deduplicated e) final filter and its checksum This PR aims to detect corruption (e.g, unexpected hardware/software corruption on data structures residing in the memory for a long time) from b) to e) and leave a) as future works for application level. - b)'s corruption is detected by verifying the xor checksum of the hash entries calculated as the entries accumulate before being added to the filter. (i.e, `XXPH3FilterBitsBuilder::MaybeVerifyHashEntriesChecksum()`) - c) - e)'s corruption is detected by verifying the hash entries indeed exists in the constructed filter by re-querying these hash entries in the filter (i.e, `FilterBitsBuilder::MaybePostVerify()`) after computing the block checksum (except for PartitionFilter, which is done right after each `FilterBitsBuilder::Finish` for impl simplicity - see code comment for more). For this stage of detection, we assume hash entries are not corrupted after checking on b) since the time interval from b) to c) is relatively short IMO. Option to enable this feature of detection is `BlockBasedTableOptions::detect_filter_construct_corruption` which is false by default. **Summary:** - Implemented new functions `XXPH3FilterBitsBuilder::MaybeVerifyHashEntriesChecksum()` and `FilterBitsBuilder::MaybePostVerify()` - Ensured hash entries, final filter and banding and their [cache reservation ](https://github.com/facebook/rocksdb/issues/9073) are released properly despite corruption - See [Filter.construction.artifacts.release.point.pdf ](https://github.com/facebook/rocksdb/files/7923487/Design.Filter.construction.artifacts.release.point.pdf) for high-level design - Bundled and refactored hash entries's related artifact in XXPH3FilterBitsBuilder into `HashEntriesInfo` for better control on lifetime of these artifact during `SwapEntires`, `ResetEntries` - Ensured RocksDB block-based table builder calls `FilterBitsBuilder::MaybePostVerify()` after constructing the filter by `FilterBitsBuilder::Finish()` - When encountering such filter construction corruption, stop writing the filter content to files and mark such a block-based table building non-ok by storing the corruption status in the builder. Pull Request resolved: https://github.com/facebook/rocksdb/pull/9342 Test Plan: - Added new unit test `DBFilterConstructionCorruptionTestWithParam.DetectCorruption` - Included this new feature in `DBFilterConstructionReserveMemoryTestWithParam.ReserveMemory` as this feature heavily touch ReserveMemory's impl - For fallback case, I run `./filter_bench -impl=3 -detect_filter_construct_corruption=true -reserve_table_builder_memory=true -strict_capacity_limit=true -quick -runs 10 | grep 'Build avg'` to make sure nothing break. - Added to `filter_bench`: increased filter construction time by **30%**, mostly by `MaybePostVerify()` - FastLocalBloom - Before change: `./filter_bench -impl=2 -quick -runs 10 | grep 'Build avg'`: **28.86643s** - After change: - `./filter_bench -impl=2 -detect_filter_construct_corruption=false -quick -runs 10 | grep 'Build avg'` (expect a tiny increase due to MaybePostVerify is always called regardless): **27.6644s (-4% perf improvement might be due to now we don't drop bloom hash entry in `AddAllEntries` along iteration but in bulk later, same with the bypassing-MaybePostVerify case below)** - `./filter_bench -impl=2 -detect_filter_construct_corruption=true -quick -runs 10 | grep 'Build avg'` (expect acceptable increase): **34.41159s (+20%)** - `./filter_bench -impl=2 -detect_filter_construct_corruption=true -quick -runs 10 | grep 'Build avg'` (by-passing MaybePostVerify, expect minor increase): **27.13431s (-6%)** - Standard128Ribbon - Before change: `./filter_bench -impl=3 -quick -runs 10 | grep 'Build avg'`: **122.5384s** - After change: - `./filter_bench -impl=3 -detect_filter_construct_corruption=false -quick -runs 10 | grep 'Build avg'` (expect a tiny increase due to MaybePostVerify is always called regardless - verified by removing MaybePostVerify under this case and found only +-1ns difference): **124.3588s (+2%)** - `./filter_bench -impl=3 -detect_filter_construct_corruption=true -quick -runs 10 | grep 'Build avg'`(expect acceptable increase): **159.4946s (+30%)** - `./filter_bench -impl=3 -detect_filter_construct_corruption=true -quick -runs 10 | grep 'Build avg'`(by-passing MaybePostVerify, expect minor increase) : **125.258s (+2%)** - Added to `db_stress`: `make crash_test`, `./db_stress --detect_filter_construct_corruption=true` - Manually smoke-tested: manually corrupted the filter construction in some db level tests with basic PUT and background flush. As expected, the error did get returned to users in subsequent PUT and Flush status. Reviewed By: pdillinger Differential Revision: D33746928 Pulled By: hx235 fbshipit-source-id: cb056426be5a7debc1cd16f23bc250f36a08ca57
593 lines
20 KiB
C++
593 lines
20 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
//
|
|
// The test uses an array to compare against values written to the database.
|
|
// Keys written to the array are in 1:1 correspondence to the actual values in
|
|
// the database according to the formula in the function GenerateValue.
|
|
|
|
// Space is reserved in the array from 0 to FLAGS_max_key and values are
|
|
// randomly written/deleted/read from those positions. During verification we
|
|
// compare all the positions in the array. To shorten/elongate the running
|
|
// time, you could change the settings: FLAGS_max_key, FLAGS_ops_per_thread,
|
|
// (sometimes also FLAGS_threads).
|
|
//
|
|
// NOTE that if FLAGS_test_batches_snapshots is set, the test will have
|
|
// different behavior. See comment of the flag for details.
|
|
|
|
#ifdef GFLAGS
|
|
#pragma once
|
|
#include <fcntl.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <sys/types.h>
|
|
|
|
#include <algorithm>
|
|
#include <array>
|
|
#include <chrono>
|
|
#include <cinttypes>
|
|
#include <exception>
|
|
#include <queue>
|
|
#include <thread>
|
|
|
|
#include "db/db_impl/db_impl.h"
|
|
#include "db/version_set.h"
|
|
#include "db_stress_tool/db_stress_env_wrapper.h"
|
|
#include "db_stress_tool/db_stress_listener.h"
|
|
#include "db_stress_tool/db_stress_shared_state.h"
|
|
#include "db_stress_tool/db_stress_test_base.h"
|
|
#include "logging/logging.h"
|
|
#include "monitoring/histogram.h"
|
|
#include "options/options_helper.h"
|
|
#include "port/port.h"
|
|
#include "rocksdb/cache.h"
|
|
#include "rocksdb/env.h"
|
|
#include "rocksdb/slice.h"
|
|
#include "rocksdb/slice_transform.h"
|
|
#include "rocksdb/statistics.h"
|
|
#include "rocksdb/utilities/backup_engine.h"
|
|
#include "rocksdb/utilities/checkpoint.h"
|
|
#include "rocksdb/utilities/db_ttl.h"
|
|
#include "rocksdb/utilities/debug.h"
|
|
#include "rocksdb/utilities/options_util.h"
|
|
#include "rocksdb/utilities/transaction.h"
|
|
#include "rocksdb/utilities/transaction_db.h"
|
|
#include "rocksdb/write_batch.h"
|
|
#include "test_util/testutil.h"
|
|
#include "util/coding.h"
|
|
#include "util/compression.h"
|
|
#include "util/crc32c.h"
|
|
#include "util/gflags_compat.h"
|
|
#include "util/mutexlock.h"
|
|
#include "util/random.h"
|
|
#include "util/string_util.h"
|
|
#include "utilities/blob_db/blob_db.h"
|
|
#include "utilities/merge_operators.h"
|
|
|
|
using GFLAGS_NAMESPACE::ParseCommandLineFlags;
|
|
using GFLAGS_NAMESPACE::RegisterFlagValidator;
|
|
using GFLAGS_NAMESPACE::SetUsageMessage;
|
|
|
|
DECLARE_uint64(seed);
|
|
DECLARE_bool(read_only);
|
|
DECLARE_int64(max_key);
|
|
DECLARE_double(hot_key_alpha);
|
|
DECLARE_int32(max_key_len);
|
|
DECLARE_string(key_len_percent_dist);
|
|
DECLARE_int32(key_window_scale_factor);
|
|
DECLARE_int32(column_families);
|
|
DECLARE_string(options_file);
|
|
DECLARE_int64(active_width);
|
|
DECLARE_bool(test_batches_snapshots);
|
|
DECLARE_bool(atomic_flush);
|
|
DECLARE_bool(test_cf_consistency);
|
|
DECLARE_bool(test_multi_ops_txns);
|
|
DECLARE_int32(threads);
|
|
DECLARE_int32(ttl);
|
|
DECLARE_int32(value_size_mult);
|
|
DECLARE_int32(compaction_readahead_size);
|
|
DECLARE_bool(enable_pipelined_write);
|
|
DECLARE_bool(verify_before_write);
|
|
DECLARE_bool(histogram);
|
|
DECLARE_bool(destroy_db_initially);
|
|
DECLARE_bool(verbose);
|
|
DECLARE_bool(progress_reports);
|
|
DECLARE_uint64(db_write_buffer_size);
|
|
DECLARE_int32(write_buffer_size);
|
|
DECLARE_int32(max_write_buffer_number);
|
|
DECLARE_int32(min_write_buffer_number_to_merge);
|
|
DECLARE_int32(max_write_buffer_number_to_maintain);
|
|
DECLARE_int64(max_write_buffer_size_to_maintain);
|
|
DECLARE_double(memtable_prefix_bloom_size_ratio);
|
|
DECLARE_bool(memtable_whole_key_filtering);
|
|
DECLARE_int32(open_files);
|
|
DECLARE_int64(compressed_cache_size);
|
|
DECLARE_int32(compaction_style);
|
|
DECLARE_int32(num_levels);
|
|
DECLARE_int32(level0_file_num_compaction_trigger);
|
|
DECLARE_int32(level0_slowdown_writes_trigger);
|
|
DECLARE_int32(level0_stop_writes_trigger);
|
|
DECLARE_int32(block_size);
|
|
DECLARE_int32(format_version);
|
|
DECLARE_int32(index_block_restart_interval);
|
|
DECLARE_int32(max_background_compactions);
|
|
DECLARE_int32(num_bottom_pri_threads);
|
|
DECLARE_int32(compaction_thread_pool_adjust_interval);
|
|
DECLARE_int32(compaction_thread_pool_variations);
|
|
DECLARE_int32(max_background_flushes);
|
|
DECLARE_int32(universal_size_ratio);
|
|
DECLARE_int32(universal_min_merge_width);
|
|
DECLARE_int32(universal_max_merge_width);
|
|
DECLARE_int32(universal_max_size_amplification_percent);
|
|
DECLARE_int32(clear_column_family_one_in);
|
|
DECLARE_int32(get_live_files_one_in);
|
|
DECLARE_int32(get_sorted_wal_files_one_in);
|
|
DECLARE_int32(get_current_wal_file_one_in);
|
|
DECLARE_int32(set_options_one_in);
|
|
DECLARE_int32(set_in_place_one_in);
|
|
DECLARE_int64(cache_size);
|
|
DECLARE_int32(cache_numshardbits);
|
|
DECLARE_bool(cache_index_and_filter_blocks);
|
|
DECLARE_int32(top_level_index_pinning);
|
|
DECLARE_int32(partition_pinning);
|
|
DECLARE_int32(unpartitioned_pinning);
|
|
DECLARE_bool(use_clock_cache);
|
|
DECLARE_uint64(subcompactions);
|
|
DECLARE_uint64(periodic_compaction_seconds);
|
|
DECLARE_uint64(compaction_ttl);
|
|
DECLARE_bool(allow_concurrent_memtable_write);
|
|
DECLARE_double(experimental_mempurge_threshold);
|
|
DECLARE_bool(enable_write_thread_adaptive_yield);
|
|
DECLARE_int32(reopen);
|
|
DECLARE_double(bloom_bits);
|
|
DECLARE_bool(use_block_based_filter);
|
|
DECLARE_int32(ribbon_starting_level);
|
|
DECLARE_bool(partition_filters);
|
|
DECLARE_bool(optimize_filters_for_memory);
|
|
DECLARE_bool(detect_filter_construct_corruption);
|
|
DECLARE_int32(index_type);
|
|
DECLARE_string(db);
|
|
DECLARE_string(secondaries_base);
|
|
DECLARE_bool(test_secondary);
|
|
DECLARE_string(expected_values_dir);
|
|
DECLARE_bool(verify_checksum);
|
|
DECLARE_bool(mmap_read);
|
|
DECLARE_bool(mmap_write);
|
|
DECLARE_bool(use_direct_reads);
|
|
DECLARE_bool(use_direct_io_for_flush_and_compaction);
|
|
DECLARE_bool(mock_direct_io);
|
|
DECLARE_bool(statistics);
|
|
DECLARE_bool(sync);
|
|
DECLARE_bool(use_fsync);
|
|
DECLARE_int32(kill_random_test);
|
|
DECLARE_string(kill_exclude_prefixes);
|
|
DECLARE_bool(disable_wal);
|
|
DECLARE_uint64(recycle_log_file_num);
|
|
DECLARE_int64(target_file_size_base);
|
|
DECLARE_int32(target_file_size_multiplier);
|
|
DECLARE_uint64(max_bytes_for_level_base);
|
|
DECLARE_double(max_bytes_for_level_multiplier);
|
|
DECLARE_int32(range_deletion_width);
|
|
DECLARE_uint64(rate_limiter_bytes_per_sec);
|
|
DECLARE_bool(rate_limit_bg_reads);
|
|
DECLARE_uint64(sst_file_manager_bytes_per_sec);
|
|
DECLARE_uint64(sst_file_manager_bytes_per_truncate);
|
|
DECLARE_bool(use_txn);
|
|
DECLARE_uint64(txn_write_policy);
|
|
DECLARE_bool(unordered_write);
|
|
DECLARE_int32(backup_one_in);
|
|
DECLARE_uint64(backup_max_size);
|
|
DECLARE_int32(checkpoint_one_in);
|
|
DECLARE_int32(ingest_external_file_one_in);
|
|
DECLARE_int32(ingest_external_file_width);
|
|
DECLARE_int32(compact_files_one_in);
|
|
DECLARE_int32(compact_range_one_in);
|
|
DECLARE_int32(mark_for_compaction_one_file_in);
|
|
DECLARE_int32(flush_one_in);
|
|
DECLARE_int32(pause_background_one_in);
|
|
DECLARE_int32(compact_range_width);
|
|
DECLARE_int32(acquire_snapshot_one_in);
|
|
DECLARE_bool(compare_full_db_state_snapshot);
|
|
DECLARE_uint64(snapshot_hold_ops);
|
|
DECLARE_bool(long_running_snapshots);
|
|
DECLARE_bool(use_multiget);
|
|
DECLARE_int32(readpercent);
|
|
DECLARE_int32(prefixpercent);
|
|
DECLARE_int32(writepercent);
|
|
DECLARE_int32(delpercent);
|
|
DECLARE_int32(delrangepercent);
|
|
DECLARE_int32(nooverwritepercent);
|
|
DECLARE_int32(iterpercent);
|
|
DECLARE_uint64(num_iterations);
|
|
DECLARE_int32(customopspercent);
|
|
DECLARE_string(compression_type);
|
|
DECLARE_string(bottommost_compression_type);
|
|
DECLARE_int32(compression_max_dict_bytes);
|
|
DECLARE_int32(compression_zstd_max_train_bytes);
|
|
DECLARE_int32(compression_parallel_threads);
|
|
DECLARE_uint64(compression_max_dict_buffer_bytes);
|
|
DECLARE_string(checksum_type);
|
|
DECLARE_string(env_uri);
|
|
DECLARE_string(fs_uri);
|
|
DECLARE_uint64(ops_per_thread);
|
|
DECLARE_uint64(log2_keys_per_lock);
|
|
DECLARE_uint64(max_manifest_file_size);
|
|
DECLARE_bool(in_place_update);
|
|
DECLARE_int32(secondary_catch_up_one_in);
|
|
DECLARE_string(memtablerep);
|
|
DECLARE_int32(prefix_size);
|
|
DECLARE_bool(use_merge);
|
|
DECLARE_bool(use_full_merge_v1);
|
|
DECLARE_int32(sync_wal_one_in);
|
|
DECLARE_bool(avoid_unnecessary_blocking_io);
|
|
DECLARE_bool(write_dbid_to_manifest);
|
|
DECLARE_bool(avoid_flush_during_recovery);
|
|
DECLARE_uint64(max_write_batch_group_size_bytes);
|
|
DECLARE_bool(level_compaction_dynamic_level_bytes);
|
|
DECLARE_int32(verify_checksum_one_in);
|
|
DECLARE_int32(verify_db_one_in);
|
|
DECLARE_int32(continuous_verification_interval);
|
|
DECLARE_int32(get_property_one_in);
|
|
DECLARE_string(file_checksum_impl);
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
// Options for StackableDB-based BlobDB
|
|
DECLARE_bool(use_blob_db);
|
|
DECLARE_uint64(blob_db_min_blob_size);
|
|
DECLARE_uint64(blob_db_bytes_per_sync);
|
|
DECLARE_uint64(blob_db_file_size);
|
|
DECLARE_bool(blob_db_enable_gc);
|
|
DECLARE_double(blob_db_gc_cutoff);
|
|
#endif // !ROCKSDB_LITE
|
|
|
|
// Options for integrated BlobDB
|
|
DECLARE_bool(allow_setting_blob_options_dynamically);
|
|
DECLARE_bool(enable_blob_files);
|
|
DECLARE_uint64(min_blob_size);
|
|
DECLARE_uint64(blob_file_size);
|
|
DECLARE_string(blob_compression_type);
|
|
DECLARE_bool(enable_blob_garbage_collection);
|
|
DECLARE_double(blob_garbage_collection_age_cutoff);
|
|
DECLARE_double(blob_garbage_collection_force_threshold);
|
|
DECLARE_uint64(blob_compaction_readahead_size);
|
|
|
|
DECLARE_int32(approximate_size_one_in);
|
|
DECLARE_bool(sync_fault_injection);
|
|
|
|
DECLARE_bool(best_efforts_recovery);
|
|
DECLARE_bool(skip_verifydb);
|
|
DECLARE_bool(enable_compaction_filter);
|
|
DECLARE_bool(paranoid_file_checks);
|
|
DECLARE_bool(fail_if_options_file_error);
|
|
DECLARE_uint64(batch_protection_bytes_per_key);
|
|
|
|
DECLARE_uint64(user_timestamp_size);
|
|
DECLARE_string(secondary_cache_uri);
|
|
DECLARE_int32(secondary_cache_fault_one_in);
|
|
|
|
DECLARE_int32(prepopulate_block_cache);
|
|
|
|
constexpr long KB = 1024;
|
|
constexpr int kRandomValueMaxFactor = 3;
|
|
constexpr int kValueMaxLen = 100;
|
|
|
|
// wrapped posix environment
|
|
extern ROCKSDB_NAMESPACE::Env* db_stress_env;
|
|
extern ROCKSDB_NAMESPACE::Env* db_stress_listener_env;
|
|
#ifndef NDEBUG
|
|
namespace ROCKSDB_NAMESPACE {
|
|
class FaultInjectionTestFS;
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
extern std::shared_ptr<ROCKSDB_NAMESPACE::FaultInjectionTestFS> fault_fs_guard;
|
|
#endif
|
|
|
|
extern enum ROCKSDB_NAMESPACE::CompressionType compression_type_e;
|
|
extern enum ROCKSDB_NAMESPACE::CompressionType bottommost_compression_type_e;
|
|
extern enum ROCKSDB_NAMESPACE::ChecksumType checksum_type_e;
|
|
|
|
enum RepFactory { kSkipList, kHashSkipList, kVectorRep };
|
|
|
|
inline enum RepFactory StringToRepFactory(const char* ctype) {
|
|
assert(ctype);
|
|
|
|
if (!strcasecmp(ctype, "skip_list"))
|
|
return kSkipList;
|
|
else if (!strcasecmp(ctype, "prefix_hash"))
|
|
return kHashSkipList;
|
|
else if (!strcasecmp(ctype, "vector"))
|
|
return kVectorRep;
|
|
|
|
fprintf(stdout, "Cannot parse memreptable %s\n", ctype);
|
|
return kSkipList;
|
|
}
|
|
|
|
extern enum RepFactory FLAGS_rep_factory;
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
inline enum ROCKSDB_NAMESPACE::CompressionType StringToCompressionType(
|
|
const char* ctype) {
|
|
assert(ctype);
|
|
|
|
ROCKSDB_NAMESPACE::CompressionType ret_compression_type;
|
|
|
|
if (!strcasecmp(ctype, "disable")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kDisableCompressionOption;
|
|
} else if (!strcasecmp(ctype, "none")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kNoCompression;
|
|
} else if (!strcasecmp(ctype, "snappy")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kSnappyCompression;
|
|
} else if (!strcasecmp(ctype, "zlib")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kZlibCompression;
|
|
} else if (!strcasecmp(ctype, "bzip2")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kBZip2Compression;
|
|
} else if (!strcasecmp(ctype, "lz4")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kLZ4Compression;
|
|
} else if (!strcasecmp(ctype, "lz4hc")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kLZ4HCCompression;
|
|
} else if (!strcasecmp(ctype, "xpress")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kXpressCompression;
|
|
} else if (!strcasecmp(ctype, "zstd")) {
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kZSTD;
|
|
} else {
|
|
fprintf(stderr, "Cannot parse compression type '%s'\n", ctype);
|
|
ret_compression_type =
|
|
ROCKSDB_NAMESPACE::kSnappyCompression; // default value
|
|
}
|
|
if (ret_compression_type != ROCKSDB_NAMESPACE::kDisableCompressionOption &&
|
|
!CompressionTypeSupported(ret_compression_type)) {
|
|
// Use no compression will be more portable but considering this is
|
|
// only a stress test and snappy is widely available. Use snappy here.
|
|
ret_compression_type = ROCKSDB_NAMESPACE::kSnappyCompression;
|
|
}
|
|
return ret_compression_type;
|
|
}
|
|
|
|
inline enum ROCKSDB_NAMESPACE::ChecksumType StringToChecksumType(
|
|
const char* ctype) {
|
|
assert(ctype);
|
|
auto iter = ROCKSDB_NAMESPACE::checksum_type_string_map.find(ctype);
|
|
if (iter != ROCKSDB_NAMESPACE::checksum_type_string_map.end()) {
|
|
return iter->second;
|
|
}
|
|
fprintf(stderr, "Cannot parse checksum type '%s'\n", ctype);
|
|
return ROCKSDB_NAMESPACE::kCRC32c;
|
|
}
|
|
|
|
inline std::string ChecksumTypeToString(ROCKSDB_NAMESPACE::ChecksumType ctype) {
|
|
auto iter = std::find_if(
|
|
ROCKSDB_NAMESPACE::checksum_type_string_map.begin(),
|
|
ROCKSDB_NAMESPACE::checksum_type_string_map.end(),
|
|
[&](const std::pair<std::string, ROCKSDB_NAMESPACE::ChecksumType>&
|
|
name_and_enum_val) { return name_and_enum_val.second == ctype; });
|
|
assert(iter != ROCKSDB_NAMESPACE::checksum_type_string_map.end());
|
|
return iter->first;
|
|
}
|
|
|
|
inline std::vector<std::string> SplitString(std::string src) {
|
|
std::vector<std::string> ret;
|
|
if (src.empty()) {
|
|
return ret;
|
|
}
|
|
size_t pos = 0;
|
|
size_t pos_comma;
|
|
while ((pos_comma = src.find(',', pos)) != std::string::npos) {
|
|
ret.push_back(src.substr(pos, pos_comma - pos));
|
|
pos = pos_comma + 1;
|
|
}
|
|
ret.push_back(src.substr(pos, src.length()));
|
|
return ret;
|
|
}
|
|
|
|
#ifdef _MSC_VER
|
|
#pragma warning(push)
|
|
// truncation of constant value on static_cast
|
|
#pragma warning(disable : 4309)
|
|
#endif
|
|
inline bool GetNextPrefix(const ROCKSDB_NAMESPACE::Slice& src, std::string* v) {
|
|
std::string ret = src.ToString();
|
|
for (int i = static_cast<int>(ret.size()) - 1; i >= 0; i--) {
|
|
if (ret[i] != static_cast<char>(255)) {
|
|
ret[i] = ret[i] + 1;
|
|
break;
|
|
} else if (i != 0) {
|
|
ret[i] = 0;
|
|
} else {
|
|
// all FF. No next prefix
|
|
return false;
|
|
}
|
|
}
|
|
*v = ret;
|
|
return true;
|
|
}
|
|
#ifdef _MSC_VER
|
|
#pragma warning(pop)
|
|
#endif
|
|
|
|
// Append `val` to `*key` in fixed-width big-endian format
|
|
extern inline void AppendIntToString(uint64_t val, std::string* key) {
|
|
// PutFixed64 uses little endian
|
|
PutFixed64(key, val);
|
|
// Reverse to get big endian
|
|
char* int_data = &((*key)[key->size() - sizeof(uint64_t)]);
|
|
for (size_t i = 0; i < sizeof(uint64_t) / 2; ++i) {
|
|
std::swap(int_data[i], int_data[sizeof(uint64_t) - 1 - i]);
|
|
}
|
|
}
|
|
|
|
// A struct for maintaining the parameters for generating variable length keys
|
|
struct KeyGenContext {
|
|
// Number of adjacent keys in one cycle of key lengths
|
|
uint64_t window;
|
|
// Number of keys of each possible length in a given window
|
|
std::vector<uint64_t> weights;
|
|
};
|
|
extern KeyGenContext key_gen_ctx;
|
|
|
|
// Generate a variable length key string from the given int64 val. The
|
|
// order of the keys is preserved. The key could be anywhere from 8 to
|
|
// max_key_len * 8 bytes.
|
|
// The algorithm picks the length based on the
|
|
// offset of the val within a configured window and the distribution of the
|
|
// number of keys of various lengths in that window. For example, if x, y, x are
|
|
// the weights assigned to each possible key length, the keys generated would be
|
|
// - {0}...{x-1}
|
|
// {(x-1),0}..{(x-1),(y-1)},{(x-1),(y-1),0}..{(x-1),(y-1),(z-1)} and so on.
|
|
// Additionally, a trailer of 0-7 bytes could be appended.
|
|
extern inline std::string Key(int64_t val) {
|
|
uint64_t window = key_gen_ctx.window;
|
|
size_t levels = key_gen_ctx.weights.size();
|
|
std::string key;
|
|
// Over-reserve and for now do not bother `shrink_to_fit()` since the key
|
|
// strings are transient.
|
|
key.reserve(FLAGS_max_key_len * 8);
|
|
|
|
uint64_t window_idx = static_cast<uint64_t>(val) / window;
|
|
uint64_t offset = static_cast<uint64_t>(val) % window;
|
|
for (size_t level = 0; level < levels; ++level) {
|
|
uint64_t weight = key_gen_ctx.weights[level];
|
|
uint64_t pfx;
|
|
if (level == 0) {
|
|
pfx = window_idx * weight;
|
|
} else {
|
|
pfx = 0;
|
|
}
|
|
pfx += offset >= weight ? weight - 1 : offset;
|
|
AppendIntToString(pfx, &key);
|
|
if (offset < weight) {
|
|
// Use the bottom 3 bits of offset as the number of trailing 'x's in the
|
|
// key. If the next key is going to be of the next level, then skip the
|
|
// trailer as it would break ordering. If the key length is already at max,
|
|
// skip the trailer.
|
|
if (offset < weight - 1 && level < levels - 1) {
|
|
size_t trailer_len = offset & 0x7;
|
|
key.append(trailer_len, 'x');
|
|
}
|
|
break;
|
|
}
|
|
offset -= weight;
|
|
}
|
|
|
|
return key;
|
|
}
|
|
|
|
// Given a string key, map it to an index into the expected values buffer
|
|
extern inline bool GetIntVal(std::string big_endian_key, uint64_t* key_p) {
|
|
size_t size_key = big_endian_key.size();
|
|
std::vector<uint64_t> prefixes;
|
|
|
|
assert(size_key <= key_gen_ctx.weights.size() * sizeof(uint64_t));
|
|
|
|
std::string little_endian_key;
|
|
little_endian_key.resize(size_key);
|
|
for (size_t start = 0; start + sizeof(uint64_t) <= size_key;
|
|
start += sizeof(uint64_t)) {
|
|
size_t end = start + sizeof(uint64_t);
|
|
for (size_t i = 0; i < sizeof(uint64_t); ++i) {
|
|
little_endian_key[start + i] = big_endian_key[end - 1 - i];
|
|
}
|
|
Slice little_endian_slice =
|
|
Slice(&little_endian_key[start], sizeof(uint64_t));
|
|
uint64_t pfx;
|
|
if (!GetFixed64(&little_endian_slice, &pfx)) {
|
|
return false;
|
|
}
|
|
prefixes.emplace_back(pfx);
|
|
}
|
|
|
|
uint64_t key = 0;
|
|
for (size_t i = 0; i < prefixes.size(); ++i) {
|
|
uint64_t pfx = prefixes[i];
|
|
key += (pfx / key_gen_ctx.weights[i]) * key_gen_ctx.window +
|
|
pfx % key_gen_ctx.weights[i];
|
|
if (i < prefixes.size() - 1) {
|
|
// The encoding writes a `key_gen_ctx.weights[i] - 1` that counts for
|
|
// `key_gen_ctx.weights[i]` when there are more prefixes to come. So we
|
|
// need to add back the one here as we're at a non-last prefix.
|
|
++key;
|
|
}
|
|
}
|
|
*key_p = key;
|
|
return true;
|
|
}
|
|
|
|
// Given a string prefix, map it to the first corresponding index in the
|
|
// expected values buffer.
|
|
inline bool GetFirstIntValInPrefix(std::string big_endian_prefix,
|
|
uint64_t* key_p) {
|
|
size_t size_key = big_endian_prefix.size();
|
|
// Pad with zeros to make it a multiple of 8. This function may be called
|
|
// with a prefix, in which case we return the first index that falls
|
|
// inside or outside that prefix, dependeing on whether the prefix is
|
|
// the start of upper bound of a scan
|
|
unsigned int pad = sizeof(uint64_t) - (size_key % sizeof(uint64_t));
|
|
if (pad < sizeof(uint64_t)) {
|
|
big_endian_prefix.append(pad, '\0');
|
|
}
|
|
return GetIntVal(std::move(big_endian_prefix), key_p);
|
|
}
|
|
|
|
extern inline uint64_t GetPrefixKeyCount(const std::string& prefix,
|
|
const std::string& ub) {
|
|
uint64_t start = 0;
|
|
uint64_t end = 0;
|
|
|
|
if (!GetFirstIntValInPrefix(prefix, &start) ||
|
|
!GetFirstIntValInPrefix(ub, &end)) {
|
|
return 0;
|
|
}
|
|
|
|
return end - start;
|
|
}
|
|
|
|
extern inline std::string StringToHex(const std::string& str) {
|
|
std::string result = "0x";
|
|
result.append(Slice(str).ToString(true));
|
|
return result;
|
|
}
|
|
|
|
// Unified output format for double parameters
|
|
extern inline std::string FormatDoubleParam(double param) {
|
|
return std::to_string(param);
|
|
}
|
|
|
|
// Make sure that double parameter is a value we can reproduce by
|
|
// re-inputting the value printed.
|
|
extern inline void SanitizeDoubleParam(double* param) {
|
|
*param = std::atof(FormatDoubleParam(*param).c_str());
|
|
}
|
|
|
|
extern void PoolSizeChangeThread(void* v);
|
|
|
|
extern void DbVerificationThread(void* v);
|
|
|
|
extern void PrintKeyValue(int cf, uint64_t key, const char* value, size_t sz);
|
|
|
|
extern int64_t GenerateOneKey(ThreadState* thread, uint64_t iteration);
|
|
|
|
extern std::vector<int64_t> GenerateNKeys(ThreadState* thread, int num_keys,
|
|
uint64_t iteration);
|
|
|
|
extern size_t GenerateValue(uint32_t rand, char* v, size_t max_sz);
|
|
extern uint32_t GetValueBase(Slice s);
|
|
|
|
extern StressTest* CreateCfConsistencyStressTest();
|
|
extern StressTest* CreateBatchedOpsStressTest();
|
|
extern StressTest* CreateNonBatchedOpsStressTest();
|
|
extern StressTest* CreateMultiOpsTxnsStressTest();
|
|
extern void CheckAndSetOptionsForMultiOpsTxnStressTest();
|
|
extern void InitializeHotKeyGenerator(double alpha);
|
|
extern int64_t GetOneHotKeyID(double rand_seed, int64_t max_key);
|
|
|
|
extern std::string GenerateTimestampForRead();
|
|
extern std::string NowNanosStr();
|
|
|
|
std::shared_ptr<FileChecksumGenFactory> GetFileChecksumImpl(
|
|
const std::string& name);
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
#endif // GFLAGS
|