2012-10-03 18:58:45 +02:00
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
2013-02-23 20:11:16 +01:00
|
|
|
//
|
2013-08-13 22:58:02 +02:00
|
|
|
// The test uses an array to compare against values written to the database.
|
|
|
|
// Keys written to the array are in 1:1 correspondence to the actual values in
|
|
|
|
// the database according to the formula in the function GenerateValue.
|
|
|
|
|
|
|
|
// Space is reserved in the array from 0 to FLAGS_max_key and values are
|
|
|
|
// randomly written/deleted/read from those positions. During verification we
|
|
|
|
// compare all the positions in the array. To shorten/elongate the running
|
|
|
|
// time, you could change the settings: FLAGS_max_key, FLAGS_ops_per_thread,
|
|
|
|
// (sometimes also FLAGS_threads).
|
|
|
|
//
|
|
|
|
// NOTE that if FLAGS_test_batches_snapshots is set, the test will have
|
|
|
|
// different behavior. See comment of the flag for details.
|
2013-02-21 02:37:13 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
#include <sys/types.h>
|
|
|
|
#include <stdio.h>
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include "db/db_impl.h"
|
|
|
|
#include "db/version_set.h"
|
|
|
|
#include "db/db_statistics.h"
|
2013-08-23 17:38:13 +02:00
|
|
|
#include "rocksdb/cache.h"
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
#include "utilities/utility_db.h"
|
2013-08-23 17:38:13 +02:00
|
|
|
#include "rocksdb/env.h"
|
|
|
|
#include "rocksdb/write_batch.h"
|
|
|
|
#include "rocksdb/statistics.h"
|
2012-10-03 18:58:45 +02:00
|
|
|
#include "port/port.h"
|
2013-08-15 01:58:36 +02:00
|
|
|
#include "util/coding.h"
|
2012-10-03 18:58:45 +02:00
|
|
|
#include "util/crc32c.h"
|
|
|
|
#include "util/histogram.h"
|
|
|
|
#include "util/mutexlock.h"
|
|
|
|
#include "util/random.h"
|
|
|
|
#include "util/testutil.h"
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
#include "util/logging.h"
|
|
|
|
#include "utilities/ttl/db_ttl.h"
|
2012-10-03 18:58:45 +02:00
|
|
|
#include "hdfs/env_hdfs.h"
|
Benchmarking for Merge Operator
Summary:
Updated db_bench and utilities/merge_operators.h to allow for dynamic benchmarking
of merge operators in db_bench. Added a new test (--benchmarks=mergerandom), which performs
a bunch of random Merge() operations over random keys. Also added a "--merge_operator=" flag
so that the tester can easily benchmark different merge operators. Currently supports
the PutOperator and UInt64Add operator. Support for stringappend or list append may come later.
Test Plan:
1. make db_bench
2. Test the PutOperator (simulating Put) as follows:
./db_bench --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom --merge_operator=put
--threads=2
3. Test the UInt64AddOperator (simulating numeric addition) similarly:
./db_bench --value_size=8 --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom
--merge_operator=uint64add --threads=2
Reviewers: haobo, dhruba, zshao, MarkCallaghan
Reviewed By: haobo
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11535
2013-08-16 02:13:07 +02:00
|
|
|
#include "utilities/merge_operators.h"
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
static const long KB = 1024;
|
|
|
|
|
|
|
|
// Seed for PRNG
|
|
|
|
static uint32_t FLAGS_seed = 2341234;
|
|
|
|
|
|
|
|
// Max number of key/values to place in database
|
2013-08-13 22:58:02 +02:00
|
|
|
static long FLAGS_max_key = 1 * KB * KB * KB;
|
|
|
|
|
|
|
|
// If set, the test uses MultiGet(), MultiPut() and MultiDelete() which
|
|
|
|
// read/write/delete multiple keys in a batch. In this mode, we do not verify
|
|
|
|
// db content by comparing the content with the pre-allocated array. Instead,
|
|
|
|
// we do partial verification inside MultiGet() by checking various values in
|
|
|
|
// a batch. Benefit of this mode:
|
|
|
|
// (a) No need to acquire mutexes during writes (less cache flushes
|
|
|
|
// in multi-core leading to speed up)
|
2013-02-21 00:57:27 +01:00
|
|
|
// (b) No long validation at the end (more speed up)
|
2013-08-13 22:58:02 +02:00
|
|
|
// (c) Test snapshot and atomicity of batch writes
|
2013-02-21 00:57:27 +01:00
|
|
|
static bool FLAGS_test_batches_snapshots = false;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Number of concurrent threads to run.
|
|
|
|
static int FLAGS_threads = 32;
|
|
|
|
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
// Opens the db with this ttl value if this is not -1
|
|
|
|
// Carefully specify a large value such that verifications on deleted
|
2013-08-13 22:58:02 +02:00
|
|
|
// values don't fail
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
static int FLAGS_ttl = -1;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Size of each value will be this number times rand_int(1,3) bytes
|
|
|
|
static int FLAGS_value_size_mult = 8;
|
|
|
|
|
|
|
|
static bool FLAGS_verify_before_write = false;
|
|
|
|
|
|
|
|
// Print histogram of operation timings
|
|
|
|
static bool FLAGS_histogram = false;
|
|
|
|
|
2013-04-08 21:35:40 +02:00
|
|
|
// Destroys the database dir before start if this is true
|
|
|
|
static bool FLAGS_destroy_db_initially = true;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
static bool FLAGS_verbose = false;
|
|
|
|
|
|
|
|
// Number of bytes to buffer in memtable before compacting
|
|
|
|
// (initialized to default value by "main")
|
|
|
|
static int FLAGS_write_buffer_size = 0;
|
|
|
|
|
2013-01-11 02:18:50 +01:00
|
|
|
// The number of in-memory memtables.
|
2012-10-19 23:00:53 +02:00
|
|
|
// Each memtable is of size FLAGS_write_buffer_size.
|
|
|
|
// This is initialized to default value of 2 in "main" function.
|
|
|
|
static int FLAGS_max_write_buffer_number = 0;
|
|
|
|
|
2013-09-10 01:06:10 +02:00
|
|
|
// The minimum number of write buffers that will be merged together
|
|
|
|
// before writing to storage. This is cheap because it is an
|
|
|
|
// in-memory merge. If this feature is not enabled, then all these
|
|
|
|
// write buffers are flushed to L0 as separate files and this increases
|
|
|
|
// read amplification because a get request has to check in all of these
|
|
|
|
// files. Also, an in-memory merge may result in writing less
|
|
|
|
// data to storage if there are duplicate records in each of these
|
|
|
|
// individual write buffers.
|
|
|
|
static int FLAGS_min_write_buffer_number_to_merge = 0;
|
|
|
|
|
2012-10-19 23:00:53 +02:00
|
|
|
// The maximum number of concurrent background compactions
|
|
|
|
// that can occur in parallel.
|
|
|
|
// This is initialized to default value of 1 in "main" function.
|
|
|
|
static int FLAGS_max_background_compactions = 0;
|
|
|
|
|
2013-06-14 07:09:08 +02:00
|
|
|
// This is initialized to default value of false
|
2013-10-04 06:49:15 +02:00
|
|
|
static rocksdb::CompactionStyle FLAGS_compaction_style = rocksdb::kCompactionStyleLevel;
|
2013-06-14 07:09:08 +02:00
|
|
|
|
2013-09-10 01:06:10 +02:00
|
|
|
// The ratio of file sizes that trigger compaction in universal style
|
|
|
|
static unsigned int FLAGS_universal_size_ratio = 0;
|
|
|
|
|
|
|
|
// The minimum number of files to compact in universal style compaction
|
|
|
|
static unsigned int FLAGS_universal_min_merge_width = 0;
|
|
|
|
|
|
|
|
// The max number of files to compact in universal style compaction
|
|
|
|
static unsigned int FLAGS_universal_max_merge_width = 0;
|
|
|
|
|
|
|
|
// The max size amplification for universal style compaction
|
|
|
|
static unsigned int FLAGS_universal_max_size_amplification_percent = 0;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Number of bytes to use as a cache of uncompressed data.
|
|
|
|
static long FLAGS_cache_size = 2 * KB * KB * KB;
|
|
|
|
|
|
|
|
// Number of bytes in a block.
|
|
|
|
static int FLAGS_block_size = 4 * KB;
|
|
|
|
|
2012-11-09 22:04:12 +01:00
|
|
|
// Number of times database reopens
|
|
|
|
static int FLAGS_reopen = 10;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Maximum number of files to keep open at the same time (use default if == 0)
|
|
|
|
static int FLAGS_open_files = 0;
|
|
|
|
|
|
|
|
// Bloom filter bits per key.
|
|
|
|
// Negative means use default settings.
|
|
|
|
static int FLAGS_bloom_bits = 10;
|
|
|
|
|
|
|
|
// Use the db with the following name.
|
2013-02-15 20:53:17 +01:00
|
|
|
static const char* FLAGS_db = nullptr;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
// Verify checksum for every block read from storage
|
|
|
|
static bool FLAGS_verify_checksum = false;
|
|
|
|
|
2013-06-18 01:13:32 +02:00
|
|
|
// Allow reads to occur via mmap-ing files
|
2013-10-04 06:49:15 +02:00
|
|
|
static bool FLAGS_use_mmap_reads = rocksdb::EnvOptions().use_mmap_reads;
|
2013-06-18 01:13:32 +02:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Database statistics
|
2013-10-04 06:49:15 +02:00
|
|
|
static std::shared_ptr<rocksdb::Statistics> dbstats;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
// Sync all writes to disk
|
|
|
|
static bool FLAGS_sync = false;
|
|
|
|
|
|
|
|
// If true, do not wait until data is synced to disk.
|
|
|
|
static bool FLAGS_disable_data_sync = false;
|
|
|
|
|
|
|
|
// If true, issue fsync instead of fdatasync
|
|
|
|
static bool FLAGS_use_fsync = false;
|
|
|
|
|
2013-04-05 08:49:43 +02:00
|
|
|
// If non-zero, kill at various points in source code with probability 1/this
|
|
|
|
static int FLAGS_kill_random_test = 0;
|
2013-10-05 07:32:05 +02:00
|
|
|
extern int rocksdb_kill_odds;
|
2013-04-05 08:49:43 +02:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// If true, do not write WAL for write.
|
|
|
|
static bool FLAGS_disable_wal = false;
|
|
|
|
|
2013-08-13 22:58:02 +02:00
|
|
|
// Target level-1 file size for compaction
|
2012-10-03 18:58:45 +02:00
|
|
|
static int FLAGS_target_file_size_base = 64 * KB;
|
|
|
|
|
2013-08-13 22:58:02 +02:00
|
|
|
// A multiplier to compute targe level-N file size (N >= 2)
|
2012-10-03 18:58:45 +02:00
|
|
|
static int FLAGS_target_file_size_multiplier = 1;
|
|
|
|
|
2013-08-13 22:58:02 +02:00
|
|
|
// Max bytes for level-1
|
2012-11-09 03:45:19 +01:00
|
|
|
static uint64_t FLAGS_max_bytes_for_level_base = 256 * KB;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
2013-08-13 22:58:02 +02:00
|
|
|
// A multiplier to compute max bytes for level-N (N >= 2)
|
2012-10-03 18:58:45 +02:00
|
|
|
static int FLAGS_max_bytes_for_level_multiplier = 2;
|
|
|
|
|
|
|
|
// Number of files in level-0 that will trigger put stop.
|
|
|
|
static int FLAGS_level0_stop_writes_trigger = 12;
|
|
|
|
|
|
|
|
// Number of files in level-0 that will slow down writes.
|
|
|
|
static int FLAGS_level0_slowdown_writes_trigger = 8;
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
// Ratio of reads to total workload (expressed as a percentage)
|
|
|
|
static unsigned int FLAGS_readpercent = 10;
|
|
|
|
|
2013-08-15 01:58:36 +02:00
|
|
|
// Ratio of prefix iterators to total workload (expressed as a percentage)
|
2013-09-20 01:47:24 +02:00
|
|
|
static unsigned int FLAGS_prefixpercent = 20;
|
2013-08-15 01:58:36 +02:00
|
|
|
|
|
|
|
// Ratio of deletes to total workload (expressed as a percentage)
|
2013-09-20 01:47:24 +02:00
|
|
|
static unsigned int FLAGS_writepercent = 45;
|
2013-08-15 01:58:36 +02:00
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
// Ratio of deletes to total workload (expressed as a percentage)
|
2013-08-15 01:58:36 +02:00
|
|
|
static unsigned int FLAGS_delpercent = 15;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
2013-09-20 01:47:24 +02:00
|
|
|
// Ratio of iterations to total workload (expressed as a percentage)
|
|
|
|
static unsigned int FLAGS_iterpercent = 10;
|
|
|
|
|
|
|
|
// Number of iterations per MultiIterate run
|
|
|
|
static uint32_t FLAGS_num_iterations = 10;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Option to disable compation triggered by read.
|
|
|
|
static int FLAGS_disable_seek_compaction = false;
|
|
|
|
|
2012-10-19 23:00:53 +02:00
|
|
|
// Option to delete obsolete files periodically
|
|
|
|
// Default: 0 which means that obsolete files are
|
|
|
|
// deleted after every compaction run.
|
|
|
|
static uint64_t FLAGS_delete_obsolete_files_period_micros = 0;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Algorithm to use to compress the database
|
2013-10-04 06:49:15 +02:00
|
|
|
static enum rocksdb::CompressionType FLAGS_compression_type =
|
|
|
|
rocksdb::kSnappyCompression;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
// posix or hdfs environment
|
2013-10-04 06:49:15 +02:00
|
|
|
static rocksdb::Env* FLAGS_env = rocksdb::Env::Default();
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
// Number of operations per thread.
|
|
|
|
static uint32_t FLAGS_ops_per_thread = 600000;
|
|
|
|
|
|
|
|
// Log2 of number of keys per lock
|
|
|
|
static uint32_t FLAGS_log2_keys_per_lock = 2; // implies 2^2 keys per lock
|
|
|
|
|
2013-03-06 21:54:55 +01:00
|
|
|
// Percentage of times we want to purge redundant keys in memory before flushing
|
|
|
|
static uint32_t FLAGS_purge_redundant_percent = 50;
|
|
|
|
|
2013-07-26 21:57:01 +02:00
|
|
|
// On true, deletes use KeyMayExist to drop the delete if key not present
|
2013-07-13 01:56:52 +02:00
|
|
|
static bool FLAGS_filter_deletes = false;
|
2013-07-06 03:49:18 +02:00
|
|
|
|
2013-06-17 22:51:12 +02:00
|
|
|
// Level0 compaction start trigger
|
|
|
|
static int FLAGS_level0_file_num_compaction_trigger = 0;
|
|
|
|
|
2013-08-23 08:10:02 +02:00
|
|
|
enum RepFactory {
|
|
|
|
kSkipList,
|
|
|
|
kPrefixHash,
|
|
|
|
kUnsorted,
|
|
|
|
kVectorRep
|
|
|
|
};
|
|
|
|
|
|
|
|
static enum RepFactory FLAGS_rep_factory;
|
|
|
|
|
|
|
|
// Control the prefix size for PrefixHashRep
|
|
|
|
static bool FLAGS_prefix_size = 0;
|
|
|
|
|
Benchmarking for Merge Operator
Summary:
Updated db_bench and utilities/merge_operators.h to allow for dynamic benchmarking
of merge operators in db_bench. Added a new test (--benchmarks=mergerandom), which performs
a bunch of random Merge() operations over random keys. Also added a "--merge_operator=" flag
so that the tester can easily benchmark different merge operators. Currently supports
the PutOperator and UInt64Add operator. Support for stringappend or list append may come later.
Test Plan:
1. make db_bench
2. Test the PutOperator (simulating Put) as follows:
./db_bench --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom --merge_operator=put
--threads=2
3. Test the UInt64AddOperator (simulating numeric addition) similarly:
./db_bench --value_size=8 --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom
--merge_operator=uint64add --threads=2
Reviewers: haobo, dhruba, zshao, MarkCallaghan
Reviewed By: haobo
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11535
2013-08-16 02:13:07 +02:00
|
|
|
// On true, replaces all writes with a Merge that behaves like a Put
|
|
|
|
static bool FLAGS_use_merge_put = false;
|
|
|
|
|
2013-10-04 06:49:15 +02:00
|
|
|
namespace rocksdb {
|
2012-10-03 18:58:45 +02:00
|
|
|
|
2013-08-15 01:58:36 +02:00
|
|
|
// convert long to a big-endian slice key
|
|
|
|
static std::string Key(long val) {
|
|
|
|
std::string little_endian_key;
|
|
|
|
std::string big_endian_key;
|
|
|
|
PutFixed64(&little_endian_key, val);
|
|
|
|
assert(little_endian_key.size() == sizeof(val));
|
|
|
|
big_endian_key.resize(sizeof(val));
|
|
|
|
for (int i=0; i<(int)sizeof(val); i++) {
|
|
|
|
big_endian_key[i] = little_endian_key[sizeof(val) - 1 - i];
|
|
|
|
}
|
|
|
|
return big_endian_key;
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
class StressTest;
|
|
|
|
namespace {
|
|
|
|
|
|
|
|
class Stats {
|
|
|
|
private:
|
|
|
|
double start_;
|
|
|
|
double finish_;
|
|
|
|
double seconds_;
|
|
|
|
long done_;
|
2013-08-15 01:58:36 +02:00
|
|
|
long gets_;
|
|
|
|
long prefixes_;
|
2012-10-03 18:58:45 +02:00
|
|
|
long writes_;
|
2012-11-08 00:35:08 +01:00
|
|
|
long deletes_;
|
2013-08-15 01:58:36 +02:00
|
|
|
long iterator_size_sums_;
|
2013-02-21 00:57:27 +01:00
|
|
|
long founds_;
|
2013-09-20 01:47:24 +02:00
|
|
|
long iterations_;
|
2013-02-21 00:57:27 +01:00
|
|
|
long errors_;
|
2012-10-03 18:58:45 +02:00
|
|
|
int next_report_;
|
|
|
|
size_t bytes_;
|
|
|
|
double last_op_finish_;
|
2013-02-15 20:53:17 +01:00
|
|
|
HistogramImpl hist_;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
public:
|
|
|
|
Stats() { }
|
|
|
|
|
|
|
|
void Start() {
|
|
|
|
next_report_ = 100;
|
|
|
|
hist_.Clear();
|
|
|
|
done_ = 0;
|
2013-08-15 01:58:36 +02:00
|
|
|
gets_ = 0;
|
|
|
|
prefixes_ = 0;
|
2012-10-03 18:58:45 +02:00
|
|
|
writes_ = 0;
|
2012-11-08 00:35:08 +01:00
|
|
|
deletes_ = 0;
|
2013-08-15 01:58:36 +02:00
|
|
|
iterator_size_sums_ = 0;
|
2013-02-21 00:57:27 +01:00
|
|
|
founds_ = 0;
|
2013-09-20 01:47:24 +02:00
|
|
|
iterations_ = 0;
|
2013-02-21 00:57:27 +01:00
|
|
|
errors_ = 0;
|
2012-10-03 18:58:45 +02:00
|
|
|
bytes_ = 0;
|
|
|
|
seconds_ = 0;
|
|
|
|
start_ = FLAGS_env->NowMicros();
|
|
|
|
last_op_finish_ = start_;
|
|
|
|
finish_ = start_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void Merge(const Stats& other) {
|
|
|
|
hist_.Merge(other.hist_);
|
|
|
|
done_ += other.done_;
|
2013-08-15 01:58:36 +02:00
|
|
|
gets_ += other.gets_;
|
|
|
|
prefixes_ += other.prefixes_;
|
2012-10-03 18:58:45 +02:00
|
|
|
writes_ += other.writes_;
|
2012-11-08 00:35:08 +01:00
|
|
|
deletes_ += other.deletes_;
|
2013-08-15 01:58:36 +02:00
|
|
|
iterator_size_sums_ += other.iterator_size_sums_;
|
2013-02-23 20:11:16 +01:00
|
|
|
founds_ += other.founds_;
|
2013-09-20 01:47:24 +02:00
|
|
|
iterations_ += other.iterations_;
|
2013-02-23 20:11:16 +01:00
|
|
|
errors_ += other.errors_;
|
2012-10-03 18:58:45 +02:00
|
|
|
bytes_ += other.bytes_;
|
|
|
|
seconds_ += other.seconds_;
|
|
|
|
if (other.start_ < start_) start_ = other.start_;
|
|
|
|
if (other.finish_ > finish_) finish_ = other.finish_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void Stop() {
|
|
|
|
finish_ = FLAGS_env->NowMicros();
|
|
|
|
seconds_ = (finish_ - start_) * 1e-6;
|
|
|
|
}
|
|
|
|
|
|
|
|
void FinishedSingleOp() {
|
|
|
|
if (FLAGS_histogram) {
|
|
|
|
double now = FLAGS_env->NowMicros();
|
|
|
|
double micros = now - last_op_finish_;
|
|
|
|
hist_.Add(micros);
|
|
|
|
if (micros > 20000) {
|
2013-03-13 07:20:14 +01:00
|
|
|
fprintf(stdout, "long op: %.1f micros%30s\r", micros, "");
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
last_op_finish_ = now;
|
|
|
|
}
|
|
|
|
|
|
|
|
done_++;
|
|
|
|
if (done_ >= next_report_) {
|
|
|
|
if (next_report_ < 1000) next_report_ += 100;
|
|
|
|
else if (next_report_ < 5000) next_report_ += 500;
|
|
|
|
else if (next_report_ < 10000) next_report_ += 1000;
|
|
|
|
else if (next_report_ < 50000) next_report_ += 5000;
|
|
|
|
else if (next_report_ < 100000) next_report_ += 10000;
|
|
|
|
else if (next_report_ < 500000) next_report_ += 50000;
|
|
|
|
else next_report_ += 100000;
|
2013-03-13 07:20:14 +01:00
|
|
|
fprintf(stdout, "... finished %ld ops%30s\r", done_, "");
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
void AddBytesForWrites(int nwrites, size_t nbytes) {
|
|
|
|
writes_ += nwrites;
|
|
|
|
bytes_ += nbytes;
|
|
|
|
}
|
|
|
|
|
2013-02-23 20:11:16 +01:00
|
|
|
void AddGets(int ngets, int nfounds) {
|
|
|
|
founds_ += nfounds;
|
|
|
|
gets_ += ngets;
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
|
2013-08-15 01:58:36 +02:00
|
|
|
void AddPrefixes(int nprefixes, int count) {
|
|
|
|
prefixes_ += nprefixes;
|
|
|
|
iterator_size_sums_ += count;
|
|
|
|
}
|
|
|
|
|
2013-09-20 01:47:24 +02:00
|
|
|
void AddIterations(int n) {
|
|
|
|
iterations_ += n;
|
|
|
|
}
|
|
|
|
|
2013-08-15 01:58:36 +02:00
|
|
|
void AddDeletes(int n) {
|
|
|
|
deletes_ += n;
|
|
|
|
}
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
void AddErrors(int n) {
|
|
|
|
errors_ += n;
|
2012-11-08 00:35:08 +01:00
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
void Report(const char* name) {
|
|
|
|
std::string extra;
|
|
|
|
if (bytes_ < 1 || done_ < 1) {
|
|
|
|
fprintf(stderr, "No writes or ops?\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
double elapsed = (finish_ - start_) * 1e-6;
|
|
|
|
double bytes_mb = bytes_ / 1048576.0;
|
|
|
|
double rate = bytes_mb / elapsed;
|
|
|
|
double throughput = (double)done_/elapsed;
|
|
|
|
|
|
|
|
fprintf(stdout, "%-12s: ", name);
|
|
|
|
fprintf(stdout, "%.3f micros/op %ld ops/sec\n",
|
|
|
|
seconds_ * 1e6 / done_, (long)throughput);
|
|
|
|
fprintf(stdout, "%-12s: Wrote %.2f MB (%.2f MB/sec) (%ld%% of %ld ops)\n",
|
|
|
|
"", bytes_mb, rate, (100*writes_)/done_, done_);
|
2013-02-23 20:11:16 +01:00
|
|
|
fprintf(stdout, "%-12s: Wrote %ld times\n", "", writes_);
|
2012-11-08 00:35:08 +01:00
|
|
|
fprintf(stdout, "%-12s: Deleted %ld times\n", "", deletes_);
|
2013-08-13 22:58:02 +02:00
|
|
|
fprintf(stdout, "%-12s: %ld read and %ld found the key\n", "",
|
|
|
|
gets_, founds_);
|
2013-08-15 01:58:36 +02:00
|
|
|
fprintf(stdout, "%-12s: Prefix scanned %ld times\n", "", prefixes_);
|
|
|
|
fprintf(stdout, "%-12s: Iterator size sum is %ld\n", "",
|
|
|
|
iterator_size_sums_);
|
2013-09-20 01:47:24 +02:00
|
|
|
fprintf(stdout, "%-12s: Iterated %ld times\n", "", iterations_);
|
2013-02-21 00:57:27 +01:00
|
|
|
fprintf(stdout, "%-12s: Got errors %ld times\n", "", errors_);
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
if (FLAGS_histogram) {
|
|
|
|
fprintf(stdout, "Microseconds per op:\n%s\n", hist_.ToString().c_str());
|
|
|
|
}
|
|
|
|
fflush(stdout);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// State shared by all concurrent executions of the same benchmark.
|
|
|
|
class SharedState {
|
|
|
|
public:
|
|
|
|
static const uint32_t SENTINEL = 0xffffffff;
|
|
|
|
|
2013-02-15 20:53:17 +01:00
|
|
|
explicit SharedState(StressTest* stress_test) :
|
2012-10-03 18:58:45 +02:00
|
|
|
cv_(&mu_),
|
|
|
|
seed_(FLAGS_seed),
|
|
|
|
max_key_(FLAGS_max_key),
|
|
|
|
log2_keys_per_lock_(FLAGS_log2_keys_per_lock),
|
|
|
|
num_threads_(FLAGS_threads),
|
|
|
|
num_initialized_(0),
|
|
|
|
num_populated_(0),
|
2012-11-09 22:04:12 +01:00
|
|
|
vote_reopen_(0),
|
2012-10-03 18:58:45 +02:00
|
|
|
num_done_(0),
|
|
|
|
start_(false),
|
|
|
|
start_verify_(false),
|
|
|
|
stress_test_(stress_test) {
|
2013-02-21 00:57:27 +01:00
|
|
|
if (FLAGS_test_batches_snapshots) {
|
|
|
|
key_locks_ = nullptr;
|
|
|
|
values_ = nullptr;
|
|
|
|
fprintf(stdout, "No lock creation because test_batches_snapshots set\n");
|
|
|
|
return;
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
values_ = new uint32_t[max_key_];
|
|
|
|
for (long i = 0; i < max_key_; i++) {
|
|
|
|
values_[i] = SENTINEL;
|
|
|
|
}
|
2013-02-21 00:57:27 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
long num_locks = (max_key_ >> log2_keys_per_lock_);
|
|
|
|
if (max_key_ & ((1 << log2_keys_per_lock_) - 1)) {
|
|
|
|
num_locks ++;
|
|
|
|
}
|
|
|
|
fprintf(stdout, "Creating %ld locks\n", num_locks);
|
|
|
|
key_locks_ = new port::Mutex[num_locks];
|
|
|
|
}
|
|
|
|
|
|
|
|
~SharedState() {
|
|
|
|
delete[] values_;
|
|
|
|
delete[] key_locks_;
|
|
|
|
}
|
|
|
|
|
|
|
|
port::Mutex* GetMutex() {
|
|
|
|
return &mu_;
|
|
|
|
}
|
|
|
|
|
|
|
|
port::CondVar* GetCondVar() {
|
|
|
|
return &cv_;
|
|
|
|
}
|
|
|
|
|
|
|
|
StressTest* GetStressTest() const {
|
|
|
|
return stress_test_;
|
|
|
|
}
|
|
|
|
|
|
|
|
long GetMaxKey() const {
|
|
|
|
return max_key_;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint32_t GetNumThreads() const {
|
|
|
|
return num_threads_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void IncInitialized() {
|
|
|
|
num_initialized_++;
|
|
|
|
}
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
void IncOperated() {
|
2012-10-03 18:58:45 +02:00
|
|
|
num_populated_++;
|
|
|
|
}
|
|
|
|
|
|
|
|
void IncDone() {
|
|
|
|
num_done_++;
|
|
|
|
}
|
|
|
|
|
2012-11-09 22:04:12 +01:00
|
|
|
void IncVotedReopen() {
|
|
|
|
vote_reopen_ = (vote_reopen_ + 1) % num_threads_;
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
bool AllInitialized() const {
|
|
|
|
return num_initialized_ >= num_threads_;
|
|
|
|
}
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
bool AllOperated() const {
|
2012-10-03 18:58:45 +02:00
|
|
|
return num_populated_ >= num_threads_;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AllDone() const {
|
|
|
|
return num_done_ >= num_threads_;
|
|
|
|
}
|
|
|
|
|
2012-11-09 22:04:12 +01:00
|
|
|
bool AllVotedReopen() {
|
|
|
|
return (vote_reopen_ == 0);
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
void SetStart() {
|
|
|
|
start_ = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
void SetStartVerify() {
|
|
|
|
start_verify_ = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool Started() const {
|
|
|
|
return start_;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool VerifyStarted() const {
|
|
|
|
return start_verify_;
|
|
|
|
}
|
|
|
|
|
|
|
|
port::Mutex* GetMutexForKey(long key) {
|
|
|
|
return &key_locks_[key >> log2_keys_per_lock_];
|
|
|
|
}
|
|
|
|
|
|
|
|
void Put(long key, uint32_t value_base) {
|
|
|
|
values_[key] = value_base;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint32_t Get(long key) const {
|
|
|
|
return values_[key];
|
|
|
|
}
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
void Delete(long key) const {
|
|
|
|
values_[key] = SENTINEL;
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
uint32_t GetSeed() const {
|
|
|
|
return seed_;
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
port::Mutex mu_;
|
|
|
|
port::CondVar cv_;
|
|
|
|
const uint32_t seed_;
|
|
|
|
const long max_key_;
|
|
|
|
const uint32_t log2_keys_per_lock_;
|
|
|
|
const int num_threads_;
|
|
|
|
long num_initialized_;
|
|
|
|
long num_populated_;
|
2012-11-09 22:04:12 +01:00
|
|
|
long vote_reopen_;
|
2012-10-03 18:58:45 +02:00
|
|
|
long num_done_;
|
|
|
|
bool start_;
|
|
|
|
bool start_verify_;
|
|
|
|
StressTest* stress_test_;
|
|
|
|
|
|
|
|
uint32_t *values_;
|
|
|
|
port::Mutex *key_locks_;
|
|
|
|
|
|
|
|
};
|
|
|
|
|
|
|
|
// Per-thread state for concurrent executions of the same benchmark.
|
|
|
|
struct ThreadState {
|
|
|
|
uint32_t tid; // 0..n-1
|
|
|
|
Random rand; // Has different seeds for different threads
|
|
|
|
SharedState* shared;
|
|
|
|
Stats stats;
|
|
|
|
|
|
|
|
ThreadState(uint32_t index, SharedState *shared)
|
|
|
|
: tid(index),
|
|
|
|
rand(1000 + index + shared->GetSeed()),
|
|
|
|
shared(shared) {
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
} // namespace
|
|
|
|
|
|
|
|
class StressTest {
|
|
|
|
public:
|
|
|
|
StressTest()
|
|
|
|
: cache_(NewLRUCache(FLAGS_cache_size)),
|
|
|
|
filter_policy_(FLAGS_bloom_bits >= 0
|
|
|
|
? NewBloomFilterPolicy(FLAGS_bloom_bits)
|
2013-02-15 20:53:17 +01:00
|
|
|
: nullptr),
|
2013-08-15 01:58:36 +02:00
|
|
|
prefix_extractor_(NewFixedPrefixTransform(
|
|
|
|
FLAGS_test_batches_snapshots ?
|
|
|
|
sizeof(long) : sizeof(long)-1)),
|
2013-02-15 20:53:17 +01:00
|
|
|
db_(nullptr),
|
2012-11-17 00:28:14 +01:00
|
|
|
num_times_reopened_(0) {
|
2013-04-08 21:35:40 +02:00
|
|
|
if (FLAGS_destroy_db_initially) {
|
|
|
|
std::vector<std::string> files;
|
|
|
|
FLAGS_env->GetChildren(FLAGS_db, &files);
|
|
|
|
for (unsigned int i = 0; i < files.size(); i++) {
|
|
|
|
if (Slice(files[i]).starts_with("heap-")) {
|
|
|
|
FLAGS_env->DeleteFile(std::string(FLAGS_db) + "/" + files[i]);
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
2013-04-08 21:35:40 +02:00
|
|
|
DestroyDB(FLAGS_db, Options());
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
~StressTest() {
|
|
|
|
delete db_;
|
|
|
|
delete filter_policy_;
|
2013-08-15 01:58:36 +02:00
|
|
|
delete prefix_extractor_;
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
void Run() {
|
|
|
|
PrintEnv();
|
|
|
|
Open();
|
|
|
|
SharedState shared(this);
|
|
|
|
uint32_t n = shared.GetNumThreads();
|
|
|
|
|
|
|
|
std::vector<ThreadState*> threads(n);
|
|
|
|
for (uint32_t i = 0; i < n; i++) {
|
|
|
|
threads[i] = new ThreadState(i, &shared);
|
|
|
|
FLAGS_env->StartThread(ThreadBody, threads[i]);
|
|
|
|
}
|
|
|
|
// Each thread goes through the following states:
|
2012-11-08 00:35:08 +01:00
|
|
|
// initializing -> wait for others to init -> read/populate/depopulate
|
|
|
|
// wait for others to operate -> verify -> done
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
{
|
|
|
|
MutexLock l(shared.GetMutex());
|
|
|
|
while (!shared.AllInitialized()) {
|
|
|
|
shared.GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
|
2012-11-17 00:28:14 +01:00
|
|
|
double now = FLAGS_env->NowMicros();
|
|
|
|
fprintf(stdout, "%s Starting database operations\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str());
|
2012-11-08 00:35:08 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
shared.SetStart();
|
|
|
|
shared.GetCondVar()->SignalAll();
|
2012-11-08 00:35:08 +01:00
|
|
|
while (!shared.AllOperated()) {
|
2012-10-03 18:58:45 +02:00
|
|
|
shared.GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
|
2012-11-17 00:28:14 +01:00
|
|
|
now = FLAGS_env->NowMicros();
|
2013-02-23 20:11:16 +01:00
|
|
|
if (FLAGS_test_batches_snapshots) {
|
|
|
|
fprintf(stdout, "%s Limited verification already done during gets\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str());
|
|
|
|
} else {
|
|
|
|
fprintf(stdout, "%s Starting verification\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str());
|
|
|
|
}
|
2012-11-17 00:28:14 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
shared.SetStartVerify();
|
|
|
|
shared.GetCondVar()->SignalAll();
|
|
|
|
while (!shared.AllDone()) {
|
|
|
|
shared.GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-11-06 21:02:18 +01:00
|
|
|
for (unsigned int i = 1; i < n; i++) {
|
2012-10-03 18:58:45 +02:00
|
|
|
threads[0]->stats.Merge(threads[i]->stats);
|
|
|
|
}
|
|
|
|
threads[0]->stats.Report("Stress Test");
|
|
|
|
|
2012-11-06 21:02:18 +01:00
|
|
|
for (unsigned int i = 0; i < n; i++) {
|
2012-10-03 18:58:45 +02:00
|
|
|
delete threads[i];
|
2013-02-15 20:53:17 +01:00
|
|
|
threads[i] = nullptr;
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
2012-11-17 00:28:14 +01:00
|
|
|
double now = FLAGS_env->NowMicros();
|
2013-02-23 20:11:16 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
|
|
|
fprintf(stdout, "%s Verification successful\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str());
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
PrintStatistics();
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
|
|
|
|
static void ThreadBody(void* v) {
|
|
|
|
ThreadState* thread = reinterpret_cast<ThreadState*>(v);
|
|
|
|
SharedState* shared = thread->shared;
|
|
|
|
|
|
|
|
{
|
|
|
|
MutexLock l(shared->GetMutex());
|
|
|
|
shared->IncInitialized();
|
|
|
|
if (shared->AllInitialized()) {
|
|
|
|
shared->GetCondVar()->SignalAll();
|
|
|
|
}
|
|
|
|
while (!shared->Started()) {
|
|
|
|
shared->GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
}
|
2012-11-08 00:35:08 +01:00
|
|
|
thread->shared->GetStressTest()->OperateDb(thread);
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
{
|
|
|
|
MutexLock l(shared->GetMutex());
|
2012-11-08 00:35:08 +01:00
|
|
|
shared->IncOperated();
|
|
|
|
if (shared->AllOperated()) {
|
2012-10-03 18:58:45 +02:00
|
|
|
shared->GetCondVar()->SignalAll();
|
|
|
|
}
|
|
|
|
while (!shared->VerifyStarted()) {
|
|
|
|
shared->GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
2013-10-02 08:19:51 +02:00
|
|
|
thread->shared->GetStressTest()->VerifyDb(thread);
|
2013-02-21 00:57:27 +01:00
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
{
|
|
|
|
MutexLock l(shared->GetMutex());
|
|
|
|
shared->IncDone();
|
|
|
|
if (shared->AllDone()) {
|
|
|
|
shared->GetCondVar()->SignalAll();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
// Given a key K and value V, this puts ("0"+K, "0"+V), ("1"+K, "1"+V), ...
|
|
|
|
// ("9"+K, "9"+V) in DB atomically i.e in a single batch.
|
|
|
|
// Also refer MultiGet.
|
|
|
|
Status MultiPut(ThreadState* thread,
|
|
|
|
const WriteOptions& writeoptions,
|
|
|
|
const Slice& key, const Slice& value, size_t sz) {
|
|
|
|
std::string keys[10] = {"9", "8", "7", "6", "5",
|
|
|
|
"4", "3", "2", "1", "0"};
|
|
|
|
std::string values[10] = {"9", "8", "7", "6", "5",
|
|
|
|
"4", "3", "2", "1", "0"};
|
|
|
|
Slice value_slices[10];
|
|
|
|
WriteBatch batch;
|
|
|
|
Status s;
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
keys[i] += key.ToString();
|
|
|
|
values[i] += value.ToString();
|
|
|
|
value_slices[i] = values[i];
|
Benchmarking for Merge Operator
Summary:
Updated db_bench and utilities/merge_operators.h to allow for dynamic benchmarking
of merge operators in db_bench. Added a new test (--benchmarks=mergerandom), which performs
a bunch of random Merge() operations over random keys. Also added a "--merge_operator=" flag
so that the tester can easily benchmark different merge operators. Currently supports
the PutOperator and UInt64Add operator. Support for stringappend or list append may come later.
Test Plan:
1. make db_bench
2. Test the PutOperator (simulating Put) as follows:
./db_bench --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom --merge_operator=put
--threads=2
3. Test the UInt64AddOperator (simulating numeric addition) similarly:
./db_bench --value_size=8 --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom
--merge_operator=uint64add --threads=2
Reviewers: haobo, dhruba, zshao, MarkCallaghan
Reviewed By: haobo
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11535
2013-08-16 02:13:07 +02:00
|
|
|
if (FLAGS_use_merge_put) {
|
|
|
|
batch.Merge(keys[i], value_slices[i]);
|
|
|
|
} else {
|
|
|
|
batch.Put(keys[i], value_slices[i]);
|
|
|
|
}
|
2013-02-21 00:57:27 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
s = db_->Write(writeoptions, &batch);
|
|
|
|
if (!s.ok()) {
|
|
|
|
fprintf(stderr, "multiput error: %s\n", s.ToString().c_str());
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
} else {
|
|
|
|
// we did 10 writes each of size sz + 1
|
|
|
|
thread->stats.AddBytesForWrites(10, (sz + 1) * 10);
|
|
|
|
}
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Given a key K, this deletes ("0"+K), ("1"+K),... ("9"+K)
|
|
|
|
// in DB atomically i.e in a single batch. Also refer MultiGet.
|
|
|
|
Status MultiDelete(ThreadState* thread,
|
|
|
|
const WriteOptions& writeoptions,
|
|
|
|
const Slice& key) {
|
|
|
|
std::string keys[10] = {"9", "7", "5", "3", "1",
|
|
|
|
"8", "6", "4", "2", "0"};
|
|
|
|
|
|
|
|
WriteBatch batch;
|
|
|
|
Status s;
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
keys[i] += key.ToString();
|
|
|
|
batch.Delete(keys[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
s = db_->Write(writeoptions, &batch);
|
|
|
|
if (!s.ok()) {
|
|
|
|
fprintf(stderr, "multidelete error: %s\n", s.ToString().c_str());
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
} else {
|
|
|
|
thread->stats.AddDeletes(10);
|
|
|
|
}
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Given a key K, this gets values for "0"+K, "1"+K,..."9"+K
|
|
|
|
// in the same snapshot, and verifies that all the values are of the form
|
|
|
|
// "0"+V, "1"+V,..."9"+V.
|
|
|
|
// ASSUMES that MultiPut was used to put (K, V) into the DB.
|
|
|
|
Status MultiGet(ThreadState* thread,
|
|
|
|
const ReadOptions& readoptions,
|
|
|
|
const Slice& key, std::string* value) {
|
|
|
|
std::string keys[10] = {"0", "1", "2", "3", "4", "5", "6", "7", "8", "9"};
|
|
|
|
Slice key_slices[10];
|
|
|
|
std::string values[10];
|
|
|
|
ReadOptions readoptionscopy = readoptions;
|
|
|
|
readoptionscopy.snapshot = db_->GetSnapshot();
|
|
|
|
Status s;
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
keys[i] += key.ToString();
|
|
|
|
key_slices[i] = keys[i];
|
|
|
|
s = db_->Get(readoptionscopy, key_slices[i], value);
|
|
|
|
if (!s.ok() && !s.IsNotFound()) {
|
|
|
|
fprintf(stderr, "get error: %s\n", s.ToString().c_str());
|
|
|
|
values[i] = "";
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
// we continue after error rather than exiting so that we can
|
|
|
|
// find more errors if any
|
|
|
|
} else if (s.IsNotFound()) {
|
|
|
|
values[i] = "";
|
2013-02-23 20:11:16 +01:00
|
|
|
thread->stats.AddGets(1, 0);
|
2013-02-21 00:57:27 +01:00
|
|
|
} else {
|
|
|
|
values[i] = *value;
|
|
|
|
|
|
|
|
char expected_prefix = (keys[i])[0];
|
|
|
|
char actual_prefix = (values[i])[0];
|
|
|
|
if (actual_prefix != expected_prefix) {
|
2013-04-05 08:49:43 +02:00
|
|
|
fprintf(stderr, "error expected prefix = %c actual = %c\n",
|
2013-02-21 00:57:27 +01:00
|
|
|
expected_prefix, actual_prefix);
|
|
|
|
}
|
|
|
|
(values[i])[0] = ' '; // blank out the differing character
|
2013-02-23 20:11:16 +01:00
|
|
|
thread->stats.AddGets(1, 1);
|
2013-02-21 00:57:27 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
db_->ReleaseSnapshot(readoptionscopy.snapshot);
|
|
|
|
|
|
|
|
// Now that we retrieved all values, check that they all match
|
|
|
|
for (int i = 1; i < 10; i++) {
|
|
|
|
if (values[i] != values[0]) {
|
2013-04-05 08:49:43 +02:00
|
|
|
fprintf(stderr, "error : inconsistent values for key %s: %s, %s\n",
|
2013-02-21 00:57:27 +01:00
|
|
|
key.ToString().c_str(), values[0].c_str(),
|
|
|
|
values[i].c_str());
|
|
|
|
// we continue after error rather than exiting so that we can
|
|
|
|
// find more errors if any
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2013-08-15 01:58:36 +02:00
|
|
|
// Given a prefix P, this does prefix scans for "0"+P, "1"+P,..."9"+P
|
|
|
|
// in the same snapshot. Each of these 10 scans returns a series of
|
|
|
|
// values; each series should be the same length, and it is verified
|
|
|
|
// for each index i that all the i'th values are of the form "0"+V,
|
|
|
|
// "1"+V,..."9"+V.
|
|
|
|
// ASSUMES that MultiPut was used to put (K, V)
|
|
|
|
Status MultiPrefixScan(ThreadState* thread,
|
|
|
|
const ReadOptions& readoptions,
|
|
|
|
const Slice& prefix) {
|
|
|
|
std::string prefixes[10] = {"0", "1", "2", "3", "4",
|
|
|
|
"5", "6", "7", "8", "9"};
|
|
|
|
Slice prefix_slices[10];
|
|
|
|
ReadOptions readoptionscopy[10];
|
|
|
|
const Snapshot* snapshot = db_->GetSnapshot();
|
|
|
|
Iterator* iters[10];
|
|
|
|
Status s = Status::OK();
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
prefixes[i] += prefix.ToString();
|
|
|
|
prefix_slices[i] = prefixes[i];
|
|
|
|
readoptionscopy[i] = readoptions;
|
|
|
|
readoptionscopy[i].prefix = &prefix_slices[i];
|
|
|
|
readoptionscopy[i].snapshot = snapshot;
|
|
|
|
iters[i] = db_->NewIterator(readoptionscopy[i]);
|
|
|
|
iters[i]->SeekToFirst();
|
|
|
|
}
|
|
|
|
|
|
|
|
int count = 0;
|
|
|
|
while (iters[0]->Valid()) {
|
|
|
|
count++;
|
|
|
|
std::string values[10];
|
|
|
|
// get list of all values for this iteration
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
// no iterator should finish before the first one
|
|
|
|
assert(iters[i]->Valid());
|
|
|
|
values[i] = iters[i]->value().ToString();
|
|
|
|
|
|
|
|
char expected_first = (prefixes[i])[0];
|
|
|
|
char actual_first = (values[i])[0];
|
|
|
|
|
|
|
|
if (actual_first != expected_first) {
|
|
|
|
fprintf(stderr, "error expected first = %c actual = %c\n",
|
|
|
|
expected_first, actual_first);
|
|
|
|
}
|
|
|
|
(values[i])[0] = ' '; // blank out the differing character
|
|
|
|
}
|
|
|
|
// make sure all values are equivalent
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
if (values[i] != values[0]) {
|
|
|
|
fprintf(stderr, "error : inconsistent values for prefix %s: %s, %s\n",
|
|
|
|
prefix.ToString().c_str(), values[0].c_str(),
|
|
|
|
values[i].c_str());
|
|
|
|
// we continue after error rather than exiting so that we can
|
|
|
|
// find more errors if any
|
|
|
|
}
|
|
|
|
iters[i]->Next();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// cleanup iterators and snapshot
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
// if the first iterator finished, they should have all finished
|
|
|
|
assert(!iters[i]->Valid());
|
|
|
|
assert(iters[i]->status().ok());
|
|
|
|
delete iters[i];
|
|
|
|
}
|
|
|
|
db_->ReleaseSnapshot(snapshot);
|
|
|
|
|
|
|
|
if (s.ok()) {
|
|
|
|
thread->stats.AddPrefixes(1, count);
|
|
|
|
} else {
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2013-09-20 01:47:24 +02:00
|
|
|
// Given a key K, this creates an iterator which scans to K and then
|
|
|
|
// does a random sequence of Next/Prev operations.
|
|
|
|
Status MultiIterate(ThreadState* thread,
|
|
|
|
const ReadOptions& readoptions,
|
|
|
|
const Slice& key) {
|
|
|
|
Status s;
|
|
|
|
const Snapshot* snapshot = db_->GetSnapshot();
|
|
|
|
ReadOptions readoptionscopy = readoptions;
|
|
|
|
readoptionscopy.snapshot = snapshot;
|
|
|
|
unique_ptr<Iterator> iter(db_->NewIterator(readoptionscopy));
|
|
|
|
|
|
|
|
iter->Seek(key);
|
|
|
|
for (long i = 0; i < FLAGS_num_iterations && iter->Valid(); i++) {
|
|
|
|
if (thread->rand.OneIn(2)) {
|
|
|
|
iter->Next();
|
|
|
|
} else {
|
|
|
|
iter->Prev();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s.ok()) {
|
|
|
|
thread->stats.AddIterations(1);
|
|
|
|
} else {
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
db_->ReleaseSnapshot(snapshot);
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
void OperateDb(ThreadState* thread) {
|
2012-10-03 18:58:45 +02:00
|
|
|
ReadOptions read_opts(FLAGS_verify_checksum, true);
|
|
|
|
WriteOptions write_opts;
|
2012-11-08 00:35:08 +01:00
|
|
|
char value[100];
|
2012-10-03 18:58:45 +02:00
|
|
|
long max_key = thread->shared->GetMaxKey();
|
|
|
|
std::string from_db;
|
|
|
|
if (FLAGS_sync) {
|
|
|
|
write_opts.sync = true;
|
|
|
|
}
|
|
|
|
write_opts.disableWAL = FLAGS_disable_wal;
|
2013-09-20 01:47:24 +02:00
|
|
|
const int prefixBound = (int)FLAGS_readpercent + (int)FLAGS_prefixpercent;
|
|
|
|
const int writeBound = prefixBound + (int)FLAGS_writepercent;
|
|
|
|
const int delBound = writeBound + (int)FLAGS_delpercent;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
thread->stats.Start();
|
2012-11-08 00:35:08 +01:00
|
|
|
for (long i = 0; i < FLAGS_ops_per_thread; i++) {
|
2012-11-09 22:04:12 +01:00
|
|
|
if(i != 0 && (i % (FLAGS_ops_per_thread / (FLAGS_reopen + 1))) == 0) {
|
|
|
|
{
|
2012-11-17 00:28:14 +01:00
|
|
|
thread->stats.FinishedSingleOp();
|
2012-11-09 22:04:12 +01:00
|
|
|
MutexLock l(thread->shared->GetMutex());
|
|
|
|
thread->shared->IncVotedReopen();
|
|
|
|
if (thread->shared->AllVotedReopen()) {
|
|
|
|
thread->shared->GetStressTest()->Reopen();
|
|
|
|
thread->shared->GetCondVar()->SignalAll();
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
thread->shared->GetCondVar()->Wait();
|
|
|
|
}
|
2013-02-23 20:11:16 +01:00
|
|
|
// Commenting this out as we don't want to reset stats on each open.
|
|
|
|
// thread->stats.Start();
|
2012-11-09 22:04:12 +01:00
|
|
|
}
|
|
|
|
}
|
2013-02-21 00:57:27 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
long rand_key = thread->rand.Next() % max_key;
|
2013-08-15 01:58:36 +02:00
|
|
|
std::string keystr = Key(rand_key);
|
|
|
|
Slice key = keystr;
|
|
|
|
int prob_op = thread->rand.Uniform(100);
|
|
|
|
|
|
|
|
if (prob_op >= 0 && prob_op < (int)FLAGS_readpercent) {
|
2013-09-20 01:47:24 +02:00
|
|
|
// OPERATION read
|
2013-02-23 20:11:16 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
|
|
|
Status s = db_->Get(read_opts, key, &from_db);
|
|
|
|
if (s.ok()) {
|
|
|
|
// found case
|
|
|
|
thread->stats.AddGets(1, 1);
|
|
|
|
} else if (s.IsNotFound()) {
|
|
|
|
// not found case
|
|
|
|
thread->stats.AddGets(1, 0);
|
|
|
|
} else {
|
|
|
|
// errors case
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
MultiGet(thread, read_opts, key, &from_db);
|
|
|
|
}
|
2013-09-20 01:47:24 +02:00
|
|
|
} else if ((int)FLAGS_readpercent <= prob_op && prob_op < prefixBound) {
|
|
|
|
// OPERATION prefix scan
|
2013-08-15 01:58:36 +02:00
|
|
|
// keys are longs (e.g., 8 bytes), so we let prefixes be
|
|
|
|
// everything except the last byte. So there will be 2^8=256
|
|
|
|
// keys per prefix.
|
|
|
|
Slice prefix = Slice(key.data(), key.size() - 1);
|
2013-02-21 00:57:27 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
2013-08-15 01:58:36 +02:00
|
|
|
read_opts.prefix = &prefix;
|
|
|
|
Iterator* iter = db_->NewIterator(read_opts);
|
|
|
|
int count = 0;
|
|
|
|
for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
|
|
|
|
assert(iter->key().starts_with(prefix));
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
assert(count <= 256);
|
|
|
|
if (iter->status().ok()) {
|
|
|
|
thread->stats.AddPrefixes(1, count);
|
|
|
|
} else {
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
}
|
|
|
|
delete iter;
|
2013-02-21 00:57:27 +01:00
|
|
|
} else {
|
2013-08-15 01:58:36 +02:00
|
|
|
MultiPrefixScan(thread, read_opts, prefix);
|
2012-11-08 00:35:08 +01:00
|
|
|
}
|
2013-09-20 01:47:24 +02:00
|
|
|
} else if (prefixBound <= prob_op && prob_op < writeBound) {
|
|
|
|
// OPERATION write
|
2012-10-03 18:58:45 +02:00
|
|
|
uint32_t value_base = thread->rand.Next();
|
|
|
|
size_t sz = GenerateValue(value_base, value, sizeof(value));
|
|
|
|
Slice v(value, sz);
|
2013-02-21 00:57:27 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
2012-10-03 18:58:45 +02:00
|
|
|
MutexLock l(thread->shared->GetMutexForKey(rand_key));
|
|
|
|
if (FLAGS_verify_before_write) {
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
std::string keystr = Key(rand_key);
|
|
|
|
Slice k = keystr;
|
|
|
|
Status s = db_->Get(read_opts, k, &from_db);
|
|
|
|
VerifyValue(rand_key,
|
|
|
|
read_opts,
|
|
|
|
*(thread->shared),
|
|
|
|
from_db,
|
|
|
|
s,
|
|
|
|
true);
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
thread->shared->Put(rand_key, value_base);
|
Benchmarking for Merge Operator
Summary:
Updated db_bench and utilities/merge_operators.h to allow for dynamic benchmarking
of merge operators in db_bench. Added a new test (--benchmarks=mergerandom), which performs
a bunch of random Merge() operations over random keys. Also added a "--merge_operator=" flag
so that the tester can easily benchmark different merge operators. Currently supports
the PutOperator and UInt64Add operator. Support for stringappend or list append may come later.
Test Plan:
1. make db_bench
2. Test the PutOperator (simulating Put) as follows:
./db_bench --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom --merge_operator=put
--threads=2
3. Test the UInt64AddOperator (simulating numeric addition) similarly:
./db_bench --value_size=8 --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom
--merge_operator=uint64add --threads=2
Reviewers: haobo, dhruba, zshao, MarkCallaghan
Reviewed By: haobo
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11535
2013-08-16 02:13:07 +02:00
|
|
|
if (FLAGS_use_merge_put) {
|
|
|
|
db_->Merge(write_opts, key, v);
|
|
|
|
} else {
|
|
|
|
db_->Put(write_opts, key, v);
|
|
|
|
}
|
2013-02-21 00:57:27 +01:00
|
|
|
thread->stats.AddBytesForWrites(1, sz);
|
|
|
|
} else {
|
|
|
|
MultiPut(thread, write_opts, key, v, sz);
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
PrintKeyValue(rand_key, value, sz);
|
2013-09-20 01:47:24 +02:00
|
|
|
} else if (writeBound <= prob_op && prob_op < delBound) {
|
|
|
|
// OPERATION delete
|
2013-08-15 01:58:36 +02:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
|
|
|
MutexLock l(thread->shared->GetMutexForKey(rand_key));
|
|
|
|
thread->shared->Delete(rand_key);
|
|
|
|
db_->Delete(write_opts, key);
|
|
|
|
thread->stats.AddDeletes(1);
|
|
|
|
} else {
|
|
|
|
MultiDelete(thread, write_opts, key);
|
|
|
|
}
|
2013-09-20 01:47:24 +02:00
|
|
|
} else {
|
|
|
|
// OPERATION iterate
|
|
|
|
MultiIterate(thread, read_opts, key);
|
2013-08-15 01:58:36 +02:00
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
thread->stats.FinishedSingleOp();
|
|
|
|
}
|
2013-09-20 01:47:24 +02:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
thread->stats.Stop();
|
|
|
|
}
|
|
|
|
|
2013-10-02 08:19:51 +02:00
|
|
|
void VerifyDb(ThreadState* thread) const {
|
2012-10-03 18:58:45 +02:00
|
|
|
ReadOptions options(FLAGS_verify_checksum, true);
|
2013-10-02 08:19:51 +02:00
|
|
|
const SharedState& shared = *(thread->shared);
|
|
|
|
static const long max_key = shared.GetMaxKey();
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
static const long keys_per_thread = max_key / shared.GetNumThreads();
|
2013-10-02 08:19:51 +02:00
|
|
|
long start = keys_per_thread * thread->tid;
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
long end = start + keys_per_thread;
|
2013-10-02 08:19:51 +02:00
|
|
|
if (thread->tid == shared.GetNumThreads() - 1) {
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
end = max_key;
|
|
|
|
}
|
2013-10-02 08:19:51 +02:00
|
|
|
if (!thread->rand.OneIn(2)) {
|
|
|
|
// Use iterator to verify this range
|
|
|
|
unique_ptr<Iterator> iter(db_->NewIterator(options));
|
|
|
|
iter->Seek(Key(start));
|
|
|
|
for (long i = start; i < end; i++) {
|
|
|
|
std::string from_db;
|
|
|
|
std::string keystr = Key(i);
|
|
|
|
Slice k = keystr;
|
|
|
|
Status s = iter->status();
|
|
|
|
if (iter->Valid()) {
|
|
|
|
if (iter->key().compare(k) > 0) {
|
|
|
|
s = Status::NotFound(Slice());
|
|
|
|
} else if (iter->key().compare(k) == 0) {
|
|
|
|
from_db = iter->value().ToString();
|
|
|
|
iter->Next();
|
|
|
|
} else if (iter->key().compare(k) < 0) {
|
|
|
|
VerificationAbort("An out of range key was found", i);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// The iterator found no value for the key in question, so do not
|
|
|
|
// move to the next item in the iterator
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
s = Status::NotFound(Slice());
|
|
|
|
}
|
2013-10-02 08:19:51 +02:00
|
|
|
VerifyValue(i, options, shared, from_db, s, true);
|
|
|
|
if (from_db.length()) {
|
|
|
|
PrintKeyValue(i, from_db.data(), from_db.length());
|
|
|
|
}
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
}
|
2013-10-02 08:19:51 +02:00
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Use Get to verify this range
|
|
|
|
for (long i = start; i < end; i++) {
|
|
|
|
std::string from_db;
|
|
|
|
std::string keystr = Key(i);
|
|
|
|
Slice k = keystr;
|
|
|
|
Status s = db_->Get(options, k, &from_db);
|
|
|
|
VerifyValue(i, options, shared, from_db, s, true);
|
|
|
|
if (from_db.length()) {
|
|
|
|
PrintKeyValue(i, from_db.data(), from_db.length());
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-10-13 02:00:25 +02:00
|
|
|
void VerificationAbort(std::string msg, long key) const {
|
2012-11-08 00:35:08 +01:00
|
|
|
fprintf(stderr, "Verification failed for key %ld: %s\n",
|
2012-10-13 02:00:25 +02:00
|
|
|
key, msg.c_str());
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
void VerifyValue(long key,
|
|
|
|
const ReadOptions &opts,
|
|
|
|
const SharedState &shared,
|
|
|
|
const std::string &value_from_db,
|
|
|
|
Status s,
|
|
|
|
bool strict=false) const {
|
|
|
|
// compare value_from_db with the value in the shared state
|
2012-11-08 00:35:08 +01:00
|
|
|
char value[100];
|
2012-10-03 18:58:45 +02:00
|
|
|
uint32_t value_base = shared.Get(key);
|
|
|
|
if (value_base == SharedState::SENTINEL && !strict) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
if (s.ok()) {
|
2012-10-03 18:58:45 +02:00
|
|
|
if (value_base == SharedState::SENTINEL) {
|
|
|
|
VerificationAbort("Unexpected value found", key);
|
|
|
|
}
|
2013-02-21 02:37:13 +01:00
|
|
|
size_t sz = GenerateValue(value_base, value, sizeof(value));
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
if (value_from_db.length() != sz) {
|
2012-10-03 18:58:45 +02:00
|
|
|
VerificationAbort("Length of value read is not equal", key);
|
|
|
|
}
|
Phase 2 of iterator stress test
Summary: Using an iterator instead of the Get method, each thread goes through a portion of the database and verifies values by comparing to the shared state.
Test Plan:
./db_stress --db=/tmp/tmppp --max_key=10000 --ops_per_thread=10000
To test some basic cases, the following lines can be added (each set in turn) to the verifyDb method with the following expected results:
// Should abort with "Unexpected value found"
shared.Delete(start);
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start));
// Should succeed
WriteOptions write_opts;
shared.Delete(start);
db_->Delete(write_opts, Key(start));
// Should abort with "Value not found"
WriteOptions write_opts;
db_->Delete(write_opts, Key(start + (end-start)/2));
// Should abort with "Value not found"
db_->Delete(write_opts, Key(end-1));
// Should abort with "Unexpected value"
shared.Delete(end-1);
// Should abort with "Unexpected value"
shared.Delete(start + (end-start)/2);
// Should abort with "Value not found"
db_->Delete(write_opts, Key(start));
shared.Delete(start);
db_->Delete(write_opts, Key(end-1));
db_->Delete(write_opts, Key(end-2));
To test the out of range abort, change the key in the for loop to Key(i+1), so that the key defined by the index i is now outside of the supposed range of the database.
Reviewers: emayanke
Reviewed By: emayanke
CC: dhruba, xjin
Differential Revision: https://reviews.facebook.net/D13071
2013-10-01 01:48:00 +02:00
|
|
|
if (memcmp(value_from_db.data(), value, sz) != 0) {
|
2012-10-03 18:58:45 +02:00
|
|
|
VerificationAbort("Contents of value read don't match", key);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (value_base != SharedState::SENTINEL) {
|
|
|
|
VerificationAbort("Value not found", key);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void PrintKeyValue(uint32_t key, const char *value, size_t sz) {
|
|
|
|
if (!FLAGS_verbose) return;
|
2012-11-06 21:02:18 +01:00
|
|
|
fprintf(stdout, "%u ==> (%u) ", key, (unsigned int)sz);
|
2012-10-03 18:58:45 +02:00
|
|
|
for (size_t i=0; i<sz; i++) {
|
|
|
|
fprintf(stdout, "%X", value[i]);
|
|
|
|
}
|
|
|
|
fprintf(stdout, "\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
static size_t GenerateValue(uint32_t rand, char *v, size_t max_sz) {
|
|
|
|
size_t value_sz = ((rand % 3) + 1) * FLAGS_value_size_mult;
|
|
|
|
assert(value_sz <= max_sz && value_sz >= sizeof(uint32_t));
|
|
|
|
*((uint32_t*)v) = rand;
|
|
|
|
for (size_t i=sizeof(uint32_t); i < value_sz; i++) {
|
|
|
|
v[i] = (char)(rand ^ i);
|
|
|
|
}
|
2013-05-20 23:45:22 +02:00
|
|
|
v[value_sz] = '\0';
|
2012-10-03 18:58:45 +02:00
|
|
|
return value_sz; // the size of the value set.
|
|
|
|
}
|
|
|
|
|
|
|
|
void PrintEnv() const {
|
|
|
|
fprintf(stdout, "LevelDB version : %d.%d\n",
|
|
|
|
kMajorVersion, kMinorVersion);
|
|
|
|
fprintf(stdout, "Number of threads : %d\n", FLAGS_threads);
|
2012-11-06 21:02:18 +01:00
|
|
|
fprintf(stdout, "Ops per thread : %d\n", FLAGS_ops_per_thread);
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
std::string ttl_state("unused");
|
|
|
|
if (FLAGS_ttl > 0) {
|
|
|
|
ttl_state = NumberToString(FLAGS_ttl);
|
|
|
|
}
|
|
|
|
fprintf(stdout, "Time to live(sec) : %s\n", ttl_state.c_str());
|
2012-11-08 00:35:08 +01:00
|
|
|
fprintf(stdout, "Read percentage : %d\n", FLAGS_readpercent);
|
2013-08-15 01:58:36 +02:00
|
|
|
fprintf(stdout, "Prefix percentage : %d\n", FLAGS_prefixpercent);
|
|
|
|
fprintf(stdout, "Write percentage : %d\n", FLAGS_writepercent);
|
|
|
|
fprintf(stdout, "Delete percentage : %d\n", FLAGS_delpercent);
|
2013-09-20 01:47:24 +02:00
|
|
|
fprintf(stdout, "Iterate percentage : %d\n", FLAGS_iterpercent);
|
2013-03-13 07:20:14 +01:00
|
|
|
fprintf(stdout, "Write-buffer-size : %d\n", FLAGS_write_buffer_size);
|
2013-09-20 01:47:24 +02:00
|
|
|
fprintf(stdout, "Iterations : %d\n", FLAGS_num_iterations);
|
2012-10-03 18:58:45 +02:00
|
|
|
fprintf(stdout, "Max key : %ld\n", FLAGS_max_key);
|
2013-08-13 22:58:02 +02:00
|
|
|
fprintf(stdout, "Ratio #ops/#keys : %f\n",
|
|
|
|
(1.0 * FLAGS_ops_per_thread * FLAGS_threads)/FLAGS_max_key);
|
2012-11-09 22:04:12 +01:00
|
|
|
fprintf(stdout, "Num times DB reopens: %d\n", FLAGS_reopen);
|
2013-02-21 00:57:27 +01:00
|
|
|
fprintf(stdout, "Batches/snapshots : %d\n",
|
|
|
|
FLAGS_test_batches_snapshots);
|
2013-08-13 22:58:02 +02:00
|
|
|
fprintf(stdout, "Purge redundant %% : %d\n",
|
2013-03-06 21:54:55 +01:00
|
|
|
FLAGS_purge_redundant_percent);
|
2013-07-06 03:49:18 +02:00
|
|
|
fprintf(stdout, "Deletes use filter : %d\n",
|
2013-07-13 01:56:52 +02:00
|
|
|
FLAGS_filter_deletes);
|
2012-11-06 21:02:18 +01:00
|
|
|
fprintf(stdout, "Num keys per lock : %d\n",
|
2012-10-03 18:58:45 +02:00
|
|
|
1 << FLAGS_log2_keys_per_lock);
|
|
|
|
|
2013-01-17 19:04:45 +01:00
|
|
|
const char* compression = "";
|
2012-10-03 18:58:45 +02:00
|
|
|
switch (FLAGS_compression_type) {
|
2013-10-04 06:49:15 +02:00
|
|
|
case rocksdb::kNoCompression:
|
2013-01-17 19:04:45 +01:00
|
|
|
compression = "none";
|
2012-10-03 18:58:45 +02:00
|
|
|
break;
|
2013-10-04 06:49:15 +02:00
|
|
|
case rocksdb::kSnappyCompression:
|
2013-01-17 19:04:45 +01:00
|
|
|
compression = "snappy";
|
2012-10-03 18:58:45 +02:00
|
|
|
break;
|
2013-10-04 06:49:15 +02:00
|
|
|
case rocksdb::kZlibCompression:
|
2013-01-17 19:04:45 +01:00
|
|
|
compression = "zlib";
|
2012-10-03 18:58:45 +02:00
|
|
|
break;
|
2013-10-04 06:49:15 +02:00
|
|
|
case rocksdb::kBZip2Compression:
|
2013-01-17 19:04:45 +01:00
|
|
|
compression = "bzip2";
|
2012-10-03 18:58:45 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
fprintf(stdout, "Compression : %s\n", compression);
|
2013-08-23 08:10:02 +02:00
|
|
|
|
|
|
|
const char* memtablerep = "";
|
|
|
|
switch (FLAGS_rep_factory) {
|
|
|
|
case kSkipList:
|
|
|
|
memtablerep = "skip_list";
|
|
|
|
break;
|
|
|
|
case kPrefixHash:
|
|
|
|
memtablerep = "prefix_hash";
|
|
|
|
break;
|
|
|
|
case kUnsorted:
|
|
|
|
memtablerep = "unsorted";
|
|
|
|
break;
|
|
|
|
case kVectorRep:
|
|
|
|
memtablerep = "vector";
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
fprintf(stdout, "Memtablerep : %s\n", memtablerep);
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
fprintf(stdout, "------------------------------------------------\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
void Open() {
|
2013-02-15 20:53:17 +01:00
|
|
|
assert(db_ == nullptr);
|
2012-10-03 18:58:45 +02:00
|
|
|
Options options;
|
|
|
|
options.block_cache = cache_;
|
|
|
|
options.write_buffer_size = FLAGS_write_buffer_size;
|
2012-10-19 23:00:53 +02:00
|
|
|
options.max_write_buffer_number = FLAGS_max_write_buffer_number;
|
2013-09-10 01:06:10 +02:00
|
|
|
options.min_write_buffer_number_to_merge =
|
|
|
|
FLAGS_min_write_buffer_number_to_merge;
|
2012-10-19 23:00:53 +02:00
|
|
|
options.max_background_compactions = FLAGS_max_background_compactions;
|
2013-07-04 00:32:49 +02:00
|
|
|
options.compaction_style = FLAGS_compaction_style;
|
2012-10-03 18:58:45 +02:00
|
|
|
options.block_size = FLAGS_block_size;
|
|
|
|
options.filter_policy = filter_policy_;
|
2013-08-15 01:58:36 +02:00
|
|
|
options.prefix_extractor = prefix_extractor_;
|
2012-10-03 18:58:45 +02:00
|
|
|
options.max_open_files = FLAGS_open_files;
|
|
|
|
options.statistics = dbstats;
|
|
|
|
options.env = FLAGS_env;
|
|
|
|
options.disableDataSync = FLAGS_disable_data_sync;
|
|
|
|
options.use_fsync = FLAGS_use_fsync;
|
2013-06-18 01:13:32 +02:00
|
|
|
options.allow_mmap_reads = FLAGS_use_mmap_reads;
|
2013-10-05 07:32:05 +02:00
|
|
|
rocksdb_kill_odds = FLAGS_kill_random_test;
|
2012-10-03 18:58:45 +02:00
|
|
|
options.target_file_size_base = FLAGS_target_file_size_base;
|
|
|
|
options.target_file_size_multiplier = FLAGS_target_file_size_multiplier;
|
|
|
|
options.max_bytes_for_level_base = FLAGS_max_bytes_for_level_base;
|
|
|
|
options.max_bytes_for_level_multiplier =
|
|
|
|
FLAGS_max_bytes_for_level_multiplier;
|
|
|
|
options.level0_stop_writes_trigger = FLAGS_level0_stop_writes_trigger;
|
|
|
|
options.level0_slowdown_writes_trigger =
|
|
|
|
FLAGS_level0_slowdown_writes_trigger;
|
2013-06-17 22:51:12 +02:00
|
|
|
options.level0_file_num_compaction_trigger =
|
|
|
|
FLAGS_level0_file_num_compaction_trigger;
|
2012-10-03 18:58:45 +02:00
|
|
|
options.compression = FLAGS_compression_type;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options.disable_seek_compaction = FLAGS_disable_seek_compaction;
|
2012-10-19 23:00:53 +02:00
|
|
|
options.delete_obsolete_files_period_micros =
|
|
|
|
FLAGS_delete_obsolete_files_period_micros;
|
2013-01-11 02:18:50 +01:00
|
|
|
options.max_manifest_file_size = 1024;
|
2013-07-13 01:56:52 +02:00
|
|
|
options.filter_deletes = FLAGS_filter_deletes;
|
2013-08-23 08:10:02 +02:00
|
|
|
if ((FLAGS_prefix_size == 0) == (FLAGS_rep_factory == kPrefixHash)) {
|
|
|
|
fprintf(stderr,
|
|
|
|
"prefix_size should be non-zero iff memtablerep == prefix_hash\n");
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
switch (FLAGS_rep_factory) {
|
|
|
|
case kPrefixHash:
|
|
|
|
options.memtable_factory.reset(
|
|
|
|
new PrefixHashRepFactory(NewFixedPrefixTransform(FLAGS_prefix_size))
|
|
|
|
);
|
|
|
|
break;
|
|
|
|
case kUnsorted:
|
|
|
|
options.memtable_factory.reset(
|
|
|
|
new UnsortedRepFactory()
|
|
|
|
);
|
|
|
|
break;
|
|
|
|
case kSkipList:
|
|
|
|
// no need to do anything
|
|
|
|
break;
|
|
|
|
case kVectorRep:
|
|
|
|
options.memtable_factory.reset(
|
|
|
|
new VectorRepFactory()
|
|
|
|
);
|
|
|
|
break;
|
|
|
|
}
|
2013-03-06 21:54:55 +01:00
|
|
|
static Random purge_percent(1000); // no benefit from non-determinism here
|
|
|
|
if (purge_percent.Uniform(100) < FLAGS_purge_redundant_percent - 1) {
|
|
|
|
options.purge_redundant_kvs_while_flush = false;
|
|
|
|
}
|
2013-06-18 01:13:32 +02:00
|
|
|
|
Benchmarking for Merge Operator
Summary:
Updated db_bench and utilities/merge_operators.h to allow for dynamic benchmarking
of merge operators in db_bench. Added a new test (--benchmarks=mergerandom), which performs
a bunch of random Merge() operations over random keys. Also added a "--merge_operator=" flag
so that the tester can easily benchmark different merge operators. Currently supports
the PutOperator and UInt64Add operator. Support for stringappend or list append may come later.
Test Plan:
1. make db_bench
2. Test the PutOperator (simulating Put) as follows:
./db_bench --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom --merge_operator=put
--threads=2
3. Test the UInt64AddOperator (simulating numeric addition) similarly:
./db_bench --value_size=8 --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom
--merge_operator=uint64add --threads=2
Reviewers: haobo, dhruba, zshao, MarkCallaghan
Reviewed By: haobo
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11535
2013-08-16 02:13:07 +02:00
|
|
|
if (FLAGS_use_merge_put) {
|
2013-08-20 22:35:28 +02:00
|
|
|
options.merge_operator = MergeOperators::CreatePutOperator();
|
Benchmarking for Merge Operator
Summary:
Updated db_bench and utilities/merge_operators.h to allow for dynamic benchmarking
of merge operators in db_bench. Added a new test (--benchmarks=mergerandom), which performs
a bunch of random Merge() operations over random keys. Also added a "--merge_operator=" flag
so that the tester can easily benchmark different merge operators. Currently supports
the PutOperator and UInt64Add operator. Support for stringappend or list append may come later.
Test Plan:
1. make db_bench
2. Test the PutOperator (simulating Put) as follows:
./db_bench --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom --merge_operator=put
--threads=2
3. Test the UInt64AddOperator (simulating numeric addition) similarly:
./db_bench --value_size=8 --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom
--merge_operator=uint64add --threads=2
Reviewers: haobo, dhruba, zshao, MarkCallaghan
Reviewed By: haobo
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11535
2013-08-16 02:13:07 +02:00
|
|
|
}
|
|
|
|
|
2013-09-10 01:06:10 +02:00
|
|
|
// set universal style compaction configurations, if applicable
|
|
|
|
if (FLAGS_universal_size_ratio != 0) {
|
|
|
|
options.compaction_options_universal.size_ratio =
|
|
|
|
FLAGS_universal_size_ratio;
|
|
|
|
}
|
|
|
|
if (FLAGS_universal_min_merge_width != 0) {
|
|
|
|
options.compaction_options_universal.min_merge_width =
|
|
|
|
FLAGS_universal_min_merge_width;
|
|
|
|
}
|
|
|
|
if (FLAGS_universal_max_merge_width != 0) {
|
|
|
|
options.compaction_options_universal.max_merge_width =
|
|
|
|
FLAGS_universal_max_merge_width;
|
|
|
|
}
|
|
|
|
if (FLAGS_universal_max_size_amplification_percent != 0) {
|
|
|
|
options.compaction_options_universal.max_size_amplification_percent =
|
|
|
|
FLAGS_universal_max_size_amplification_percent;
|
|
|
|
}
|
|
|
|
|
2013-06-18 01:13:32 +02:00
|
|
|
fprintf(stdout, "DB path: [%s]\n", FLAGS_db);
|
|
|
|
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
Status s;
|
|
|
|
if (FLAGS_ttl == -1) {
|
|
|
|
s = DB::Open(options, FLAGS_db, &db_);
|
|
|
|
} else {
|
2013-08-06 02:55:44 +02:00
|
|
|
s = UtilityDB::OpenTtlDB(options, FLAGS_db, &sdb_, FLAGS_ttl);
|
|
|
|
db_ = sdb_;
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
if (!s.ok()) {
|
|
|
|
fprintf(stderr, "open error: %s\n", s.ToString().c_str());
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-11-09 22:04:12 +01:00
|
|
|
void Reopen() {
|
2012-11-17 00:28:14 +01:00
|
|
|
// do not close the db. Just delete the lock file. This
|
|
|
|
// simulates a crash-recovery kind of situation.
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
if (FLAGS_ttl != -1) {
|
|
|
|
((DBWithTTL*) db_)->TEST_Destroy_DBWithTtl();
|
|
|
|
} else {
|
|
|
|
((DBImpl*) db_)->TEST_Destroy_DBImpl();
|
|
|
|
}
|
2013-02-15 20:53:17 +01:00
|
|
|
db_ = nullptr;
|
2012-11-17 00:28:14 +01:00
|
|
|
|
|
|
|
num_times_reopened_++;
|
|
|
|
double now = FLAGS_env->NowMicros();
|
|
|
|
fprintf(stdout, "%s Reopening database for the %dth time\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str(),
|
|
|
|
num_times_reopened_);
|
2012-11-09 22:04:12 +01:00
|
|
|
Open();
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
void PrintStatistics() {
|
|
|
|
if (dbstats) {
|
2013-07-10 22:17:51 +02:00
|
|
|
fprintf(stdout, "STATISTICS:\n%s\n", dbstats->ToString().c_str());
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
2013-01-20 11:07:13 +01:00
|
|
|
shared_ptr<Cache> cache_;
|
2012-10-03 18:58:45 +02:00
|
|
|
const FilterPolicy* filter_policy_;
|
2013-08-15 01:58:36 +02:00
|
|
|
const SliceTransform* prefix_extractor_;
|
2012-10-03 18:58:45 +02:00
|
|
|
DB* db_;
|
2013-08-06 02:55:44 +02:00
|
|
|
StackableDB* sdb_;
|
2012-11-17 00:28:14 +01:00
|
|
|
int num_times_reopened_;
|
2012-10-03 18:58:45 +02:00
|
|
|
};
|
|
|
|
|
2013-10-04 06:49:15 +02:00
|
|
|
} // namespace rocksdb
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
2013-10-04 06:49:15 +02:00
|
|
|
FLAGS_write_buffer_size = rocksdb::Options().write_buffer_size;
|
|
|
|
FLAGS_max_write_buffer_number = rocksdb::Options().max_write_buffer_number;
|
2013-09-10 01:06:10 +02:00
|
|
|
FLAGS_min_write_buffer_number_to_merge =
|
2013-10-04 06:49:15 +02:00
|
|
|
rocksdb::Options().min_write_buffer_number_to_merge;
|
|
|
|
FLAGS_open_files = rocksdb::Options().max_open_files;
|
2012-10-19 23:00:53 +02:00
|
|
|
FLAGS_max_background_compactions =
|
2013-10-04 06:49:15 +02:00
|
|
|
rocksdb::Options().max_background_compactions;
|
2013-07-04 00:32:49 +02:00
|
|
|
FLAGS_compaction_style =
|
2013-10-04 06:49:15 +02:00
|
|
|
rocksdb::Options().compaction_style;
|
2013-06-17 22:51:12 +02:00
|
|
|
FLAGS_level0_file_num_compaction_trigger =
|
2013-10-04 06:49:15 +02:00
|
|
|
rocksdb::Options().level0_file_num_compaction_trigger;
|
2013-06-17 22:51:12 +02:00
|
|
|
FLAGS_level0_slowdown_writes_trigger =
|
2013-10-04 06:49:15 +02:00
|
|
|
rocksdb::Options().level0_slowdown_writes_trigger;
|
2013-06-17 22:51:12 +02:00
|
|
|
FLAGS_level0_stop_writes_trigger =
|
2013-10-04 06:49:15 +02:00
|
|
|
rocksdb::Options().level0_stop_writes_trigger;
|
2012-10-03 18:58:45 +02:00
|
|
|
// Compression test code above refers to FLAGS_block_size
|
2013-10-04 06:49:15 +02:00
|
|
|
FLAGS_block_size = rocksdb::Options().block_size;
|
2012-10-03 18:58:45 +02:00
|
|
|
std::string default_db_path;
|
|
|
|
|
|
|
|
for (int i = 1; i < argc; i++) {
|
|
|
|
int n;
|
|
|
|
uint32_t u;
|
|
|
|
long l;
|
|
|
|
char junk;
|
|
|
|
char hdfsname[2048];
|
|
|
|
|
|
|
|
if (sscanf(argv[i], "--seed=%uf%c", &u, &junk) == 1) {
|
|
|
|
FLAGS_seed = u;
|
|
|
|
} else if (sscanf(argv[i], "--max_key=%ld%c", &l, &junk) == 1) {
|
|
|
|
FLAGS_max_key = l;
|
|
|
|
} else if (sscanf(argv[i], "--log2_keys_per_lock=%u%c", &u, &junk) == 1) {
|
|
|
|
FLAGS_log2_keys_per_lock = u;
|
|
|
|
} else if (sscanf(argv[i], "--ops_per_thread=%u%c", &u, &junk) == 1) {
|
|
|
|
FLAGS_ops_per_thread = u;
|
|
|
|
} else if (sscanf(argv[i], "--verbose=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_verbose = n;
|
|
|
|
} else if (sscanf(argv[i], "--histogram=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_histogram = n;
|
2013-04-08 21:35:40 +02:00
|
|
|
} else if (sscanf(argv[i], "--destroy_db_initially=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_destroy_db_initially = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--verify_before_write=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_verify_before_write = n;
|
2013-02-21 00:57:27 +01:00
|
|
|
} else if (sscanf(argv[i], "--test_batches_snapshots=%d%c", &n, &junk) == 1
|
|
|
|
&& (n == 0 || n == 1)) {
|
|
|
|
FLAGS_test_batches_snapshots = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--threads=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_threads = n;
|
Timestamp and TTL Wrapper for rocksdb
Summary:
When opened with DBTimestamp::Open call, timestamps are prepended to and stripped from the value during subsequent Put and Get calls respectively. The Timestamp is used to discard values in Get and custom compaction filter which have exceeded their TTL which is specified during Open.
Have made a temporary change to Makefile to let us test with the temporary file TestTime.cc. Have also changed the private members of db_impl.h to protected to let them be inherited by the new class DBTimestamp
Test Plan: make db_timestamp; TestTime.cc(will not check it in) shows how to use the apis currently, but I will write unit-tests shortly
Reviewers: dhruba, vamsi, haobo, sheki, heyongqiang, vkrest
Reviewed By: vamsi
CC: zshao, xjin, vkrest, MarkCallaghan
Differential Revision: https://reviews.facebook.net/D10311
2013-04-15 22:33:13 +02:00
|
|
|
} else if (sscanf(argv[i], "--ttl=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_ttl = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--value_size_mult=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_value_size_mult = n;
|
|
|
|
} else if (sscanf(argv[i], "--write_buffer_size=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_write_buffer_size = n;
|
2012-10-19 23:00:53 +02:00
|
|
|
} else if (sscanf(argv[i], "--max_write_buffer_number=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_max_write_buffer_number = n;
|
2013-09-10 01:06:10 +02:00
|
|
|
} else if (sscanf(argv[i], "--min_write_buffer_number_to_merge=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_min_write_buffer_number_to_merge = n;
|
2012-10-19 23:00:53 +02:00
|
|
|
} else if (sscanf(argv[i], "--max_background_compactions=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_max_background_compactions = n;
|
2013-07-04 00:32:49 +02:00
|
|
|
} else if (sscanf(argv[i], "--compaction_style=%d%c", &n, &junk) == 1) {
|
2013-10-04 06:49:15 +02:00
|
|
|
FLAGS_compaction_style = (rocksdb::CompactionStyle)n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--cache_size=%ld%c", &l, &junk) == 1) {
|
|
|
|
FLAGS_cache_size = l;
|
|
|
|
} else if (sscanf(argv[i], "--block_size=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_block_size = n;
|
2012-11-09 22:04:12 +01:00
|
|
|
} else if (sscanf(argv[i], "--reopen=%d%c", &n, &junk) == 1 && n >= 0) {
|
|
|
|
FLAGS_reopen = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--bloom_bits=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_bloom_bits = n;
|
|
|
|
} else if (sscanf(argv[i], "--open_files=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_open_files = n;
|
|
|
|
} else if (strncmp(argv[i], "--db=", 5) == 0) {
|
|
|
|
FLAGS_db = argv[i] + 5;
|
|
|
|
} else if (sscanf(argv[i], "--verify_checksum=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_verify_checksum = n;
|
2013-06-18 01:13:32 +02:00
|
|
|
} else if (sscanf(argv[i], "--mmap_read=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_use_mmap_reads = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--statistics=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
if (n == 1) {
|
2013-10-04 06:49:15 +02:00
|
|
|
dbstats = rocksdb::CreateDBStatistics();
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
} else if (sscanf(argv[i], "--sync=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_sync = n;
|
2012-11-08 00:35:08 +01:00
|
|
|
} else if (sscanf(argv[i], "--readpercent=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_readpercent = n;
|
2013-08-15 01:58:36 +02:00
|
|
|
} else if (sscanf(argv[i], "--prefixpercent=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_prefixpercent = n;
|
|
|
|
} else if (sscanf(argv[i], "--writepercent=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_writepercent = n;
|
2013-09-20 01:47:24 +02:00
|
|
|
} else if (sscanf(argv[i], "--iterpercent=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_iterpercent = n;
|
2012-11-08 00:35:08 +01:00
|
|
|
} else if (sscanf(argv[i], "--delpercent=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_delpercent = n;
|
2013-09-20 01:47:24 +02:00
|
|
|
} else if (sscanf(argv[i], "--numiterations=%u%c", &u, &junk) == 1) {
|
|
|
|
FLAGS_num_iterations = u;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--disable_data_sync=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_disable_data_sync = n;
|
|
|
|
} else if (sscanf(argv[i], "--use_fsync=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_use_fsync = n;
|
2013-04-05 08:49:43 +02:00
|
|
|
} else if (sscanf(argv[i], "--kill_random_test=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n >= 0)) {
|
|
|
|
FLAGS_kill_random_test = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--disable_wal=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_disable_wal = n;
|
|
|
|
} else if (sscanf(argv[i], "--hdfs=%s", hdfsname) == 1) {
|
2013-10-04 06:49:15 +02:00
|
|
|
FLAGS_env = new rocksdb::HdfsEnv(hdfsname);
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--target_file_size_base=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_target_file_size_base = n;
|
|
|
|
} else if ( sscanf(argv[i], "--target_file_size_multiplier=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_target_file_size_multiplier = n;
|
|
|
|
} else if (
|
2012-11-09 03:45:19 +01:00
|
|
|
sscanf(argv[i], "--max_bytes_for_level_base=%ld%c", &l, &junk) == 1) {
|
|
|
|
FLAGS_max_bytes_for_level_base = l;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--max_bytes_for_level_multiplier=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_max_bytes_for_level_multiplier = n;
|
|
|
|
} else if (sscanf(argv[i],"--level0_stop_writes_trigger=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_level0_stop_writes_trigger = n;
|
|
|
|
} else if (sscanf(argv[i],"--level0_slowdown_writes_trigger=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_level0_slowdown_writes_trigger = n;
|
2013-06-17 22:51:12 +02:00
|
|
|
} else if (sscanf(argv[i],"--level0_file_num_compaction_trigger=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_level0_file_num_compaction_trigger = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (strncmp(argv[i], "--compression_type=", 19) == 0) {
|
|
|
|
const char* ctype = argv[i] + 19;
|
|
|
|
if (!strcasecmp(ctype, "none"))
|
2013-10-04 06:49:15 +02:00
|
|
|
FLAGS_compression_type = rocksdb::kNoCompression;
|
2012-10-03 18:58:45 +02:00
|
|
|
else if (!strcasecmp(ctype, "snappy"))
|
2013-10-04 06:49:15 +02:00
|
|
|
FLAGS_compression_type = rocksdb::kSnappyCompression;
|
2012-10-03 18:58:45 +02:00
|
|
|
else if (!strcasecmp(ctype, "zlib"))
|
2013-10-04 06:49:15 +02:00
|
|
|
FLAGS_compression_type = rocksdb::kZlibCompression;
|
2012-10-03 18:58:45 +02:00
|
|
|
else if (!strcasecmp(ctype, "bzip2"))
|
2013-10-04 06:49:15 +02:00
|
|
|
FLAGS_compression_type = rocksdb::kBZip2Compression;
|
2012-10-03 18:58:45 +02:00
|
|
|
else {
|
|
|
|
fprintf(stdout, "Cannot parse %s\n", argv[i]);
|
|
|
|
}
|
2013-08-23 08:10:02 +02:00
|
|
|
} else if (strncmp(argv[i], "--memtablerep=", 14) == 0) {
|
|
|
|
const char* ctype = argv[i] + 14;
|
|
|
|
if (!strcasecmp(ctype, "skip_list"))
|
|
|
|
FLAGS_rep_factory = kSkipList;
|
|
|
|
else if (!strcasecmp(ctype, "prefix_hash"))
|
|
|
|
FLAGS_rep_factory = kPrefixHash;
|
|
|
|
else if (!strcasecmp(ctype, "unsorted"))
|
|
|
|
FLAGS_rep_factory = kUnsorted;
|
|
|
|
else if (!strcasecmp(ctype, "vector"))
|
|
|
|
FLAGS_rep_factory = kVectorRep;
|
|
|
|
else {
|
|
|
|
fprintf(stdout, "Cannot parse %s\n", argv[i]);
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--disable_seek_compaction=%d%c", &n, &junk) == 1
|
|
|
|
&& (n == 0 || n == 1)) {
|
|
|
|
FLAGS_disable_seek_compaction = n;
|
2012-10-19 23:00:53 +02:00
|
|
|
} else if (sscanf(argv[i], "--delete_obsolete_files_period_micros=%ld%c",
|
|
|
|
&l, &junk) == 1) {
|
|
|
|
FLAGS_delete_obsolete_files_period_micros = n;
|
2013-03-06 21:54:55 +01:00
|
|
|
} else if (sscanf(argv[i], "--purge_redundant_percent=%d%c", &n, &junk) == 1
|
|
|
|
&& (n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_purge_redundant_percent = n;
|
2013-07-13 01:56:52 +02:00
|
|
|
} else if (sscanf(argv[i], "--filter_deletes=%d%c", &n, &junk)
|
2013-07-06 03:49:18 +02:00
|
|
|
== 1 && (n == 0 || n == 1)) {
|
2013-07-13 01:56:52 +02:00
|
|
|
FLAGS_filter_deletes = n;
|
2013-08-23 08:10:02 +02:00
|
|
|
} else if (sscanf(argv[i], "--prefix_size=%d%c", &n, &junk) == 1 &&
|
|
|
|
n >= 0 && n < 2000000000) {
|
|
|
|
FLAGS_prefix_size = n;
|
Benchmarking for Merge Operator
Summary:
Updated db_bench and utilities/merge_operators.h to allow for dynamic benchmarking
of merge operators in db_bench. Added a new test (--benchmarks=mergerandom), which performs
a bunch of random Merge() operations over random keys. Also added a "--merge_operator=" flag
so that the tester can easily benchmark different merge operators. Currently supports
the PutOperator and UInt64Add operator. Support for stringappend or list append may come later.
Test Plan:
1. make db_bench
2. Test the PutOperator (simulating Put) as follows:
./db_bench --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom --merge_operator=put
--threads=2
3. Test the UInt64AddOperator (simulating numeric addition) similarly:
./db_bench --value_size=8 --benchmarks=fillrandom,readrandom,updaterandom,readrandom,mergerandom,readrandom
--merge_operator=uint64add --threads=2
Reviewers: haobo, dhruba, zshao, MarkCallaghan
Reviewed By: haobo
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11535
2013-08-16 02:13:07 +02:00
|
|
|
} else if (sscanf(argv[i], "--use_merge=%d%c", &n, &junk)
|
|
|
|
== 1 && (n == 0 || n == 1)) {
|
|
|
|
FLAGS_use_merge_put = n;
|
2013-09-10 01:06:10 +02:00
|
|
|
} else if (sscanf(argv[i], "--universal_size_ratio=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_universal_size_ratio = n;
|
|
|
|
} else if (sscanf(argv[i], "--universal_min_merge_width=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_universal_min_merge_width = n;
|
|
|
|
} else if (sscanf(argv[i], "--universal_max_merge_width=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_universal_max_merge_width = n;
|
|
|
|
} else if (sscanf(argv[i],
|
|
|
|
"--universal_max_size_amplification_percent=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_universal_max_size_amplification_percent = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else {
|
|
|
|
fprintf(stderr, "Invalid flag '%s'\n", argv[i]);
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-11-15 07:00:11 +01:00
|
|
|
// The number of background threads should be at least as much the
|
|
|
|
// max number of concurrent compactions.
|
|
|
|
FLAGS_env->SetBackgroundThreads(FLAGS_max_background_compactions);
|
|
|
|
|
2013-08-15 01:58:36 +02:00
|
|
|
if ((FLAGS_readpercent + FLAGS_prefixpercent +
|
2013-09-20 01:47:24 +02:00
|
|
|
FLAGS_writepercent + FLAGS_delpercent + FLAGS_iterpercent) != 100) {
|
|
|
|
fprintf(stderr,
|
|
|
|
"Error: Read+Prefix+Write+Delete+Iterate percents != 100!\n");
|
2012-11-08 00:35:08 +01:00
|
|
|
exit(1);
|
|
|
|
}
|
2013-05-21 20:27:23 +02:00
|
|
|
if (FLAGS_disable_wal == 1 && FLAGS_reopen > 0) {
|
|
|
|
fprintf(stderr, "Error: Db cannot reopen safely with disable_wal set!\n");
|
|
|
|
exit(1);
|
|
|
|
}
|
2013-04-08 20:50:44 +02:00
|
|
|
if ((unsigned)FLAGS_reopen >= FLAGS_ops_per_thread) {
|
|
|
|
fprintf(stderr, "Error: #DB-reopens should be < ops_per_thread\n"
|
|
|
|
"Provided reopens = %d and ops_per_thread = %u\n", FLAGS_reopen,
|
|
|
|
FLAGS_ops_per_thread);
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Choose a location for the test database if none given with --db=<path>
|
2013-02-15 20:53:17 +01:00
|
|
|
if (FLAGS_db == nullptr) {
|
2013-10-04 06:49:15 +02:00
|
|
|
rocksdb::Env::Default()->GetTestDirectory(&default_db_path);
|
2012-10-03 18:58:45 +02:00
|
|
|
default_db_path += "/dbstress";
|
|
|
|
FLAGS_db = default_db_path.c_str();
|
|
|
|
}
|
|
|
|
|
2013-10-04 06:49:15 +02:00
|
|
|
rocksdb::StressTest stress;
|
2012-10-03 18:58:45 +02:00
|
|
|
stress.Run();
|
|
|
|
return 0;
|
|
|
|
}
|