2012-10-03 18:58:45 +02:00
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
2013-02-23 20:11:16 +01:00
|
|
|
//
|
2013-02-21 02:37:13 +01:00
|
|
|
//The test uses an array to compare against values written to the database.
|
|
|
|
//Keys written to the array are in 1:1 correspondence to the actual values in
|
|
|
|
//the database according to the formula in the functino GenerateValue
|
|
|
|
|
|
|
|
//Space is reserved in the array from 0 to FLAGS_max_key and values are randomly
|
|
|
|
//written/deleted/read from those positions. During verification we compare all
|
|
|
|
//the positions in the array. Therefore to shorten/elongate the amount of time
|
|
|
|
//that this test runs for, you should change the settings:
|
|
|
|
//FLAGS_max_key, FLAGS_ops_per_thread, (sometimes also FLAGS_threads)
|
2013-02-23 20:11:16 +01:00
|
|
|
//NOTE that if FLAGS_test_batches_snapshots is set, the test behaves a little
|
|
|
|
//differently. See comment header for the flag.
|
2013-02-21 02:37:13 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
#include <sys/types.h>
|
|
|
|
#include <stdio.h>
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include "db/db_impl.h"
|
|
|
|
#include "db/version_set.h"
|
|
|
|
#include "db/db_statistics.h"
|
|
|
|
#include "leveldb/cache.h"
|
|
|
|
#include "leveldb/db.h"
|
|
|
|
#include "leveldb/env.h"
|
|
|
|
#include "leveldb/write_batch.h"
|
|
|
|
#include "leveldb/statistics.h"
|
|
|
|
#include "port/port.h"
|
|
|
|
#include "util/crc32c.h"
|
|
|
|
#include "util/histogram.h"
|
|
|
|
#include "util/mutexlock.h"
|
|
|
|
#include "util/random.h"
|
|
|
|
#include "util/testutil.h"
|
|
|
|
#include "hdfs/env_hdfs.h"
|
|
|
|
|
|
|
|
static const long KB = 1024;
|
|
|
|
|
|
|
|
// Seed for PRNG
|
|
|
|
static uint32_t FLAGS_seed = 2341234;
|
|
|
|
|
|
|
|
// Max number of key/values to place in database
|
|
|
|
static long FLAGS_max_key = 2 * KB * KB * KB;
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
// If set, the test uses MultiGet, MultiPut and MultiDelete that
|
|
|
|
// do a different kind of validation during the test itself,
|
|
|
|
// rather than at the end. This is meant to solve the following
|
|
|
|
// problems at the expense of doing less degree of validation.
|
|
|
|
// (a) No need to acquire mutexes during writes (less cache flushes
|
|
|
|
// in multi-core leading to speed up)
|
|
|
|
// (b) No long validation at the end (more speed up)
|
|
|
|
// (c) Also test snapshot and atomicity of batch writes
|
|
|
|
static bool FLAGS_test_batches_snapshots = false;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Number of concurrent threads to run.
|
|
|
|
static int FLAGS_threads = 32;
|
|
|
|
|
|
|
|
// Size of each value will be this number times rand_int(1,3) bytes
|
|
|
|
static int FLAGS_value_size_mult = 8;
|
|
|
|
|
|
|
|
static bool FLAGS_verify_before_write = false;
|
|
|
|
|
|
|
|
// Print histogram of operation timings
|
|
|
|
static bool FLAGS_histogram = false;
|
|
|
|
|
|
|
|
static bool FLAGS_verbose = false;
|
|
|
|
|
|
|
|
// Number of bytes to buffer in memtable before compacting
|
|
|
|
// (initialized to default value by "main")
|
|
|
|
static int FLAGS_write_buffer_size = 0;
|
|
|
|
|
2013-01-11 02:18:50 +01:00
|
|
|
// The number of in-memory memtables.
|
2012-10-19 23:00:53 +02:00
|
|
|
// Each memtable is of size FLAGS_write_buffer_size.
|
|
|
|
// This is initialized to default value of 2 in "main" function.
|
|
|
|
static int FLAGS_max_write_buffer_number = 0;
|
|
|
|
|
|
|
|
// The maximum number of concurrent background compactions
|
|
|
|
// that can occur in parallel.
|
|
|
|
// This is initialized to default value of 1 in "main" function.
|
|
|
|
static int FLAGS_max_background_compactions = 0;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Number of bytes to use as a cache of uncompressed data.
|
|
|
|
static long FLAGS_cache_size = 2 * KB * KB * KB;
|
|
|
|
|
|
|
|
// Number of bytes in a block.
|
|
|
|
static int FLAGS_block_size = 4 * KB;
|
|
|
|
|
2012-11-09 22:04:12 +01:00
|
|
|
// Number of times database reopens
|
|
|
|
static int FLAGS_reopen = 10;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Maximum number of files to keep open at the same time (use default if == 0)
|
|
|
|
static int FLAGS_open_files = 0;
|
|
|
|
|
|
|
|
// Bloom filter bits per key.
|
|
|
|
// Negative means use default settings.
|
|
|
|
static int FLAGS_bloom_bits = 10;
|
|
|
|
|
|
|
|
// Use the db with the following name.
|
2013-02-15 20:53:17 +01:00
|
|
|
static const char* FLAGS_db = nullptr;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
// Verify checksum for every block read from storage
|
|
|
|
static bool FLAGS_verify_checksum = false;
|
|
|
|
|
|
|
|
// Database statistics
|
2013-03-27 19:27:39 +01:00
|
|
|
static std::shared_ptr<leveldb::Statistics> dbstats;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
// Sync all writes to disk
|
|
|
|
static bool FLAGS_sync = false;
|
|
|
|
|
|
|
|
// If true, do not wait until data is synced to disk.
|
|
|
|
static bool FLAGS_disable_data_sync = false;
|
|
|
|
|
|
|
|
// If true, issue fsync instead of fdatasync
|
|
|
|
static bool FLAGS_use_fsync = false;
|
|
|
|
|
|
|
|
// If true, do not write WAL for write.
|
|
|
|
static bool FLAGS_disable_wal = false;
|
|
|
|
|
|
|
|
// Target level-0 file size for compaction
|
|
|
|
static int FLAGS_target_file_size_base = 64 * KB;
|
|
|
|
|
|
|
|
// A multiplier to compute targe level-N file size
|
|
|
|
static int FLAGS_target_file_size_multiplier = 1;
|
|
|
|
|
|
|
|
// Max bytes for level-0
|
2012-11-09 03:45:19 +01:00
|
|
|
static uint64_t FLAGS_max_bytes_for_level_base = 256 * KB;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
// A multiplier to compute max bytes for level-N
|
|
|
|
static int FLAGS_max_bytes_for_level_multiplier = 2;
|
|
|
|
|
|
|
|
// Number of files in level-0 that will trigger put stop.
|
|
|
|
static int FLAGS_level0_stop_writes_trigger = 12;
|
|
|
|
|
|
|
|
// Number of files in level-0 that will slow down writes.
|
|
|
|
static int FLAGS_level0_slowdown_writes_trigger = 8;
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
// Ratio of reads to total workload (expressed as a percentage)
|
|
|
|
static unsigned int FLAGS_readpercent = 10;
|
|
|
|
|
|
|
|
// Ratio of deletes to total workload (expressed as a percentage)
|
|
|
|
static unsigned int FLAGS_delpercent = 30;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
// Option to disable compation triggered by read.
|
|
|
|
static int FLAGS_disable_seek_compaction = false;
|
|
|
|
|
2012-10-19 23:00:53 +02:00
|
|
|
// Option to delete obsolete files periodically
|
|
|
|
// Default: 0 which means that obsolete files are
|
|
|
|
// deleted after every compaction run.
|
|
|
|
static uint64_t FLAGS_delete_obsolete_files_period_micros = 0;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Algorithm to use to compress the database
|
|
|
|
static enum leveldb::CompressionType FLAGS_compression_type =
|
|
|
|
leveldb::kSnappyCompression;
|
|
|
|
|
|
|
|
// posix or hdfs environment
|
|
|
|
static leveldb::Env* FLAGS_env = leveldb::Env::Default();
|
|
|
|
|
|
|
|
// Number of operations per thread.
|
|
|
|
static uint32_t FLAGS_ops_per_thread = 600000;
|
|
|
|
|
|
|
|
// Log2 of number of keys per lock
|
|
|
|
static uint32_t FLAGS_log2_keys_per_lock = 2; // implies 2^2 keys per lock
|
|
|
|
|
2013-03-06 21:54:55 +01:00
|
|
|
// Percentage of times we want to purge redundant keys in memory before flushing
|
|
|
|
static uint32_t FLAGS_purge_redundant_percent = 50;
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
extern bool useOsBuffer;
|
|
|
|
extern bool useFsReadAhead;
|
|
|
|
extern bool useMmapRead;
|
|
|
|
|
|
|
|
namespace leveldb {
|
|
|
|
|
|
|
|
class StressTest;
|
|
|
|
namespace {
|
|
|
|
|
|
|
|
class Stats {
|
|
|
|
private:
|
|
|
|
double start_;
|
|
|
|
double finish_;
|
|
|
|
double seconds_;
|
|
|
|
long done_;
|
|
|
|
long writes_;
|
2012-11-08 00:35:08 +01:00
|
|
|
long deletes_;
|
2013-02-23 20:11:16 +01:00
|
|
|
long gets_;
|
2013-02-21 00:57:27 +01:00
|
|
|
long founds_;
|
|
|
|
long errors_;
|
2012-10-03 18:58:45 +02:00
|
|
|
int next_report_;
|
|
|
|
size_t bytes_;
|
|
|
|
double last_op_finish_;
|
2013-02-15 20:53:17 +01:00
|
|
|
HistogramImpl hist_;
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
public:
|
|
|
|
Stats() { }
|
|
|
|
|
|
|
|
void Start() {
|
|
|
|
next_report_ = 100;
|
|
|
|
hist_.Clear();
|
|
|
|
done_ = 0;
|
|
|
|
writes_ = 0;
|
2012-11-08 00:35:08 +01:00
|
|
|
deletes_ = 0;
|
2013-02-23 20:11:16 +01:00
|
|
|
gets_ = 0;
|
2013-02-21 00:57:27 +01:00
|
|
|
founds_ = 0;
|
|
|
|
errors_ = 0;
|
2012-10-03 18:58:45 +02:00
|
|
|
bytes_ = 0;
|
|
|
|
seconds_ = 0;
|
|
|
|
start_ = FLAGS_env->NowMicros();
|
|
|
|
last_op_finish_ = start_;
|
|
|
|
finish_ = start_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void Merge(const Stats& other) {
|
|
|
|
hist_.Merge(other.hist_);
|
|
|
|
done_ += other.done_;
|
|
|
|
writes_ += other.writes_;
|
2012-11-08 00:35:08 +01:00
|
|
|
deletes_ += other.deletes_;
|
2013-02-23 20:11:16 +01:00
|
|
|
gets_ += other.gets_;
|
|
|
|
founds_ += other.founds_;
|
|
|
|
errors_ += other.errors_;
|
2012-10-03 18:58:45 +02:00
|
|
|
bytes_ += other.bytes_;
|
|
|
|
seconds_ += other.seconds_;
|
|
|
|
if (other.start_ < start_) start_ = other.start_;
|
|
|
|
if (other.finish_ > finish_) finish_ = other.finish_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void Stop() {
|
|
|
|
finish_ = FLAGS_env->NowMicros();
|
|
|
|
seconds_ = (finish_ - start_) * 1e-6;
|
|
|
|
}
|
|
|
|
|
|
|
|
void FinishedSingleOp() {
|
|
|
|
if (FLAGS_histogram) {
|
|
|
|
double now = FLAGS_env->NowMicros();
|
|
|
|
double micros = now - last_op_finish_;
|
|
|
|
hist_.Add(micros);
|
|
|
|
if (micros > 20000) {
|
|
|
|
fprintf(stderr, "long op: %.1f micros%30s\r", micros, "");
|
|
|
|
fflush(stderr);
|
|
|
|
}
|
|
|
|
last_op_finish_ = now;
|
|
|
|
}
|
|
|
|
|
|
|
|
done_++;
|
|
|
|
if (done_ >= next_report_) {
|
|
|
|
if (next_report_ < 1000) next_report_ += 100;
|
|
|
|
else if (next_report_ < 5000) next_report_ += 500;
|
|
|
|
else if (next_report_ < 10000) next_report_ += 1000;
|
|
|
|
else if (next_report_ < 50000) next_report_ += 5000;
|
|
|
|
else if (next_report_ < 100000) next_report_ += 10000;
|
|
|
|
else if (next_report_ < 500000) next_report_ += 50000;
|
|
|
|
else next_report_ += 100000;
|
|
|
|
fprintf(stderr, "... finished %ld ops%30s\r", done_, "");
|
|
|
|
fflush(stderr);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
void AddBytesForWrites(int nwrites, size_t nbytes) {
|
|
|
|
writes_ += nwrites;
|
|
|
|
bytes_ += nbytes;
|
|
|
|
}
|
|
|
|
|
|
|
|
void AddDeletes(int n) {
|
|
|
|
deletes_ += n;
|
|
|
|
}
|
|
|
|
|
2013-02-23 20:11:16 +01:00
|
|
|
void AddGets(int ngets, int nfounds) {
|
|
|
|
founds_ += nfounds;
|
|
|
|
gets_ += ngets;
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
void AddErrors(int n) {
|
|
|
|
errors_ += n;
|
2012-11-08 00:35:08 +01:00
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
void Report(const char* name) {
|
|
|
|
std::string extra;
|
|
|
|
if (bytes_ < 1 || done_ < 1) {
|
|
|
|
fprintf(stderr, "No writes or ops?\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
double elapsed = (finish_ - start_) * 1e-6;
|
|
|
|
double bytes_mb = bytes_ / 1048576.0;
|
|
|
|
double rate = bytes_mb / elapsed;
|
|
|
|
double throughput = (double)done_/elapsed;
|
|
|
|
|
|
|
|
fprintf(stdout, "%-12s: ", name);
|
|
|
|
fprintf(stdout, "%.3f micros/op %ld ops/sec\n",
|
|
|
|
seconds_ * 1e6 / done_, (long)throughput);
|
|
|
|
fprintf(stdout, "%-12s: Wrote %.2f MB (%.2f MB/sec) (%ld%% of %ld ops)\n",
|
|
|
|
"", bytes_mb, rate, (100*writes_)/done_, done_);
|
2013-02-23 20:11:16 +01:00
|
|
|
fprintf(stdout, "%-12s: Wrote %ld times\n", "", writes_);
|
2012-11-08 00:35:08 +01:00
|
|
|
fprintf(stdout, "%-12s: Deleted %ld times\n", "", deletes_);
|
2013-02-23 20:11:16 +01:00
|
|
|
fprintf(stdout, "%-12s: %ld/%ld gets found the key\n", "", founds_, gets_);
|
2013-02-21 00:57:27 +01:00
|
|
|
fprintf(stdout, "%-12s: Got errors %ld times\n", "", errors_);
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
if (FLAGS_histogram) {
|
|
|
|
fprintf(stdout, "Microseconds per op:\n%s\n", hist_.ToString().c_str());
|
|
|
|
}
|
|
|
|
fflush(stdout);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// State shared by all concurrent executions of the same benchmark.
|
|
|
|
class SharedState {
|
|
|
|
public:
|
|
|
|
static const uint32_t SENTINEL = 0xffffffff;
|
|
|
|
|
2013-02-15 20:53:17 +01:00
|
|
|
explicit SharedState(StressTest* stress_test) :
|
2012-10-03 18:58:45 +02:00
|
|
|
cv_(&mu_),
|
|
|
|
seed_(FLAGS_seed),
|
|
|
|
max_key_(FLAGS_max_key),
|
|
|
|
log2_keys_per_lock_(FLAGS_log2_keys_per_lock),
|
|
|
|
num_threads_(FLAGS_threads),
|
|
|
|
num_initialized_(0),
|
|
|
|
num_populated_(0),
|
2012-11-09 22:04:12 +01:00
|
|
|
vote_reopen_(0),
|
2012-10-03 18:58:45 +02:00
|
|
|
num_done_(0),
|
|
|
|
start_(false),
|
|
|
|
start_verify_(false),
|
|
|
|
stress_test_(stress_test) {
|
2013-02-21 00:57:27 +01:00
|
|
|
if (FLAGS_test_batches_snapshots) {
|
|
|
|
key_locks_ = nullptr;
|
|
|
|
values_ = nullptr;
|
|
|
|
fprintf(stdout, "No lock creation because test_batches_snapshots set\n");
|
|
|
|
return;
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
values_ = new uint32_t[max_key_];
|
|
|
|
for (long i = 0; i < max_key_; i++) {
|
|
|
|
values_[i] = SENTINEL;
|
|
|
|
}
|
2013-02-21 00:57:27 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
long num_locks = (max_key_ >> log2_keys_per_lock_);
|
|
|
|
if (max_key_ & ((1 << log2_keys_per_lock_) - 1)) {
|
|
|
|
num_locks ++;
|
|
|
|
}
|
|
|
|
fprintf(stdout, "Creating %ld locks\n", num_locks);
|
|
|
|
key_locks_ = new port::Mutex[num_locks];
|
|
|
|
}
|
|
|
|
|
|
|
|
~SharedState() {
|
|
|
|
delete[] values_;
|
|
|
|
delete[] key_locks_;
|
|
|
|
}
|
|
|
|
|
|
|
|
port::Mutex* GetMutex() {
|
|
|
|
return &mu_;
|
|
|
|
}
|
|
|
|
|
|
|
|
port::CondVar* GetCondVar() {
|
|
|
|
return &cv_;
|
|
|
|
}
|
|
|
|
|
|
|
|
StressTest* GetStressTest() const {
|
|
|
|
return stress_test_;
|
|
|
|
}
|
|
|
|
|
|
|
|
long GetMaxKey() const {
|
|
|
|
return max_key_;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint32_t GetNumThreads() const {
|
|
|
|
return num_threads_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void IncInitialized() {
|
|
|
|
num_initialized_++;
|
|
|
|
}
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
void IncOperated() {
|
2012-10-03 18:58:45 +02:00
|
|
|
num_populated_++;
|
|
|
|
}
|
|
|
|
|
|
|
|
void IncDone() {
|
|
|
|
num_done_++;
|
|
|
|
}
|
|
|
|
|
2012-11-09 22:04:12 +01:00
|
|
|
void IncVotedReopen() {
|
|
|
|
vote_reopen_ = (vote_reopen_ + 1) % num_threads_;
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
bool AllInitialized() const {
|
|
|
|
return num_initialized_ >= num_threads_;
|
|
|
|
}
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
bool AllOperated() const {
|
2012-10-03 18:58:45 +02:00
|
|
|
return num_populated_ >= num_threads_;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AllDone() const {
|
|
|
|
return num_done_ >= num_threads_;
|
|
|
|
}
|
|
|
|
|
2012-11-09 22:04:12 +01:00
|
|
|
bool AllVotedReopen() {
|
|
|
|
return (vote_reopen_ == 0);
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
void SetStart() {
|
|
|
|
start_ = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
void SetStartVerify() {
|
|
|
|
start_verify_ = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool Started() const {
|
|
|
|
return start_;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool VerifyStarted() const {
|
|
|
|
return start_verify_;
|
|
|
|
}
|
|
|
|
|
|
|
|
port::Mutex* GetMutexForKey(long key) {
|
|
|
|
return &key_locks_[key >> log2_keys_per_lock_];
|
|
|
|
}
|
|
|
|
|
|
|
|
void Put(long key, uint32_t value_base) {
|
|
|
|
values_[key] = value_base;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint32_t Get(long key) const {
|
|
|
|
return values_[key];
|
|
|
|
}
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
void Delete(long key) const {
|
|
|
|
values_[key] = SENTINEL;
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
uint32_t GetSeed() const {
|
|
|
|
return seed_;
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
port::Mutex mu_;
|
|
|
|
port::CondVar cv_;
|
|
|
|
const uint32_t seed_;
|
|
|
|
const long max_key_;
|
|
|
|
const uint32_t log2_keys_per_lock_;
|
|
|
|
const int num_threads_;
|
|
|
|
long num_initialized_;
|
|
|
|
long num_populated_;
|
2012-11-09 22:04:12 +01:00
|
|
|
long vote_reopen_;
|
2012-10-03 18:58:45 +02:00
|
|
|
long num_done_;
|
|
|
|
bool start_;
|
|
|
|
bool start_verify_;
|
|
|
|
StressTest* stress_test_;
|
|
|
|
|
|
|
|
uint32_t *values_;
|
|
|
|
port::Mutex *key_locks_;
|
|
|
|
|
|
|
|
};
|
|
|
|
|
|
|
|
// Per-thread state for concurrent executions of the same benchmark.
|
|
|
|
struct ThreadState {
|
|
|
|
uint32_t tid; // 0..n-1
|
|
|
|
Random rand; // Has different seeds for different threads
|
|
|
|
SharedState* shared;
|
|
|
|
Stats stats;
|
|
|
|
|
|
|
|
ThreadState(uint32_t index, SharedState *shared)
|
|
|
|
: tid(index),
|
|
|
|
rand(1000 + index + shared->GetSeed()),
|
|
|
|
shared(shared) {
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
} // namespace
|
|
|
|
|
|
|
|
class StressTest {
|
|
|
|
public:
|
|
|
|
StressTest()
|
|
|
|
: cache_(NewLRUCache(FLAGS_cache_size)),
|
|
|
|
filter_policy_(FLAGS_bloom_bits >= 0
|
|
|
|
? NewBloomFilterPolicy(FLAGS_bloom_bits)
|
2013-02-15 20:53:17 +01:00
|
|
|
: nullptr),
|
|
|
|
db_(nullptr),
|
2012-11-17 00:28:14 +01:00
|
|
|
num_times_reopened_(0) {
|
2012-10-03 18:58:45 +02:00
|
|
|
std::vector<std::string> files;
|
|
|
|
FLAGS_env->GetChildren(FLAGS_db, &files);
|
2012-11-06 21:02:18 +01:00
|
|
|
for (unsigned int i = 0; i < files.size(); i++) {
|
2012-10-03 18:58:45 +02:00
|
|
|
if (Slice(files[i]).starts_with("heap-")) {
|
|
|
|
FLAGS_env->DeleteFile(std::string(FLAGS_db) + "/" + files[i]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
DestroyDB(FLAGS_db, Options());
|
|
|
|
}
|
|
|
|
|
|
|
|
~StressTest() {
|
|
|
|
delete db_;
|
|
|
|
delete filter_policy_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void Run() {
|
|
|
|
PrintEnv();
|
|
|
|
Open();
|
|
|
|
SharedState shared(this);
|
|
|
|
uint32_t n = shared.GetNumThreads();
|
|
|
|
|
|
|
|
std::vector<ThreadState*> threads(n);
|
|
|
|
for (uint32_t i = 0; i < n; i++) {
|
|
|
|
threads[i] = new ThreadState(i, &shared);
|
|
|
|
FLAGS_env->StartThread(ThreadBody, threads[i]);
|
|
|
|
}
|
|
|
|
// Each thread goes through the following states:
|
2012-11-08 00:35:08 +01:00
|
|
|
// initializing -> wait for others to init -> read/populate/depopulate
|
|
|
|
// wait for others to operate -> verify -> done
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
{
|
|
|
|
MutexLock l(shared.GetMutex());
|
|
|
|
while (!shared.AllInitialized()) {
|
|
|
|
shared.GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
|
2012-11-17 00:28:14 +01:00
|
|
|
double now = FLAGS_env->NowMicros();
|
|
|
|
fprintf(stdout, "%s Starting database operations\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str());
|
2012-11-08 00:35:08 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
shared.SetStart();
|
|
|
|
shared.GetCondVar()->SignalAll();
|
2012-11-08 00:35:08 +01:00
|
|
|
while (!shared.AllOperated()) {
|
2012-10-03 18:58:45 +02:00
|
|
|
shared.GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
|
2012-11-17 00:28:14 +01:00
|
|
|
now = FLAGS_env->NowMicros();
|
2013-02-23 20:11:16 +01:00
|
|
|
if (FLAGS_test_batches_snapshots) {
|
|
|
|
fprintf(stdout, "%s Limited verification already done during gets\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str());
|
|
|
|
} else {
|
|
|
|
fprintf(stdout, "%s Starting verification\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str());
|
|
|
|
}
|
2012-11-17 00:28:14 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
shared.SetStartVerify();
|
|
|
|
shared.GetCondVar()->SignalAll();
|
|
|
|
while (!shared.AllDone()) {
|
|
|
|
shared.GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-11-06 21:02:18 +01:00
|
|
|
for (unsigned int i = 1; i < n; i++) {
|
2012-10-03 18:58:45 +02:00
|
|
|
threads[0]->stats.Merge(threads[i]->stats);
|
|
|
|
}
|
|
|
|
threads[0]->stats.Report("Stress Test");
|
|
|
|
|
2012-11-06 21:02:18 +01:00
|
|
|
for (unsigned int i = 0; i < n; i++) {
|
2012-10-03 18:58:45 +02:00
|
|
|
delete threads[i];
|
2013-02-15 20:53:17 +01:00
|
|
|
threads[i] = nullptr;
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
2012-11-17 00:28:14 +01:00
|
|
|
double now = FLAGS_env->NowMicros();
|
2013-02-23 20:11:16 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
|
|
|
fprintf(stdout, "%s Verification successful\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str());
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
PrintStatistics();
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
|
|
|
|
static void ThreadBody(void* v) {
|
|
|
|
ThreadState* thread = reinterpret_cast<ThreadState*>(v);
|
|
|
|
SharedState* shared = thread->shared;
|
|
|
|
|
|
|
|
{
|
|
|
|
MutexLock l(shared->GetMutex());
|
|
|
|
shared->IncInitialized();
|
|
|
|
if (shared->AllInitialized()) {
|
|
|
|
shared->GetCondVar()->SignalAll();
|
|
|
|
}
|
|
|
|
while (!shared->Started()) {
|
|
|
|
shared->GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
}
|
2012-11-08 00:35:08 +01:00
|
|
|
thread->shared->GetStressTest()->OperateDb(thread);
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
{
|
|
|
|
MutexLock l(shared->GetMutex());
|
2012-11-08 00:35:08 +01:00
|
|
|
shared->IncOperated();
|
|
|
|
if (shared->AllOperated()) {
|
2012-10-03 18:58:45 +02:00
|
|
|
shared->GetCondVar()->SignalAll();
|
|
|
|
}
|
|
|
|
while (!shared->VerifyStarted()) {
|
|
|
|
shared->GetCondVar()->Wait();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
|
|
|
thread->shared->GetStressTest()->VerifyDb(*(thread->shared),
|
|
|
|
thread->tid);
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
|
|
|
|
{
|
|
|
|
MutexLock l(shared->GetMutex());
|
|
|
|
shared->IncDone();
|
|
|
|
if (shared->AllDone()) {
|
|
|
|
shared->GetCondVar()->SignalAll();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2013-02-21 00:57:27 +01:00
|
|
|
// Given a key K and value V, this puts ("0"+K, "0"+V), ("1"+K, "1"+V), ...
|
|
|
|
// ("9"+K, "9"+V) in DB atomically i.e in a single batch.
|
|
|
|
// Also refer MultiGet.
|
|
|
|
Status MultiPut(ThreadState* thread,
|
|
|
|
const WriteOptions& writeoptions,
|
|
|
|
const Slice& key, const Slice& value, size_t sz) {
|
|
|
|
std::string keys[10] = {"9", "8", "7", "6", "5",
|
|
|
|
"4", "3", "2", "1", "0"};
|
|
|
|
std::string values[10] = {"9", "8", "7", "6", "5",
|
|
|
|
"4", "3", "2", "1", "0"};
|
|
|
|
Slice value_slices[10];
|
|
|
|
WriteBatch batch;
|
|
|
|
Status s;
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
keys[i] += key.ToString();
|
|
|
|
values[i] += value.ToString();
|
|
|
|
value_slices[i] = values[i];
|
|
|
|
batch.Put(keys[i], value_slices[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
s = db_->Write(writeoptions, &batch);
|
|
|
|
if (!s.ok()) {
|
|
|
|
fprintf(stderr, "multiput error: %s\n", s.ToString().c_str());
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
} else {
|
|
|
|
// we did 10 writes each of size sz + 1
|
|
|
|
thread->stats.AddBytesForWrites(10, (sz + 1) * 10);
|
|
|
|
}
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Given a key K, this deletes ("0"+K), ("1"+K),... ("9"+K)
|
|
|
|
// in DB atomically i.e in a single batch. Also refer MultiGet.
|
|
|
|
Status MultiDelete(ThreadState* thread,
|
|
|
|
const WriteOptions& writeoptions,
|
|
|
|
const Slice& key) {
|
|
|
|
std::string keys[10] = {"9", "7", "5", "3", "1",
|
|
|
|
"8", "6", "4", "2", "0"};
|
|
|
|
|
|
|
|
WriteBatch batch;
|
|
|
|
Status s;
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
keys[i] += key.ToString();
|
|
|
|
batch.Delete(keys[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
s = db_->Write(writeoptions, &batch);
|
|
|
|
if (!s.ok()) {
|
|
|
|
fprintf(stderr, "multidelete error: %s\n", s.ToString().c_str());
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
} else {
|
|
|
|
thread->stats.AddDeletes(10);
|
|
|
|
}
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Given a key K, this gets values for "0"+K, "1"+K,..."9"+K
|
|
|
|
// in the same snapshot, and verifies that all the values are of the form
|
|
|
|
// "0"+V, "1"+V,..."9"+V.
|
|
|
|
// ASSUMES that MultiPut was used to put (K, V) into the DB.
|
|
|
|
Status MultiGet(ThreadState* thread,
|
|
|
|
const ReadOptions& readoptions,
|
|
|
|
const Slice& key, std::string* value) {
|
|
|
|
std::string keys[10] = {"0", "1", "2", "3", "4", "5", "6", "7", "8", "9"};
|
|
|
|
Slice key_slices[10];
|
|
|
|
std::string values[10];
|
|
|
|
ReadOptions readoptionscopy = readoptions;
|
|
|
|
readoptionscopy.snapshot = db_->GetSnapshot();
|
|
|
|
Status s;
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
keys[i] += key.ToString();
|
|
|
|
key_slices[i] = keys[i];
|
|
|
|
s = db_->Get(readoptionscopy, key_slices[i], value);
|
|
|
|
if (!s.ok() && !s.IsNotFound()) {
|
|
|
|
fprintf(stderr, "get error: %s\n", s.ToString().c_str());
|
|
|
|
values[i] = "";
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
// we continue after error rather than exiting so that we can
|
|
|
|
// find more errors if any
|
|
|
|
} else if (s.IsNotFound()) {
|
|
|
|
values[i] = "";
|
2013-02-23 20:11:16 +01:00
|
|
|
thread->stats.AddGets(1, 0);
|
2013-02-21 00:57:27 +01:00
|
|
|
} else {
|
|
|
|
values[i] = *value;
|
|
|
|
|
|
|
|
char expected_prefix = (keys[i])[0];
|
|
|
|
char actual_prefix = (values[i])[0];
|
|
|
|
if (actual_prefix != expected_prefix) {
|
|
|
|
fprintf(stderr, "expected prefix = %c actual = %c\n",
|
|
|
|
expected_prefix, actual_prefix);
|
|
|
|
}
|
|
|
|
(values[i])[0] = ' '; // blank out the differing character
|
2013-02-23 20:11:16 +01:00
|
|
|
thread->stats.AddGets(1, 1);
|
2013-02-21 00:57:27 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
db_->ReleaseSnapshot(readoptionscopy.snapshot);
|
|
|
|
|
|
|
|
// Now that we retrieved all values, check that they all match
|
|
|
|
for (int i = 1; i < 10; i++) {
|
|
|
|
if (values[i] != values[0]) {
|
|
|
|
fprintf(stderr, "inconsistent values for key %s: %s, %s\n",
|
|
|
|
key.ToString().c_str(), values[0].c_str(),
|
|
|
|
values[i].c_str());
|
|
|
|
// we continue after error rather than exiting so that we can
|
|
|
|
// find more errors if any
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
void OperateDb(ThreadState* thread) {
|
2012-10-03 18:58:45 +02:00
|
|
|
ReadOptions read_opts(FLAGS_verify_checksum, true);
|
|
|
|
WriteOptions write_opts;
|
2012-11-08 00:35:08 +01:00
|
|
|
char value[100];
|
2012-10-03 18:58:45 +02:00
|
|
|
long max_key = thread->shared->GetMaxKey();
|
|
|
|
std::string from_db;
|
|
|
|
if (FLAGS_sync) {
|
|
|
|
write_opts.sync = true;
|
|
|
|
}
|
|
|
|
write_opts.disableWAL = FLAGS_disable_wal;
|
|
|
|
|
|
|
|
thread->stats.Start();
|
2012-11-08 00:35:08 +01:00
|
|
|
for (long i = 0; i < FLAGS_ops_per_thread; i++) {
|
2012-11-09 22:04:12 +01:00
|
|
|
if(i != 0 && (i % (FLAGS_ops_per_thread / (FLAGS_reopen + 1))) == 0) {
|
|
|
|
{
|
2012-11-17 00:28:14 +01:00
|
|
|
thread->stats.FinishedSingleOp();
|
2012-11-09 22:04:12 +01:00
|
|
|
MutexLock l(thread->shared->GetMutex());
|
|
|
|
thread->shared->IncVotedReopen();
|
|
|
|
if (thread->shared->AllVotedReopen()) {
|
|
|
|
thread->shared->GetStressTest()->Reopen();
|
|
|
|
thread->shared->GetCondVar()->SignalAll();
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
thread->shared->GetCondVar()->Wait();
|
|
|
|
}
|
2013-02-23 20:11:16 +01:00
|
|
|
// Commenting this out as we don't want to reset stats on each open.
|
|
|
|
// thread->stats.Start();
|
2012-11-09 22:04:12 +01:00
|
|
|
}
|
|
|
|
}
|
2013-02-21 00:57:27 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
long rand_key = thread->rand.Next() % max_key;
|
|
|
|
Slice key((char*)&rand_key, sizeof(rand_key));
|
2012-11-08 00:35:08 +01:00
|
|
|
//Read:10%;Delete:30%;Write:60%
|
|
|
|
unsigned int probability_operation = thread->rand.Uniform(100);
|
|
|
|
if (probability_operation < FLAGS_readpercent) {
|
|
|
|
// read load
|
2013-02-23 20:11:16 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
|
|
|
Status s = db_->Get(read_opts, key, &from_db);
|
|
|
|
if (s.ok()) {
|
|
|
|
// found case
|
|
|
|
thread->stats.AddGets(1, 1);
|
|
|
|
} else if (s.IsNotFound()) {
|
|
|
|
// not found case
|
|
|
|
thread->stats.AddGets(1, 0);
|
|
|
|
} else {
|
|
|
|
// errors case
|
|
|
|
thread->stats.AddErrors(1);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
MultiGet(thread, read_opts, key, &from_db);
|
|
|
|
}
|
2012-11-08 00:35:08 +01:00
|
|
|
} else if (probability_operation < FLAGS_delpercent + FLAGS_readpercent) {
|
|
|
|
//introduce delete load
|
2013-02-21 00:57:27 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
2012-11-08 00:35:08 +01:00
|
|
|
MutexLock l(thread->shared->GetMutexForKey(rand_key));
|
|
|
|
thread->shared->Delete(rand_key);
|
|
|
|
db_->Delete(write_opts, key);
|
2013-02-21 00:57:27 +01:00
|
|
|
thread->stats.AddDeletes(1);
|
|
|
|
} else {
|
|
|
|
MultiDelete(thread, write_opts, key);
|
2012-11-08 00:35:08 +01:00
|
|
|
}
|
2013-02-21 00:57:27 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
} else {
|
2012-11-08 00:35:08 +01:00
|
|
|
// write load
|
2012-10-03 18:58:45 +02:00
|
|
|
uint32_t value_base = thread->rand.Next();
|
|
|
|
size_t sz = GenerateValue(value_base, value, sizeof(value));
|
|
|
|
Slice v(value, sz);
|
2013-02-21 00:57:27 +01:00
|
|
|
if (!FLAGS_test_batches_snapshots) {
|
2012-10-03 18:58:45 +02:00
|
|
|
MutexLock l(thread->shared->GetMutexForKey(rand_key));
|
|
|
|
if (FLAGS_verify_before_write) {
|
2012-11-08 00:35:08 +01:00
|
|
|
VerifyValue(rand_key, read_opts, *(thread->shared), &from_db, true);
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
thread->shared->Put(rand_key, value_base);
|
|
|
|
db_->Put(write_opts, key, v);
|
2013-02-21 00:57:27 +01:00
|
|
|
thread->stats.AddBytesForWrites(1, sz);
|
|
|
|
} else {
|
|
|
|
MultiPut(thread, write_opts, key, v, sz);
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
2013-02-21 00:57:27 +01:00
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
PrintKeyValue(rand_key, value, sz);
|
|
|
|
}
|
|
|
|
thread->stats.FinishedSingleOp();
|
|
|
|
}
|
|
|
|
thread->stats.Stop();
|
|
|
|
}
|
|
|
|
|
|
|
|
void VerifyDb(const SharedState &shared, long start) const {
|
|
|
|
ReadOptions options(FLAGS_verify_checksum, true);
|
|
|
|
long max_key = shared.GetMaxKey();
|
|
|
|
long step = shared.GetNumThreads();
|
|
|
|
for (long i = start; i < max_key; i+= step) {
|
|
|
|
std::string from_db;
|
2012-11-08 00:35:08 +01:00
|
|
|
VerifyValue(i, options, shared, &from_db, true);
|
2012-10-03 18:58:45 +02:00
|
|
|
if (from_db.length()) {
|
|
|
|
PrintKeyValue(i, from_db.data(), from_db.length());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-10-13 02:00:25 +02:00
|
|
|
void VerificationAbort(std::string msg, long key) const {
|
2012-11-08 00:35:08 +01:00
|
|
|
fprintf(stderr, "Verification failed for key %ld: %s\n",
|
2012-10-13 02:00:25 +02:00
|
|
|
key, msg.c_str());
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
void VerifyValue(long key, const ReadOptions &opts, const SharedState &shared,
|
|
|
|
std::string *value_from_db, bool strict=false) const {
|
|
|
|
Slice k((char*)&key, sizeof(key));
|
2012-11-08 00:35:08 +01:00
|
|
|
char value[100];
|
2012-10-03 18:58:45 +02:00
|
|
|
uint32_t value_base = shared.Get(key);
|
|
|
|
if (value_base == SharedState::SENTINEL && !strict) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (db_->Get(opts, k, value_from_db).ok()) {
|
|
|
|
if (value_base == SharedState::SENTINEL) {
|
|
|
|
VerificationAbort("Unexpected value found", key);
|
|
|
|
}
|
2013-02-21 02:37:13 +01:00
|
|
|
size_t sz = GenerateValue(value_base, value, sizeof(value));
|
2012-10-03 18:58:45 +02:00
|
|
|
if (value_from_db->length() != sz) {
|
|
|
|
VerificationAbort("Length of value read is not equal", key);
|
|
|
|
}
|
|
|
|
if (memcmp(value_from_db->data(), value, sz) != 0) {
|
|
|
|
VerificationAbort("Contents of value read don't match", key);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (value_base != SharedState::SENTINEL) {
|
|
|
|
VerificationAbort("Value not found", key);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void PrintKeyValue(uint32_t key, const char *value, size_t sz) {
|
|
|
|
if (!FLAGS_verbose) return;
|
2012-11-06 21:02:18 +01:00
|
|
|
fprintf(stdout, "%u ==> (%u) ", key, (unsigned int)sz);
|
2012-10-03 18:58:45 +02:00
|
|
|
for (size_t i=0; i<sz; i++) {
|
|
|
|
fprintf(stdout, "%X", value[i]);
|
|
|
|
}
|
|
|
|
fprintf(stdout, "\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
static size_t GenerateValue(uint32_t rand, char *v, size_t max_sz) {
|
|
|
|
size_t value_sz = ((rand % 3) + 1) * FLAGS_value_size_mult;
|
|
|
|
assert(value_sz <= max_sz && value_sz >= sizeof(uint32_t));
|
|
|
|
*((uint32_t*)v) = rand;
|
|
|
|
for (size_t i=sizeof(uint32_t); i < value_sz; i++) {
|
|
|
|
v[i] = (char)(rand ^ i);
|
|
|
|
}
|
|
|
|
return value_sz; // the size of the value set.
|
|
|
|
}
|
|
|
|
|
|
|
|
void PrintEnv() const {
|
|
|
|
fprintf(stdout, "LevelDB version : %d.%d\n",
|
|
|
|
kMajorVersion, kMinorVersion);
|
|
|
|
fprintf(stdout, "Number of threads : %d\n", FLAGS_threads);
|
2012-11-06 21:02:18 +01:00
|
|
|
fprintf(stdout, "Ops per thread : %d\n", FLAGS_ops_per_thread);
|
2012-11-08 00:35:08 +01:00
|
|
|
fprintf(stdout, "Read percentage : %d\n", FLAGS_readpercent);
|
|
|
|
fprintf(stdout, "Delete percentage : %d\n", FLAGS_delpercent);
|
2012-10-03 18:58:45 +02:00
|
|
|
fprintf(stdout, "Max key : %ld\n", FLAGS_max_key);
|
2013-02-21 00:57:27 +01:00
|
|
|
fprintf(stdout, "Ratio #ops/#keys : %ld\n",
|
|
|
|
(FLAGS_ops_per_thread * FLAGS_threads)/FLAGS_max_key);
|
2012-11-09 22:04:12 +01:00
|
|
|
fprintf(stdout, "Num times DB reopens: %d\n", FLAGS_reopen);
|
2013-02-21 00:57:27 +01:00
|
|
|
fprintf(stdout, "Batches/snapshots : %d\n",
|
|
|
|
FLAGS_test_batches_snapshots);
|
2013-03-06 21:54:55 +01:00
|
|
|
fprintf(stdout, "Purge redundant %% : %d\n",
|
|
|
|
FLAGS_purge_redundant_percent);
|
2012-11-06 21:02:18 +01:00
|
|
|
fprintf(stdout, "Num keys per lock : %d\n",
|
2012-10-03 18:58:45 +02:00
|
|
|
1 << FLAGS_log2_keys_per_lock);
|
|
|
|
|
2013-01-17 19:04:45 +01:00
|
|
|
const char* compression = "";
|
2012-10-03 18:58:45 +02:00
|
|
|
switch (FLAGS_compression_type) {
|
|
|
|
case leveldb::kNoCompression:
|
2013-01-17 19:04:45 +01:00
|
|
|
compression = "none";
|
2012-10-03 18:58:45 +02:00
|
|
|
break;
|
|
|
|
case leveldb::kSnappyCompression:
|
2013-01-17 19:04:45 +01:00
|
|
|
compression = "snappy";
|
2012-10-03 18:58:45 +02:00
|
|
|
break;
|
|
|
|
case leveldb::kZlibCompression:
|
2013-01-17 19:04:45 +01:00
|
|
|
compression = "zlib";
|
2012-10-03 18:58:45 +02:00
|
|
|
break;
|
|
|
|
case leveldb::kBZip2Compression:
|
2013-01-17 19:04:45 +01:00
|
|
|
compression = "bzip2";
|
2012-10-03 18:58:45 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
fprintf(stdout, "Compression : %s\n", compression);
|
|
|
|
fprintf(stdout, "------------------------------------------------\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
void Open() {
|
2013-02-15 20:53:17 +01:00
|
|
|
assert(db_ == nullptr);
|
2012-10-03 18:58:45 +02:00
|
|
|
Options options;
|
|
|
|
options.block_cache = cache_;
|
|
|
|
options.write_buffer_size = FLAGS_write_buffer_size;
|
2012-10-19 23:00:53 +02:00
|
|
|
options.max_write_buffer_number = FLAGS_max_write_buffer_number;
|
|
|
|
options.max_background_compactions = FLAGS_max_background_compactions;
|
2012-10-03 18:58:45 +02:00
|
|
|
options.block_size = FLAGS_block_size;
|
|
|
|
options.filter_policy = filter_policy_;
|
|
|
|
options.max_open_files = FLAGS_open_files;
|
|
|
|
options.statistics = dbstats;
|
|
|
|
options.env = FLAGS_env;
|
|
|
|
options.disableDataSync = FLAGS_disable_data_sync;
|
|
|
|
options.use_fsync = FLAGS_use_fsync;
|
|
|
|
options.target_file_size_base = FLAGS_target_file_size_base;
|
|
|
|
options.target_file_size_multiplier = FLAGS_target_file_size_multiplier;
|
|
|
|
options.max_bytes_for_level_base = FLAGS_max_bytes_for_level_base;
|
|
|
|
options.max_bytes_for_level_multiplier =
|
|
|
|
FLAGS_max_bytes_for_level_multiplier;
|
|
|
|
options.level0_stop_writes_trigger = FLAGS_level0_stop_writes_trigger;
|
|
|
|
options.level0_slowdown_writes_trigger =
|
|
|
|
FLAGS_level0_slowdown_writes_trigger;
|
|
|
|
options.compression = FLAGS_compression_type;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options.disable_seek_compaction = FLAGS_disable_seek_compaction;
|
2012-10-19 23:00:53 +02:00
|
|
|
options.delete_obsolete_files_period_micros =
|
|
|
|
FLAGS_delete_obsolete_files_period_micros;
|
2013-01-11 02:18:50 +01:00
|
|
|
options.max_manifest_file_size = 1024;
|
2013-03-06 21:54:55 +01:00
|
|
|
static Random purge_percent(1000); // no benefit from non-determinism here
|
|
|
|
if (purge_percent.Uniform(100) < FLAGS_purge_redundant_percent - 1) {
|
|
|
|
options.purge_redundant_kvs_while_flush = false;
|
|
|
|
}
|
2012-10-03 18:58:45 +02:00
|
|
|
Status s = DB::Open(options, FLAGS_db, &db_);
|
|
|
|
if (!s.ok()) {
|
|
|
|
fprintf(stderr, "open error: %s\n", s.ToString().c_str());
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-11-09 22:04:12 +01:00
|
|
|
void Reopen() {
|
2012-11-17 00:28:14 +01:00
|
|
|
// do not close the db. Just delete the lock file. This
|
|
|
|
// simulates a crash-recovery kind of situation.
|
|
|
|
((DBImpl*) db_)->TEST_Destroy_DBImpl();
|
2013-02-15 20:53:17 +01:00
|
|
|
db_ = nullptr;
|
2012-11-17 00:28:14 +01:00
|
|
|
|
|
|
|
num_times_reopened_++;
|
|
|
|
double now = FLAGS_env->NowMicros();
|
|
|
|
fprintf(stdout, "%s Reopening database for the %dth time\n",
|
|
|
|
FLAGS_env->TimeToString((uint64_t) now/1000000).c_str(),
|
|
|
|
num_times_reopened_);
|
2012-11-09 22:04:12 +01:00
|
|
|
Open();
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
void PrintStatistics() {
|
|
|
|
if (dbstats) {
|
|
|
|
fprintf(stdout, "File opened:%ld closed:%ld errors:%ld\n",
|
2013-02-25 22:58:34 +01:00
|
|
|
dbstats->getTickerCount(NO_FILE_OPENS),
|
|
|
|
dbstats->getTickerCount(NO_FILE_CLOSES),
|
|
|
|
dbstats->getTickerCount(NO_FILE_ERRORS));
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
2013-01-20 11:07:13 +01:00
|
|
|
shared_ptr<Cache> cache_;
|
2012-10-03 18:58:45 +02:00
|
|
|
const FilterPolicy* filter_policy_;
|
|
|
|
DB* db_;
|
2012-11-17 00:28:14 +01:00
|
|
|
int num_times_reopened_;
|
2012-10-03 18:58:45 +02:00
|
|
|
};
|
|
|
|
|
|
|
|
} // namespace leveldb
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
|
|
FLAGS_write_buffer_size = leveldb::Options().write_buffer_size;
|
2012-10-19 23:00:53 +02:00
|
|
|
FLAGS_max_write_buffer_number = leveldb::Options().max_write_buffer_number;
|
2012-10-03 18:58:45 +02:00
|
|
|
FLAGS_open_files = leveldb::Options().max_open_files;
|
2012-10-19 23:00:53 +02:00
|
|
|
FLAGS_max_background_compactions =
|
|
|
|
leveldb::Options().max_background_compactions;
|
2012-10-03 18:58:45 +02:00
|
|
|
// Compression test code above refers to FLAGS_block_size
|
|
|
|
FLAGS_block_size = leveldb::Options().block_size;
|
|
|
|
std::string default_db_path;
|
|
|
|
|
|
|
|
for (int i = 1; i < argc; i++) {
|
|
|
|
int n;
|
|
|
|
uint32_t u;
|
|
|
|
long l;
|
|
|
|
char junk;
|
|
|
|
char hdfsname[2048];
|
|
|
|
|
|
|
|
if (sscanf(argv[i], "--seed=%uf%c", &u, &junk) == 1) {
|
|
|
|
FLAGS_seed = u;
|
|
|
|
} else if (sscanf(argv[i], "--max_key=%ld%c", &l, &junk) == 1) {
|
|
|
|
FLAGS_max_key = l;
|
|
|
|
} else if (sscanf(argv[i], "--log2_keys_per_lock=%u%c", &u, &junk) == 1) {
|
|
|
|
FLAGS_log2_keys_per_lock = u;
|
|
|
|
} else if (sscanf(argv[i], "--ops_per_thread=%u%c", &u, &junk) == 1) {
|
|
|
|
FLAGS_ops_per_thread = u;
|
|
|
|
} else if (sscanf(argv[i], "--verbose=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_verbose = n;
|
|
|
|
} else if (sscanf(argv[i], "--histogram=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_histogram = n;
|
|
|
|
} else if (sscanf(argv[i], "--verify_before_write=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_verify_before_write = n;
|
2013-02-21 00:57:27 +01:00
|
|
|
} else if (sscanf(argv[i], "--test_batches_snapshots=%d%c", &n, &junk) == 1
|
|
|
|
&& (n == 0 || n == 1)) {
|
|
|
|
FLAGS_test_batches_snapshots = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--threads=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_threads = n;
|
|
|
|
} else if (sscanf(argv[i], "--value_size_mult=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_value_size_mult = n;
|
|
|
|
} else if (sscanf(argv[i], "--write_buffer_size=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_write_buffer_size = n;
|
2012-10-19 23:00:53 +02:00
|
|
|
} else if (sscanf(argv[i], "--max_write_buffer_number=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_max_write_buffer_number = n;
|
|
|
|
} else if (sscanf(argv[i], "--max_background_compactions=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_max_background_compactions = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--cache_size=%ld%c", &l, &junk) == 1) {
|
|
|
|
FLAGS_cache_size = l;
|
|
|
|
} else if (sscanf(argv[i], "--block_size=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_block_size = n;
|
2012-11-09 22:04:12 +01:00
|
|
|
} else if (sscanf(argv[i], "--reopen=%d%c", &n, &junk) == 1 && n >= 0) {
|
|
|
|
FLAGS_reopen = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--bloom_bits=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_bloom_bits = n;
|
|
|
|
} else if (sscanf(argv[i], "--open_files=%d%c", &n, &junk) == 1) {
|
|
|
|
FLAGS_open_files = n;
|
|
|
|
} else if (strncmp(argv[i], "--db=", 5) == 0) {
|
|
|
|
FLAGS_db = argv[i] + 5;
|
|
|
|
} else if (sscanf(argv[i], "--verify_checksum=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_verify_checksum = n;
|
|
|
|
} else if (sscanf(argv[i], "--bufferedio=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
useOsBuffer = n;
|
|
|
|
} else if (sscanf(argv[i], "--mmap_read=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
useMmapRead = n;
|
|
|
|
} else if (sscanf(argv[i], "--readhead=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
useFsReadAhead = n;
|
|
|
|
} else if (sscanf(argv[i], "--statistics=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
if (n == 1) {
|
2013-03-27 19:27:39 +01:00
|
|
|
dbstats.reset(new leveldb::DBStatistics());
|
2012-10-03 18:58:45 +02:00
|
|
|
}
|
|
|
|
} else if (sscanf(argv[i], "--sync=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_sync = n;
|
2012-11-08 00:35:08 +01:00
|
|
|
} else if (sscanf(argv[i], "--readpercent=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_readpercent = n;
|
|
|
|
} else if (sscanf(argv[i], "--delpercent=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_delpercent = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--disable_data_sync=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_disable_data_sync = n;
|
|
|
|
} else if (sscanf(argv[i], "--use_fsync=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_use_fsync = n;
|
|
|
|
} else if (sscanf(argv[i], "--disable_wal=%d%c", &n, &junk) == 1 &&
|
|
|
|
(n == 0 || n == 1)) {
|
|
|
|
FLAGS_disable_wal = n;
|
|
|
|
} else if (sscanf(argv[i], "--hdfs=%s", hdfsname) == 1) {
|
|
|
|
FLAGS_env = new leveldb::HdfsEnv(hdfsname);
|
|
|
|
} else if (sscanf(argv[i], "--target_file_size_base=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_target_file_size_base = n;
|
|
|
|
} else if ( sscanf(argv[i], "--target_file_size_multiplier=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_target_file_size_multiplier = n;
|
|
|
|
} else if (
|
2012-11-09 03:45:19 +01:00
|
|
|
sscanf(argv[i], "--max_bytes_for_level_base=%ld%c", &l, &junk) == 1) {
|
|
|
|
FLAGS_max_bytes_for_level_base = l;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else if (sscanf(argv[i], "--max_bytes_for_level_multiplier=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_max_bytes_for_level_multiplier = n;
|
|
|
|
} else if (sscanf(argv[i],"--level0_stop_writes_trigger=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_level0_stop_writes_trigger = n;
|
|
|
|
} else if (sscanf(argv[i],"--level0_slowdown_writes_trigger=%d%c",
|
|
|
|
&n, &junk) == 1) {
|
|
|
|
FLAGS_level0_slowdown_writes_trigger = n;
|
|
|
|
} else if (strncmp(argv[i], "--compression_type=", 19) == 0) {
|
|
|
|
const char* ctype = argv[i] + 19;
|
|
|
|
if (!strcasecmp(ctype, "none"))
|
|
|
|
FLAGS_compression_type = leveldb::kNoCompression;
|
|
|
|
else if (!strcasecmp(ctype, "snappy"))
|
|
|
|
FLAGS_compression_type = leveldb::kSnappyCompression;
|
|
|
|
else if (!strcasecmp(ctype, "zlib"))
|
|
|
|
FLAGS_compression_type = leveldb::kZlibCompression;
|
|
|
|
else if (!strcasecmp(ctype, "bzip2"))
|
|
|
|
FLAGS_compression_type = leveldb::kBZip2Compression;
|
|
|
|
else {
|
|
|
|
fprintf(stdout, "Cannot parse %s\n", argv[i]);
|
|
|
|
}
|
|
|
|
} else if (sscanf(argv[i], "--disable_seek_compaction=%d%c", &n, &junk) == 1
|
|
|
|
&& (n == 0 || n == 1)) {
|
|
|
|
FLAGS_disable_seek_compaction = n;
|
2012-10-19 23:00:53 +02:00
|
|
|
} else if (sscanf(argv[i], "--delete_obsolete_files_period_micros=%ld%c",
|
|
|
|
&l, &junk) == 1) {
|
|
|
|
FLAGS_delete_obsolete_files_period_micros = n;
|
2013-03-06 21:54:55 +01:00
|
|
|
} else if (sscanf(argv[i], "--purge_redundant_percent=%d%c", &n, &junk) == 1
|
|
|
|
&& (n >= 0 && n <= 100)) {
|
|
|
|
FLAGS_purge_redundant_percent = n;
|
2012-10-03 18:58:45 +02:00
|
|
|
} else {
|
|
|
|
fprintf(stderr, "Invalid flag '%s'\n", argv[i]);
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-11-15 07:00:11 +01:00
|
|
|
// The number of background threads should be at least as much the
|
|
|
|
// max number of concurrent compactions.
|
|
|
|
FLAGS_env->SetBackgroundThreads(FLAGS_max_background_compactions);
|
|
|
|
|
2012-11-08 00:35:08 +01:00
|
|
|
if ((FLAGS_readpercent + FLAGS_delpercent) > 100) {
|
|
|
|
fprintf(stderr, "Error: Read + Delete percents > 100!\n");
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
|
2012-10-03 18:58:45 +02:00
|
|
|
// Choose a location for the test database if none given with --db=<path>
|
2013-02-15 20:53:17 +01:00
|
|
|
if (FLAGS_db == nullptr) {
|
2012-10-03 18:58:45 +02:00
|
|
|
leveldb::Env::Default()->GetTestDirectory(&default_db_path);
|
|
|
|
default_db_path += "/dbstress";
|
|
|
|
FLAGS_db = default_db_path.c_str();
|
|
|
|
}
|
|
|
|
|
|
|
|
leveldb::StressTest stress;
|
|
|
|
stress.Run();
|
|
|
|
return 0;
|
|
|
|
}
|