Merge branch 'master' into columnfamilies
Conflicts: db/db_impl.cc tools/db_stress.cc
This commit is contained in:
commit
dff9214165
@ -20,6 +20,7 @@
|
|||||||
* Added new option -- verify_checksums_in_compaction
|
* Added new option -- verify_checksums_in_compaction
|
||||||
* Chagned Options.prefix_extractor from raw pointer to shared_ptr (take ownership)
|
* Chagned Options.prefix_extractor from raw pointer to shared_ptr (take ownership)
|
||||||
Changed HashSkipListRepFactory and HashLinkListRepFactory constructor to not take SliceTransform object (use Options.prefix_extractor implicitly)
|
Changed HashSkipListRepFactory and HashLinkListRepFactory constructor to not take SliceTransform object (use Options.prefix_extractor implicitly)
|
||||||
|
* Added Env::GetThreadPoolQueueLen(), which returns the waiting queue length of thread pools
|
||||||
|
|
||||||
### New Features
|
### New Features
|
||||||
* If we find one truncated record at the end of the MANIFEST or WAL files,
|
* If we find one truncated record at the end of the MANIFEST or WAL files,
|
||||||
|
@ -226,6 +226,7 @@ DBImpl::DBImpl(const DBOptions& options, const std::string& dbname)
|
|||||||
logfile_number_(0),
|
logfile_number_(0),
|
||||||
default_cf_handle_(nullptr),
|
default_cf_handle_(nullptr),
|
||||||
tmp_batch_(),
|
tmp_batch_(),
|
||||||
|
bg_schedule_needed_(false),
|
||||||
bg_compaction_scheduled_(0),
|
bg_compaction_scheduled_(0),
|
||||||
bg_manual_only_(0),
|
bg_manual_only_(0),
|
||||||
bg_flush_scheduled_(0),
|
bg_flush_scheduled_(0),
|
||||||
@ -1732,6 +1733,7 @@ Status DBImpl::TEST_WaitForCompact() {
|
|||||||
|
|
||||||
void DBImpl::MaybeScheduleFlushOrCompaction() {
|
void DBImpl::MaybeScheduleFlushOrCompaction() {
|
||||||
mutex_.AssertHeld();
|
mutex_.AssertHeld();
|
||||||
|
bg_schedule_needed_ = false;
|
||||||
if (bg_work_gate_closed_) {
|
if (bg_work_gate_closed_) {
|
||||||
// gate closed for backgrond work
|
// gate closed for backgrond work
|
||||||
} else if (shutting_down_.Acquire_Load()) {
|
} else if (shutting_down_.Acquire_Load()) {
|
||||||
@ -1752,6 +1754,8 @@ void DBImpl::MaybeScheduleFlushOrCompaction() {
|
|||||||
if (bg_flush_scheduled_ < options_.max_background_flushes) {
|
if (bg_flush_scheduled_ < options_.max_background_flushes) {
|
||||||
bg_flush_scheduled_++;
|
bg_flush_scheduled_++;
|
||||||
env_->Schedule(&DBImpl::BGWorkFlush, this, Env::Priority::HIGH);
|
env_->Schedule(&DBImpl::BGWorkFlush, this, Env::Priority::HIGH);
|
||||||
|
} else {
|
||||||
|
bg_schedule_needed_ = true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
bool is_compaction_needed = false;
|
bool is_compaction_needed = false;
|
||||||
@ -1767,11 +1771,13 @@ void DBImpl::MaybeScheduleFlushOrCompaction() {
|
|||||||
// Do it only if max_background_compactions hasn't been reached and, in case
|
// Do it only if max_background_compactions hasn't been reached and, in case
|
||||||
// bg_manual_only_ > 0, if it's a manual compaction.
|
// bg_manual_only_ > 0, if it's a manual compaction.
|
||||||
if ((manual_compaction_ || is_compaction_needed) &&
|
if ((manual_compaction_ || is_compaction_needed) &&
|
||||||
bg_compaction_scheduled_ < options_.max_background_compactions &&
|
|
||||||
(!bg_manual_only_ || manual_compaction_)) {
|
(!bg_manual_only_ || manual_compaction_)) {
|
||||||
|
if (bg_compaction_scheduled_ < options_.max_background_compactions) {
|
||||||
bg_compaction_scheduled_++;
|
bg_compaction_scheduled_++;
|
||||||
env_->Schedule(&DBImpl::BGWorkCompaction, this, Env::Priority::LOW);
|
env_->Schedule(&DBImpl::BGWorkCompaction, this, Env::Priority::LOW);
|
||||||
|
} else {
|
||||||
|
bg_schedule_needed_ = true;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1850,20 +1856,34 @@ void DBImpl::BackgroundCallFlush() {
|
|||||||
// to delete all obsolete files and we force FindObsoleteFiles()
|
// to delete all obsolete files and we force FindObsoleteFiles()
|
||||||
FindObsoleteFiles(deletion_state, !s.ok());
|
FindObsoleteFiles(deletion_state, !s.ok());
|
||||||
// delete unnecessary files if any, this is done outside the mutex
|
// delete unnecessary files if any, this is done outside the mutex
|
||||||
if (deletion_state.HaveSomethingToDelete()) {
|
if (deletion_state.HaveSomethingToDelete() || !log_buffer.IsEmpty()) {
|
||||||
mutex_.Unlock();
|
mutex_.Unlock();
|
||||||
|
// Have to flush the info logs before bg_flush_scheduled_--
|
||||||
|
// because if bg_flush_scheduled_ becomes 0 and the lock is
|
||||||
|
// released, the deconstructor of DB can kick in and destroy all the
|
||||||
|
// states of DB so info_log might not be available after that point.
|
||||||
|
// It also applies to access other states that DB owns.
|
||||||
log_buffer.FlushBufferToLog();
|
log_buffer.FlushBufferToLog();
|
||||||
PurgeObsoleteFiles(deletion_state);
|
if (deletion_state.HaveSomethingToDelete()) {
|
||||||
|
PurgeObsoleteFiles(deletion_state);
|
||||||
|
}
|
||||||
mutex_.Lock();
|
mutex_.Lock();
|
||||||
}
|
}
|
||||||
|
|
||||||
bg_flush_scheduled_--;
|
bg_flush_scheduled_--;
|
||||||
if (madeProgress) {
|
// Any time the mutex is released After finding the work to do, another
|
||||||
|
// thread might execute MaybeScheduleFlushOrCompaction(). It is possible
|
||||||
|
// that there is a pending job but it is not scheduled because of the
|
||||||
|
// max thread limit.
|
||||||
|
if (madeProgress || bg_schedule_needed_) {
|
||||||
MaybeScheduleFlushOrCompaction();
|
MaybeScheduleFlushOrCompaction();
|
||||||
}
|
}
|
||||||
bg_cv_.SignalAll();
|
bg_cv_.SignalAll();
|
||||||
|
// IMPORTANT: there should be no code after calling SignalAll. This call may
|
||||||
|
// signal the DB destructor that it's OK to proceed with destruction. In
|
||||||
|
// that case, all DB variables will be dealloacated and referencing them
|
||||||
|
// will cause trouble.
|
||||||
}
|
}
|
||||||
log_buffer.FlushBufferToLog();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -1913,10 +1933,17 @@ void DBImpl::BackgroundCallCompaction() {
|
|||||||
FindObsoleteFiles(deletion_state, !s.ok());
|
FindObsoleteFiles(deletion_state, !s.ok());
|
||||||
|
|
||||||
// delete unnecessary files if any, this is done outside the mutex
|
// delete unnecessary files if any, this is done outside the mutex
|
||||||
if (deletion_state.HaveSomethingToDelete()) {
|
if (deletion_state.HaveSomethingToDelete() || !log_buffer.IsEmpty()) {
|
||||||
mutex_.Unlock();
|
mutex_.Unlock();
|
||||||
|
// Have to flush the info logs before bg_compaction_scheduled_--
|
||||||
|
// because if bg_flush_scheduled_ becomes 0 and the lock is
|
||||||
|
// released, the deconstructor of DB can kick in and destroy all the
|
||||||
|
// states of DB so info_log might not be available after that point.
|
||||||
|
// It also applies to access other states that DB owns.
|
||||||
log_buffer.FlushBufferToLog();
|
log_buffer.FlushBufferToLog();
|
||||||
PurgeObsoleteFiles(deletion_state);
|
if (deletion_state.HaveSomethingToDelete()) {
|
||||||
|
PurgeObsoleteFiles(deletion_state);
|
||||||
|
}
|
||||||
mutex_.Lock();
|
mutex_.Lock();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1927,12 +1954,20 @@ void DBImpl::BackgroundCallCompaction() {
|
|||||||
// Previous compaction may have produced too many files in a level,
|
// Previous compaction may have produced too many files in a level,
|
||||||
// So reschedule another compaction if we made progress in the
|
// So reschedule another compaction if we made progress in the
|
||||||
// last compaction.
|
// last compaction.
|
||||||
if (madeProgress) {
|
//
|
||||||
|
// Also, any time the mutex is released After finding the work to do,
|
||||||
|
// another thread might execute MaybeScheduleFlushOrCompaction(). It is
|
||||||
|
// possible that there is a pending job but it is not scheduled because of
|
||||||
|
// the max thread limit.
|
||||||
|
if (madeProgress || bg_schedule_needed_) {
|
||||||
MaybeScheduleFlushOrCompaction();
|
MaybeScheduleFlushOrCompaction();
|
||||||
}
|
}
|
||||||
bg_cv_.SignalAll();
|
bg_cv_.SignalAll();
|
||||||
|
// IMPORTANT: there should be no code after calling SignalAll. This call may
|
||||||
|
// signal the DB destructor that it's OK to proceed with destruction. In
|
||||||
|
// that case, all DB variables will be dealloacated and referencing them
|
||||||
|
// will cause trouble.
|
||||||
}
|
}
|
||||||
log_buffer.FlushBufferToLog();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
Status DBImpl::BackgroundCompaction(bool* madeProgress,
|
Status DBImpl::BackgroundCompaction(bool* madeProgress,
|
||||||
|
@ -410,6 +410,10 @@ class DBImpl : public DB {
|
|||||||
// part of ongoing compactions.
|
// part of ongoing compactions.
|
||||||
std::set<uint64_t> pending_outputs_;
|
std::set<uint64_t> pending_outputs_;
|
||||||
|
|
||||||
|
// At least one compaction or flush job is pending but not yet scheduled
|
||||||
|
// because of the max background thread limit.
|
||||||
|
bool bg_schedule_needed_;
|
||||||
|
|
||||||
// count how many background compactions are running or have been scheduled
|
// count how many background compactions are running or have been scheduled
|
||||||
int bg_compaction_scheduled_;
|
int bg_compaction_scheduled_;
|
||||||
|
|
||||||
|
@ -110,6 +110,11 @@ class HdfsEnv : public Env {
|
|||||||
|
|
||||||
virtual void WaitForJoin() { posixEnv->WaitForJoin(); }
|
virtual void WaitForJoin() { posixEnv->WaitForJoin(); }
|
||||||
|
|
||||||
|
virtual unsigned int GetThreadPoolQueueLen(Priority pri = LOW) const
|
||||||
|
override {
|
||||||
|
return posixEnv->GetThreadPoolQueueLen(pri);
|
||||||
|
}
|
||||||
|
|
||||||
virtual Status GetTestDirectory(std::string* path) {
|
virtual Status GetTestDirectory(std::string* path) {
|
||||||
return posixEnv->GetTestDirectory(path);
|
return posixEnv->GetTestDirectory(path);
|
||||||
}
|
}
|
||||||
@ -292,6 +297,10 @@ class HdfsEnv : public Env {
|
|||||||
|
|
||||||
virtual void WaitForJoin() {}
|
virtual void WaitForJoin() {}
|
||||||
|
|
||||||
|
virtual unsigned int GetThreadPoolQueueLen(Priority pri = LOW) const {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
virtual Status GetTestDirectory(std::string* path) {return notsup;}
|
virtual Status GetTestDirectory(std::string* path) {return notsup;}
|
||||||
|
|
||||||
virtual uint64_t NowMicros() {return 0;}
|
virtual uint64_t NowMicros() {return 0;}
|
||||||
|
@ -210,6 +210,11 @@ class Env {
|
|||||||
// Wait for all threads started by StartThread to terminate.
|
// Wait for all threads started by StartThread to terminate.
|
||||||
virtual void WaitForJoin() = 0;
|
virtual void WaitForJoin() = 0;
|
||||||
|
|
||||||
|
// Get thread pool queue length for specific thrad pool.
|
||||||
|
virtual unsigned int GetThreadPoolQueueLen(Priority pri = LOW) const {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
// *path is set to a temporary directory that can be used for testing. It may
|
// *path is set to a temporary directory that can be used for testing. It may
|
||||||
// or many not have just been created. The directory may or may not differ
|
// or many not have just been created. The directory may or may not differ
|
||||||
// between runs of the same process, but subsequent calls will return the
|
// between runs of the same process, but subsequent calls will return the
|
||||||
@ -702,6 +707,9 @@ class EnvWrapper : public Env {
|
|||||||
return target_->StartThread(f, a);
|
return target_->StartThread(f, a);
|
||||||
}
|
}
|
||||||
void WaitForJoin() { return target_->WaitForJoin(); }
|
void WaitForJoin() { return target_->WaitForJoin(); }
|
||||||
|
virtual unsigned int GetThreadPoolQueueLen(Priority pri = LOW) const {
|
||||||
|
return target_->GetThreadPoolQueueLen(pri);
|
||||||
|
}
|
||||||
virtual Status GetTestDirectory(std::string* path) {
|
virtual Status GetTestDirectory(std::string* path) {
|
||||||
return target_->GetTestDirectory(path);
|
return target_->GetTestDirectory(path);
|
||||||
}
|
}
|
||||||
|
@ -93,6 +93,8 @@ def main(argv):
|
|||||||
--max_background_compactions=20
|
--max_background_compactions=20
|
||||||
--max_bytes_for_level_base=10485760
|
--max_bytes_for_level_base=10485760
|
||||||
--filter_deletes=%s
|
--filter_deletes=%s
|
||||||
|
--memtablerep=prefix_hash
|
||||||
|
--prefix_size=7
|
||||||
""" % (ops_per_thread,
|
""" % (ops_per_thread,
|
||||||
threads,
|
threads,
|
||||||
write_buf_size,
|
write_buf_size,
|
||||||
@ -108,16 +110,23 @@ def main(argv):
|
|||||||
print("Running db_stress with pid=%d: %s\n\n"
|
print("Running db_stress with pid=%d: %s\n\n"
|
||||||
% (child.pid, cmd))
|
% (child.pid, cmd))
|
||||||
|
|
||||||
|
stop_early = False
|
||||||
while time.time() < killtime:
|
while time.time() < killtime:
|
||||||
time.sleep(10)
|
if child.poll() is not None:
|
||||||
|
print("WARNING: db_stress ended before kill: exitcode=%d\n"
|
||||||
|
% child.returncode)
|
||||||
|
stop_early = True
|
||||||
|
break
|
||||||
|
time.sleep(1)
|
||||||
|
|
||||||
if child.poll() is not None:
|
if not stop_early:
|
||||||
print("WARNING: db_stress ended before kill: exitcode=%d\n"
|
if child.poll() is not None:
|
||||||
% child.returncode)
|
print("WARNING: db_stress ended before kill: exitcode=%d\n"
|
||||||
else:
|
% child.returncode)
|
||||||
child.kill()
|
else:
|
||||||
print("KILLED %d\n" % child.pid)
|
child.kill()
|
||||||
time.sleep(1) # time to stabilize after a kill
|
print("KILLED %d\n" % child.pid)
|
||||||
|
time.sleep(1) # time to stabilize after a kill
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
line = child.stderr.readline().strip()
|
line = child.stderr.readline().strip()
|
||||||
|
@ -107,6 +107,8 @@ def main(argv):
|
|||||||
--max_background_compactions=20
|
--max_background_compactions=20
|
||||||
--max_bytes_for_level_base=10485760
|
--max_bytes_for_level_base=10485760
|
||||||
--filter_deletes=%s
|
--filter_deletes=%s
|
||||||
|
--memtablerep=prefix_hash
|
||||||
|
--prefix_size=7
|
||||||
%s
|
%s
|
||||||
""" % (random.randint(0, 1),
|
""" % (random.randint(0, 1),
|
||||||
threads,
|
threads,
|
||||||
|
@ -334,19 +334,19 @@ enum RepFactory StringToRepFactory(const char* ctype) {
|
|||||||
return kSkipList;
|
return kSkipList;
|
||||||
}
|
}
|
||||||
static enum RepFactory FLAGS_rep_factory;
|
static enum RepFactory FLAGS_rep_factory;
|
||||||
DEFINE_string(memtablerep, "skip_list", "");
|
DEFINE_string(memtablerep, "prefix_hash", "");
|
||||||
|
|
||||||
static bool ValidatePrefixSize(const char* flagname, int32_t value) {
|
static bool ValidatePrefixSize(const char* flagname, int32_t value) {
|
||||||
if (value < 0 || value>=2000000000) {
|
if (value < 0 || value > 8) {
|
||||||
fprintf(stderr, "Invalid value for --%s: %d. 0<= PrefixSize <=2000000000\n",
|
fprintf(stderr, "Invalid value for --%s: %d. 0 <= PrefixSize <= 8\n",
|
||||||
flagname, value);
|
flagname, value);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
DEFINE_int32(prefix_size, 0, "Control the prefix size for HashSkipListRep");
|
DEFINE_int32(prefix_size, 7, "Control the prefix size for HashSkipListRep");
|
||||||
static const bool FLAGS_prefix_size_dummy __attribute__((unused)) =
|
static const bool FLAGS_prefix_size_dummy =
|
||||||
google::RegisterFlagValidator(&FLAGS_prefix_size, &ValidatePrefixSize);
|
google::RegisterFlagValidator(&FLAGS_prefix_size, &ValidatePrefixSize);
|
||||||
|
|
||||||
DEFINE_bool(use_merge, false, "On true, replaces all writes with a Merge "
|
DEFINE_bool(use_merge, false, "On true, replaces all writes with a Merge "
|
||||||
"that behaves like a Put");
|
"that behaves like a Put");
|
||||||
@ -951,15 +951,15 @@ class StressTest {
|
|||||||
return s;
|
return s;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Given a prefix P, this does prefix scans for "0"+P, "1"+P,..."9"+P
|
// Given a key, this does prefix scans for "0"+P, "1"+P,..."9"+P
|
||||||
// in the same snapshot. Each of these 10 scans returns a series of
|
// in the same snapshot where P is the first FLAGS_prefix_size - 1 bytes
|
||||||
// values; each series should be the same length, and it is verified
|
// of the key. Each of these 10 scans returns a series of values;
|
||||||
// for each index i that all the i'th values are of the form "0"+V,
|
// each series should be the same length, and it is verified for each
|
||||||
// "1"+V,..."9"+V.
|
// index i that all the i'th values are of the form "0"+V, "1"+V,..."9"+V.
|
||||||
// ASSUMES that MultiPut was used to put (K, V)
|
// ASSUMES that MultiPut was used to put (K, V)
|
||||||
Status MultiPrefixScan(ThreadState* thread, const ReadOptions& readoptions,
|
Status MultiPrefixScan(ThreadState* thread, const ReadOptions& readoptions,
|
||||||
ColumnFamilyHandle* column_family,
|
ColumnFamilyHandle* column_family,
|
||||||
const Slice& prefix) {
|
const Slice& key) {
|
||||||
std::string prefixes[10] = {"0", "1", "2", "3", "4",
|
std::string prefixes[10] = {"0", "1", "2", "3", "4",
|
||||||
"5", "6", "7", "8", "9"};
|
"5", "6", "7", "8", "9"};
|
||||||
Slice prefix_slices[10];
|
Slice prefix_slices[10];
|
||||||
@ -968,8 +968,9 @@ class StressTest {
|
|||||||
Iterator* iters[10];
|
Iterator* iters[10];
|
||||||
Status s = Status::OK();
|
Status s = Status::OK();
|
||||||
for (int i = 0; i < 10; i++) {
|
for (int i = 0; i < 10; i++) {
|
||||||
prefixes[i] += prefix.ToString();
|
prefixes[i] += key.ToString();
|
||||||
prefix_slices[i] = prefixes[i];
|
prefixes[i].resize(FLAGS_prefix_size);
|
||||||
|
prefix_slices[i] = Slice(prefixes[i]);
|
||||||
readoptionscopy[i] = readoptions;
|
readoptionscopy[i] = readoptions;
|
||||||
readoptionscopy[i].prefix = &prefix_slices[i];
|
readoptionscopy[i].prefix = &prefix_slices[i];
|
||||||
readoptionscopy[i].snapshot = snapshot;
|
readoptionscopy[i].snapshot = snapshot;
|
||||||
@ -1000,7 +1001,7 @@ class StressTest {
|
|||||||
for (int i = 0; i < 10; i++) {
|
for (int i = 0; i < 10; i++) {
|
||||||
if (values[i] != values[0]) {
|
if (values[i] != values[0]) {
|
||||||
fprintf(stderr, "error : inconsistent values for prefix %s: %s, %s\n",
|
fprintf(stderr, "error : inconsistent values for prefix %s: %s, %s\n",
|
||||||
prefix.ToString().c_str(), values[0].c_str(),
|
prefixes[i].c_str(), values[0].c_str(),
|
||||||
values[i].c_str());
|
values[i].c_str());
|
||||||
// we continue after error rather than exiting so that we can
|
// we continue after error rather than exiting so that we can
|
||||||
// find more errors if any
|
// find more errors if any
|
||||||
@ -1035,6 +1036,7 @@ class StressTest {
|
|||||||
const Snapshot* snapshot = db_->GetSnapshot();
|
const Snapshot* snapshot = db_->GetSnapshot();
|
||||||
ReadOptions readoptionscopy = readoptions;
|
ReadOptions readoptionscopy = readoptions;
|
||||||
readoptionscopy.snapshot = snapshot;
|
readoptionscopy.snapshot = snapshot;
|
||||||
|
readoptionscopy.prefix_seek = FLAGS_prefix_size > 0;
|
||||||
unique_ptr<Iterator> iter(db_->NewIterator(readoptionscopy, column_family));
|
unique_ptr<Iterator> iter(db_->NewIterator(readoptionscopy, column_family));
|
||||||
|
|
||||||
iter->Seek(key);
|
iter->Seek(key);
|
||||||
@ -1149,19 +1151,21 @@ class StressTest {
|
|||||||
}
|
}
|
||||||
} else if ((int)FLAGS_readpercent <= prob_op && prob_op < prefixBound) {
|
} else if ((int)FLAGS_readpercent <= prob_op && prob_op < prefixBound) {
|
||||||
// OPERATION prefix scan
|
// OPERATION prefix scan
|
||||||
// keys are longs (e.g., 8 bytes), so we let prefixes be
|
// keys are 8 bytes long, prefix size is FLAGS_prefix_size. There are
|
||||||
// everything except the last byte. So there will be 2^8=256
|
// (8 - FLAGS_prefix_size) bytes besides the prefix. So there will
|
||||||
// keys per prefix.
|
// be 2 ^ ((8 - FLAGS_prefix_size) * 8) possible keys with the same
|
||||||
Slice prefix = Slice(key.data(), key.size() - 1);
|
// prefix
|
||||||
if (!FLAGS_test_batches_snapshots) {
|
if (!FLAGS_test_batches_snapshots) {
|
||||||
|
Slice prefix = Slice(key.data(), FLAGS_prefix_size);
|
||||||
read_opts.prefix = &prefix;
|
read_opts.prefix = &prefix;
|
||||||
Iterator* iter = db_->NewIterator(read_opts, column_family);
|
Iterator* iter = db_->NewIterator(read_opts, column_family);
|
||||||
int count = 0;
|
int64_t count = 0;
|
||||||
for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
|
for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
|
||||||
assert(iter->key().starts_with(prefix));
|
assert(iter->key().starts_with(prefix));
|
||||||
count++;
|
++count;
|
||||||
}
|
}
|
||||||
assert(count <= 256);
|
assert(count <=
|
||||||
|
(static_cast<int64_t>(1) << ((8 - FLAGS_prefix_size) * 8)));
|
||||||
if (iter->status().ok()) {
|
if (iter->status().ok()) {
|
||||||
thread->stats.AddPrefixes(1, count);
|
thread->stats.AddPrefixes(1, count);
|
||||||
} else {
|
} else {
|
||||||
@ -1169,7 +1173,7 @@ class StressTest {
|
|||||||
}
|
}
|
||||||
delete iter;
|
delete iter;
|
||||||
} else {
|
} else {
|
||||||
MultiPrefixScan(thread, read_opts, column_family, prefix);
|
MultiPrefixScan(thread, read_opts, column_family, key);
|
||||||
}
|
}
|
||||||
read_opts.prefix = nullptr;
|
read_opts.prefix = nullptr;
|
||||||
} else if (prefixBound <= prob_op && prob_op < writeBound) {
|
} else if (prefixBound <= prob_op && prob_op < writeBound) {
|
||||||
@ -1617,6 +1621,18 @@ int main(int argc, char** argv) {
|
|||||||
// max number of concurrent compactions.
|
// max number of concurrent compactions.
|
||||||
FLAGS_env->SetBackgroundThreads(FLAGS_max_background_compactions);
|
FLAGS_env->SetBackgroundThreads(FLAGS_max_background_compactions);
|
||||||
|
|
||||||
|
if (FLAGS_prefixpercent > 0 && FLAGS_prefix_size <= 0) {
|
||||||
|
fprintf(stderr,
|
||||||
|
"Error: prefixpercent is non-zero while prefix_size is "
|
||||||
|
"not positive!\n");
|
||||||
|
exit(1);
|
||||||
|
}
|
||||||
|
if (FLAGS_test_batches_snapshots && FLAGS_prefix_size <= 0) {
|
||||||
|
fprintf(stderr,
|
||||||
|
"Error: please specify prefix_size for "
|
||||||
|
"test_batches_snapshots test!\n");
|
||||||
|
exit(1);
|
||||||
|
}
|
||||||
if ((FLAGS_readpercent + FLAGS_prefixpercent +
|
if ((FLAGS_readpercent + FLAGS_prefixpercent +
|
||||||
FLAGS_writepercent + FLAGS_delpercent + FLAGS_iterpercent) != 100) {
|
FLAGS_writepercent + FLAGS_delpercent + FLAGS_iterpercent) != 100) {
|
||||||
fprintf(stderr,
|
fprintf(stderr,
|
||||||
|
@ -1206,6 +1206,8 @@ class PosixEnv : public Env {
|
|||||||
|
|
||||||
virtual void WaitForJoin();
|
virtual void WaitForJoin();
|
||||||
|
|
||||||
|
virtual unsigned int GetThreadPoolQueueLen(Priority pri = LOW) const override;
|
||||||
|
|
||||||
virtual Status GetTestDirectory(std::string* result) {
|
virtual Status GetTestDirectory(std::string* result) {
|
||||||
const char* env = getenv("TEST_TMPDIR");
|
const char* env = getenv("TEST_TMPDIR");
|
||||||
if (env && env[0] != '\0') {
|
if (env && env[0] != '\0') {
|
||||||
@ -1370,12 +1372,12 @@ class PosixEnv : public Env {
|
|||||||
|
|
||||||
class ThreadPool {
|
class ThreadPool {
|
||||||
public:
|
public:
|
||||||
|
ThreadPool()
|
||||||
ThreadPool() :
|
: total_threads_limit_(1),
|
||||||
total_threads_limit_(1),
|
bgthreads_(0),
|
||||||
bgthreads_(0),
|
queue_(),
|
||||||
queue_(),
|
queue_len_(0),
|
||||||
exit_all_threads_(false) {
|
exit_all_threads_(false) {
|
||||||
PthreadCall("mutex_init", pthread_mutex_init(&mu_, nullptr));
|
PthreadCall("mutex_init", pthread_mutex_init(&mu_, nullptr));
|
||||||
PthreadCall("cvar_init", pthread_cond_init(&bgsignal_, nullptr));
|
PthreadCall("cvar_init", pthread_cond_init(&bgsignal_, nullptr));
|
||||||
}
|
}
|
||||||
@ -1405,6 +1407,7 @@ class PosixEnv : public Env {
|
|||||||
void (*function)(void*) = queue_.front().function;
|
void (*function)(void*) = queue_.front().function;
|
||||||
void* arg = queue_.front().arg;
|
void* arg = queue_.front().arg;
|
||||||
queue_.pop_front();
|
queue_.pop_front();
|
||||||
|
queue_len_.store(queue_.size(), std::memory_order_relaxed);
|
||||||
|
|
||||||
PthreadCall("unlock", pthread_mutex_unlock(&mu_));
|
PthreadCall("unlock", pthread_mutex_unlock(&mu_));
|
||||||
(*function)(arg);
|
(*function)(arg);
|
||||||
@ -1459,6 +1462,7 @@ class PosixEnv : public Env {
|
|||||||
queue_.push_back(BGItem());
|
queue_.push_back(BGItem());
|
||||||
queue_.back().function = function;
|
queue_.back().function = function;
|
||||||
queue_.back().arg = arg;
|
queue_.back().arg = arg;
|
||||||
|
queue_len_.store(queue_.size(), std::memory_order_relaxed);
|
||||||
|
|
||||||
// always wake up at least one waiting thread.
|
// always wake up at least one waiting thread.
|
||||||
PthreadCall("signal", pthread_cond_signal(&bgsignal_));
|
PthreadCall("signal", pthread_cond_signal(&bgsignal_));
|
||||||
@ -1466,6 +1470,10 @@ class PosixEnv : public Env {
|
|||||||
PthreadCall("unlock", pthread_mutex_unlock(&mu_));
|
PthreadCall("unlock", pthread_mutex_unlock(&mu_));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
unsigned int GetQueueLen() const {
|
||||||
|
return queue_len_.load(std::memory_order_relaxed);
|
||||||
|
}
|
||||||
|
|
||||||
private:
|
private:
|
||||||
// Entry per Schedule() call
|
// Entry per Schedule() call
|
||||||
struct BGItem { void* arg; void (*function)(void*); };
|
struct BGItem { void* arg; void (*function)(void*); };
|
||||||
@ -1476,6 +1484,7 @@ class PosixEnv : public Env {
|
|||||||
int total_threads_limit_;
|
int total_threads_limit_;
|
||||||
std::vector<pthread_t> bgthreads_;
|
std::vector<pthread_t> bgthreads_;
|
||||||
BGQueue queue_;
|
BGQueue queue_;
|
||||||
|
std::atomic_uint queue_len_; // Queue length. Used for stats reporting
|
||||||
bool exit_all_threads_;
|
bool exit_all_threads_;
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -1498,6 +1507,11 @@ void PosixEnv::Schedule(void (*function)(void*), void* arg, Priority pri) {
|
|||||||
thread_pools_[pri].Schedule(function, arg);
|
thread_pools_[pri].Schedule(function, arg);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
unsigned int PosixEnv::GetThreadPoolQueueLen(Priority pri) const {
|
||||||
|
assert(pri >= Priority::LOW && pri <= Priority::HIGH);
|
||||||
|
return thread_pools_[pri].GetQueueLen();
|
||||||
|
}
|
||||||
|
|
||||||
namespace {
|
namespace {
|
||||||
struct StartThreadState {
|
struct StartThreadState {
|
||||||
void (*user_function)(void*);
|
void (*user_function)(void*);
|
||||||
|
@ -172,17 +172,30 @@ TEST(EnvPosixTest, TwoPools) {
|
|||||||
env_->SetBackgroundThreads(kLowPoolSize);
|
env_->SetBackgroundThreads(kLowPoolSize);
|
||||||
env_->SetBackgroundThreads(kHighPoolSize, Env::Priority::HIGH);
|
env_->SetBackgroundThreads(kHighPoolSize, Env::Priority::HIGH);
|
||||||
|
|
||||||
|
ASSERT_EQ(0, env_->GetThreadPoolQueueLen(Env::Priority::LOW));
|
||||||
|
ASSERT_EQ(0, env_->GetThreadPoolQueueLen(Env::Priority::HIGH));
|
||||||
|
|
||||||
// schedule same number of jobs in each pool
|
// schedule same number of jobs in each pool
|
||||||
for (int i = 0; i < kJobs; i++) {
|
for (int i = 0; i < kJobs; i++) {
|
||||||
env_->Schedule(&CB::Run, &low_pool_job);
|
env_->Schedule(&CB::Run, &low_pool_job);
|
||||||
env_->Schedule(&CB::Run, &high_pool_job, Env::Priority::HIGH);
|
env_->Schedule(&CB::Run, &high_pool_job, Env::Priority::HIGH);
|
||||||
}
|
}
|
||||||
|
// Wait a short while for the jobs to be dispatched.
|
||||||
|
Env::Default()->SleepForMicroseconds(kDelayMicros);
|
||||||
|
ASSERT_EQ(kJobs - kLowPoolSize, env_->GetThreadPoolQueueLen());
|
||||||
|
ASSERT_EQ(kJobs - kLowPoolSize,
|
||||||
|
env_->GetThreadPoolQueueLen(Env::Priority::LOW));
|
||||||
|
ASSERT_EQ(kJobs - kHighPoolSize,
|
||||||
|
env_->GetThreadPoolQueueLen(Env::Priority::HIGH));
|
||||||
|
|
||||||
// wait for all jobs to finish
|
// wait for all jobs to finish
|
||||||
while (low_pool_job.NumFinished() < kJobs ||
|
while (low_pool_job.NumFinished() < kJobs ||
|
||||||
high_pool_job.NumFinished() < kJobs) {
|
high_pool_job.NumFinished() < kJobs) {
|
||||||
env_->SleepForMicroseconds(kDelayMicros);
|
env_->SleepForMicroseconds(kDelayMicros);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
ASSERT_EQ(0, env_->GetThreadPoolQueueLen(Env::Priority::LOW));
|
||||||
|
ASSERT_EQ(0, env_->GetThreadPoolQueueLen(Env::Priority::HIGH));
|
||||||
}
|
}
|
||||||
|
|
||||||
bool IsSingleVarint(const std::string& s) {
|
bool IsSingleVarint(const std::string& s) {
|
||||||
|
@ -23,6 +23,8 @@ class LogBuffer {
|
|||||||
// Add a log entry to the buffer.
|
// Add a log entry to the buffer.
|
||||||
void AddLogToBuffer(const char* format, va_list ap);
|
void AddLogToBuffer(const char* format, va_list ap);
|
||||||
|
|
||||||
|
size_t IsEmpty() const { return logs_.empty(); }
|
||||||
|
|
||||||
// Flush all buffered log to the info log.
|
// Flush all buffered log to the info log.
|
||||||
void FlushBufferToLog();
|
void FlushBufferToLog();
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user