2016-02-10 00:12:00 +01:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2017-07-16 01:03:42 +02:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2016-01-21 00:17:52 +01:00
|
|
|
//
|
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
|
|
|
|
#include <stdio.h>
|
|
|
|
|
|
|
|
#include <algorithm>
|
|
|
|
#include <string>
|
|
|
|
|
|
|
|
#include "db/db_test_util.h"
|
|
|
|
#include "port/stack_trace.h"
|
2018-03-02 02:50:54 +01:00
|
|
|
#include "rocksdb/listener.h"
|
2016-01-21 00:17:52 +01:00
|
|
|
#include "rocksdb/options.h"
|
|
|
|
#include "rocksdb/perf_context.h"
|
|
|
|
#include "rocksdb/perf_level.h"
|
|
|
|
#include "rocksdb/table.h"
|
|
|
|
#include "util/random.h"
|
2016-04-21 03:46:54 +02:00
|
|
|
#include "util/string_util.h"
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
namespace rocksdb {
|
|
|
|
|
|
|
|
class DBPropertiesTest : public DBTestBase {
|
|
|
|
public:
|
|
|
|
DBPropertiesTest() : DBTestBase("/db_properties_test") {}
|
|
|
|
};
|
|
|
|
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
TEST_F(DBPropertiesTest, Empty) {
|
|
|
|
do {
|
|
|
|
Options options;
|
|
|
|
options.env = env_;
|
|
|
|
options.write_buffer_size = 100000; // Small write buffer
|
2016-11-16 18:24:52 +01:00
|
|
|
options.allow_concurrent_memtable_write = false;
|
2016-01-21 00:17:52 +01:00
|
|
|
options = CurrentOptions(options);
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
|
|
|
|
std::string num;
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-active-mem-table", &num));
|
|
|
|
ASSERT_EQ("0", num);
|
|
|
|
|
|
|
|
ASSERT_OK(Put(1, "foo", "v1"));
|
|
|
|
ASSERT_EQ("v1", Get(1, "foo"));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-active-mem-table", &num));
|
|
|
|
ASSERT_EQ("1", num);
|
|
|
|
|
|
|
|
// Block sync calls
|
|
|
|
env_->delay_sstable_sync_.store(true, std::memory_order_release);
|
|
|
|
Put(1, "k1", std::string(100000, 'x')); // Fill memtable
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-active-mem-table", &num));
|
|
|
|
ASSERT_EQ("2", num);
|
|
|
|
|
|
|
|
Put(1, "k2", std::string(100000, 'y')); // Trigger compaction
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-active-mem-table", &num));
|
|
|
|
ASSERT_EQ("1", num);
|
|
|
|
|
|
|
|
ASSERT_EQ("v1", Get(1, "foo"));
|
|
|
|
// Release sync calls
|
|
|
|
env_->delay_sstable_sync_.store(false, std::memory_order_release);
|
|
|
|
|
|
|
|
ASSERT_OK(db_->DisableFileDeletions());
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetProperty("rocksdb.is-file-deletions-enabled", &num));
|
2018-03-22 06:07:55 +01:00
|
|
|
ASSERT_EQ("0", num);
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(db_->DisableFileDeletions());
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetProperty("rocksdb.is-file-deletions-enabled", &num));
|
2018-03-22 06:07:55 +01:00
|
|
|
ASSERT_EQ("0", num);
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(db_->DisableFileDeletions());
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetProperty("rocksdb.is-file-deletions-enabled", &num));
|
2018-03-22 06:07:55 +01:00
|
|
|
ASSERT_EQ("0", num);
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(db_->EnableFileDeletions(false));
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetProperty("rocksdb.is-file-deletions-enabled", &num));
|
2018-03-22 06:07:55 +01:00
|
|
|
ASSERT_EQ("0", num);
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(db_->EnableFileDeletions());
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetProperty("rocksdb.is-file-deletions-enabled", &num));
|
2018-03-22 06:07:55 +01:00
|
|
|
ASSERT_EQ("1", num);
|
2016-01-21 00:17:52 +01:00
|
|
|
} while (ChangeOptions());
|
|
|
|
}
|
|
|
|
|
2016-03-01 18:34:50 +01:00
|
|
|
TEST_F(DBPropertiesTest, CurrentVersionNumber) {
|
|
|
|
uint64_t v1, v2, v3;
|
2016-03-03 22:18:56 +01:00
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.current-super-version-number", &v1));
|
2016-03-01 18:34:50 +01:00
|
|
|
Put("12345678", "");
|
2016-03-03 22:18:56 +01:00
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.current-super-version-number", &v2));
|
2016-03-01 18:34:50 +01:00
|
|
|
Flush();
|
2016-03-03 22:18:56 +01:00
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.current-super-version-number", &v3));
|
2016-03-01 18:34:50 +01:00
|
|
|
|
|
|
|
ASSERT_EQ(v1, v2);
|
|
|
|
ASSERT_GT(v3, v2);
|
|
|
|
}
|
|
|
|
|
2016-01-21 00:17:52 +01:00
|
|
|
TEST_F(DBPropertiesTest, GetAggregatedIntPropertyTest) {
|
|
|
|
const int kKeySize = 100;
|
|
|
|
const int kValueSize = 500;
|
|
|
|
const int kKeyNum = 100;
|
|
|
|
|
|
|
|
Options options;
|
|
|
|
options.env = env_;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options.write_buffer_size = (kKeySize + kValueSize) * kKeyNum / 10;
|
|
|
|
// Make them never flush
|
|
|
|
options.min_write_buffer_number_to_merge = 1000;
|
|
|
|
options.max_write_buffer_number = 1000;
|
|
|
|
options = CurrentOptions(options);
|
|
|
|
CreateAndReopenWithCF({"one", "two", "three", "four"}, options);
|
|
|
|
|
|
|
|
Random rnd(301);
|
|
|
|
for (auto* handle : handles_) {
|
|
|
|
for (int i = 0; i < kKeyNum; ++i) {
|
|
|
|
db_->Put(WriteOptions(), handle, RandomString(&rnd, kKeySize),
|
|
|
|
RandomString(&rnd, kValueSize));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t manual_sum = 0;
|
|
|
|
uint64_t api_sum = 0;
|
|
|
|
uint64_t value = 0;
|
|
|
|
for (auto* handle : handles_) {
|
|
|
|
ASSERT_TRUE(
|
|
|
|
db_->GetIntProperty(handle, DB::Properties::kSizeAllMemTables, &value));
|
|
|
|
manual_sum += value;
|
|
|
|
}
|
|
|
|
ASSERT_TRUE(db_->GetAggregatedIntProperty(DB::Properties::kSizeAllMemTables,
|
|
|
|
&api_sum));
|
|
|
|
ASSERT_GT(manual_sum, 0);
|
|
|
|
ASSERT_EQ(manual_sum, api_sum);
|
|
|
|
|
|
|
|
ASSERT_FALSE(db_->GetAggregatedIntProperty(DB::Properties::kDBStats, &value));
|
|
|
|
|
|
|
|
uint64_t before_flush_trm;
|
|
|
|
uint64_t after_flush_trm;
|
|
|
|
for (auto* handle : handles_) {
|
|
|
|
ASSERT_TRUE(db_->GetAggregatedIntProperty(
|
|
|
|
DB::Properties::kEstimateTableReadersMem, &before_flush_trm));
|
|
|
|
|
|
|
|
// Issue flush and expect larger memory usage of table readers.
|
|
|
|
db_->Flush(FlushOptions(), handle);
|
|
|
|
|
|
|
|
ASSERT_TRUE(db_->GetAggregatedIntProperty(
|
|
|
|
DB::Properties::kEstimateTableReadersMem, &after_flush_trm));
|
|
|
|
ASSERT_GT(after_flush_trm, before_flush_trm);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
void ResetTableProperties(TableProperties* tp) {
|
|
|
|
tp->data_size = 0;
|
|
|
|
tp->index_size = 0;
|
|
|
|
tp->filter_size = 0;
|
|
|
|
tp->raw_key_size = 0;
|
|
|
|
tp->raw_value_size = 0;
|
|
|
|
tp->num_data_blocks = 0;
|
|
|
|
tp->num_entries = 0;
|
2018-10-30 23:29:58 +01:00
|
|
|
tp->num_deletions = 0;
|
|
|
|
tp->num_merge_operands = 0;
|
2018-06-27 05:18:43 +02:00
|
|
|
tp->num_range_deletions = 0;
|
2016-01-21 00:17:52 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
void ParseTablePropertiesString(std::string tp_string, TableProperties* tp) {
|
|
|
|
double dummy_double;
|
|
|
|
std::replace(tp_string.begin(), tp_string.end(), ';', ' ');
|
|
|
|
std::replace(tp_string.begin(), tp_string.end(), '=', ' ');
|
|
|
|
ResetTableProperties(tp);
|
|
|
|
sscanf(tp_string.c_str(),
|
2018-10-30 23:29:58 +01:00
|
|
|
"# data blocks %" SCNu64 " # entries %" SCNu64 " # deletions %" SCNu64
|
|
|
|
" # merge operands %" SCNu64 " # range deletions %" SCNu64
|
|
|
|
" raw key size %" SCNu64
|
2016-01-21 00:17:52 +01:00
|
|
|
" raw average key size %lf "
|
|
|
|
" raw value size %" SCNu64
|
|
|
|
" raw average value size %lf "
|
2018-05-26 03:41:31 +02:00
|
|
|
" data block size %" SCNu64 " index block size (user-key? %" SCNu64
|
2018-08-10 01:49:45 +02:00
|
|
|
", delta-value? %" SCNu64 ") %" SCNu64 " filter block size %" SCNu64,
|
2018-10-30 23:29:58 +01:00
|
|
|
&tp->num_data_blocks, &tp->num_entries, &tp->num_deletions,
|
|
|
|
&tp->num_merge_operands, &tp->num_range_deletions, &tp->raw_key_size,
|
|
|
|
&dummy_double, &tp->raw_value_size, &dummy_double, &tp->data_size,
|
|
|
|
&tp->index_key_is_user_key, &tp->index_value_is_delta_encoded,
|
|
|
|
&tp->index_size, &tp->filter_size);
|
2016-01-21 00:17:52 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
void VerifySimilar(uint64_t a, uint64_t b, double bias) {
|
|
|
|
ASSERT_EQ(a == 0U, b == 0U);
|
|
|
|
if (a == 0) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
double dbl_a = static_cast<double>(a);
|
|
|
|
double dbl_b = static_cast<double>(b);
|
|
|
|
if (dbl_a > dbl_b) {
|
|
|
|
ASSERT_LT(static_cast<double>(dbl_a - dbl_b) / (dbl_a + dbl_b), bias);
|
|
|
|
} else {
|
|
|
|
ASSERT_LT(static_cast<double>(dbl_b - dbl_a) / (dbl_a + dbl_b), bias);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void VerifyTableProperties(const TableProperties& base_tp,
|
|
|
|
const TableProperties& new_tp,
|
|
|
|
double filter_size_bias = 0.1,
|
|
|
|
double index_size_bias = 0.1,
|
|
|
|
double data_size_bias = 0.1,
|
|
|
|
double num_data_blocks_bias = 0.05) {
|
|
|
|
VerifySimilar(base_tp.data_size, new_tp.data_size, data_size_bias);
|
|
|
|
VerifySimilar(base_tp.index_size, new_tp.index_size, index_size_bias);
|
|
|
|
VerifySimilar(base_tp.filter_size, new_tp.filter_size, filter_size_bias);
|
|
|
|
VerifySimilar(base_tp.num_data_blocks, new_tp.num_data_blocks,
|
|
|
|
num_data_blocks_bias);
|
2018-10-30 23:29:58 +01:00
|
|
|
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_EQ(base_tp.raw_key_size, new_tp.raw_key_size);
|
|
|
|
ASSERT_EQ(base_tp.raw_value_size, new_tp.raw_value_size);
|
|
|
|
ASSERT_EQ(base_tp.num_entries, new_tp.num_entries);
|
2018-10-30 23:29:58 +01:00
|
|
|
ASSERT_EQ(base_tp.num_deletions, new_tp.num_deletions);
|
2018-06-27 05:18:43 +02:00
|
|
|
ASSERT_EQ(base_tp.num_range_deletions, new_tp.num_range_deletions);
|
2018-10-30 23:29:58 +01:00
|
|
|
|
|
|
|
// Merge operands may become Puts, so we only have an upper bound the exact
|
|
|
|
// number of merge operands.
|
|
|
|
ASSERT_GE(base_tp.num_merge_operands, new_tp.num_merge_operands);
|
2016-01-21 00:17:52 +01:00
|
|
|
}
|
|
|
|
|
2018-08-10 01:49:45 +02:00
|
|
|
void GetExpectedTableProperties(
|
|
|
|
TableProperties* expected_tp, const int kKeySize, const int kValueSize,
|
2018-10-30 23:29:58 +01:00
|
|
|
const int kPutsPerTable, const int kDeletionsPerTable,
|
|
|
|
const int kMergeOperandsPerTable, const int kRangeDeletionsPerTable,
|
2018-08-10 01:49:45 +02:00
|
|
|
const int kTableCount, const int kBloomBitsPerKey, const size_t kBlockSize,
|
|
|
|
const bool index_key_is_user_key, const bool value_delta_encoding) {
|
2018-10-30 23:29:58 +01:00
|
|
|
const int kKeysPerTable =
|
|
|
|
kPutsPerTable + kDeletionsPerTable + kMergeOperandsPerTable;
|
|
|
|
const int kPutCount = kTableCount * kPutsPerTable;
|
|
|
|
const int kDeletionCount = kTableCount * kDeletionsPerTable;
|
|
|
|
const int kMergeCount = kTableCount * kMergeOperandsPerTable;
|
2018-06-27 05:18:43 +02:00
|
|
|
const int kRangeDeletionCount = kTableCount * kRangeDeletionsPerTable;
|
2018-10-30 23:29:58 +01:00
|
|
|
const int kKeyCount = kPutCount + kDeletionCount + kMergeCount;
|
2016-04-26 08:02:14 +02:00
|
|
|
const int kAvgSuccessorSize = kKeySize / 5;
|
2016-01-21 00:17:52 +01:00
|
|
|
const int kEncodingSavePerKey = kKeySize / 4;
|
2018-10-30 23:29:58 +01:00
|
|
|
expected_tp->raw_key_size =
|
|
|
|
(kKeyCount + kRangeDeletionCount) * (kKeySize + 8);
|
|
|
|
expected_tp->raw_value_size =
|
|
|
|
(kPutCount + kMergeCount + kRangeDeletionCount) * kValueSize;
|
2016-01-21 00:17:52 +01:00
|
|
|
expected_tp->num_entries = kKeyCount;
|
2018-10-30 23:29:58 +01:00
|
|
|
expected_tp->num_deletions = kDeletionCount;
|
|
|
|
expected_tp->num_merge_operands = kMergeCount;
|
2018-06-27 05:18:43 +02:00
|
|
|
expected_tp->num_range_deletions = kRangeDeletionCount;
|
2016-01-21 00:17:52 +01:00
|
|
|
expected_tp->num_data_blocks =
|
2018-10-30 23:29:58 +01:00
|
|
|
kTableCount * (kKeysPerTable * (kKeySize - kEncodingSavePerKey + kValueSize)) /
|
2016-01-21 00:17:52 +01:00
|
|
|
kBlockSize;
|
|
|
|
expected_tp->data_size =
|
|
|
|
kTableCount * (kKeysPerTable * (kKeySize + 8 + kValueSize));
|
|
|
|
expected_tp->index_size =
|
2018-05-26 03:41:31 +02:00
|
|
|
expected_tp->num_data_blocks *
|
2018-08-10 01:49:45 +02:00
|
|
|
(kAvgSuccessorSize + (index_key_is_user_key ? 0 : 8) -
|
|
|
|
// discount 1 byte as value size is not encoded in value delta encoding
|
|
|
|
(value_delta_encoding ? 1 : 0));
|
2016-01-21 00:17:52 +01:00
|
|
|
expected_tp->filter_size =
|
|
|
|
kTableCount * (kKeysPerTable * kBloomBitsPerKey / 8);
|
|
|
|
}
|
|
|
|
} // anonymous namespace
|
|
|
|
|
Eliminate duplicated property constants
Summary:
Before this diff, there were duplicated constants to refer to properties (user-
facing API had strings and InternalStats had an enum). I noticed these were
inconsistent in terms of which constants are provided, names of constants, and
documentation of constants. Overall it seemed annoying/error-prone to maintain
these duplicated constants.
So, this diff gets rid of InternalStats's constants and replaces them with a map
keyed on the user-facing constant. The value in that map contains a function
pointer to get the property value, so we don't need to do string matching while
holding db->mutex_. This approach has a side benefit of making many small
handler functions rather than a giant switch-statement.
Test Plan: db_properties_test passes, running "make commit-prereq -j32"
Reviewers: sdong, yhchiang, kradhakrishnan, IslamAbdelRahman, rven, anthony
Reviewed By: anthony
Subscribers: dhruba, leveldb
Differential Revision: https://reviews.facebook.net/D53253
2016-02-03 04:14:56 +01:00
|
|
|
TEST_F(DBPropertiesTest, ValidatePropertyInfo) {
|
|
|
|
for (const auto& ppt_name_and_info : InternalStats::ppt_name_to_info) {
|
|
|
|
// If C++ gets a std::string_literal, this would be better to check at
|
|
|
|
// compile-time using static_assert.
|
|
|
|
ASSERT_TRUE(ppt_name_and_info.first.empty() ||
|
|
|
|
!isdigit(ppt_name_and_info.first.back()));
|
|
|
|
|
2018-06-16 02:23:08 +02:00
|
|
|
int count = 0;
|
|
|
|
count += (ppt_name_and_info.second.handle_string == nullptr) ? 0 : 1;
|
|
|
|
count += (ppt_name_and_info.second.handle_int == nullptr) ? 0 : 1;
|
|
|
|
count += (ppt_name_and_info.second.handle_string_dbimpl == nullptr) ? 0 : 1;
|
|
|
|
ASSERT_TRUE(count == 1);
|
Eliminate duplicated property constants
Summary:
Before this diff, there were duplicated constants to refer to properties (user-
facing API had strings and InternalStats had an enum). I noticed these were
inconsistent in terms of which constants are provided, names of constants, and
documentation of constants. Overall it seemed annoying/error-prone to maintain
these duplicated constants.
So, this diff gets rid of InternalStats's constants and replaces them with a map
keyed on the user-facing constant. The value in that map contains a function
pointer to get the property value, so we don't need to do string matching while
holding db->mutex_. This approach has a side benefit of making many small
handler functions rather than a giant switch-statement.
Test Plan: db_properties_test passes, running "make commit-prereq -j32"
Reviewers: sdong, yhchiang, kradhakrishnan, IslamAbdelRahman, rven, anthony
Reviewed By: anthony
Subscribers: dhruba, leveldb
Differential Revision: https://reviews.facebook.net/D53253
2016-02-03 04:14:56 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-04-02 01:19:12 +02:00
|
|
|
TEST_F(DBPropertiesTest, ValidateSampleNumber) {
|
|
|
|
// When "max_open_files" is -1, we read all the files for
|
|
|
|
// "rocksdb.estimate-num-keys" computation, which is the ground truth.
|
|
|
|
// Otherwise, we sample 20 newest files to make an estimation.
|
|
|
|
// Formula: lastest_20_files_active_key_ratio * total_files
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
options.level0_stop_writes_trigger = 1000;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
int key = 0;
|
|
|
|
for (int files = 20; files >= 10; files -= 10) {
|
|
|
|
for (int i = 0; i < files; i++) {
|
|
|
|
int rows = files / 10;
|
|
|
|
for (int j = 0; j < rows; j++) {
|
|
|
|
db_->Put(WriteOptions(), std::to_string(++key), "foo");
|
|
|
|
}
|
|
|
|
db_->Flush(FlushOptions());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
std::string num;
|
|
|
|
Reopen(options);
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.estimate-num-keys", &num));
|
|
|
|
ASSERT_EQ("45", num);
|
|
|
|
options.max_open_files = -1;
|
|
|
|
Reopen(options);
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.estimate-num-keys", &num));
|
|
|
|
ASSERT_EQ("50", num);
|
|
|
|
}
|
|
|
|
|
2016-01-21 00:17:52 +01:00
|
|
|
TEST_F(DBPropertiesTest, AggregatedTableProperties) {
|
|
|
|
for (int kTableCount = 40; kTableCount <= 100; kTableCount += 30) {
|
2018-10-30 23:29:58 +01:00
|
|
|
const int kDeletionsPerTable = 5;
|
|
|
|
const int kMergeOperandsPerTable = 15;
|
2018-06-27 05:18:43 +02:00
|
|
|
const int kRangeDeletionsPerTable = 5;
|
2018-10-30 23:29:58 +01:00
|
|
|
const int kPutsPerTable = 100;
|
2016-01-21 00:17:52 +01:00
|
|
|
const int kKeySize = 80;
|
|
|
|
const int kValueSize = 200;
|
|
|
|
const int kBloomBitsPerKey = 20;
|
|
|
|
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.level0_file_num_compaction_trigger = 8;
|
|
|
|
options.compression = kNoCompression;
|
|
|
|
options.create_if_missing = true;
|
2018-10-30 23:29:58 +01:00
|
|
|
options.preserve_deletes = true;
|
|
|
|
options.merge_operator.reset(new TestPutOperator());
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.filter_policy.reset(
|
|
|
|
NewBloomFilterPolicy(kBloomBitsPerKey, false));
|
|
|
|
table_options.block_size = 1024;
|
|
|
|
options.table_factory.reset(new BlockBasedTableFactory(table_options));
|
|
|
|
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
2018-06-27 05:18:43 +02:00
|
|
|
// Hold open a snapshot to prevent range tombstones from being compacted
|
|
|
|
// away.
|
|
|
|
ManagedSnapshot snapshot(db_);
|
|
|
|
|
2016-01-21 00:17:52 +01:00
|
|
|
Random rnd(5632);
|
|
|
|
for (int table = 1; table <= kTableCount; ++table) {
|
2018-10-30 23:29:58 +01:00
|
|
|
for (int i = 0; i < kPutsPerTable; ++i) {
|
2016-01-21 00:17:52 +01:00
|
|
|
db_->Put(WriteOptions(), RandomString(&rnd, kKeySize),
|
|
|
|
RandomString(&rnd, kValueSize));
|
|
|
|
}
|
2018-10-30 23:29:58 +01:00
|
|
|
for (int i = 0; i < kDeletionsPerTable; i++) {
|
|
|
|
db_->Delete(WriteOptions(), RandomString(&rnd, kKeySize));
|
|
|
|
}
|
|
|
|
for (int i = 0; i < kMergeOperandsPerTable; i++) {
|
|
|
|
db_->Merge(WriteOptions(), RandomString(&rnd, kKeySize),
|
|
|
|
RandomString(&rnd, kValueSize));
|
|
|
|
}
|
2018-06-27 05:18:43 +02:00
|
|
|
for (int i = 0; i < kRangeDeletionsPerTable; i++) {
|
|
|
|
std::string start = RandomString(&rnd, kKeySize);
|
|
|
|
std::string end = start;
|
|
|
|
end.resize(kValueSize);
|
|
|
|
db_->DeleteRange(WriteOptions(), db_->DefaultColumnFamily(), start, end);
|
|
|
|
}
|
2016-01-21 00:17:52 +01:00
|
|
|
db_->Flush(FlushOptions());
|
|
|
|
}
|
|
|
|
std::string property;
|
|
|
|
db_->GetProperty(DB::Properties::kAggregatedTableProperties, &property);
|
2018-05-26 03:41:31 +02:00
|
|
|
TableProperties output_tp;
|
|
|
|
ParseTablePropertiesString(property, &output_tp);
|
|
|
|
bool index_key_is_user_key = output_tp.index_key_is_user_key > 0;
|
2018-08-10 01:49:45 +02:00
|
|
|
bool value_is_delta_encoded = output_tp.index_value_is_delta_encoded > 0;
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
TableProperties expected_tp;
|
2018-10-30 23:29:58 +01:00
|
|
|
GetExpectedTableProperties(
|
|
|
|
&expected_tp, kKeySize, kValueSize, kPutsPerTable, kDeletionsPerTable,
|
|
|
|
kMergeOperandsPerTable, kRangeDeletionsPerTable, kTableCount,
|
|
|
|
kBloomBitsPerKey, table_options.block_size, index_key_is_user_key,
|
|
|
|
value_is_delta_encoded);
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
VerifyTableProperties(expected_tp, output_tp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, ReadLatencyHistogramByLevel) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.write_buffer_size = 110 << 10;
|
|
|
|
options.level0_file_num_compaction_trigger = 6;
|
|
|
|
options.num_levels = 4;
|
|
|
|
options.compression = kNoCompression;
|
|
|
|
options.max_bytes_for_level_base = 4500 << 10;
|
|
|
|
options.target_file_size_base = 98 << 10;
|
|
|
|
options.max_write_buffer_number = 2;
|
|
|
|
options.statistics = rocksdb::CreateDBStatistics();
|
|
|
|
options.max_open_files = 100;
|
|
|
|
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.no_block_cache = true;
|
|
|
|
|
2017-04-11 17:38:22 +02:00
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
2016-01-21 00:17:52 +01:00
|
|
|
int key_index = 0;
|
|
|
|
Random rnd(301);
|
|
|
|
for (int num = 0; num < 8; num++) {
|
|
|
|
Put("foo", "bar");
|
|
|
|
GenerateNewFile(&rnd, &key_index);
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
}
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
|
|
|
|
std::string prop;
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.dbstats", &prop));
|
|
|
|
|
|
|
|
// Get() after flushes, See latency histogram tracked.
|
|
|
|
for (int key = 0; key < key_index; key++) {
|
|
|
|
Get(Key(key));
|
|
|
|
}
|
2017-04-11 17:38:22 +02:00
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.cfstats", &prop));
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
|
|
|
|
|
|
|
// Reopen and issue Get(). See thee latency tracked
|
2017-04-11 17:38:22 +02:00
|
|
|
ReopenWithColumnFamilies({"default", "pikachu"}, options);
|
2016-01-21 00:17:52 +01:00
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
for (int key = 0; key < key_index; key++) {
|
|
|
|
Get(Key(key));
|
|
|
|
}
|
2018-06-16 02:23:08 +02:00
|
|
|
|
|
|
|
// Test for getting immutable_db_options_.statistics
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(dbfull()->DefaultColumnFamily(),
|
|
|
|
"rocksdb.options-statistics", &prop));
|
|
|
|
ASSERT_NE(std::string::npos, prop.find("rocksdb.block.cache.miss"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("rocksdb.db.f.micros"));
|
|
|
|
|
2017-04-11 17:38:22 +02:00
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(dbfull()->DefaultColumnFamily(),
|
|
|
|
"rocksdb.cf-file-histogram", &prop));
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
|
|
|
|
|
|
|
// Reopen and issue iterating. See thee latency tracked
|
2017-04-11 17:38:22 +02:00
|
|
|
ReopenWithColumnFamilies({"default", "pikachu"}, options);
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.cf-file-histogram", &prop));
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
|
|
|
{
|
|
|
|
unique_ptr<Iterator> iter(db_->NewIterator(ReadOptions()));
|
|
|
|
for (iter->Seek(Key(0)); iter->Valid(); iter->Next()) {
|
|
|
|
}
|
|
|
|
}
|
2017-04-11 17:38:22 +02:00
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.cf-file-histogram", &prop));
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
|
|
|
|
2017-04-11 17:38:22 +02:00
|
|
|
// CF 1 should show no histogram.
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetProperty(handles_[1], "rocksdb.cf-file-histogram", &prop));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
|
|
|
// put something and read it back , CF 1 should show histogram.
|
|
|
|
Put(1, "foo", "bar");
|
|
|
|
Flush(1);
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
ASSERT_EQ("bar", Get(1, "foo"));
|
|
|
|
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetProperty(handles_[1], "rocksdb.cf-file-histogram", &prop));
|
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
|
|
|
|
2016-01-21 00:17:52 +01:00
|
|
|
// options.max_open_files preloads table readers.
|
|
|
|
options.max_open_files = -1;
|
2017-04-11 17:38:22 +02:00
|
|
|
ReopenWithColumnFamilies({"default", "pikachu"}, options);
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(dbfull()->DefaultColumnFamily(),
|
|
|
|
"rocksdb.cf-file-histogram", &prop));
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
|
|
|
for (int key = 0; key < key_index; key++) {
|
|
|
|
Get(Key(key));
|
|
|
|
}
|
2017-04-11 17:38:22 +02:00
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.cfstats", &prop));
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_NE(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
2017-04-19 01:30:51 +02:00
|
|
|
|
|
|
|
// Clear internal stats
|
|
|
|
dbfull()->ResetStats();
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.cfstats", &prop));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 0 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 1 read latency histogram"));
|
|
|
|
ASSERT_EQ(std::string::npos, prop.find("** Level 2 read latency histogram"));
|
2016-01-21 00:17:52 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, AggregatedTablePropertiesAtLevel) {
|
|
|
|
const int kTableCount = 100;
|
2018-10-30 23:29:58 +01:00
|
|
|
const int kDeletionsPerTable = 2;
|
|
|
|
const int kMergeOperandsPerTable = 2;
|
2018-06-27 05:18:43 +02:00
|
|
|
const int kRangeDeletionsPerTable = 2;
|
2018-10-30 23:29:58 +01:00
|
|
|
const int kPutsPerTable = 10;
|
2016-01-21 00:17:52 +01:00
|
|
|
const int kKeySize = 50;
|
|
|
|
const int kValueSize = 400;
|
|
|
|
const int kMaxLevel = 7;
|
|
|
|
const int kBloomBitsPerKey = 20;
|
|
|
|
Random rnd(301);
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.level0_file_num_compaction_trigger = 8;
|
|
|
|
options.compression = kNoCompression;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options.level0_file_num_compaction_trigger = 2;
|
|
|
|
options.target_file_size_base = 8192;
|
|
|
|
options.max_bytes_for_level_base = 10000;
|
|
|
|
options.max_bytes_for_level_multiplier = 2;
|
|
|
|
// This ensures there no compaction happening when we call GetProperty().
|
|
|
|
options.disable_auto_compactions = true;
|
2018-10-30 23:29:58 +01:00
|
|
|
options.preserve_deletes = true;
|
|
|
|
options.merge_operator.reset(new TestPutOperator());
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.filter_policy.reset(
|
|
|
|
NewBloomFilterPolicy(kBloomBitsPerKey, false));
|
|
|
|
table_options.block_size = 1024;
|
|
|
|
options.table_factory.reset(new BlockBasedTableFactory(table_options));
|
|
|
|
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
2018-06-27 05:18:43 +02:00
|
|
|
// Hold open a snapshot to prevent range tombstones from being compacted away.
|
|
|
|
ManagedSnapshot snapshot(db_);
|
|
|
|
|
2016-01-21 00:17:52 +01:00
|
|
|
std::string level_tp_strings[kMaxLevel];
|
|
|
|
std::string tp_string;
|
|
|
|
TableProperties level_tps[kMaxLevel];
|
|
|
|
TableProperties tp, sum_tp, expected_tp;
|
|
|
|
for (int table = 1; table <= kTableCount; ++table) {
|
2018-10-30 23:29:58 +01:00
|
|
|
for (int i = 0; i < kPutsPerTable; ++i) {
|
2016-01-21 00:17:52 +01:00
|
|
|
db_->Put(WriteOptions(), RandomString(&rnd, kKeySize),
|
|
|
|
RandomString(&rnd, kValueSize));
|
|
|
|
}
|
2018-10-30 23:29:58 +01:00
|
|
|
for (int i = 0; i < kDeletionsPerTable; i++) {
|
|
|
|
db_->Delete(WriteOptions(), RandomString(&rnd, kKeySize));
|
|
|
|
}
|
|
|
|
for (int i = 0; i < kMergeOperandsPerTable; i++) {
|
|
|
|
db_->Merge(WriteOptions(), RandomString(&rnd, kKeySize),
|
|
|
|
RandomString(&rnd, kValueSize));
|
|
|
|
}
|
2018-06-27 05:18:43 +02:00
|
|
|
for (int i = 0; i < kRangeDeletionsPerTable; i++) {
|
|
|
|
std::string start = RandomString(&rnd, kKeySize);
|
|
|
|
std::string end = start;
|
|
|
|
end.resize(kValueSize);
|
|
|
|
db_->DeleteRange(WriteOptions(), db_->DefaultColumnFamily(), start, end);
|
|
|
|
}
|
2016-01-21 00:17:52 +01:00
|
|
|
db_->Flush(FlushOptions());
|
|
|
|
db_->CompactRange(CompactRangeOptions(), nullptr, nullptr);
|
|
|
|
ResetTableProperties(&sum_tp);
|
|
|
|
for (int level = 0; level < kMaxLevel; ++level) {
|
|
|
|
db_->GetProperty(
|
|
|
|
DB::Properties::kAggregatedTablePropertiesAtLevel + ToString(level),
|
|
|
|
&level_tp_strings[level]);
|
|
|
|
ParseTablePropertiesString(level_tp_strings[level], &level_tps[level]);
|
|
|
|
sum_tp.data_size += level_tps[level].data_size;
|
|
|
|
sum_tp.index_size += level_tps[level].index_size;
|
|
|
|
sum_tp.filter_size += level_tps[level].filter_size;
|
|
|
|
sum_tp.raw_key_size += level_tps[level].raw_key_size;
|
|
|
|
sum_tp.raw_value_size += level_tps[level].raw_value_size;
|
|
|
|
sum_tp.num_data_blocks += level_tps[level].num_data_blocks;
|
|
|
|
sum_tp.num_entries += level_tps[level].num_entries;
|
2018-10-30 23:29:58 +01:00
|
|
|
sum_tp.num_deletions += level_tps[level].num_deletions;
|
|
|
|
sum_tp.num_merge_operands += level_tps[level].num_merge_operands;
|
2018-06-27 05:18:43 +02:00
|
|
|
sum_tp.num_range_deletions += level_tps[level].num_range_deletions;
|
2016-01-21 00:17:52 +01:00
|
|
|
}
|
|
|
|
db_->GetProperty(DB::Properties::kAggregatedTableProperties, &tp_string);
|
|
|
|
ParseTablePropertiesString(tp_string, &tp);
|
2018-05-26 03:41:31 +02:00
|
|
|
bool index_key_is_user_key = tp.index_key_is_user_key > 0;
|
2018-08-10 01:49:45 +02:00
|
|
|
bool value_is_delta_encoded = tp.index_value_is_delta_encoded > 0;
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_EQ(sum_tp.data_size, tp.data_size);
|
|
|
|
ASSERT_EQ(sum_tp.index_size, tp.index_size);
|
|
|
|
ASSERT_EQ(sum_tp.filter_size, tp.filter_size);
|
|
|
|
ASSERT_EQ(sum_tp.raw_key_size, tp.raw_key_size);
|
|
|
|
ASSERT_EQ(sum_tp.raw_value_size, tp.raw_value_size);
|
|
|
|
ASSERT_EQ(sum_tp.num_data_blocks, tp.num_data_blocks);
|
|
|
|
ASSERT_EQ(sum_tp.num_entries, tp.num_entries);
|
2018-10-30 23:29:58 +01:00
|
|
|
ASSERT_EQ(sum_tp.num_deletions, tp.num_deletions);
|
|
|
|
ASSERT_EQ(sum_tp.num_merge_operands, tp.num_merge_operands);
|
2018-06-27 05:18:43 +02:00
|
|
|
ASSERT_EQ(sum_tp.num_range_deletions, tp.num_range_deletions);
|
2016-01-21 00:17:52 +01:00
|
|
|
if (table > 3) {
|
2018-10-30 23:29:58 +01:00
|
|
|
GetExpectedTableProperties(
|
|
|
|
&expected_tp, kKeySize, kValueSize, kPutsPerTable, kDeletionsPerTable,
|
|
|
|
kMergeOperandsPerTable, kRangeDeletionsPerTable, table,
|
|
|
|
kBloomBitsPerKey, table_options.block_size, index_key_is_user_key,
|
|
|
|
value_is_delta_encoded);
|
2016-01-21 00:17:52 +01:00
|
|
|
// Gives larger bias here as index block size, filter block size,
|
|
|
|
// and data block size become much harder to estimate in this test.
|
2018-06-27 05:18:43 +02:00
|
|
|
VerifyTableProperties(expected_tp, tp, 0.5, 0.4, 0.4, 0.25);
|
2016-01-21 00:17:52 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, NumImmutableMemTable) {
|
|
|
|
do {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
WriteOptions writeOpt = WriteOptions();
|
|
|
|
writeOpt.disableWAL = true;
|
|
|
|
options.max_write_buffer_number = 4;
|
|
|
|
options.min_write_buffer_number_to_merge = 3;
|
2016-01-21 19:59:36 +01:00
|
|
|
options.max_write_buffer_number_to_maintain = 4;
|
2016-01-21 00:17:52 +01:00
|
|
|
options.write_buffer_size = 1000000;
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
|
|
|
|
std::string big_value(1000000 * 2, 'x');
|
|
|
|
std::string num;
|
2016-11-16 05:05:36 +01:00
|
|
|
uint64_t value;
|
2016-01-21 00:17:52 +01:00
|
|
|
SetPerfLevel(kEnableTime);
|
|
|
|
ASSERT_TRUE(GetPerfLevel() == kEnableTime);
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, handles_[1], "k1", big_value));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(handles_[1],
|
|
|
|
"rocksdb.num-immutable-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "0");
|
2016-01-21 19:59:36 +01:00
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], DB::Properties::kNumImmutableMemTableFlushed, &num));
|
|
|
|
ASSERT_EQ(num, "0");
|
2016-01-21 00:17:52 +01:00
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-active-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
2017-06-03 02:12:39 +02:00
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
Get(1, "k1");
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_EQ(1, static_cast<int>(get_perf_context()->get_from_memtable_count));
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, handles_[1], "k2", big_value));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(handles_[1],
|
|
|
|
"rocksdb.num-immutable-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-active-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-imm-mem-tables", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
|
|
|
|
2017-06-03 02:12:39 +02:00
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
Get(1, "k1");
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_EQ(2, static_cast<int>(get_perf_context()->get_from_memtable_count));
|
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
Get(1, "k2");
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_EQ(1, static_cast<int>(get_perf_context()->get_from_memtable_count));
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, handles_[1], "k3", big_value));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.cur-size-active-mem-table", &num));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(handles_[1],
|
|
|
|
"rocksdb.num-immutable-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "2");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-active-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-imm-mem-tables", &num));
|
|
|
|
ASSERT_EQ(num, "2");
|
2017-06-03 02:12:39 +02:00
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
Get(1, "k2");
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_EQ(2, static_cast<int>(get_perf_context()->get_from_memtable_count));
|
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
Get(1, "k3");
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_EQ(1, static_cast<int>(get_perf_context()->get_from_memtable_count));
|
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
Get(1, "k1");
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_EQ(3, static_cast<int>(get_perf_context()->get_from_memtable_count));
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(handles_[1],
|
|
|
|
"rocksdb.num-immutable-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "0");
|
2016-01-21 19:59:36 +01:00
|
|
|
ASSERT_TRUE(dbfull()->GetProperty(
|
|
|
|
handles_[1], DB::Properties::kNumImmutableMemTableFlushed, &num));
|
|
|
|
ASSERT_EQ(num, "3");
|
2016-11-16 05:05:36 +01:00
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
handles_[1], "rocksdb.cur-size-active-mem-table", &value));
|
2016-11-22 23:06:54 +01:00
|
|
|
// "192" is the size of the metadata of two empty skiplists, this would
|
2016-10-27 19:07:28 +02:00
|
|
|
// break if we change the default skiplist implementation
|
2016-11-22 23:06:54 +01:00
|
|
|
ASSERT_GE(value, 192);
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
uint64_t int_num;
|
|
|
|
uint64_t base_total_size;
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
handles_[1], "rocksdb.estimate-num-keys", &base_total_size));
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Delete(writeOpt, handles_[1], "k2"));
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, handles_[1], "k3", ""));
|
|
|
|
ASSERT_OK(dbfull()->Delete(writeOpt, handles_[1], "k3"));
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
handles_[1], "rocksdb.num-deletes-active-mem-table", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 2U);
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-active-mem-table", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 3U);
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, handles_[1], "k2", big_value));
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, handles_[1], "k2", big_value));
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
handles_[1], "rocksdb.num-entries-imm-mem-tables", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 4U);
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
handles_[1], "rocksdb.num-deletes-imm-mem-tables", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 2U);
|
|
|
|
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
handles_[1], "rocksdb.estimate-num-keys", &int_num));
|
|
|
|
ASSERT_EQ(int_num, base_total_size + 1);
|
|
|
|
|
|
|
|
SetPerfLevel(kDisable);
|
|
|
|
ASSERT_TRUE(GetPerfLevel() == kDisable);
|
|
|
|
} while (ChangeCompactOptions());
|
|
|
|
}
|
|
|
|
|
2016-08-25 00:32:01 +02:00
|
|
|
// TODO(techdept) : Disabled flaky test #12863555
|
|
|
|
TEST_F(DBPropertiesTest, DISABLED_GetProperty) {
|
2016-01-21 00:17:52 +01:00
|
|
|
// Set sizes to both background thread pool to be 1 and block them.
|
|
|
|
env_->SetBackgroundThreads(1, Env::HIGH);
|
|
|
|
env_->SetBackgroundThreads(1, Env::LOW);
|
|
|
|
test::SleepingBackgroundTask sleeping_task_low;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask, &sleeping_task_low,
|
|
|
|
Env::Priority::LOW);
|
|
|
|
test::SleepingBackgroundTask sleeping_task_high;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_high, Env::Priority::HIGH);
|
|
|
|
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
WriteOptions writeOpt = WriteOptions();
|
|
|
|
writeOpt.disableWAL = true;
|
|
|
|
options.compaction_style = kCompactionStyleUniversal;
|
|
|
|
options.level0_file_num_compaction_trigger = 1;
|
|
|
|
options.compaction_options_universal.size_ratio = 50;
|
|
|
|
options.max_background_compactions = 1;
|
|
|
|
options.max_background_flushes = 1;
|
|
|
|
options.max_write_buffer_number = 10;
|
|
|
|
options.min_write_buffer_number_to_merge = 1;
|
|
|
|
options.max_write_buffer_number_to_maintain = 0;
|
|
|
|
options.write_buffer_size = 1000000;
|
|
|
|
Reopen(options);
|
|
|
|
|
|
|
|
std::string big_value(1000000 * 2, 'x');
|
|
|
|
std::string num;
|
|
|
|
uint64_t int_num;
|
|
|
|
SetPerfLevel(kEnableTime);
|
|
|
|
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.estimate-table-readers-mem", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 0U);
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.estimate-live-data-size", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 0U);
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k1", big_value));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.num-immutable-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "0");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.mem-table-flush-pending", &num));
|
|
|
|
ASSERT_EQ(num, "0");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.compaction-pending", &num));
|
|
|
|
ASSERT_EQ(num, "0");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.estimate-num-keys", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
2017-06-03 02:12:39 +02:00
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k2", big_value));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.num-immutable-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
|
|
|
ASSERT_OK(dbfull()->Delete(writeOpt, "k-non-existing"));
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k3", big_value));
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.num-immutable-mem-table", &num));
|
|
|
|
ASSERT_EQ(num, "2");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.mem-table-flush-pending", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.compaction-pending", &num));
|
|
|
|
ASSERT_EQ(num, "0");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.estimate-num-keys", &num));
|
|
|
|
ASSERT_EQ(num, "2");
|
|
|
|
// Verify the same set of properties through GetIntProperty
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.num-immutable-mem-table", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 2U);
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.mem-table-flush-pending", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 1U);
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty("rocksdb.compaction-pending", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 0U);
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty("rocksdb.estimate-num-keys", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 2U);
|
|
|
|
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.estimate-table-readers-mem", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 0U);
|
|
|
|
|
|
|
|
sleeping_task_high.WakeUp();
|
|
|
|
sleeping_task_high.WaitUntilDone();
|
|
|
|
dbfull()->TEST_WaitForFlushMemTable();
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k4", big_value));
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k5", big_value));
|
|
|
|
dbfull()->TEST_WaitForFlushMemTable();
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.mem-table-flush-pending", &num));
|
|
|
|
ASSERT_EQ(num, "0");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.compaction-pending", &num));
|
|
|
|
ASSERT_EQ(num, "1");
|
|
|
|
ASSERT_TRUE(dbfull()->GetProperty("rocksdb.estimate-num-keys", &num));
|
|
|
|
ASSERT_EQ(num, "4");
|
|
|
|
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.estimate-table-readers-mem", &int_num));
|
|
|
|
ASSERT_GT(int_num, 0U);
|
|
|
|
|
|
|
|
sleeping_task_low.WakeUp();
|
|
|
|
sleeping_task_low.WaitUntilDone();
|
|
|
|
|
|
|
|
// Wait for compaction to be done. This is important because otherwise RocksDB
|
|
|
|
// might schedule a compaction when reopening the database, failing assertion
|
|
|
|
// (A) as a result.
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
options.max_open_files = 10;
|
|
|
|
Reopen(options);
|
|
|
|
// After reopening, no table reader is loaded, so no memory for table readers
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.estimate-table-readers-mem", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 0U); // (A)
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty("rocksdb.estimate-num-keys", &int_num));
|
|
|
|
ASSERT_GT(int_num, 0U);
|
|
|
|
|
|
|
|
// After reading a key, at least one table reader is loaded.
|
|
|
|
Get("k5");
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.estimate-table-readers-mem", &int_num));
|
|
|
|
ASSERT_GT(int_num, 0U);
|
|
|
|
|
|
|
|
// Test rocksdb.num-live-versions
|
|
|
|
{
|
|
|
|
options.level0_file_num_compaction_trigger = 20;
|
|
|
|
Reopen(options);
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.num-live-versions", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 1U);
|
|
|
|
|
|
|
|
// Use an iterator to hold current version
|
|
|
|
std::unique_ptr<Iterator> iter1(dbfull()->NewIterator(ReadOptions()));
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k6", big_value));
|
|
|
|
Flush();
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.num-live-versions", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 2U);
|
|
|
|
|
|
|
|
// Use an iterator to hold current version
|
|
|
|
std::unique_ptr<Iterator> iter2(dbfull()->NewIterator(ReadOptions()));
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k7", big_value));
|
|
|
|
Flush();
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.num-live-versions", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 3U);
|
|
|
|
|
|
|
|
iter2.reset();
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.num-live-versions", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 2U);
|
|
|
|
|
|
|
|
iter1.reset();
|
|
|
|
ASSERT_TRUE(
|
|
|
|
dbfull()->GetIntProperty("rocksdb.num-live-versions", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 1U);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, ApproximateMemoryUsage) {
|
|
|
|
const int kNumRounds = 10;
|
|
|
|
// TODO(noetzli) kFlushesPerRound does not really correlate with how many
|
|
|
|
// flushes happen.
|
|
|
|
const int kFlushesPerRound = 10;
|
|
|
|
const int kWritesPerFlush = 10;
|
|
|
|
const int kKeySize = 100;
|
|
|
|
const int kValueSize = 1000;
|
|
|
|
Options options;
|
|
|
|
options.write_buffer_size = 1000; // small write buffer
|
|
|
|
options.min_write_buffer_number_to_merge = 4;
|
|
|
|
options.compression = kNoCompression;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options = CurrentOptions(options);
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
Random rnd(301);
|
|
|
|
|
|
|
|
std::vector<Iterator*> iters;
|
|
|
|
|
|
|
|
uint64_t active_mem;
|
|
|
|
uint64_t unflushed_mem;
|
|
|
|
uint64_t all_mem;
|
|
|
|
uint64_t prev_all_mem;
|
|
|
|
|
|
|
|
// Phase 0. The verify the initial value of all these properties are the same
|
|
|
|
// as we have no mem-tables.
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-active-mem-table", &active_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-all-mem-tables", &unflushed_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.size-all-mem-tables", &all_mem);
|
|
|
|
ASSERT_EQ(all_mem, active_mem);
|
|
|
|
ASSERT_EQ(all_mem, unflushed_mem);
|
|
|
|
|
|
|
|
// Phase 1. Simply issue Put() and expect "cur-size-all-mem-tables" equals to
|
|
|
|
// "size-all-mem-tables"
|
|
|
|
for (int r = 0; r < kNumRounds; ++r) {
|
|
|
|
for (int f = 0; f < kFlushesPerRound; ++f) {
|
|
|
|
for (int w = 0; w < kWritesPerFlush; ++w) {
|
|
|
|
Put(RandomString(&rnd, kKeySize), RandomString(&rnd, kValueSize));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Make sure that there is no flush between getting the two properties.
|
|
|
|
dbfull()->TEST_WaitForFlushMemTable();
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-all-mem-tables", &unflushed_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.size-all-mem-tables", &all_mem);
|
|
|
|
// in no iterator case, these two number should be the same.
|
|
|
|
ASSERT_EQ(unflushed_mem, all_mem);
|
|
|
|
}
|
|
|
|
prev_all_mem = all_mem;
|
|
|
|
|
|
|
|
// Phase 2. Keep issuing Put() but also create new iterators. This time we
|
|
|
|
// expect "size-all-mem-tables" > "cur-size-all-mem-tables".
|
|
|
|
for (int r = 0; r < kNumRounds; ++r) {
|
|
|
|
iters.push_back(db_->NewIterator(ReadOptions()));
|
|
|
|
for (int f = 0; f < kFlushesPerRound; ++f) {
|
|
|
|
for (int w = 0; w < kWritesPerFlush; ++w) {
|
|
|
|
Put(RandomString(&rnd, kKeySize), RandomString(&rnd, kValueSize));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Force flush to prevent flush from happening between getting the
|
|
|
|
// properties or after getting the properties and before the new round.
|
|
|
|
Flush();
|
|
|
|
|
|
|
|
// In the second round, add iterators.
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-active-mem-table", &active_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-all-mem-tables", &unflushed_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.size-all-mem-tables", &all_mem);
|
|
|
|
ASSERT_GT(all_mem, active_mem);
|
|
|
|
ASSERT_GT(all_mem, unflushed_mem);
|
|
|
|
ASSERT_GT(all_mem, prev_all_mem);
|
|
|
|
prev_all_mem = all_mem;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Phase 3. Delete iterators and expect "size-all-mem-tables" shrinks
|
|
|
|
// whenever we release an iterator.
|
|
|
|
for (auto* iter : iters) {
|
|
|
|
delete iter;
|
|
|
|
dbfull()->GetIntProperty("rocksdb.size-all-mem-tables", &all_mem);
|
|
|
|
// Expect the size shrinking
|
|
|
|
ASSERT_LT(all_mem, prev_all_mem);
|
|
|
|
prev_all_mem = all_mem;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Expect all these three counters to be the same.
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-active-mem-table", &active_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-all-mem-tables", &unflushed_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.size-all-mem-tables", &all_mem);
|
|
|
|
ASSERT_EQ(active_mem, unflushed_mem);
|
|
|
|
ASSERT_EQ(unflushed_mem, all_mem);
|
|
|
|
|
|
|
|
// Phase 5. Reopen, and expect all these three counters to be the same again.
|
|
|
|
Reopen(options);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-active-mem-table", &active_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.cur-size-all-mem-tables", &unflushed_mem);
|
|
|
|
dbfull()->GetIntProperty("rocksdb.size-all-mem-tables", &all_mem);
|
|
|
|
ASSERT_EQ(active_mem, unflushed_mem);
|
|
|
|
ASSERT_EQ(unflushed_mem, all_mem);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, EstimatePendingCompBytes) {
|
|
|
|
// Set sizes to both background thread pool to be 1 and block them.
|
|
|
|
env_->SetBackgroundThreads(1, Env::HIGH);
|
|
|
|
env_->SetBackgroundThreads(1, Env::LOW);
|
|
|
|
test::SleepingBackgroundTask sleeping_task_low;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask, &sleeping_task_low,
|
|
|
|
Env::Priority::LOW);
|
|
|
|
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
WriteOptions writeOpt = WriteOptions();
|
|
|
|
writeOpt.disableWAL = true;
|
|
|
|
options.compaction_style = kCompactionStyleLevel;
|
|
|
|
options.level0_file_num_compaction_trigger = 2;
|
|
|
|
options.max_background_compactions = 1;
|
|
|
|
options.max_background_flushes = 1;
|
|
|
|
options.max_write_buffer_number = 10;
|
|
|
|
options.min_write_buffer_number_to_merge = 1;
|
|
|
|
options.max_write_buffer_number_to_maintain = 0;
|
|
|
|
options.write_buffer_size = 1000000;
|
|
|
|
Reopen(options);
|
|
|
|
|
|
|
|
std::string big_value(1000000 * 2, 'x');
|
|
|
|
std::string num;
|
|
|
|
uint64_t int_num;
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k1", big_value));
|
|
|
|
Flush();
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
"rocksdb.estimate-pending-compaction-bytes", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 0U);
|
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k2", big_value));
|
|
|
|
Flush();
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
"rocksdb.estimate-pending-compaction-bytes", &int_num));
|
2017-06-02 02:54:06 +02:00
|
|
|
ASSERT_GT(int_num, 0U);
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
ASSERT_OK(dbfull()->Put(writeOpt, "k3", big_value));
|
|
|
|
Flush();
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
"rocksdb.estimate-pending-compaction-bytes", &int_num));
|
|
|
|
ASSERT_GT(int_num, 0U);
|
|
|
|
|
|
|
|
sleeping_task_low.WakeUp();
|
|
|
|
sleeping_task_low.WaitUntilDone();
|
|
|
|
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
|
|
"rocksdb.estimate-pending-compaction-bytes", &int_num));
|
|
|
|
ASSERT_EQ(int_num, 0U);
|
|
|
|
}
|
2016-04-21 03:46:54 +02:00
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, EstimateCompressionRatio) {
|
|
|
|
if (!Snappy_Supported()) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
const int kNumL0Files = 3;
|
|
|
|
const int kNumEntriesPerFile = 1000;
|
|
|
|
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.compression_per_level = {kNoCompression, kSnappyCompression};
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
options.num_levels = 2;
|
|
|
|
Reopen(options);
|
|
|
|
|
|
|
|
// compression ratio is -1.0 when no open files at level
|
|
|
|
ASSERT_EQ(CompressionRatioAtLevel(0), -1.0);
|
|
|
|
|
|
|
|
const std::string kVal(100, 'a');
|
|
|
|
for (int i = 0; i < kNumL0Files; ++i) {
|
|
|
|
for (int j = 0; j < kNumEntriesPerFile; ++j) {
|
|
|
|
// Put common data ("key") at end to prevent delta encoding from
|
|
|
|
// compressing the key effectively
|
|
|
|
std::string key = ToString(i) + ToString(j) + "key";
|
|
|
|
ASSERT_OK(dbfull()->Put(WriteOptions(), key, kVal));
|
|
|
|
}
|
|
|
|
Flush();
|
|
|
|
}
|
|
|
|
|
|
|
|
// no compression at L0, so ratio is less than one
|
|
|
|
ASSERT_LT(CompressionRatioAtLevel(0), 1.0);
|
|
|
|
ASSERT_GT(CompressionRatioAtLevel(0), 0.0);
|
|
|
|
ASSERT_EQ(CompressionRatioAtLevel(1), -1.0);
|
|
|
|
|
|
|
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
|
|
|
|
|
|
|
ASSERT_EQ(CompressionRatioAtLevel(0), -1.0);
|
|
|
|
// Data at L1 should be highly compressed thanks to Snappy and redundant data
|
|
|
|
// in values (ratio is 12.846 as of 4/19/2016).
|
|
|
|
ASSERT_GT(CompressionRatioAtLevel(1), 10.0);
|
|
|
|
}
|
|
|
|
|
2016-01-21 00:17:52 +01:00
|
|
|
#endif // ROCKSDB_LITE
|
|
|
|
|
|
|
|
class CountingUserTblPropCollector : public TablePropertiesCollector {
|
|
|
|
public:
|
|
|
|
const char* Name() const override { return "CountingUserTblPropCollector"; }
|
|
|
|
|
|
|
|
Status Finish(UserCollectedProperties* properties) override {
|
|
|
|
std::string encoded;
|
|
|
|
PutVarint32(&encoded, count_);
|
|
|
|
*properties = UserCollectedProperties{
|
|
|
|
{"CountingUserTblPropCollector", message_}, {"Count", encoded},
|
|
|
|
};
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
2018-03-05 22:08:17 +01:00
|
|
|
Status AddUserKey(const Slice& /*user_key*/, const Slice& /*value*/,
|
|
|
|
EntryType /*type*/, SequenceNumber /*seq*/,
|
|
|
|
uint64_t /*file_size*/) override {
|
2016-01-21 00:17:52 +01:00
|
|
|
++count_;
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
|
|
|
virtual UserCollectedProperties GetReadableProperties() const override {
|
|
|
|
return UserCollectedProperties{};
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
std::string message_ = "Rocksdb";
|
|
|
|
uint32_t count_ = 0;
|
|
|
|
};
|
|
|
|
|
|
|
|
class CountingUserTblPropCollectorFactory
|
|
|
|
: public TablePropertiesCollectorFactory {
|
|
|
|
public:
|
|
|
|
explicit CountingUserTblPropCollectorFactory(
|
|
|
|
uint32_t expected_column_family_id)
|
|
|
|
: expected_column_family_id_(expected_column_family_id),
|
|
|
|
num_created_(0) {}
|
|
|
|
virtual TablePropertiesCollector* CreateTablePropertiesCollector(
|
|
|
|
TablePropertiesCollectorFactory::Context context) override {
|
|
|
|
EXPECT_EQ(expected_column_family_id_, context.column_family_id);
|
|
|
|
num_created_++;
|
|
|
|
return new CountingUserTblPropCollector();
|
|
|
|
}
|
|
|
|
const char* Name() const override {
|
|
|
|
return "CountingUserTblPropCollectorFactory";
|
|
|
|
}
|
|
|
|
void set_expected_column_family_id(uint32_t v) {
|
|
|
|
expected_column_family_id_ = v;
|
|
|
|
}
|
|
|
|
uint32_t expected_column_family_id_;
|
|
|
|
uint32_t num_created_;
|
|
|
|
};
|
|
|
|
|
|
|
|
class CountingDeleteTabPropCollector : public TablePropertiesCollector {
|
|
|
|
public:
|
|
|
|
const char* Name() const override { return "CountingDeleteTabPropCollector"; }
|
|
|
|
|
2018-03-05 22:08:17 +01:00
|
|
|
Status AddUserKey(const Slice& /*user_key*/, const Slice& /*value*/,
|
|
|
|
EntryType type, SequenceNumber /*seq*/,
|
|
|
|
uint64_t /*file_size*/) override {
|
2016-01-21 00:17:52 +01:00
|
|
|
if (type == kEntryDelete) {
|
|
|
|
num_deletes_++;
|
|
|
|
}
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
|
|
|
bool NeedCompact() const override { return num_deletes_ > 10; }
|
|
|
|
|
|
|
|
UserCollectedProperties GetReadableProperties() const override {
|
|
|
|
return UserCollectedProperties{};
|
|
|
|
}
|
|
|
|
|
|
|
|
Status Finish(UserCollectedProperties* properties) override {
|
|
|
|
*properties =
|
|
|
|
UserCollectedProperties{{"num_delete", ToString(num_deletes_)}};
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
uint32_t num_deletes_ = 0;
|
|
|
|
};
|
|
|
|
|
|
|
|
class CountingDeleteTabPropCollectorFactory
|
|
|
|
: public TablePropertiesCollectorFactory {
|
|
|
|
public:
|
|
|
|
virtual TablePropertiesCollector* CreateTablePropertiesCollector(
|
2018-03-05 22:08:17 +01:00
|
|
|
TablePropertiesCollectorFactory::Context /*context*/) override {
|
2016-01-21 00:17:52 +01:00
|
|
|
return new CountingDeleteTabPropCollector();
|
|
|
|
}
|
|
|
|
const char* Name() const override {
|
|
|
|
return "CountingDeleteTabPropCollectorFactory";
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
TEST_F(DBPropertiesTest, GetUserDefinedTableProperties) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.level0_file_num_compaction_trigger = (1 << 30);
|
|
|
|
options.table_properties_collector_factories.resize(1);
|
|
|
|
std::shared_ptr<CountingUserTblPropCollectorFactory> collector_factory =
|
|
|
|
std::make_shared<CountingUserTblPropCollectorFactory>(0);
|
|
|
|
options.table_properties_collector_factories[0] = collector_factory;
|
|
|
|
Reopen(options);
|
|
|
|
// Create 4 tables
|
|
|
|
for (int table = 0; table < 4; ++table) {
|
|
|
|
for (int i = 0; i < 10 + table; ++i) {
|
|
|
|
db_->Put(WriteOptions(), ToString(table * 100 + i), "val");
|
|
|
|
}
|
|
|
|
db_->Flush(FlushOptions());
|
|
|
|
}
|
|
|
|
|
|
|
|
TablePropertiesCollection props;
|
|
|
|
ASSERT_OK(db_->GetPropertiesOfAllTables(&props));
|
|
|
|
ASSERT_EQ(4U, props.size());
|
|
|
|
uint32_t sum = 0;
|
|
|
|
for (const auto& item : props) {
|
|
|
|
auto& user_collected = item.second->user_collected_properties;
|
|
|
|
ASSERT_TRUE(user_collected.find("CountingUserTblPropCollector") !=
|
|
|
|
user_collected.end());
|
|
|
|
ASSERT_EQ(user_collected.at("CountingUserTblPropCollector"), "Rocksdb");
|
|
|
|
ASSERT_TRUE(user_collected.find("Count") != user_collected.end());
|
|
|
|
Slice key(user_collected.at("Count"));
|
|
|
|
uint32_t count;
|
|
|
|
ASSERT_TRUE(GetVarint32(&key, &count));
|
|
|
|
sum += count;
|
|
|
|
}
|
|
|
|
ASSERT_EQ(10u + 11u + 12u + 13u, sum);
|
|
|
|
|
|
|
|
ASSERT_GT(collector_factory->num_created_, 0U);
|
|
|
|
collector_factory->num_created_ = 0;
|
|
|
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
|
|
|
ASSERT_GT(collector_factory->num_created_, 0U);
|
|
|
|
}
|
|
|
|
#endif // ROCKSDB_LITE
|
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, UserDefinedTablePropertiesContext) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.level0_file_num_compaction_trigger = 3;
|
|
|
|
options.table_properties_collector_factories.resize(1);
|
|
|
|
std::shared_ptr<CountingUserTblPropCollectorFactory> collector_factory =
|
|
|
|
std::make_shared<CountingUserTblPropCollectorFactory>(1);
|
|
|
|
options.table_properties_collector_factories[0] = collector_factory,
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
// Create 2 files
|
|
|
|
for (int table = 0; table < 2; ++table) {
|
|
|
|
for (int i = 0; i < 10 + table; ++i) {
|
|
|
|
Put(1, ToString(table * 100 + i), "val");
|
|
|
|
}
|
|
|
|
Flush(1);
|
|
|
|
}
|
|
|
|
ASSERT_GT(collector_factory->num_created_, 0U);
|
|
|
|
|
|
|
|
collector_factory->num_created_ = 0;
|
|
|
|
// Trigger automatic compactions.
|
|
|
|
for (int table = 0; table < 3; ++table) {
|
|
|
|
for (int i = 0; i < 10 + table; ++i) {
|
|
|
|
Put(1, ToString(table * 100 + i), "val");
|
|
|
|
}
|
|
|
|
Flush(1);
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
}
|
|
|
|
ASSERT_GT(collector_factory->num_created_, 0U);
|
|
|
|
|
|
|
|
collector_factory->num_created_ = 0;
|
|
|
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr, handles_[1]);
|
|
|
|
ASSERT_GT(collector_factory->num_created_, 0U);
|
|
|
|
|
|
|
|
// Come back to write to default column family
|
|
|
|
collector_factory->num_created_ = 0;
|
|
|
|
collector_factory->set_expected_column_family_id(0); // default CF
|
|
|
|
// Create 4 tables in default column family
|
|
|
|
for (int table = 0; table < 2; ++table) {
|
|
|
|
for (int i = 0; i < 10 + table; ++i) {
|
|
|
|
Put(ToString(table * 100 + i), "val");
|
|
|
|
}
|
|
|
|
Flush();
|
|
|
|
}
|
|
|
|
ASSERT_GT(collector_factory->num_created_, 0U);
|
|
|
|
|
|
|
|
collector_factory->num_created_ = 0;
|
|
|
|
// Trigger automatic compactions.
|
|
|
|
for (int table = 0; table < 3; ++table) {
|
|
|
|
for (int i = 0; i < 10 + table; ++i) {
|
|
|
|
Put(ToString(table * 100 + i), "val");
|
|
|
|
}
|
|
|
|
Flush();
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
}
|
|
|
|
ASSERT_GT(collector_factory->num_created_, 0U);
|
|
|
|
|
|
|
|
collector_factory->num_created_ = 0;
|
|
|
|
dbfull()->TEST_CompactRange(0, nullptr, nullptr);
|
|
|
|
ASSERT_GT(collector_factory->num_created_, 0U);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
TEST_F(DBPropertiesTest, TablePropertiesNeedCompactTest) {
|
|
|
|
Random rnd(301);
|
|
|
|
|
|
|
|
Options options;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options.write_buffer_size = 4096;
|
|
|
|
options.max_write_buffer_number = 8;
|
|
|
|
options.level0_file_num_compaction_trigger = 2;
|
|
|
|
options.level0_slowdown_writes_trigger = 2;
|
|
|
|
options.level0_stop_writes_trigger = 4;
|
|
|
|
options.target_file_size_base = 2048;
|
|
|
|
options.max_bytes_for_level_base = 10240;
|
|
|
|
options.max_bytes_for_level_multiplier = 4;
|
|
|
|
options.soft_pending_compaction_bytes_limit = 1024 * 1024;
|
|
|
|
options.num_levels = 8;
|
2017-03-13 17:41:30 +01:00
|
|
|
options.env = env_;
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
std::shared_ptr<TablePropertiesCollectorFactory> collector_factory =
|
|
|
|
std::make_shared<CountingDeleteTabPropCollectorFactory>();
|
|
|
|
options.table_properties_collector_factories.resize(1);
|
|
|
|
options.table_properties_collector_factories[0] = collector_factory;
|
|
|
|
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
const int kMaxKey = 1000;
|
|
|
|
for (int i = 0; i < kMaxKey; i++) {
|
|
|
|
ASSERT_OK(Put(Key(i), RandomString(&rnd, 102)));
|
|
|
|
ASSERT_OK(Put(Key(kMaxKey + i), RandomString(&rnd, 102)));
|
|
|
|
}
|
|
|
|
Flush();
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
if (NumTableFilesAtLevel(0) == 1) {
|
|
|
|
// Clear Level 0 so that when later flush a file with deletions,
|
|
|
|
// we don't trigger an organic compaction.
|
|
|
|
ASSERT_OK(Put(Key(0), ""));
|
|
|
|
ASSERT_OK(Put(Key(kMaxKey * 2), ""));
|
|
|
|
Flush();
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
}
|
|
|
|
ASSERT_EQ(NumTableFilesAtLevel(0), 0);
|
|
|
|
|
|
|
|
{
|
|
|
|
int c = 0;
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(ReadOptions()));
|
|
|
|
iter->Seek(Key(kMaxKey - 100));
|
|
|
|
while (iter->Valid() && iter->key().compare(Key(kMaxKey + 100)) < 0) {
|
|
|
|
iter->Next();
|
|
|
|
++c;
|
|
|
|
}
|
|
|
|
ASSERT_EQ(c, 200);
|
|
|
|
}
|
|
|
|
|
|
|
|
Delete(Key(0));
|
|
|
|
for (int i = kMaxKey - 100; i < kMaxKey + 100; i++) {
|
|
|
|
Delete(Key(i));
|
|
|
|
}
|
|
|
|
Delete(Key(kMaxKey * 2));
|
|
|
|
|
|
|
|
Flush();
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
|
|
|
|
{
|
|
|
|
SetPerfLevel(kEnableCount);
|
2017-06-03 02:12:39 +02:00
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
int c = 0;
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(ReadOptions()));
|
|
|
|
iter->Seek(Key(kMaxKey - 100));
|
|
|
|
while (iter->Valid() && iter->key().compare(Key(kMaxKey + 100)) < 0) {
|
|
|
|
iter->Next();
|
|
|
|
}
|
|
|
|
ASSERT_EQ(c, 0);
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_LT(get_perf_context()->internal_delete_skipped_count, 30u);
|
|
|
|
ASSERT_LT(get_perf_context()->internal_key_skipped_count, 30u);
|
2016-01-21 00:17:52 +01:00
|
|
|
SetPerfLevel(kDisable);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, NeedCompactHintPersistentTest) {
|
|
|
|
Random rnd(301);
|
|
|
|
|
|
|
|
Options options;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options.max_write_buffer_number = 8;
|
|
|
|
options.level0_file_num_compaction_trigger = 10;
|
|
|
|
options.level0_slowdown_writes_trigger = 10;
|
|
|
|
options.level0_stop_writes_trigger = 10;
|
|
|
|
options.disable_auto_compactions = true;
|
2017-03-13 17:41:30 +01:00
|
|
|
options.env = env_;
|
2016-01-21 00:17:52 +01:00
|
|
|
|
|
|
|
std::shared_ptr<TablePropertiesCollectorFactory> collector_factory =
|
|
|
|
std::make_shared<CountingDeleteTabPropCollectorFactory>();
|
|
|
|
options.table_properties_collector_factories.resize(1);
|
|
|
|
options.table_properties_collector_factories[0] = collector_factory;
|
|
|
|
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
const int kMaxKey = 100;
|
|
|
|
for (int i = 0; i < kMaxKey; i++) {
|
|
|
|
ASSERT_OK(Put(Key(i), ""));
|
|
|
|
}
|
|
|
|
Flush();
|
|
|
|
dbfull()->TEST_WaitForFlushMemTable();
|
|
|
|
|
|
|
|
for (int i = 1; i < kMaxKey - 1; i++) {
|
|
|
|
Delete(Key(i));
|
|
|
|
}
|
|
|
|
Flush();
|
|
|
|
dbfull()->TEST_WaitForFlushMemTable();
|
|
|
|
ASSERT_EQ(NumTableFilesAtLevel(0), 2);
|
|
|
|
|
|
|
|
// Restart the DB. Although number of files didn't reach
|
|
|
|
// options.level0_file_num_compaction_trigger, compaction should
|
|
|
|
// still be triggered because of the need-compaction hint.
|
|
|
|
options.disable_auto_compactions = false;
|
|
|
|
Reopen(options);
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
ASSERT_EQ(NumTableFilesAtLevel(0), 0);
|
|
|
|
{
|
|
|
|
SetPerfLevel(kEnableCount);
|
2017-06-03 02:12:39 +02:00
|
|
|
get_perf_context()->Reset();
|
2016-01-21 00:17:52 +01:00
|
|
|
int c = 0;
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(ReadOptions()));
|
|
|
|
for (iter->Seek(Key(0)); iter->Valid(); iter->Next()) {
|
|
|
|
c++;
|
|
|
|
}
|
|
|
|
ASSERT_EQ(c, 2);
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_EQ(get_perf_context()->internal_delete_skipped_count, 0);
|
2016-01-21 00:17:52 +01:00
|
|
|
// We iterate every key twice. Is it a bug?
|
2017-06-03 02:12:39 +02:00
|
|
|
ASSERT_LE(get_perf_context()->internal_key_skipped_count, 2);
|
2016-01-21 00:17:52 +01:00
|
|
|
SetPerfLevel(kDisable);
|
|
|
|
}
|
|
|
|
}
|
2017-05-23 19:32:02 +02:00
|
|
|
|
|
|
|
TEST_F(DBPropertiesTest, EstimateNumKeysUnderflow) {
|
|
|
|
Options options;
|
|
|
|
Reopen(options);
|
|
|
|
Put("foo", "bar");
|
|
|
|
Delete("foo");
|
|
|
|
Delete("foo");
|
|
|
|
uint64_t num_keys = 0;
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty("rocksdb.estimate-num-keys", &num_keys));
|
|
|
|
ASSERT_EQ(0, num_keys);
|
|
|
|
}
|
|
|
|
|
2017-10-24 00:22:05 +02:00
|
|
|
TEST_F(DBPropertiesTest, EstimateOldestKeyTime) {
|
|
|
|
std::unique_ptr<MockTimeEnv> mock_env(new MockTimeEnv(Env::Default()));
|
|
|
|
uint64_t oldest_key_time = 0;
|
|
|
|
Options options;
|
|
|
|
options.env = mock_env.get();
|
|
|
|
|
|
|
|
// "rocksdb.estimate-oldest-key-time" only available to fifo compaction.
|
|
|
|
mock_env->set_current_time(100);
|
|
|
|
for (auto compaction : {kCompactionStyleLevel, kCompactionStyleUniversal,
|
|
|
|
kCompactionStyleNone}) {
|
|
|
|
options.compaction_style = compaction;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
ASSERT_OK(Put("foo", "bar"));
|
|
|
|
ASSERT_FALSE(dbfull()->GetIntProperty(
|
|
|
|
DB::Properties::kEstimateOldestKeyTime, &oldest_key_time));
|
|
|
|
}
|
|
|
|
|
|
|
|
options.compaction_style = kCompactionStyleFIFO;
|
|
|
|
options.compaction_options_fifo.ttl = 300;
|
|
|
|
options.compaction_options_fifo.allow_compaction = false;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
mock_env->set_current_time(100);
|
|
|
|
ASSERT_OK(Put("k1", "v1"));
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(DB::Properties::kEstimateOldestKeyTime,
|
|
|
|
&oldest_key_time));
|
|
|
|
ASSERT_EQ(100, oldest_key_time);
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_EQ("1", FilesPerLevel());
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(DB::Properties::kEstimateOldestKeyTime,
|
|
|
|
&oldest_key_time));
|
|
|
|
ASSERT_EQ(100, oldest_key_time);
|
|
|
|
|
|
|
|
mock_env->set_current_time(200);
|
|
|
|
ASSERT_OK(Put("k2", "v2"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_EQ("2", FilesPerLevel());
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(DB::Properties::kEstimateOldestKeyTime,
|
|
|
|
&oldest_key_time));
|
|
|
|
ASSERT_EQ(100, oldest_key_time);
|
|
|
|
|
|
|
|
mock_env->set_current_time(300);
|
|
|
|
ASSERT_OK(Put("k3", "v3"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_EQ("3", FilesPerLevel());
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(DB::Properties::kEstimateOldestKeyTime,
|
|
|
|
&oldest_key_time));
|
|
|
|
ASSERT_EQ(100, oldest_key_time);
|
|
|
|
|
|
|
|
mock_env->set_current_time(450);
|
|
|
|
ASSERT_OK(dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
ASSERT_EQ("2", FilesPerLevel());
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(DB::Properties::kEstimateOldestKeyTime,
|
|
|
|
&oldest_key_time));
|
|
|
|
ASSERT_EQ(200, oldest_key_time);
|
|
|
|
|
|
|
|
mock_env->set_current_time(550);
|
|
|
|
ASSERT_OK(dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
ASSERT_EQ("1", FilesPerLevel());
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(DB::Properties::kEstimateOldestKeyTime,
|
|
|
|
&oldest_key_time));
|
|
|
|
ASSERT_EQ(300, oldest_key_time);
|
|
|
|
|
|
|
|
mock_env->set_current_time(650);
|
|
|
|
ASSERT_OK(dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
ASSERT_EQ("", FilesPerLevel());
|
|
|
|
ASSERT_FALSE(dbfull()->GetIntProperty(DB::Properties::kEstimateOldestKeyTime,
|
|
|
|
&oldest_key_time));
|
|
|
|
|
|
|
|
// Close before mock_env destructs.
|
|
|
|
Close();
|
|
|
|
}
|
|
|
|
|
2018-03-02 02:50:54 +01:00
|
|
|
TEST_F(DBPropertiesTest, SstFilesSize) {
|
|
|
|
struct TestListener : public EventListener {
|
|
|
|
void OnCompactionCompleted(DB* db,
|
|
|
|
const CompactionJobInfo& /*info*/) override {
|
|
|
|
assert(callback_triggered == false);
|
|
|
|
assert(size_before_compaction > 0);
|
|
|
|
callback_triggered = true;
|
|
|
|
uint64_t total_sst_size = 0;
|
|
|
|
uint64_t live_sst_size = 0;
|
|
|
|
bool ok = db->GetIntProperty(DB::Properties::kTotalSstFilesSize,
|
|
|
|
&total_sst_size);
|
|
|
|
ASSERT_TRUE(ok);
|
|
|
|
// total_sst_size include files before and after compaction.
|
|
|
|
ASSERT_GT(total_sst_size, size_before_compaction);
|
|
|
|
ok =
|
|
|
|
db->GetIntProperty(DB::Properties::kLiveSstFilesSize, &live_sst_size);
|
|
|
|
ASSERT_TRUE(ok);
|
|
|
|
// live_sst_size only include files after compaction.
|
|
|
|
ASSERT_GT(live_sst_size, 0);
|
|
|
|
ASSERT_LT(live_sst_size, size_before_compaction);
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t size_before_compaction = 0;
|
|
|
|
bool callback_triggered = false;
|
|
|
|
};
|
|
|
|
std::shared_ptr<TestListener> listener = std::make_shared<TestListener>();
|
|
|
|
|
|
|
|
Options options;
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
options.listeners.push_back(listener);
|
|
|
|
Reopen(options);
|
|
|
|
|
|
|
|
for (int i = 0; i < 10; i++) {
|
|
|
|
ASSERT_OK(Put("key" + ToString(i), std::string(1000, 'v')));
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
for (int i = 0; i < 5; i++) {
|
|
|
|
ASSERT_OK(Delete("key" + ToString(i)));
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
uint64_t sst_size;
|
|
|
|
bool ok = db_->GetIntProperty(DB::Properties::kTotalSstFilesSize, &sst_size);
|
|
|
|
ASSERT_TRUE(ok);
|
|
|
|
ASSERT_GT(sst_size, 0);
|
|
|
|
listener->size_before_compaction = sst_size;
|
|
|
|
// Compact to clean all keys and trigger listener.
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
ASSERT_TRUE(listener->callback_triggered);
|
|
|
|
}
|
|
|
|
|
2018-04-19 06:35:12 +02:00
|
|
|
TEST_F(DBPropertiesTest, BlockCacheProperties) {
|
|
|
|
Options options;
|
|
|
|
uint64_t value;
|
|
|
|
|
|
|
|
// Block cache properties are not available for tables other than
|
|
|
|
// block-based table.
|
|
|
|
options.table_factory.reset(NewPlainTableFactory());
|
|
|
|
Reopen(options);
|
|
|
|
ASSERT_FALSE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCacheCapacity, &value));
|
|
|
|
ASSERT_FALSE(db_->GetIntProperty(DB::Properties::kBlockCacheUsage, &value));
|
|
|
|
ASSERT_FALSE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCachePinnedUsage, &value));
|
|
|
|
|
|
|
|
options.table_factory.reset(NewCuckooTableFactory());
|
|
|
|
Reopen(options);
|
|
|
|
ASSERT_FALSE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCacheCapacity, &value));
|
|
|
|
ASSERT_FALSE(db_->GetIntProperty(DB::Properties::kBlockCacheUsage, &value));
|
|
|
|
ASSERT_FALSE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCachePinnedUsage, &value));
|
|
|
|
|
|
|
|
// Block cache properties are not available if block cache is not used.
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.no_block_cache = true;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
Reopen(options);
|
|
|
|
ASSERT_FALSE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCacheCapacity, &value));
|
|
|
|
ASSERT_FALSE(db_->GetIntProperty(DB::Properties::kBlockCacheUsage, &value));
|
|
|
|
ASSERT_FALSE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCachePinnedUsage, &value));
|
|
|
|
|
|
|
|
// Test with empty block cache.
|
|
|
|
constexpr size_t kCapacity = 100;
|
|
|
|
auto block_cache = NewLRUCache(kCapacity, 0 /*num_shard_bits*/);
|
|
|
|
table_options.block_cache = block_cache;
|
|
|
|
table_options.no_block_cache = false;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
Reopen(options);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheCapacity, &value));
|
|
|
|
ASSERT_EQ(kCapacity, value);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheUsage, &value));
|
|
|
|
ASSERT_EQ(0, value);
|
|
|
|
ASSERT_TRUE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCachePinnedUsage, &value));
|
|
|
|
ASSERT_EQ(0, value);
|
|
|
|
|
|
|
|
// Insert unpinned item to the cache and check size.
|
|
|
|
constexpr size_t kSize1 = 50;
|
|
|
|
block_cache->Insert("item1", nullptr /*value*/, kSize1, nullptr /*deleter*/);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheCapacity, &value));
|
|
|
|
ASSERT_EQ(kCapacity, value);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheUsage, &value));
|
|
|
|
ASSERT_EQ(kSize1, value);
|
|
|
|
ASSERT_TRUE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCachePinnedUsage, &value));
|
|
|
|
ASSERT_EQ(0, value);
|
|
|
|
|
|
|
|
// Insert pinned item to the cache and check size.
|
|
|
|
constexpr size_t kSize2 = 30;
|
|
|
|
Cache::Handle* item2 = nullptr;
|
|
|
|
block_cache->Insert("item2", nullptr /*value*/, kSize2, nullptr /*deleter*/,
|
|
|
|
&item2);
|
|
|
|
ASSERT_NE(nullptr, item2);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheCapacity, &value));
|
|
|
|
ASSERT_EQ(kCapacity, value);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheUsage, &value));
|
|
|
|
ASSERT_EQ(kSize1 + kSize2, value);
|
|
|
|
ASSERT_TRUE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCachePinnedUsage, &value));
|
|
|
|
ASSERT_EQ(kSize2, value);
|
|
|
|
|
|
|
|
// Insert another pinned item to make the cache over-sized.
|
|
|
|
constexpr size_t kSize3 = 80;
|
|
|
|
Cache::Handle* item3 = nullptr;
|
|
|
|
block_cache->Insert("item3", nullptr /*value*/, kSize3, nullptr /*deleter*/,
|
|
|
|
&item3);
|
|
|
|
ASSERT_NE(nullptr, item2);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheCapacity, &value));
|
|
|
|
ASSERT_EQ(kCapacity, value);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheUsage, &value));
|
|
|
|
// Item 1 is evicted.
|
|
|
|
ASSERT_EQ(kSize2 + kSize3, value);
|
|
|
|
ASSERT_TRUE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCachePinnedUsage, &value));
|
|
|
|
ASSERT_EQ(kSize2 + kSize3, value);
|
|
|
|
|
|
|
|
// Check size after release.
|
|
|
|
block_cache->Release(item2);
|
|
|
|
block_cache->Release(item3);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheCapacity, &value));
|
|
|
|
ASSERT_EQ(kCapacity, value);
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty(DB::Properties::kBlockCacheUsage, &value));
|
|
|
|
// item2 will be evicted, while item3 remain in cache after release.
|
|
|
|
ASSERT_EQ(kSize3, value);
|
|
|
|
ASSERT_TRUE(
|
|
|
|
db_->GetIntProperty(DB::Properties::kBlockCachePinnedUsage, &value));
|
|
|
|
ASSERT_EQ(0, value);
|
|
|
|
}
|
|
|
|
|
2016-01-21 00:17:52 +01:00
|
|
|
#endif // ROCKSDB_LITE
|
|
|
|
} // namespace rocksdb
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
|
|
rocksdb::port::InstallStackTraceHandler();
|
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
|
|
return RUN_ALL_TESTS();
|
|
|
|
}
|