74544d582f
Summary: Note: This PR is the 4th part of a bigger PR stack (https://github.com/facebook/rocksdb/pull/9073) and will rebase/merge only after the first three PRs (https://github.com/facebook/rocksdb/pull/9070, https://github.com/facebook/rocksdb/pull/9071, https://github.com/facebook/rocksdb/pull/9130) merge. **Context:** Similar to https://github.com/facebook/rocksdb/pull/8428, this PR is to track memory usage during (new) Bloom Filter (i.e,FastLocalBloom) and Ribbon Filter (i.e, Ribbon128) construction, moving toward the goal of [single global memory limit using block cache capacity](https://github.com/facebook/rocksdb/wiki/Projects-Being-Developed#improving-memory-efficiency). It also constrains the size of the banding portion of Ribbon Filter during construction by falling back to Bloom Filter if that banding is, at some point, larger than the available space in the cache under `LRUCacheOptions::strict_capacity_limit=true`. The option to turn on this feature is `BlockBasedTableOptions::reserve_table_builder_memory = true` which by default is set to `false`. We [decided](https://github.com/facebook/rocksdb/pull/9073#discussion_r741548409) not to have separate option for separate memory user in table building therefore their memory accounting are all bundled under one general option. **Summary:** - Reserved/released cache for creation/destruction of three main memory users with the passed-in `FilterBuildingContext::cache_res_mgr` during filter construction: - hash entries (i.e`hash_entries`.size(), we bucket-charge hash entries during insertion for performance), - banding (Ribbon Filter only, `bytes_coeff_rows` +`bytes_result_rows` + `bytes_backtrack`), - final filter (i.e, `mutable_buf`'s size). - Implementation details: in order to use `CacheReservationManager::CacheReservationHandle` to account final filter's memory, we have to store the `CacheReservationManager` object and `CacheReservationHandle` for final filter in `XXPH3BitsFilterBuilder` as well as explicitly delete the filter bits builder when done with the final filter in block based table. - Added option fo run `filter_bench` with this memory reservation feature Pull Request resolved: https://github.com/facebook/rocksdb/pull/9073 Test Plan: - Added new tests in `db_bloom_filter_test` to verify filter construction peak cache reservation under combination of `BlockBasedTable::Rep::FilterType` (e.g, `kFullFilter`, `kPartitionedFilter`), `BloomFilterPolicy::Mode`(e.g, `kFastLocalBloom`, `kStandard128Ribbon`, `kDeprecatedBlock`) and `BlockBasedTableOptions::reserve_table_builder_memory` - To address the concern for slow test: tests with memory reservation under `kFullFilter` + `kStandard128Ribbon` and `kPartitionedFilter` take around **3000 - 6000 ms** and others take around **1500 - 2000 ms**, in total adding **20000 - 25000 ms** to the test suit running locally - Added new test in `bloom_test` to verify Ribbon Filter fallback on large banding in FullFilter - Added test in `filter_bench` to verify that this feature does not significantly slow down Bloom/Ribbon Filter construction speed. Local result averaged over **20** run as below: - FastLocalBloom - baseline `./filter_bench -impl=2 -quick -runs 20 | grep 'Build avg'`: - **Build avg ns/key: 29.56295** (DEBUG_LEVEL=1), **29.98153** (DEBUG_LEVEL=0) - new feature (expected to be similar as above)`./filter_bench -impl=2 -quick -runs 20 -reserve_table_builder_memory=true | grep 'Build avg'`: - **Build avg ns/key: 30.99046** (DEBUG_LEVEL=1), **30.48867** (DEBUG_LEVEL=0) - new feature of RibbonFilter with fallback (expected to be similar as above) `./filter_bench -impl=2 -quick -runs 20 -reserve_table_builder_memory=true -strict_capacity_limit=true | grep 'Build avg'` : - **Build avg ns/key: 31.146975** (DEBUG_LEVEL=1), **30.08165** (DEBUG_LEVEL=0) - Ribbon128 - baseline `./filter_bench -impl=3 -quick -runs 20 | grep 'Build avg'`: - **Build avg ns/key: 129.17585** (DEBUG_LEVEL=1), **130.5225** (DEBUG_LEVEL=0) - new feature (expected to be similar as above) `./filter_bench -impl=3 -quick -runs 20 -reserve_table_builder_memory=true | grep 'Build avg' `: - **Build avg ns/key: 131.61645** (DEBUG_LEVEL=1), **132.98075** (DEBUG_LEVEL=0) - new feature of RibbonFilter with fallback (expected to be a lot faster than above due to fallback) `./filter_bench -impl=3 -quick -runs 20 -reserve_table_builder_memory=true -strict_capacity_limit=true | grep 'Build avg'` : - **Build avg ns/key: 52.032965** (DEBUG_LEVEL=1), **52.597825** (DEBUG_LEVEL=0) - And the warning message of `"Cache reservation for Ribbon filter banding failed due to cache full"` is indeed logged to console. Reviewed By: pdillinger Differential Revision: D31991348 Pulled By: hx235 fbshipit-source-id: 9336b2c60f44d530063da518ceaf56dac5f9df8e
1345 lines
45 KiB
C++
1345 lines
45 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2012 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
#ifndef GFLAGS
|
|
#include <cstdio>
|
|
int main() {
|
|
fprintf(stderr, "Please install gflags to run this test... Skipping...\n");
|
|
return 0;
|
|
}
|
|
#else
|
|
|
|
#include <array>
|
|
#include <cmath>
|
|
#include <vector>
|
|
|
|
#include "cache/cache_entry_roles.h"
|
|
#include "cache/cache_reservation_manager.h"
|
|
#include "memory/arena.h"
|
|
#include "port/jemalloc_helper.h"
|
|
#include "rocksdb/filter_policy.h"
|
|
#include "table/block_based/filter_policy_internal.h"
|
|
#include "test_util/testharness.h"
|
|
#include "test_util/testutil.h"
|
|
#include "util/gflags_compat.h"
|
|
#include "util/hash.h"
|
|
|
|
using GFLAGS_NAMESPACE::ParseCommandLineFlags;
|
|
|
|
// The test is not fully designed for bits_per_key other than 10, but with
|
|
// this parameter you can easily explore the behavior of other bits_per_key.
|
|
// See also filter_bench.
|
|
DEFINE_int32(bits_per_key, 10, "");
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
static const int kVerbose = 1;
|
|
|
|
static Slice Key(int i, char* buffer) {
|
|
std::string s;
|
|
PutFixed32(&s, static_cast<uint32_t>(i));
|
|
memcpy(buffer, s.c_str(), sizeof(i));
|
|
return Slice(buffer, sizeof(i));
|
|
}
|
|
|
|
static int NextLength(int length) {
|
|
if (length < 10) {
|
|
length += 1;
|
|
} else if (length < 100) {
|
|
length += 10;
|
|
} else if (length < 1000) {
|
|
length += 100;
|
|
} else {
|
|
length += 1000;
|
|
}
|
|
return length;
|
|
}
|
|
|
|
class BlockBasedBloomTest : public testing::Test {
|
|
private:
|
|
std::unique_ptr<const FilterPolicy> policy_;
|
|
std::string filter_;
|
|
std::vector<std::string> keys_;
|
|
|
|
public:
|
|
BlockBasedBloomTest() { ResetPolicy(); }
|
|
|
|
void Reset() {
|
|
keys_.clear();
|
|
filter_.clear();
|
|
}
|
|
|
|
void ResetPolicy(double bits_per_key) {
|
|
policy_.reset(new BloomFilterPolicy(bits_per_key,
|
|
BloomFilterPolicy::kDeprecatedBlock));
|
|
Reset();
|
|
}
|
|
|
|
void ResetPolicy() { ResetPolicy(FLAGS_bits_per_key); }
|
|
|
|
void Add(const Slice& s) {
|
|
keys_.push_back(s.ToString());
|
|
}
|
|
|
|
void Build() {
|
|
std::vector<Slice> key_slices;
|
|
for (size_t i = 0; i < keys_.size(); i++) {
|
|
key_slices.push_back(Slice(keys_[i]));
|
|
}
|
|
filter_.clear();
|
|
policy_->CreateFilter(&key_slices[0], static_cast<int>(key_slices.size()),
|
|
&filter_);
|
|
keys_.clear();
|
|
if (kVerbose >= 2) DumpFilter();
|
|
}
|
|
|
|
size_t FilterSize() const {
|
|
return filter_.size();
|
|
}
|
|
|
|
Slice FilterData() const { return Slice(filter_); }
|
|
|
|
void DumpFilter() {
|
|
fprintf(stderr, "F(");
|
|
for (size_t i = 0; i+1 < filter_.size(); i++) {
|
|
const unsigned int c = static_cast<unsigned int>(filter_[i]);
|
|
for (int j = 0; j < 8; j++) {
|
|
fprintf(stderr, "%c", (c & (1 <<j)) ? '1' : '.');
|
|
}
|
|
}
|
|
fprintf(stderr, ")\n");
|
|
}
|
|
|
|
bool Matches(const Slice& s) {
|
|
if (!keys_.empty()) {
|
|
Build();
|
|
}
|
|
return policy_->KeyMayMatch(s, filter_);
|
|
}
|
|
|
|
double FalsePositiveRate() {
|
|
char buffer[sizeof(int)];
|
|
int result = 0;
|
|
for (int i = 0; i < 10000; i++) {
|
|
if (Matches(Key(i + 1000000000, buffer))) {
|
|
result++;
|
|
}
|
|
}
|
|
return result / 10000.0;
|
|
}
|
|
};
|
|
|
|
TEST_F(BlockBasedBloomTest, EmptyFilter) {
|
|
ASSERT_TRUE(! Matches("hello"));
|
|
ASSERT_TRUE(! Matches("world"));
|
|
}
|
|
|
|
TEST_F(BlockBasedBloomTest, Small) {
|
|
Add("hello");
|
|
Add("world");
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
ASSERT_TRUE(! Matches("x"));
|
|
ASSERT_TRUE(! Matches("foo"));
|
|
}
|
|
|
|
TEST_F(BlockBasedBloomTest, VaryingLengths) {
|
|
char buffer[sizeof(int)];
|
|
|
|
// Count number of filters that significantly exceed the false positive rate
|
|
int mediocre_filters = 0;
|
|
int good_filters = 0;
|
|
|
|
for (int length = 1; length <= 10000; length = NextLength(length)) {
|
|
Reset();
|
|
for (int i = 0; i < length; i++) {
|
|
Add(Key(i, buffer));
|
|
}
|
|
Build();
|
|
|
|
ASSERT_LE(FilterSize(), (size_t)((length * FLAGS_bits_per_key / 8) + 40))
|
|
<< length;
|
|
|
|
// All added keys must match
|
|
for (int i = 0; i < length; i++) {
|
|
ASSERT_TRUE(Matches(Key(i, buffer)))
|
|
<< "Length " << length << "; key " << i;
|
|
}
|
|
|
|
// Check false positive rate
|
|
double rate = FalsePositiveRate();
|
|
if (kVerbose >= 1) {
|
|
fprintf(stderr, "False positives: %5.2f%% @ length = %6d ; bytes = %6d\n",
|
|
rate*100.0, length, static_cast<int>(FilterSize()));
|
|
}
|
|
if (FLAGS_bits_per_key == 10) {
|
|
ASSERT_LE(rate, 0.02); // Must not be over 2%
|
|
if (rate > 0.0125) {
|
|
mediocre_filters++; // Allowed, but not too often
|
|
} else {
|
|
good_filters++;
|
|
}
|
|
}
|
|
}
|
|
if (FLAGS_bits_per_key == 10 && kVerbose >= 1) {
|
|
fprintf(stderr, "Filters: %d good, %d mediocre\n",
|
|
good_filters, mediocre_filters);
|
|
}
|
|
ASSERT_LE(mediocre_filters, good_filters/5);
|
|
}
|
|
|
|
// Ensure the implementation doesn't accidentally change in an
|
|
// incompatible way
|
|
TEST_F(BlockBasedBloomTest, Schema) {
|
|
char buffer[sizeof(int)];
|
|
|
|
ResetPolicy(8); // num_probes = 5
|
|
for (int key = 0; key < 87; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
ASSERT_EQ(BloomHash(FilterData()), 3589896109U);
|
|
|
|
ResetPolicy(9); // num_probes = 6
|
|
for (int key = 0; key < 87; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
ASSERT_EQ(BloomHash(FilterData()), 969445585U);
|
|
|
|
ResetPolicy(11); // num_probes = 7
|
|
for (int key = 0; key < 87; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
ASSERT_EQ(BloomHash(FilterData()), 1694458207U);
|
|
|
|
ResetPolicy(10); // num_probes = 6
|
|
for (int key = 0; key < 87; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
ASSERT_EQ(BloomHash(FilterData()), 2373646410U);
|
|
|
|
ResetPolicy(10);
|
|
for (int key = /*CHANGED*/ 1; key < 87; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
ASSERT_EQ(BloomHash(FilterData()), 1908442116U);
|
|
|
|
ResetPolicy(10);
|
|
for (int key = 1; key < /*CHANGED*/ 88; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
ASSERT_EQ(BloomHash(FilterData()), 3057004015U);
|
|
|
|
// With new fractional bits_per_key, check that we are rounding to
|
|
// whole bits per key for old Bloom filters.
|
|
ResetPolicy(9.5); // Treated as 10
|
|
for (int key = 1; key < 88; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
ASSERT_EQ(BloomHash(FilterData()), /*SAME*/ 3057004015U);
|
|
|
|
ResetPolicy(10.499); // Treated as 10
|
|
for (int key = 1; key < 88; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
ASSERT_EQ(BloomHash(FilterData()), /*SAME*/ 3057004015U);
|
|
|
|
ResetPolicy();
|
|
}
|
|
|
|
// Different bits-per-byte
|
|
|
|
class FullBloomTest : public testing::TestWithParam<BloomFilterPolicy::Mode> {
|
|
protected:
|
|
BlockBasedTableOptions table_options_;
|
|
|
|
private:
|
|
std::shared_ptr<const FilterPolicy>& policy_;
|
|
std::unique_ptr<FilterBitsBuilder> bits_builder_;
|
|
std::unique_ptr<FilterBitsReader> bits_reader_;
|
|
std::unique_ptr<const char[]> buf_;
|
|
size_t filter_size_;
|
|
|
|
public:
|
|
FullBloomTest() : policy_(table_options_.filter_policy), filter_size_(0) {
|
|
ResetPolicy();
|
|
}
|
|
|
|
BuiltinFilterBitsBuilder* GetBuiltinFilterBitsBuilder() {
|
|
// Throws on bad cast
|
|
return &dynamic_cast<BuiltinFilterBitsBuilder&>(*bits_builder_);
|
|
}
|
|
|
|
const BloomFilterPolicy* GetBloomFilterPolicy() {
|
|
// Throws on bad cast
|
|
return &dynamic_cast<const BloomFilterPolicy&>(*policy_);
|
|
}
|
|
|
|
void Reset() {
|
|
bits_builder_.reset(BloomFilterPolicy::GetBuilderFromContext(
|
|
FilterBuildingContext(table_options_)));
|
|
bits_reader_.reset(nullptr);
|
|
buf_.reset(nullptr);
|
|
filter_size_ = 0;
|
|
}
|
|
|
|
void ResetPolicy(double bits_per_key) {
|
|
policy_.reset(new BloomFilterPolicy(bits_per_key, GetParam()));
|
|
Reset();
|
|
}
|
|
|
|
void ResetPolicy() { ResetPolicy(FLAGS_bits_per_key); }
|
|
|
|
void Add(const Slice& s) {
|
|
bits_builder_->AddKey(s);
|
|
}
|
|
|
|
void OpenRaw(const Slice& s) {
|
|
bits_reader_.reset(policy_->GetFilterBitsReader(s));
|
|
}
|
|
|
|
void Build() {
|
|
Slice filter = bits_builder_->Finish(&buf_);
|
|
bits_reader_.reset(policy_->GetFilterBitsReader(filter));
|
|
filter_size_ = filter.size();
|
|
}
|
|
|
|
size_t FilterSize() const {
|
|
return filter_size_;
|
|
}
|
|
|
|
Slice FilterData() { return Slice(buf_.get(), filter_size_); }
|
|
|
|
int GetNumProbesFromFilterData() {
|
|
assert(filter_size_ >= 5);
|
|
int8_t raw_num_probes = static_cast<int8_t>(buf_.get()[filter_size_ - 5]);
|
|
if (raw_num_probes == -1) { // New bloom filter marker
|
|
return static_cast<uint8_t>(buf_.get()[filter_size_ - 3]);
|
|
} else {
|
|
return raw_num_probes;
|
|
}
|
|
}
|
|
|
|
int GetRibbonSeedFromFilterData() {
|
|
assert(filter_size_ >= 5);
|
|
// Check for ribbon marker
|
|
assert(-2 == static_cast<int8_t>(buf_.get()[filter_size_ - 5]));
|
|
return static_cast<uint8_t>(buf_.get()[filter_size_ - 4]);
|
|
}
|
|
|
|
bool Matches(const Slice& s) {
|
|
if (bits_reader_ == nullptr) {
|
|
Build();
|
|
}
|
|
return bits_reader_->MayMatch(s);
|
|
}
|
|
|
|
// Provides a kind of fingerprint on the Bloom filter's
|
|
// behavior, for reasonbly high FP rates.
|
|
uint64_t PackedMatches() {
|
|
char buffer[sizeof(int)];
|
|
uint64_t result = 0;
|
|
for (int i = 0; i < 64; i++) {
|
|
if (Matches(Key(i + 12345, buffer))) {
|
|
result |= uint64_t{1} << i;
|
|
}
|
|
}
|
|
return result;
|
|
}
|
|
|
|
// Provides a kind of fingerprint on the Bloom filter's
|
|
// behavior, for lower FP rates.
|
|
std::string FirstFPs(int count) {
|
|
char buffer[sizeof(int)];
|
|
std::string rv;
|
|
int fp_count = 0;
|
|
for (int i = 0; i < 1000000; i++) {
|
|
// Pack four match booleans into each hexadecimal digit
|
|
if (Matches(Key(i + 1000000, buffer))) {
|
|
++fp_count;
|
|
rv += std::to_string(i);
|
|
if (fp_count == count) {
|
|
break;
|
|
}
|
|
rv += ',';
|
|
}
|
|
}
|
|
return rv;
|
|
}
|
|
|
|
double FalsePositiveRate() {
|
|
char buffer[sizeof(int)];
|
|
int result = 0;
|
|
for (int i = 0; i < 10000; i++) {
|
|
if (Matches(Key(i + 1000000000, buffer))) {
|
|
result++;
|
|
}
|
|
}
|
|
return result / 10000.0;
|
|
}
|
|
};
|
|
|
|
TEST_P(FullBloomTest, FilterSize) {
|
|
// In addition to checking the consistency of space computation, we are
|
|
// checking that denoted and computed doubles are interpreted as expected
|
|
// as bits_per_key values.
|
|
bool some_computed_less_than_denoted = false;
|
|
// Note: enforced minimum is 1 bit per key (1000 millibits), and enforced
|
|
// maximum is 100 bits per key (100000 millibits).
|
|
for (auto bpk :
|
|
std::vector<std::pair<double, int> >{{-HUGE_VAL, 1000},
|
|
{-INFINITY, 1000},
|
|
{0.0, 1000},
|
|
{1.234, 1234},
|
|
{3.456, 3456},
|
|
{9.5, 9500},
|
|
{10.0, 10000},
|
|
{10.499, 10499},
|
|
{21.345, 21345},
|
|
{99.999, 99999},
|
|
{1234.0, 100000},
|
|
{HUGE_VAL, 100000},
|
|
{INFINITY, 100000},
|
|
{NAN, 100000}}) {
|
|
ResetPolicy(bpk.first);
|
|
auto bfp = GetBloomFilterPolicy();
|
|
EXPECT_EQ(bpk.second, bfp->GetMillibitsPerKey());
|
|
EXPECT_EQ((bpk.second + 500) / 1000, bfp->GetWholeBitsPerKey());
|
|
|
|
double computed = bpk.first;
|
|
// This transforms e.g. 9.5 -> 9.499999999999998, which we still
|
|
// round to 10 for whole bits per key.
|
|
computed += 0.5;
|
|
computed /= 1234567.0;
|
|
computed *= 1234567.0;
|
|
computed -= 0.5;
|
|
some_computed_less_than_denoted |= (computed < bpk.first);
|
|
ResetPolicy(computed);
|
|
bfp = GetBloomFilterPolicy();
|
|
EXPECT_EQ(bpk.second, bfp->GetMillibitsPerKey());
|
|
EXPECT_EQ((bpk.second + 500) / 1000, bfp->GetWholeBitsPerKey());
|
|
|
|
auto bits_builder = GetBuiltinFilterBitsBuilder();
|
|
|
|
size_t n = 1;
|
|
size_t space = 0;
|
|
for (; n < 1000000; n += 1 + n / 1000) {
|
|
// Ensure consistency between CalculateSpace and ApproximateNumEntries
|
|
space = bits_builder->CalculateSpace(n);
|
|
size_t n2 = bits_builder->ApproximateNumEntries(space);
|
|
EXPECT_GE(n2, n);
|
|
size_t space2 = bits_builder->CalculateSpace(n2);
|
|
if (n > 12000 && GetParam() == BloomFilterPolicy::kStandard128Ribbon) {
|
|
// TODO(peterd): better approximation?
|
|
EXPECT_GE(space2, space);
|
|
EXPECT_LE(space2 * 0.998, space * 1.0);
|
|
} else {
|
|
EXPECT_EQ(space2, space);
|
|
}
|
|
}
|
|
// Until size_t overflow
|
|
for (; n < (n + n / 3); n += n / 3) {
|
|
// Ensure space computation is not overflowing; capped is OK
|
|
size_t space2 = bits_builder->CalculateSpace(n);
|
|
EXPECT_GE(space2, space);
|
|
space = space2;
|
|
}
|
|
}
|
|
// Check that the compiler hasn't optimized our computation into nothing
|
|
EXPECT_TRUE(some_computed_less_than_denoted);
|
|
ResetPolicy();
|
|
}
|
|
|
|
TEST_P(FullBloomTest, FullEmptyFilter) {
|
|
// Empty filter is not match, at this level
|
|
ASSERT_TRUE(!Matches("hello"));
|
|
ASSERT_TRUE(!Matches("world"));
|
|
}
|
|
|
|
TEST_P(FullBloomTest, FullSmall) {
|
|
Add("hello");
|
|
Add("world");
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
ASSERT_TRUE(!Matches("x"));
|
|
ASSERT_TRUE(!Matches("foo"));
|
|
}
|
|
|
|
TEST_P(FullBloomTest, FullVaryingLengths) {
|
|
char buffer[sizeof(int)];
|
|
|
|
// Count number of filters that significantly exceed the false positive rate
|
|
int mediocre_filters = 0;
|
|
int good_filters = 0;
|
|
|
|
for (int length = 1; length <= 10000; length = NextLength(length)) {
|
|
Reset();
|
|
for (int i = 0; i < length; i++) {
|
|
Add(Key(i, buffer));
|
|
}
|
|
Build();
|
|
|
|
EXPECT_LE(FilterSize(), (size_t)((length * FLAGS_bits_per_key / 8) +
|
|
CACHE_LINE_SIZE * 2 + 5));
|
|
|
|
// All added keys must match
|
|
for (int i = 0; i < length; i++) {
|
|
ASSERT_TRUE(Matches(Key(i, buffer)))
|
|
<< "Length " << length << "; key " << i;
|
|
}
|
|
|
|
// Check false positive rate
|
|
double rate = FalsePositiveRate();
|
|
if (kVerbose >= 1) {
|
|
fprintf(stderr, "False positives: %5.2f%% @ length = %6d ; bytes = %6d\n",
|
|
rate*100.0, length, static_cast<int>(FilterSize()));
|
|
}
|
|
if (FLAGS_bits_per_key == 10) {
|
|
EXPECT_LE(rate, 0.02); // Must not be over 2%
|
|
if (rate > 0.0125) {
|
|
mediocre_filters++; // Allowed, but not too often
|
|
} else {
|
|
good_filters++;
|
|
}
|
|
}
|
|
}
|
|
if (kVerbose >= 1) {
|
|
fprintf(stderr, "Filters: %d good, %d mediocre\n",
|
|
good_filters, mediocre_filters);
|
|
}
|
|
EXPECT_LE(mediocre_filters, good_filters / 5);
|
|
}
|
|
|
|
TEST_P(FullBloomTest, OptimizeForMemory) {
|
|
char buffer[sizeof(int)];
|
|
for (bool offm : {true, false}) {
|
|
table_options_.optimize_filters_for_memory = offm;
|
|
ResetPolicy();
|
|
Random32 rnd(12345);
|
|
uint64_t total_size = 0;
|
|
uint64_t total_mem = 0;
|
|
int64_t total_keys = 0;
|
|
double total_fp_rate = 0;
|
|
constexpr int nfilters = 100;
|
|
for (int i = 0; i < nfilters; ++i) {
|
|
int nkeys = static_cast<int>(rnd.Uniformish(10000)) + 100;
|
|
Reset();
|
|
for (int j = 0; j < nkeys; ++j) {
|
|
Add(Key(j, buffer));
|
|
}
|
|
Build();
|
|
size_t size = FilterData().size();
|
|
total_size += size;
|
|
// optimize_filters_for_memory currently depends on malloc_usable_size
|
|
// but we run the rest of the test to ensure no bad behavior without it.
|
|
#ifdef ROCKSDB_MALLOC_USABLE_SIZE
|
|
size = malloc_usable_size(const_cast<char*>(FilterData().data()));
|
|
#endif // ROCKSDB_MALLOC_USABLE_SIZE
|
|
total_mem += size;
|
|
total_keys += nkeys;
|
|
total_fp_rate += FalsePositiveRate();
|
|
}
|
|
if (FLAGS_bits_per_key == 10) {
|
|
EXPECT_LE(total_fp_rate / double{nfilters}, 0.011);
|
|
EXPECT_GE(total_fp_rate / double{nfilters},
|
|
CACHE_LINE_SIZE >= 256 ? 0.007 : 0.008);
|
|
}
|
|
|
|
int64_t ex_min_total_size = int64_t{FLAGS_bits_per_key} * total_keys / 8;
|
|
if (GetParam() == BloomFilterPolicy::kStandard128Ribbon) {
|
|
// ~ 30% savings vs. Bloom filter
|
|
ex_min_total_size = 7 * ex_min_total_size / 10;
|
|
}
|
|
EXPECT_GE(static_cast<int64_t>(total_size), ex_min_total_size);
|
|
|
|
int64_t blocked_bloom_overhead = nfilters * (CACHE_LINE_SIZE + 5);
|
|
if (GetParam() == BloomFilterPolicy::kLegacyBloom) {
|
|
// this config can add extra cache line to make odd number
|
|
blocked_bloom_overhead += nfilters * CACHE_LINE_SIZE;
|
|
}
|
|
|
|
EXPECT_GE(total_mem, total_size);
|
|
|
|
// optimize_filters_for_memory not implemented with legacy Bloom
|
|
if (offm && GetParam() != BloomFilterPolicy::kLegacyBloom) {
|
|
// This value can include a small extra penalty for kExtraPadding
|
|
fprintf(stderr, "Internal fragmentation (optimized): %g%%\n",
|
|
(total_mem - total_size) * 100.0 / total_size);
|
|
// Less than 1% internal fragmentation
|
|
EXPECT_LE(total_mem, total_size * 101 / 100);
|
|
// Up to 2% storage penalty
|
|
EXPECT_LE(static_cast<int64_t>(total_size),
|
|
ex_min_total_size * 102 / 100 + blocked_bloom_overhead);
|
|
} else {
|
|
fprintf(stderr, "Internal fragmentation (not optimized): %g%%\n",
|
|
(total_mem - total_size) * 100.0 / total_size);
|
|
// TODO: add control checks for more allocators?
|
|
#ifdef ROCKSDB_JEMALLOC
|
|
fprintf(stderr, "Jemalloc detected? %d\n", HasJemalloc());
|
|
if (HasJemalloc()) {
|
|
#ifdef ROCKSDB_MALLOC_USABLE_SIZE
|
|
// More than 5% internal fragmentation
|
|
EXPECT_GE(total_mem, total_size * 105 / 100);
|
|
#endif // ROCKSDB_MALLOC_USABLE_SIZE
|
|
}
|
|
#endif // ROCKSDB_JEMALLOC
|
|
// No storage penalty, just usual overhead
|
|
EXPECT_LE(static_cast<int64_t>(total_size),
|
|
ex_min_total_size + blocked_bloom_overhead);
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST(FullBloomFilterConstructionReserveMemTest,
|
|
RibbonFilterFallBackOnLargeBanding) {
|
|
constexpr std::size_t kCacheCapacity =
|
|
8 * CacheReservationManager::GetDummyEntrySize();
|
|
constexpr std::size_t num_entries_for_cache_full = kCacheCapacity / 8;
|
|
|
|
for (bool reserve_builder_mem : {true, false}) {
|
|
bool will_fall_back = reserve_builder_mem;
|
|
|
|
BlockBasedTableOptions table_options;
|
|
table_options.reserve_table_builder_memory = reserve_builder_mem;
|
|
LRUCacheOptions lo;
|
|
lo.capacity = kCacheCapacity;
|
|
lo.num_shard_bits = 0; // 2^0 shard
|
|
lo.strict_capacity_limit = true;
|
|
std::shared_ptr<Cache> cache(NewLRUCache(lo));
|
|
table_options.block_cache = cache;
|
|
table_options.filter_policy.reset(new BloomFilterPolicy(
|
|
FLAGS_bits_per_key, BloomFilterPolicy::Mode::kStandard128Ribbon));
|
|
FilterBuildingContext ctx(table_options);
|
|
std::unique_ptr<FilterBitsBuilder> filter_bits_builder(
|
|
table_options.filter_policy->GetBuilderWithContext(ctx));
|
|
|
|
char key_buffer[sizeof(int)];
|
|
for (std::size_t i = 0; i < num_entries_for_cache_full; ++i) {
|
|
filter_bits_builder->AddKey(Key(static_cast<int>(i), key_buffer));
|
|
}
|
|
|
|
std::unique_ptr<const char[]> buf;
|
|
Slice filter = filter_bits_builder->Finish(&buf);
|
|
|
|
// To verify Ribbon Filter fallbacks to Bloom Filter properly
|
|
// based on cache reservation result
|
|
// See BloomFilterPolicy::GetBloomBitsReader re: metadata
|
|
// -1 = Marker for newer Bloom implementations
|
|
// -2 = Marker for Standard128 Ribbon
|
|
if (will_fall_back) {
|
|
EXPECT_EQ(filter.data()[filter.size() - 5], static_cast<char>(-1));
|
|
} else {
|
|
EXPECT_EQ(filter.data()[filter.size() - 5], static_cast<char>(-2));
|
|
}
|
|
}
|
|
}
|
|
|
|
namespace {
|
|
inline uint32_t SelectByCacheLineSize(uint32_t for64, uint32_t for128,
|
|
uint32_t for256) {
|
|
(void)for64;
|
|
(void)for128;
|
|
(void)for256;
|
|
#if CACHE_LINE_SIZE == 64
|
|
return for64;
|
|
#elif CACHE_LINE_SIZE == 128
|
|
return for128;
|
|
#elif CACHE_LINE_SIZE == 256
|
|
return for256;
|
|
#else
|
|
#error "CACHE_LINE_SIZE unknown or unrecognized"
|
|
#endif
|
|
}
|
|
} // namespace
|
|
|
|
// Ensure the implementation doesn't accidentally change in an
|
|
// incompatible way. This test doesn't check the reading side
|
|
// (FirstFPs/PackedMatches) for LegacyBloom because it requires the
|
|
// ability to read filters generated using other cache line sizes.
|
|
// See RawSchema.
|
|
TEST_P(FullBloomTest, Schema) {
|
|
#define EXPECT_EQ_Bloom(a, b) \
|
|
{ \
|
|
if (GetParam() != BloomFilterPolicy::kStandard128Ribbon) { \
|
|
EXPECT_EQ(a, b); \
|
|
} \
|
|
}
|
|
#define EXPECT_EQ_Ribbon(a, b) \
|
|
{ \
|
|
if (GetParam() == BloomFilterPolicy::kStandard128Ribbon) { \
|
|
EXPECT_EQ(a, b); \
|
|
} \
|
|
}
|
|
#define EXPECT_EQ_FastBloom(a, b) \
|
|
{ \
|
|
if (GetParam() == BloomFilterPolicy::kFastLocalBloom) { \
|
|
EXPECT_EQ(a, b); \
|
|
} \
|
|
}
|
|
#define EXPECT_EQ_LegacyBloom(a, b) \
|
|
{ \
|
|
if (GetParam() == BloomFilterPolicy::kLegacyBloom) { \
|
|
EXPECT_EQ(a, b); \
|
|
} \
|
|
}
|
|
#define EXPECT_EQ_NotLegacy(a, b) \
|
|
{ \
|
|
if (GetParam() != BloomFilterPolicy::kLegacyBloom) { \
|
|
EXPECT_EQ(a, b); \
|
|
} \
|
|
}
|
|
|
|
char buffer[sizeof(int)];
|
|
|
|
// First do a small number of keys, where Ribbon config will fall back on
|
|
// fast Bloom filter and generate the same data
|
|
ResetPolicy(5); // num_probes = 3
|
|
for (int key = 0; key < 87; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ(GetNumProbesFromFilterData(), 3);
|
|
|
|
EXPECT_EQ_NotLegacy(BloomHash(FilterData()), 4130687756U);
|
|
|
|
EXPECT_EQ_NotLegacy("31,38,40,43,61,83,86,112,125,131", FirstFPs(10));
|
|
|
|
// Now use enough keys so that changing bits / key by 1 is guaranteed to
|
|
// change number of allocated cache lines. So keys > max cache line bits.
|
|
|
|
// Note that the first attempted Ribbon seed is determined by the hash
|
|
// of the first key added (for pseudorandomness in practice, determinism in
|
|
// testing)
|
|
|
|
ResetPolicy(2); // num_probes = 1
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 1);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(1567096579, 1964771444, 2659542661U));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 3817481309U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 1705851228U);
|
|
|
|
EXPECT_EQ_FastBloom("11,13,17,25,29,30,35,37,45,53", FirstFPs(10));
|
|
EXPECT_EQ_Ribbon("3,8,10,17,19,20,23,28,31,32", FirstFPs(10));
|
|
|
|
ResetPolicy(3); // num_probes = 2
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 2);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(2707206547U, 2571983456U, 218344685));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 2807269961U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 1095342358U);
|
|
|
|
EXPECT_EQ_FastBloom("4,15,17,24,27,28,29,53,63,70", FirstFPs(10));
|
|
EXPECT_EQ_Ribbon("3,17,20,28,32,33,36,43,49,54", FirstFPs(10));
|
|
|
|
ResetPolicy(5); // num_probes = 3
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 3);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(515748486, 94611728, 2436112214U));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 204628445U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 3971337699U);
|
|
|
|
EXPECT_EQ_FastBloom("15,24,29,39,53,87,89,100,103,104", FirstFPs(10));
|
|
EXPECT_EQ_Ribbon("3,33,36,43,67,70,76,78,84,102", FirstFPs(10));
|
|
|
|
ResetPolicy(8); // num_probes = 5
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 5);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(1302145999, 2811644657U, 756553699));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 355564975U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 3651449053U);
|
|
|
|
EXPECT_EQ_FastBloom("16,60,66,126,220,238,244,256,265,287", FirstFPs(10));
|
|
EXPECT_EQ_Ribbon("33,187,203,296,300,322,411,419,547,582", FirstFPs(10));
|
|
|
|
ResetPolicy(9); // num_probes = 6
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 6);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(2092755149, 661139132, 1182970461));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 2137566013U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 1005676675U);
|
|
|
|
EXPECT_EQ_FastBloom("156,367,791,872,945,1015,1139,1159,1265", FirstFPs(9));
|
|
EXPECT_EQ_Ribbon("33,187,203,296,411,419,604,612,615,619", FirstFPs(10));
|
|
|
|
ResetPolicy(11); // num_probes = 7
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 7);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(3755609649U, 1812694762, 1449142939));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 2561502687U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 3129900846U);
|
|
|
|
EXPECT_EQ_FastBloom("34,74,130,236,643,882,962,1015,1035,1110", FirstFPs(10));
|
|
EXPECT_EQ_Ribbon("411,419,623,665,727,794,955,1052,1323,1330", FirstFPs(10));
|
|
|
|
// This used to be 9 probes, but 8 is a better choice for speed,
|
|
// especially with SIMD groups of 8 probes, with essentially no
|
|
// change in FP rate.
|
|
// FP rate @ 9 probes, old Bloom: 0.4321%
|
|
// FP rate @ 9 probes, new Bloom: 0.1846%
|
|
// FP rate @ 8 probes, new Bloom: 0.1843%
|
|
ResetPolicy(14); // num_probes = 8 (new), 9 (old)
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_LegacyBloom(GetNumProbesFromFilterData(), 9);
|
|
EXPECT_EQ_FastBloom(GetNumProbesFromFilterData(), 8);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(178861123, 379087593, 2574136516U));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 3709876890U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 1855638875U);
|
|
|
|
EXPECT_EQ_FastBloom("130,240,522,565,989,2002,2526,3147,3543", FirstFPs(9));
|
|
EXPECT_EQ_Ribbon("665,727,1323,1755,3866,4232,4442,4492,4736", FirstFPs(9));
|
|
|
|
// This used to be 11 probes, but 9 is a better choice for speed
|
|
// AND accuracy.
|
|
// FP rate @ 11 probes, old Bloom: 0.3571%
|
|
// FP rate @ 11 probes, new Bloom: 0.0884%
|
|
// FP rate @ 9 probes, new Bloom: 0.0843%
|
|
ResetPolicy(16); // num_probes = 9 (new), 11 (old)
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_LegacyBloom(GetNumProbesFromFilterData(), 11);
|
|
EXPECT_EQ_FastBloom(GetNumProbesFromFilterData(), 9);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(1129406313, 3049154394U, 1727750964));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 1087138490U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 459379967U);
|
|
|
|
EXPECT_EQ_FastBloom("3299,3611,3916,6620,7822,8079,8482,8942", FirstFPs(8));
|
|
EXPECT_EQ_Ribbon("727,1323,1755,4442,4736,5386,6974,7154,8222", FirstFPs(9));
|
|
|
|
ResetPolicy(10); // num_probes = 6, but different memory ratio vs. 9
|
|
for (int key = 0; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 6);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 61);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(1478976371, 2910591341U, 1182970461));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 2498541272U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 1273231667U);
|
|
|
|
EXPECT_EQ_FastBloom("16,126,133,422,466,472,813,1002,1035", FirstFPs(9));
|
|
EXPECT_EQ_Ribbon("296,411,419,612,619,623,630,665,686,727", FirstFPs(10));
|
|
|
|
ResetPolicy(10);
|
|
for (int key = /*CHANGED*/ 1; key < 2087; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 6);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), /*CHANGED*/ 184);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(4205696321U, 1132081253U, 2385981855U));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 2058382345U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 3007790572U);
|
|
|
|
EXPECT_EQ_FastBloom("16,126,133,422,466,472,813,1002,1035", FirstFPs(9));
|
|
EXPECT_EQ_Ribbon("33,152,383,497,589,633,737,781,911,990", FirstFPs(10));
|
|
|
|
ResetPolicy(10);
|
|
for (int key = 1; key < /*CHANGED*/ 2088; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 6);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 184);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
SelectByCacheLineSize(2885052954U, 769447944, 4175124908U));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 23699164U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 1942323379U);
|
|
|
|
EXPECT_EQ_FastBloom("16,126,133,422,466,472,813,1002,1035", FirstFPs(9));
|
|
EXPECT_EQ_Ribbon("33,95,360,589,737,911,990,1048,1081,1414", FirstFPs(10));
|
|
|
|
// With new fractional bits_per_key, check that we are rounding to
|
|
// whole bits per key for old Bloom filters but fractional for
|
|
// new Bloom filter.
|
|
ResetPolicy(9.5);
|
|
for (int key = 1; key < 2088; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_Bloom(GetNumProbesFromFilterData(), 6);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 184);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
/*SAME*/ SelectByCacheLineSize(2885052954U, 769447944, 4175124908U));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 3166884174U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 1148258663U);
|
|
|
|
EXPECT_EQ_FastBloom("126,156,367,444,458,791,813,976,1015", FirstFPs(9));
|
|
EXPECT_EQ_Ribbon("33,54,95,360,589,693,737,911,990,1048", FirstFPs(10));
|
|
|
|
ResetPolicy(10.499);
|
|
for (int key = 1; key < 2088; key++) {
|
|
Add(Key(key, buffer));
|
|
}
|
|
Build();
|
|
EXPECT_EQ_LegacyBloom(GetNumProbesFromFilterData(), 6);
|
|
EXPECT_EQ_FastBloom(GetNumProbesFromFilterData(), 7);
|
|
EXPECT_EQ_Ribbon(GetRibbonSeedFromFilterData(), 184);
|
|
|
|
EXPECT_EQ_LegacyBloom(
|
|
BloomHash(FilterData()),
|
|
/*SAME*/ SelectByCacheLineSize(2885052954U, 769447944, 4175124908U));
|
|
EXPECT_EQ_FastBloom(BloomHash(FilterData()), 4098502778U);
|
|
EXPECT_EQ_Ribbon(BloomHash(FilterData()), 792138188U);
|
|
|
|
EXPECT_EQ_FastBloom("16,236,240,472,1015,1045,1111,1409,1465", FirstFPs(9));
|
|
EXPECT_EQ_Ribbon("33,95,360,589,737,990,1048,1081,1414,1643", FirstFPs(10));
|
|
|
|
ResetPolicy();
|
|
}
|
|
|
|
// A helper class for testing custom or corrupt filter bits as read by
|
|
// built-in FilterBitsReaders.
|
|
struct RawFilterTester {
|
|
// Buffer, from which we always return a tail Slice, so the
|
|
// last five bytes are always the metadata bytes.
|
|
std::array<char, 3000> data_;
|
|
// Points five bytes from the end
|
|
char* metadata_ptr_;
|
|
|
|
RawFilterTester() : metadata_ptr_(&*(data_.end() - 5)) {}
|
|
|
|
Slice ResetNoFill(uint32_t len_without_metadata, uint32_t num_lines,
|
|
uint32_t num_probes) {
|
|
metadata_ptr_[0] = static_cast<char>(num_probes);
|
|
EncodeFixed32(metadata_ptr_ + 1, num_lines);
|
|
uint32_t len = len_without_metadata + /*metadata*/ 5;
|
|
assert(len <= data_.size());
|
|
return Slice(metadata_ptr_ - len_without_metadata, len);
|
|
}
|
|
|
|
Slice Reset(uint32_t len_without_metadata, uint32_t num_lines,
|
|
uint32_t num_probes, bool fill_ones) {
|
|
data_.fill(fill_ones ? 0xff : 0);
|
|
return ResetNoFill(len_without_metadata, num_lines, num_probes);
|
|
}
|
|
|
|
Slice ResetWeirdFill(uint32_t len_without_metadata, uint32_t num_lines,
|
|
uint32_t num_probes) {
|
|
for (uint32_t i = 0; i < data_.size(); ++i) {
|
|
data_[i] = static_cast<char>(0x7b7b >> (i % 7));
|
|
}
|
|
return ResetNoFill(len_without_metadata, num_lines, num_probes);
|
|
}
|
|
};
|
|
|
|
TEST_P(FullBloomTest, RawSchema) {
|
|
RawFilterTester cft;
|
|
// Legacy Bloom configurations
|
|
// Two probes, about 3/4 bits set: ~50% "FP" rate
|
|
// One 256-byte cache line.
|
|
OpenRaw(cft.ResetWeirdFill(256, 1, 2));
|
|
EXPECT_EQ(uint64_t{11384799501900898790U}, PackedMatches());
|
|
|
|
// Two 128-byte cache lines.
|
|
OpenRaw(cft.ResetWeirdFill(256, 2, 2));
|
|
EXPECT_EQ(uint64_t{10157853359773492589U}, PackedMatches());
|
|
|
|
// Four 64-byte cache lines.
|
|
OpenRaw(cft.ResetWeirdFill(256, 4, 2));
|
|
EXPECT_EQ(uint64_t{7123594913907464682U}, PackedMatches());
|
|
|
|
// Fast local Bloom configurations (marker 255 -> -1)
|
|
// Two probes, about 3/4 bits set: ~50% "FP" rate
|
|
// Four 64-byte cache lines.
|
|
OpenRaw(cft.ResetWeirdFill(256, 2U << 8, 255));
|
|
EXPECT_EQ(uint64_t{9957045189927952471U}, PackedMatches());
|
|
|
|
// Ribbon configurations (marker 254 -> -2)
|
|
|
|
// Even though the builder never builds configurations this
|
|
// small (preferring Bloom), we can test that the configuration
|
|
// can be read, for possible future-proofing.
|
|
|
|
// 256 slots, one result column = 32 bytes (2 blocks, seed 0)
|
|
// ~50% FP rate:
|
|
// 0b0101010111110101010000110000011011011111100100001110010011101010
|
|
OpenRaw(cft.ResetWeirdFill(32, 2U << 8, 254));
|
|
EXPECT_EQ(uint64_t{6193930559317665002U}, PackedMatches());
|
|
|
|
// 256 slots, three-to-four result columns = 112 bytes
|
|
// ~ 1 in 10 FP rate:
|
|
// 0b0000000000100000000000000000000001000001000000010000101000000000
|
|
OpenRaw(cft.ResetWeirdFill(112, 2U << 8, 254));
|
|
EXPECT_EQ(uint64_t{9007200345328128U}, PackedMatches());
|
|
}
|
|
|
|
TEST_P(FullBloomTest, CorruptFilters) {
|
|
RawFilterTester cft;
|
|
|
|
for (bool fill : {false, true}) {
|
|
// Legacy Bloom configurations
|
|
// Good filter bits - returns same as fill
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 1, 6, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Good filter bits - returns same as fill
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE * 3, 3, 6, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Good filter bits - returns same as fill
|
|
// 256 is unusual but legal cache line size
|
|
OpenRaw(cft.Reset(256 * 3, 3, 6, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Good filter bits - returns same as fill
|
|
// 30 should be max num_probes
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 1, 30, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Good filter bits - returns same as fill
|
|
// 1 should be min num_probes
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 1, 1, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Type 1 trivial filter bits - returns true as if FP by zero probes
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 1, 0, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// Type 2 trivial filter bits - returns false as if built from zero keys
|
|
OpenRaw(cft.Reset(0, 0, 6, fill));
|
|
ASSERT_FALSE(Matches("hello"));
|
|
ASSERT_FALSE(Matches("world"));
|
|
|
|
// Type 2 trivial filter bits - returns false as if built from zero keys
|
|
OpenRaw(cft.Reset(0, 37, 6, fill));
|
|
ASSERT_FALSE(Matches("hello"));
|
|
ASSERT_FALSE(Matches("world"));
|
|
|
|
// Type 2 trivial filter bits - returns false as 0 size trumps 0 probes
|
|
OpenRaw(cft.Reset(0, 0, 0, fill));
|
|
ASSERT_FALSE(Matches("hello"));
|
|
ASSERT_FALSE(Matches("world"));
|
|
|
|
// Bad filter bits - returns true for safety
|
|
// No solution to 0 * x == CACHE_LINE_SIZE
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 0, 6, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// Bad filter bits - returns true for safety
|
|
// Can't have 3 * x == 4 for integer x
|
|
OpenRaw(cft.Reset(4, 3, 6, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// Bad filter bits - returns true for safety
|
|
// 97 bytes is not a power of two, so not a legal cache line size
|
|
OpenRaw(cft.Reset(97 * 3, 3, 6, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// Bad filter bits - returns true for safety
|
|
// 65 bytes is not a power of two, so not a legal cache line size
|
|
OpenRaw(cft.Reset(65 * 3, 3, 6, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// Bad filter bits - returns false as if built from zero keys
|
|
// < 5 bytes overall means missing even metadata
|
|
OpenRaw(cft.Reset(static_cast<uint32_t>(-1), 3, 6, fill));
|
|
ASSERT_FALSE(Matches("hello"));
|
|
ASSERT_FALSE(Matches("world"));
|
|
|
|
OpenRaw(cft.Reset(static_cast<uint32_t>(-5), 3, 6, fill));
|
|
ASSERT_FALSE(Matches("hello"));
|
|
ASSERT_FALSE(Matches("world"));
|
|
|
|
// Dubious filter bits - returns same as fill (for now)
|
|
// 31 is not a useful num_probes, nor generated by RocksDB unless directly
|
|
// using filter bits API without BloomFilterPolicy.
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 1, 31, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Dubious filter bits - returns same as fill (for now)
|
|
// Similar, with 127, largest positive char
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 1, 127, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Dubious filter bits - returns true (for now)
|
|
// num_probes set to 128 / -128, lowest negative char
|
|
// NB: Bug in implementation interprets this as negative and has same
|
|
// effect as zero probes, but effectively reserves negative char values
|
|
// for future use.
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 1, 128, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// Dubious filter bits - returns true (for now)
|
|
// Similar, with 253 / -3
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 1, 253, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// #########################################################
|
|
// Fast local Bloom configurations (marker 255 -> -1)
|
|
// Good config with six probes
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 6U << 8, 255, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Becomes bad/reserved config (always true) if any other byte set
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, (6U << 8) | 1U, 255, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, (6U << 8) | (1U << 16), 255, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, (6U << 8) | (1U << 24), 255, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// Good config, max 30 probes
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 30U << 8, 255, fill));
|
|
ASSERT_EQ(fill, Matches("hello"));
|
|
ASSERT_EQ(fill, Matches("world"));
|
|
|
|
// Bad/reserved config (always true) if more than 30
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 31U << 8, 255, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 33U << 8, 255, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 66U << 8, 255, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
OpenRaw(cft.Reset(CACHE_LINE_SIZE, 130U << 8, 255, fill));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
}
|
|
|
|
// #########################################################
|
|
// Ribbon configurations (marker 254 -> -2)
|
|
// ("fill" doesn't work to detect good configurations, we just
|
|
// have to rely on TN probability)
|
|
|
|
// Good: 2 blocks * 16 bytes / segment * 4 columns = 128 bytes
|
|
// seed = 123
|
|
OpenRaw(cft.Reset(128, (2U << 8) + 123U, 254, false));
|
|
ASSERT_FALSE(Matches("hello"));
|
|
ASSERT_FALSE(Matches("world"));
|
|
|
|
// Good: 2 blocks * 16 bytes / segment * 8 columns = 256 bytes
|
|
OpenRaw(cft.Reset(256, (2U << 8) + 123U, 254, false));
|
|
ASSERT_FALSE(Matches("hello"));
|
|
ASSERT_FALSE(Matches("world"));
|
|
|
|
// Surprisingly OK: 5000 blocks (640,000 slots) in only 1024 bits
|
|
// -> average close to 0 columns
|
|
OpenRaw(cft.Reset(128, (5000U << 8) + 123U, 254, false));
|
|
// *Almost* all FPs
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
// Need many queries to find a "true negative"
|
|
for (int i = 0; Matches(ToString(i)); ++i) {
|
|
ASSERT_LT(i, 1000);
|
|
}
|
|
|
|
// Bad: 1 block not allowed (for implementation detail reasons)
|
|
OpenRaw(cft.Reset(128, (1U << 8) + 123U, 254, false));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
|
|
// Bad: 0 blocks not allowed
|
|
OpenRaw(cft.Reset(128, (0U << 8) + 123U, 254, false));
|
|
ASSERT_TRUE(Matches("hello"));
|
|
ASSERT_TRUE(Matches("world"));
|
|
}
|
|
|
|
INSTANTIATE_TEST_CASE_P(Full, FullBloomTest,
|
|
testing::Values(BloomFilterPolicy::kLegacyBloom,
|
|
BloomFilterPolicy::kFastLocalBloom,
|
|
BloomFilterPolicy::kStandard128Ribbon));
|
|
|
|
static double GetEffectiveBitsPerKey(FilterBitsBuilder* builder) {
|
|
union {
|
|
uint64_t key_value;
|
|
char key_bytes[8];
|
|
};
|
|
|
|
const unsigned kNumKeys = 1000;
|
|
|
|
Slice key_slice{key_bytes, 8};
|
|
for (key_value = 0; key_value < kNumKeys; ++key_value) {
|
|
builder->AddKey(key_slice);
|
|
}
|
|
|
|
std::unique_ptr<const char[]> buf;
|
|
auto filter = builder->Finish(&buf);
|
|
return filter.size() * /*bits per byte*/ 8 / (1.0 * kNumKeys);
|
|
}
|
|
|
|
static void SetTestingLevel(int levelish, FilterBuildingContext* ctx) {
|
|
if (levelish == -1) {
|
|
// Flush is treated as level -1 for this option but actually level 0
|
|
ctx->level_at_creation = 0;
|
|
ctx->reason = TableFileCreationReason::kFlush;
|
|
} else {
|
|
ctx->level_at_creation = levelish;
|
|
ctx->reason = TableFileCreationReason::kCompaction;
|
|
}
|
|
}
|
|
|
|
TEST(RibbonTest, RibbonTestLevelThreshold) {
|
|
BlockBasedTableOptions opts;
|
|
FilterBuildingContext ctx(opts);
|
|
// A few settings
|
|
for (CompactionStyle cs : {kCompactionStyleLevel, kCompactionStyleUniversal,
|
|
kCompactionStyleFIFO, kCompactionStyleNone}) {
|
|
ctx.compaction_style = cs;
|
|
for (int bloom_before_level : {-1, 0, 1, 10}) {
|
|
std::vector<std::unique_ptr<const FilterPolicy> > policies;
|
|
policies.emplace_back(NewRibbonFilterPolicy(10, bloom_before_level));
|
|
|
|
if (bloom_before_level == -1) {
|
|
// Also test old API
|
|
policies.emplace_back(NewExperimentalRibbonFilterPolicy(10));
|
|
}
|
|
|
|
if (bloom_before_level == 0) {
|
|
// Also test old API and new API default
|
|
policies.emplace_back(NewRibbonFilterPolicy(10));
|
|
}
|
|
|
|
for (std::unique_ptr<const FilterPolicy>& policy : policies) {
|
|
// Claim to be generating filter for this level
|
|
SetTestingLevel(bloom_before_level, &ctx);
|
|
|
|
std::unique_ptr<FilterBitsBuilder> builder{
|
|
policy->GetBuilderWithContext(ctx)};
|
|
|
|
// Must be Ribbon (more space efficient than 10 bits per key)
|
|
ASSERT_LT(GetEffectiveBitsPerKey(builder.get()), 8);
|
|
|
|
if (bloom_before_level >= 0) {
|
|
// Claim to be generating filter for previous level
|
|
SetTestingLevel(bloom_before_level - 1, &ctx);
|
|
|
|
builder.reset(policy->GetBuilderWithContext(ctx));
|
|
|
|
if (cs == kCompactionStyleLevel || cs == kCompactionStyleUniversal) {
|
|
// Level is considered.
|
|
// Must be Bloom (~ 10 bits per key)
|
|
ASSERT_GT(GetEffectiveBitsPerKey(builder.get()), 9);
|
|
} else {
|
|
// Level is ignored under non-traditional compaction styles.
|
|
// Must be Ribbon (more space efficient than 10 bits per key)
|
|
ASSERT_LT(GetEffectiveBitsPerKey(builder.get()), 8);
|
|
}
|
|
}
|
|
|
|
// Like SST file writer
|
|
ctx.level_at_creation = -1;
|
|
ctx.reason = TableFileCreationReason::kMisc;
|
|
|
|
builder.reset(policy->GetBuilderWithContext(ctx));
|
|
|
|
// Must be Ribbon (more space efficient than 10 bits per key)
|
|
ASSERT_LT(GetEffectiveBitsPerKey(builder.get()), 8);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
|
|
int main(int argc, char** argv) {
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
ParseCommandLineFlags(&argc, &argv, true);
|
|
|
|
return RUN_ALL_TESTS();
|
|
}
|
|
|
|
#endif // GFLAGS
|