6d37fdb365
Summary: `BlockBasedTableBuilder::Add()` eventually calls `DataBlockHashIndexBuilder::EstimateSize()`. The previous implementation divides the `num_keys` by the `util_ratio_` to get an estimizted `num_buckets`. Such division is expensive as it happens in every `BlockBasedTableBuilder::Add()`. This diff estimates the `num_buckets` by double addition instead of double division. Specifically, in each `Add()`, we add `bucket_per_key_` (inverse of `util_ratio_`) to the current `estimiated_num_buckets_`. The cost is that we are gonna have the `estimated_num_buckets_` stored as one extra field in the DataBlockHashIndexBuilder. Pull Request resolved: https://github.com/facebook/rocksdb/pull/4293 Differential Revision: D9412472 Pulled By: fgwu fbshipit-source-id: 2925c8509a401e7bd3c1ab1d9e9c7244755c277a
94 lines
3.1 KiB
C++
94 lines
3.1 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
#include <string>
|
|
#include <vector>
|
|
|
|
#include "rocksdb/slice.h"
|
|
#include "table/data_block_hash_index.h"
|
|
#include "util/coding.h"
|
|
#include "util/hash.h"
|
|
|
|
namespace rocksdb {
|
|
|
|
void DataBlockHashIndexBuilder::Add(const Slice& key,
|
|
const size_t restart_index) {
|
|
assert(Valid());
|
|
if (restart_index > kMaxRestartSupportedByHashIndex) {
|
|
valid_ = false;
|
|
return;
|
|
}
|
|
|
|
uint32_t hash_value = GetSliceHash(key);
|
|
hash_and_restart_pairs_.emplace_back(hash_value,
|
|
static_cast<uint8_t>(restart_index));
|
|
estimated_num_buckets_ += bucket_per_key_;
|
|
}
|
|
|
|
void DataBlockHashIndexBuilder::Finish(std::string& buffer) {
|
|
assert(Valid());
|
|
uint16_t num_buckets = static_cast<uint16_t>(estimated_num_buckets_);
|
|
|
|
if (num_buckets == 0) {
|
|
num_buckets = 1; // sanity check
|
|
}
|
|
|
|
// The build-in hash cannot well distribute strings when into different
|
|
// buckets when num_buckets is power of two, resulting in high hash
|
|
// collision.
|
|
// We made the num_buckets to be odd to avoid this issue.
|
|
num_buckets |= 1;
|
|
|
|
std::vector<uint8_t> buckets(num_buckets, kNoEntry);
|
|
// write the restart_index array
|
|
for (auto& entry : hash_and_restart_pairs_) {
|
|
uint32_t hash_value = entry.first;
|
|
uint8_t restart_index = entry.second;
|
|
uint16_t buck_idx = static_cast<uint16_t>(hash_value % num_buckets);
|
|
if (buckets[buck_idx] == kNoEntry) {
|
|
buckets[buck_idx] = restart_index;
|
|
} else if (buckets[buck_idx] != restart_index) {
|
|
// same bucket cannot store two different restart_index, mark collision
|
|
buckets[buck_idx] = kCollision;
|
|
}
|
|
}
|
|
|
|
for (uint8_t restart_index : buckets) {
|
|
buffer.append(
|
|
const_cast<const char*>(reinterpret_cast<char*>(&restart_index)),
|
|
sizeof(restart_index));
|
|
}
|
|
|
|
// write NUM_BUCK
|
|
PutFixed16(&buffer, num_buckets);
|
|
|
|
assert(buffer.size() <= kMaxBlockSizeSupportedByHashIndex);
|
|
}
|
|
|
|
void DataBlockHashIndexBuilder::Reset() {
|
|
estimated_num_buckets_ = 0;
|
|
valid_ = true;
|
|
hash_and_restart_pairs_.clear();
|
|
}
|
|
|
|
void DataBlockHashIndex::Initialize(const char* data, uint16_t size,
|
|
uint16_t* map_offset) {
|
|
assert(size >= sizeof(uint16_t)); // NUM_BUCKETS
|
|
num_buckets_ = DecodeFixed16(data + size - sizeof(uint16_t));
|
|
assert(num_buckets_ > 0);
|
|
assert(size > num_buckets_ * sizeof(uint8_t));
|
|
*map_offset = static_cast<uint16_t>(size - sizeof(uint16_t) -
|
|
num_buckets_ * sizeof(uint8_t));
|
|
}
|
|
|
|
uint8_t DataBlockHashIndex::Lookup(const char* data, uint32_t map_offset,
|
|
const Slice& key) const {
|
|
uint32_t hash_value = GetSliceHash(key);
|
|
uint16_t idx = static_cast<uint16_t>(hash_value % num_buckets_);
|
|
const char* bucket_table = data + map_offset;
|
|
return static_cast<uint8_t>(*(bucket_table + idx * sizeof(uint8_t)));
|
|
}
|
|
|
|
} // namespace rocksdb
|