5efa0d6b0d
Summary: BlockCacheLookupContext only contains the caller for now. We will trace block accesses at five places: 1. BlockBasedTable::GetFilter. 2. BlockBasedTable::GetUncompressedDict. 3. BlockBasedTable::MaybeReadAndLoadToCache. (To trace access on data, index, and range deletion block.) 4. BlockBasedTable::Get. (To trace the referenced key and whether the referenced key exists in a fetched data block.) 5. BlockBasedTable::MultiGet. (To trace the referenced key and whether the referenced key exists in a fetched data block.) We create the context at: 1. BlockBasedTable::Get. (kUserGet) 2. BlockBasedTable::MultiGet. (kUserMGet) 3. BlockBasedTable::NewIterator. (either kUserIterator, kCompaction, or external SST ingestion calls this function.) 4. BlockBasedTable::Open. (kPrefetch) 5. Index/Filter::CacheDependencies. (kPrefetch) 6. BlockBasedTable::ApproximateOffsetOf. (kCompaction or kUserApproximateSize). I loaded 1 million key-value pairs into the database and ran the readrandom benchmark with a single thread. I gave the block cache 10 GB to make sure all reads hit the block cache after warmup. The throughput is comparable. Throughput of this PR: 231334 ops/s. Throughput of the master branch: 238428 ops/s. Experiment setup: RocksDB: version 6.2 Date: Mon Jun 10 10:42:51 2019 CPU: 24 * Intel Core Processor (Skylake) CPUCache: 16384 KB Keys: 20 bytes each Values: 100 bytes each (100 bytes after compression) Entries: 1000000 Prefix: 20 bytes Keys per prefix: 0 RawSize: 114.4 MB (estimated) FileSize: 114.4 MB (estimated) Write rate: 0 bytes/second Read rate: 0 ops/second Compression: NoCompression Compression sampling rate: 0 Memtablerep: skip_list Perf Level: 1 Load command: ./db_bench --benchmarks="fillseq" --key_size=20 --prefix_size=20 --keys_per_prefix=0 --value_size=100 --statistics --cache_index_and_filter_blocks --cache_size=10737418240 --disable_auto_compactions=1 --disable_wal=1 --compression_type=none --min_level_to_compress=-1 --compression_ratio=1 --num=1000000 Run command: ./db_bench --benchmarks="readrandom,stats" --use_existing_db --threads=1 --duration=120 --key_size=20 --prefix_size=20 --keys_per_prefix=0 --value_size=100 --statistics --cache_index_and_filter_blocks --cache_size=10737418240 --disable_auto_compactions=1 --disable_wal=1 --compression_type=none --min_level_to_compress=-1 --compression_ratio=1 --num=1000000 --duration=120 TODOs: 1. Create a caller for external SST file ingestion and differentiate the callers for iterator. 2. Integrate tracer to trace block cache accesses. Pull Request resolved: https://github.com/facebook/rocksdb/pull/5421 Differential Revision: D15704258 Pulled By: HaoyuHuang fbshipit-source-id: 4aa8a55f8cb1576ffb367bfa3186a91d8f06d93a
92 lines
3.0 KiB
C++
92 lines
3.0 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
#pragma once
|
|
#ifndef ROCKSDB_LITE
|
|
#include <string>
|
|
#include <memory>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
#include "db/dbformat.h"
|
|
#include "options/cf_options.h"
|
|
#include "rocksdb/env.h"
|
|
#include "rocksdb/options.h"
|
|
#include "table/table_reader.h"
|
|
#include "util/file_reader_writer.h"
|
|
|
|
namespace rocksdb {
|
|
|
|
class Arena;
|
|
class TableReader;
|
|
|
|
class CuckooTableReader: public TableReader {
|
|
public:
|
|
CuckooTableReader(const ImmutableCFOptions& ioptions,
|
|
std::unique_ptr<RandomAccessFileReader>&& file,
|
|
uint64_t file_size, const Comparator* user_comparator,
|
|
uint64_t (*get_slice_hash)(const Slice&, uint32_t,
|
|
uint64_t));
|
|
~CuckooTableReader() {}
|
|
|
|
std::shared_ptr<const TableProperties> GetTableProperties() const override {
|
|
return table_props_;
|
|
}
|
|
|
|
Status status() const { return status_; }
|
|
|
|
Status Get(const ReadOptions& readOptions, const Slice& key,
|
|
GetContext* get_context, const SliceTransform* prefix_extractor,
|
|
bool skip_filters = false) override;
|
|
|
|
InternalIterator* NewIterator(const ReadOptions&,
|
|
const SliceTransform* prefix_extractor,
|
|
Arena* arena = nullptr,
|
|
bool skip_filters = false,
|
|
bool for_compaction = false) override;
|
|
void Prepare(const Slice& target) override;
|
|
|
|
// Report an approximation of how much memory has been used.
|
|
size_t ApproximateMemoryUsage() const override;
|
|
|
|
// Following methods are not implemented for Cuckoo Table Reader
|
|
uint64_t ApproximateOffsetOf(const Slice& /*key*/,
|
|
bool /*for_compaction*/ = false) override {
|
|
return 0;
|
|
}
|
|
void SetupForCompaction() override {}
|
|
// End of methods not implemented.
|
|
|
|
private:
|
|
friend class CuckooTableIterator;
|
|
void LoadAllKeys(std::vector<std::pair<Slice, uint32_t>>* key_to_bucket_id);
|
|
std::unique_ptr<RandomAccessFileReader> file_;
|
|
Slice file_data_;
|
|
bool is_last_level_;
|
|
bool identity_as_first_hash_;
|
|
bool use_module_hash_;
|
|
std::shared_ptr<const TableProperties> table_props_;
|
|
Status status_;
|
|
uint32_t num_hash_func_;
|
|
std::string unused_key_;
|
|
uint32_t key_length_;
|
|
uint32_t user_key_length_;
|
|
uint32_t value_length_;
|
|
uint32_t bucket_length_;
|
|
uint32_t cuckoo_block_size_;
|
|
uint32_t cuckoo_block_bytes_minus_one_;
|
|
uint64_t table_size_;
|
|
const Comparator* ucomp_;
|
|
uint64_t (*get_slice_hash_)(const Slice& s, uint32_t index,
|
|
uint64_t max_num_buckets);
|
|
};
|
|
|
|
} // namespace rocksdb
|
|
#endif // ROCKSDB_LITE
|