22655a398b
Summary: The patch adds a class called `BlobFileReader` that can be used to retrieve blobs using the information available in blob references (e.g. blob file number, offset, and size). This will come in handy when implementing blob support for `Get`, `MultiGet`, and iterators, and also for compaction/garbage collection. When a `BlobFileReader` object is created (using the factory method `Create`), it first checks whether the specified file is potentially valid by comparing the file size against the combined size of the blob file header and footer (files smaller than the threshold are considered malformed). Then, it opens the file, and reads and verifies the header and footer. The verification involves magic number/CRC checks as well as checking for unexpected header/footer fields, e.g. incorrect column family ID or TTL blob files. Blobs can be retrieved using `GetBlob`. `GetBlob` validates the offset and compression type passed by the caller (because of the presence of the header and footer, the specified offset cannot be too close to the start/end of the file; also, the compression type has to match the one in the blob file header), and retrieves and potentially verifies and uncompresses the blob. In particular, when `ReadOptions::verify_checksums` is set, `BlobFileReader` reads the blob record header as well (as opposed to just the blob itself) and verifies the key/value size, the key itself, as well as the CRC of the blob record header and the key/value pair. In addition, the patch exposes the compression type from `BlobIndex` (both using an accessor and via `DebugString`), and adds a blob file read latency histogram to `InternalStats` that can be used with `BlobFileReader`. Pull Request resolved: https://github.com/facebook/rocksdb/pull/7461 Test Plan: `make check` Reviewed By: riversand963 Differential Revision: D23999219 Pulled By: ltamasi fbshipit-source-id: deb6b1160d251258b308d5156e2ec063c3e12e5e
704 lines
25 KiB
C++
704 lines
25 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
//
|
|
|
|
#pragma once
|
|
#include <map>
|
|
#include <string>
|
|
#include <vector>
|
|
|
|
#include "db/version_set.h"
|
|
|
|
class ColumnFamilyData;
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
class DBImpl;
|
|
class MemTableList;
|
|
|
|
// Config for retrieving a property's value.
|
|
struct DBPropertyInfo {
|
|
bool need_out_of_mutex;
|
|
|
|
// gcc had an internal error for initializing union of pointer-to-member-
|
|
// functions. Workaround is to populate exactly one of the following function
|
|
// pointers with a non-nullptr value.
|
|
|
|
// @param value Value-result argument for storing the property's string value
|
|
// @param suffix Argument portion of the property. For example, suffix would
|
|
// be "5" for the property "rocksdb.num-files-at-level5". So far, only
|
|
// certain string properties take an argument.
|
|
bool (InternalStats::*handle_string)(std::string* value, Slice suffix);
|
|
|
|
// @param value Value-result argument for storing the property's uint64 value
|
|
// @param db Many of the int properties rely on DBImpl methods.
|
|
// @param version Version is needed in case the property is retrieved without
|
|
// holding db mutex, which is only supported for int properties.
|
|
bool (InternalStats::*handle_int)(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
|
|
// @param props Map of general properties to populate
|
|
bool (InternalStats::*handle_map)(std::map<std::string, std::string>* props);
|
|
|
|
// handle the string type properties rely on DBImpl methods
|
|
// @param value Value-result argument for storing the property's string value
|
|
bool (DBImpl::*handle_string_dbimpl)(std::string* value);
|
|
};
|
|
|
|
extern const DBPropertyInfo* GetPropertyInfo(const Slice& property);
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
#undef SCORE
|
|
enum class LevelStatType {
|
|
INVALID = 0,
|
|
NUM_FILES,
|
|
COMPACTED_FILES,
|
|
SIZE_BYTES,
|
|
SCORE,
|
|
READ_GB,
|
|
RN_GB,
|
|
RNP1_GB,
|
|
WRITE_GB,
|
|
W_NEW_GB,
|
|
MOVED_GB,
|
|
WRITE_AMP,
|
|
READ_MBPS,
|
|
WRITE_MBPS,
|
|
COMP_SEC,
|
|
COMP_CPU_SEC,
|
|
COMP_COUNT,
|
|
AVG_SEC,
|
|
KEY_IN,
|
|
KEY_DROP,
|
|
TOTAL // total number of types
|
|
};
|
|
|
|
struct LevelStat {
|
|
// This what will be L?.property_name in the flat map returned to the user
|
|
std::string property_name;
|
|
// This will be what we will print in the header in the cli
|
|
std::string header_name;
|
|
};
|
|
|
|
class InternalStats {
|
|
public:
|
|
static const std::map<LevelStatType, LevelStat> compaction_level_stats;
|
|
|
|
enum InternalCFStatsType {
|
|
L0_FILE_COUNT_LIMIT_SLOWDOWNS,
|
|
LOCKED_L0_FILE_COUNT_LIMIT_SLOWDOWNS,
|
|
MEMTABLE_LIMIT_STOPS,
|
|
MEMTABLE_LIMIT_SLOWDOWNS,
|
|
L0_FILE_COUNT_LIMIT_STOPS,
|
|
LOCKED_L0_FILE_COUNT_LIMIT_STOPS,
|
|
PENDING_COMPACTION_BYTES_LIMIT_SLOWDOWNS,
|
|
PENDING_COMPACTION_BYTES_LIMIT_STOPS,
|
|
WRITE_STALLS_ENUM_MAX,
|
|
BYTES_FLUSHED,
|
|
BYTES_INGESTED_ADD_FILE,
|
|
INGESTED_NUM_FILES_TOTAL,
|
|
INGESTED_LEVEL0_NUM_FILES_TOTAL,
|
|
INGESTED_NUM_KEYS_TOTAL,
|
|
INTERNAL_CF_STATS_ENUM_MAX,
|
|
};
|
|
|
|
enum InternalDBStatsType {
|
|
kIntStatsWalFileBytes,
|
|
kIntStatsWalFileSynced,
|
|
kIntStatsBytesWritten,
|
|
kIntStatsNumKeysWritten,
|
|
kIntStatsWriteDoneByOther,
|
|
kIntStatsWriteDoneBySelf,
|
|
kIntStatsWriteWithWal,
|
|
kIntStatsWriteStallMicros,
|
|
kIntStatsNumMax,
|
|
};
|
|
|
|
InternalStats(int num_levels, Env* env, ColumnFamilyData* cfd)
|
|
: db_stats_{},
|
|
cf_stats_value_{},
|
|
cf_stats_count_{},
|
|
comp_stats_(num_levels),
|
|
comp_stats_by_pri_(Env::Priority::TOTAL),
|
|
file_read_latency_(num_levels),
|
|
bg_error_count_(0),
|
|
number_levels_(num_levels),
|
|
env_(env),
|
|
cfd_(cfd),
|
|
started_at_(env->NowMicros()) {}
|
|
|
|
// Per level compaction stats. comp_stats_[level] stores the stats for
|
|
// compactions that produced data for the specified "level".
|
|
struct CompactionStats {
|
|
uint64_t micros;
|
|
uint64_t cpu_micros;
|
|
|
|
// The number of bytes read from all non-output levels
|
|
uint64_t bytes_read_non_output_levels;
|
|
|
|
// The number of bytes read from the compaction output level.
|
|
uint64_t bytes_read_output_level;
|
|
|
|
// Total number of bytes written during compaction
|
|
uint64_t bytes_written;
|
|
|
|
// Total number of bytes moved to the output level
|
|
uint64_t bytes_moved;
|
|
|
|
// The number of compaction input files in all non-output levels.
|
|
int num_input_files_in_non_output_levels;
|
|
|
|
// The number of compaction input files in the output level.
|
|
int num_input_files_in_output_level;
|
|
|
|
// The number of compaction output files.
|
|
int num_output_files;
|
|
|
|
// Total incoming entries during compaction between levels N and N+1
|
|
uint64_t num_input_records;
|
|
|
|
// Accumulated diff number of entries
|
|
// (num input entries - num output entires) for compaction levels N and N+1
|
|
uint64_t num_dropped_records;
|
|
|
|
// Number of compactions done
|
|
int count;
|
|
|
|
// Number of compactions done per CompactionReason
|
|
int counts[static_cast<int>(CompactionReason::kNumOfReasons)];
|
|
|
|
explicit CompactionStats()
|
|
: micros(0),
|
|
cpu_micros(0),
|
|
bytes_read_non_output_levels(0),
|
|
bytes_read_output_level(0),
|
|
bytes_written(0),
|
|
bytes_moved(0),
|
|
num_input_files_in_non_output_levels(0),
|
|
num_input_files_in_output_level(0),
|
|
num_output_files(0),
|
|
num_input_records(0),
|
|
num_dropped_records(0),
|
|
count(0) {
|
|
int num_of_reasons = static_cast<int>(CompactionReason::kNumOfReasons);
|
|
for (int i = 0; i < num_of_reasons; i++) {
|
|
counts[i] = 0;
|
|
}
|
|
}
|
|
|
|
explicit CompactionStats(CompactionReason reason, int c)
|
|
: micros(0),
|
|
cpu_micros(0),
|
|
bytes_read_non_output_levels(0),
|
|
bytes_read_output_level(0),
|
|
bytes_written(0),
|
|
bytes_moved(0),
|
|
num_input_files_in_non_output_levels(0),
|
|
num_input_files_in_output_level(0),
|
|
num_output_files(0),
|
|
num_input_records(0),
|
|
num_dropped_records(0),
|
|
count(c) {
|
|
int num_of_reasons = static_cast<int>(CompactionReason::kNumOfReasons);
|
|
for (int i = 0; i < num_of_reasons; i++) {
|
|
counts[i] = 0;
|
|
}
|
|
int r = static_cast<int>(reason);
|
|
if (r >= 0 && r < num_of_reasons) {
|
|
counts[r] = c;
|
|
} else {
|
|
count = 0;
|
|
}
|
|
}
|
|
|
|
explicit CompactionStats(const CompactionStats& c)
|
|
: micros(c.micros),
|
|
cpu_micros(c.cpu_micros),
|
|
bytes_read_non_output_levels(c.bytes_read_non_output_levels),
|
|
bytes_read_output_level(c.bytes_read_output_level),
|
|
bytes_written(c.bytes_written),
|
|
bytes_moved(c.bytes_moved),
|
|
num_input_files_in_non_output_levels(
|
|
c.num_input_files_in_non_output_levels),
|
|
num_input_files_in_output_level(c.num_input_files_in_output_level),
|
|
num_output_files(c.num_output_files),
|
|
num_input_records(c.num_input_records),
|
|
num_dropped_records(c.num_dropped_records),
|
|
count(c.count) {
|
|
int num_of_reasons = static_cast<int>(CompactionReason::kNumOfReasons);
|
|
for (int i = 0; i < num_of_reasons; i++) {
|
|
counts[i] = c.counts[i];
|
|
}
|
|
}
|
|
|
|
CompactionStats& operator=(const CompactionStats& c) {
|
|
micros = c.micros;
|
|
cpu_micros = c.cpu_micros;
|
|
bytes_read_non_output_levels = c.bytes_read_non_output_levels;
|
|
bytes_read_output_level = c.bytes_read_output_level;
|
|
bytes_written = c.bytes_written;
|
|
bytes_moved = c.bytes_moved;
|
|
num_input_files_in_non_output_levels =
|
|
c.num_input_files_in_non_output_levels;
|
|
num_input_files_in_output_level = c.num_input_files_in_output_level;
|
|
num_output_files = c.num_output_files;
|
|
num_input_records = c.num_input_records;
|
|
num_dropped_records = c.num_dropped_records;
|
|
count = c.count;
|
|
|
|
int num_of_reasons = static_cast<int>(CompactionReason::kNumOfReasons);
|
|
for (int i = 0; i < num_of_reasons; i++) {
|
|
counts[i] = c.counts[i];
|
|
}
|
|
return *this;
|
|
}
|
|
|
|
void Clear() {
|
|
this->micros = 0;
|
|
this->cpu_micros = 0;
|
|
this->bytes_read_non_output_levels = 0;
|
|
this->bytes_read_output_level = 0;
|
|
this->bytes_written = 0;
|
|
this->bytes_moved = 0;
|
|
this->num_input_files_in_non_output_levels = 0;
|
|
this->num_input_files_in_output_level = 0;
|
|
this->num_output_files = 0;
|
|
this->num_input_records = 0;
|
|
this->num_dropped_records = 0;
|
|
this->count = 0;
|
|
int num_of_reasons = static_cast<int>(CompactionReason::kNumOfReasons);
|
|
for (int i = 0; i < num_of_reasons; i++) {
|
|
counts[i] = 0;
|
|
}
|
|
}
|
|
|
|
void Add(const CompactionStats& c) {
|
|
this->micros += c.micros;
|
|
this->cpu_micros += c.cpu_micros;
|
|
this->bytes_read_non_output_levels += c.bytes_read_non_output_levels;
|
|
this->bytes_read_output_level += c.bytes_read_output_level;
|
|
this->bytes_written += c.bytes_written;
|
|
this->bytes_moved += c.bytes_moved;
|
|
this->num_input_files_in_non_output_levels +=
|
|
c.num_input_files_in_non_output_levels;
|
|
this->num_input_files_in_output_level +=
|
|
c.num_input_files_in_output_level;
|
|
this->num_output_files += c.num_output_files;
|
|
this->num_input_records += c.num_input_records;
|
|
this->num_dropped_records += c.num_dropped_records;
|
|
this->count += c.count;
|
|
int num_of_reasons = static_cast<int>(CompactionReason::kNumOfReasons);
|
|
for (int i = 0; i< num_of_reasons; i++) {
|
|
counts[i] += c.counts[i];
|
|
}
|
|
}
|
|
|
|
void Subtract(const CompactionStats& c) {
|
|
this->micros -= c.micros;
|
|
this->cpu_micros -= c.cpu_micros;
|
|
this->bytes_read_non_output_levels -= c.bytes_read_non_output_levels;
|
|
this->bytes_read_output_level -= c.bytes_read_output_level;
|
|
this->bytes_written -= c.bytes_written;
|
|
this->bytes_moved -= c.bytes_moved;
|
|
this->num_input_files_in_non_output_levels -=
|
|
c.num_input_files_in_non_output_levels;
|
|
this->num_input_files_in_output_level -=
|
|
c.num_input_files_in_output_level;
|
|
this->num_output_files -= c.num_output_files;
|
|
this->num_input_records -= c.num_input_records;
|
|
this->num_dropped_records -= c.num_dropped_records;
|
|
this->count -= c.count;
|
|
int num_of_reasons = static_cast<int>(CompactionReason::kNumOfReasons);
|
|
for (int i = 0; i < num_of_reasons; i++) {
|
|
counts[i] -= c.counts[i];
|
|
}
|
|
}
|
|
};
|
|
|
|
void Clear() {
|
|
for (int i = 0; i < kIntStatsNumMax; i++) {
|
|
db_stats_[i].store(0);
|
|
}
|
|
for (int i = 0; i < INTERNAL_CF_STATS_ENUM_MAX; i++) {
|
|
cf_stats_count_[i] = 0;
|
|
cf_stats_value_[i] = 0;
|
|
}
|
|
for (auto& comp_stat : comp_stats_) {
|
|
comp_stat.Clear();
|
|
}
|
|
for (auto& h : file_read_latency_) {
|
|
h.Clear();
|
|
}
|
|
blob_file_read_latency_.Clear();
|
|
cf_stats_snapshot_.Clear();
|
|
db_stats_snapshot_.Clear();
|
|
bg_error_count_ = 0;
|
|
started_at_ = env_->NowMicros();
|
|
}
|
|
|
|
void AddCompactionStats(int level, Env::Priority thread_pri,
|
|
const CompactionStats& stats) {
|
|
comp_stats_[level].Add(stats);
|
|
comp_stats_by_pri_[thread_pri].Add(stats);
|
|
}
|
|
|
|
void IncBytesMoved(int level, uint64_t amount) {
|
|
comp_stats_[level].bytes_moved += amount;
|
|
}
|
|
|
|
void AddCFStats(InternalCFStatsType type, uint64_t value) {
|
|
cf_stats_value_[type] += value;
|
|
++cf_stats_count_[type];
|
|
}
|
|
|
|
void AddDBStats(InternalDBStatsType type, uint64_t value,
|
|
bool concurrent = false) {
|
|
auto& v = db_stats_[type];
|
|
if (concurrent) {
|
|
v.fetch_add(value, std::memory_order_relaxed);
|
|
} else {
|
|
v.store(v.load(std::memory_order_relaxed) + value,
|
|
std::memory_order_relaxed);
|
|
}
|
|
}
|
|
|
|
uint64_t GetDBStats(InternalDBStatsType type) {
|
|
return db_stats_[type].load(std::memory_order_relaxed);
|
|
}
|
|
|
|
HistogramImpl* GetFileReadHist(int level) {
|
|
return &file_read_latency_[level];
|
|
}
|
|
|
|
HistogramImpl* GetBlobFileReadHist() { return &blob_file_read_latency_; }
|
|
|
|
uint64_t GetBackgroundErrorCount() const { return bg_error_count_; }
|
|
|
|
uint64_t BumpAndGetBackgroundErrorCount() { return ++bg_error_count_; }
|
|
|
|
bool GetStringProperty(const DBPropertyInfo& property_info,
|
|
const Slice& property, std::string* value);
|
|
|
|
bool GetMapProperty(const DBPropertyInfo& property_info,
|
|
const Slice& property,
|
|
std::map<std::string, std::string>* value);
|
|
|
|
bool GetIntProperty(const DBPropertyInfo& property_info, uint64_t* value,
|
|
DBImpl* db);
|
|
|
|
bool GetIntPropertyOutOfMutex(const DBPropertyInfo& property_info,
|
|
Version* version, uint64_t* value);
|
|
|
|
const uint64_t* TEST_GetCFStatsValue() const { return cf_stats_value_; }
|
|
|
|
const std::vector<CompactionStats>& TEST_GetCompactionStats() const {
|
|
return comp_stats_;
|
|
}
|
|
|
|
// Store a mapping from the user-facing DB::Properties string to our
|
|
// DBPropertyInfo struct used internally for retrieving properties.
|
|
static const std::unordered_map<std::string, DBPropertyInfo> ppt_name_to_info;
|
|
|
|
private:
|
|
void DumpDBStats(std::string* value);
|
|
void DumpCFMapStats(std::map<std::string, std::string>* cf_stats);
|
|
void DumpCFMapStats(
|
|
std::map<int, std::map<LevelStatType, double>>* level_stats,
|
|
CompactionStats* compaction_stats_sum);
|
|
void DumpCFMapStatsByPriority(
|
|
std::map<int, std::map<LevelStatType, double>>* priorities_stats);
|
|
void DumpCFMapStatsIOStalls(std::map<std::string, std::string>* cf_stats);
|
|
void DumpCFStats(std::string* value);
|
|
void DumpCFStatsNoFileHistogram(std::string* value);
|
|
void DumpCFFileHistogram(std::string* value);
|
|
|
|
bool HandleBlockCacheStat(Cache** block_cache);
|
|
|
|
// Per-DB stats
|
|
std::atomic<uint64_t> db_stats_[kIntStatsNumMax];
|
|
// Per-ColumnFamily stats
|
|
uint64_t cf_stats_value_[INTERNAL_CF_STATS_ENUM_MAX];
|
|
uint64_t cf_stats_count_[INTERNAL_CF_STATS_ENUM_MAX];
|
|
// Per-ColumnFamily/level compaction stats
|
|
std::vector<CompactionStats> comp_stats_;
|
|
std::vector<CompactionStats> comp_stats_by_pri_;
|
|
std::vector<HistogramImpl> file_read_latency_;
|
|
HistogramImpl blob_file_read_latency_;
|
|
|
|
// Used to compute per-interval statistics
|
|
struct CFStatsSnapshot {
|
|
// ColumnFamily-level stats
|
|
CompactionStats comp_stats;
|
|
uint64_t ingest_bytes_flush; // Bytes written to L0 (Flush)
|
|
uint64_t stall_count; // Stall count
|
|
// Stats from compaction jobs - bytes written, bytes read, duration.
|
|
uint64_t compact_bytes_write;
|
|
uint64_t compact_bytes_read;
|
|
uint64_t compact_micros;
|
|
double seconds_up;
|
|
|
|
// AddFile specific stats
|
|
uint64_t ingest_bytes_addfile; // Total Bytes ingested
|
|
uint64_t ingest_files_addfile; // Total number of files ingested
|
|
uint64_t ingest_l0_files_addfile; // Total number of files ingested to L0
|
|
uint64_t ingest_keys_addfile; // Total number of keys ingested
|
|
|
|
CFStatsSnapshot()
|
|
: ingest_bytes_flush(0),
|
|
stall_count(0),
|
|
compact_bytes_write(0),
|
|
compact_bytes_read(0),
|
|
compact_micros(0),
|
|
seconds_up(0),
|
|
ingest_bytes_addfile(0),
|
|
ingest_files_addfile(0),
|
|
ingest_l0_files_addfile(0),
|
|
ingest_keys_addfile(0) {}
|
|
|
|
void Clear() {
|
|
comp_stats.Clear();
|
|
ingest_bytes_flush = 0;
|
|
stall_count = 0;
|
|
compact_bytes_write = 0;
|
|
compact_bytes_read = 0;
|
|
compact_micros = 0;
|
|
seconds_up = 0;
|
|
ingest_bytes_addfile = 0;
|
|
ingest_files_addfile = 0;
|
|
ingest_l0_files_addfile = 0;
|
|
ingest_keys_addfile = 0;
|
|
}
|
|
} cf_stats_snapshot_;
|
|
|
|
struct DBStatsSnapshot {
|
|
// DB-level stats
|
|
uint64_t ingest_bytes; // Bytes written by user
|
|
uint64_t wal_bytes; // Bytes written to WAL
|
|
uint64_t wal_synced; // Number of times WAL is synced
|
|
uint64_t write_with_wal; // Number of writes that request WAL
|
|
// These count the number of writes processed by the calling thread or
|
|
// another thread.
|
|
uint64_t write_other;
|
|
uint64_t write_self;
|
|
// Total number of keys written. write_self and write_other measure number
|
|
// of write requests written, Each of the write request can contain updates
|
|
// to multiple keys. num_keys_written is total number of keys updated by all
|
|
// those writes.
|
|
uint64_t num_keys_written;
|
|
// Total time writes delayed by stalls.
|
|
uint64_t write_stall_micros;
|
|
double seconds_up;
|
|
|
|
DBStatsSnapshot()
|
|
: ingest_bytes(0),
|
|
wal_bytes(0),
|
|
wal_synced(0),
|
|
write_with_wal(0),
|
|
write_other(0),
|
|
write_self(0),
|
|
num_keys_written(0),
|
|
write_stall_micros(0),
|
|
seconds_up(0) {}
|
|
|
|
void Clear() {
|
|
ingest_bytes = 0;
|
|
wal_bytes = 0;
|
|
wal_synced = 0;
|
|
write_with_wal = 0;
|
|
write_other = 0;
|
|
write_self = 0;
|
|
num_keys_written = 0;
|
|
write_stall_micros = 0;
|
|
seconds_up = 0;
|
|
}
|
|
} db_stats_snapshot_;
|
|
|
|
// Handler functions for getting property values. They use "value" as a value-
|
|
// result argument, and return true upon successfully setting "value".
|
|
bool HandleNumFilesAtLevel(std::string* value, Slice suffix);
|
|
bool HandleCompressionRatioAtLevelPrefix(std::string* value, Slice suffix);
|
|
bool HandleLevelStats(std::string* value, Slice suffix);
|
|
bool HandleStats(std::string* value, Slice suffix);
|
|
bool HandleCFMapStats(std::map<std::string, std::string>* compaction_stats);
|
|
bool HandleCFStats(std::string* value, Slice suffix);
|
|
bool HandleCFStatsNoFileHistogram(std::string* value, Slice suffix);
|
|
bool HandleCFFileHistogram(std::string* value, Slice suffix);
|
|
bool HandleDBStats(std::string* value, Slice suffix);
|
|
bool HandleSsTables(std::string* value, Slice suffix);
|
|
bool HandleAggregatedTableProperties(std::string* value, Slice suffix);
|
|
bool HandleAggregatedTablePropertiesAtLevel(std::string* value, Slice suffix);
|
|
bool HandleNumImmutableMemTable(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleNumImmutableMemTableFlushed(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleMemTableFlushPending(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleNumRunningFlushes(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleCompactionPending(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleNumRunningCompactions(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleBackgroundErrors(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleCurSizeActiveMemTable(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleCurSizeAllMemTables(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleSizeAllMemTables(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleNumEntriesActiveMemTable(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleNumEntriesImmMemTables(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleNumDeletesActiveMemTable(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleNumDeletesImmMemTables(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleEstimateNumKeys(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleNumSnapshots(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleOldestSnapshotTime(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleOldestSnapshotSequence(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleNumLiveVersions(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleCurrentSuperVersionNumber(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleIsFileDeletionsEnabled(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleBaseLevel(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleTotalSstFilesSize(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleLiveSstFilesSize(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleEstimatePendingCompactionBytes(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleEstimateTableReadersMem(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleEstimateLiveDataSize(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleMinLogNumberToKeep(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleMinObsoleteSstNumberToKeep(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleActualDelayedWriteRate(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleIsWriteStopped(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleEstimateOldestKeyTime(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
bool HandleBlockCacheCapacity(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleBlockCacheUsage(uint64_t* value, DBImpl* db, Version* version);
|
|
bool HandleBlockCachePinnedUsage(uint64_t* value, DBImpl* db,
|
|
Version* version);
|
|
// Total number of background errors encountered. Every time a flush task
|
|
// or compaction task fails, this counter is incremented. The failure can
|
|
// be caused by any possible reason, including file system errors, out of
|
|
// resources, or input file corruption. Failing when retrying the same flush
|
|
// or compaction will cause the counter to increase too.
|
|
uint64_t bg_error_count_;
|
|
|
|
const int number_levels_;
|
|
Env* env_;
|
|
ColumnFamilyData* cfd_;
|
|
uint64_t started_at_;
|
|
};
|
|
|
|
#else
|
|
|
|
class InternalStats {
|
|
public:
|
|
enum InternalCFStatsType {
|
|
L0_FILE_COUNT_LIMIT_SLOWDOWNS,
|
|
LOCKED_L0_FILE_COUNT_LIMIT_SLOWDOWNS,
|
|
MEMTABLE_LIMIT_STOPS,
|
|
MEMTABLE_LIMIT_SLOWDOWNS,
|
|
L0_FILE_COUNT_LIMIT_STOPS,
|
|
LOCKED_L0_FILE_COUNT_LIMIT_STOPS,
|
|
PENDING_COMPACTION_BYTES_LIMIT_SLOWDOWNS,
|
|
PENDING_COMPACTION_BYTES_LIMIT_STOPS,
|
|
WRITE_STALLS_ENUM_MAX,
|
|
BYTES_FLUSHED,
|
|
BYTES_INGESTED_ADD_FILE,
|
|
INGESTED_NUM_FILES_TOTAL,
|
|
INGESTED_LEVEL0_NUM_FILES_TOTAL,
|
|
INGESTED_NUM_KEYS_TOTAL,
|
|
INTERNAL_CF_STATS_ENUM_MAX,
|
|
};
|
|
|
|
enum InternalDBStatsType {
|
|
kIntStatsWalFileBytes,
|
|
kIntStatsWalFileSynced,
|
|
kIntStatsBytesWritten,
|
|
kIntStatsNumKeysWritten,
|
|
kIntStatsWriteDoneByOther,
|
|
kIntStatsWriteDoneBySelf,
|
|
kIntStatsWriteWithWal,
|
|
kIntStatsWriteStallMicros,
|
|
kIntStatsNumMax,
|
|
};
|
|
|
|
InternalStats(int /*num_levels*/, Env* /*env*/, ColumnFamilyData* /*cfd*/) {}
|
|
|
|
struct CompactionStats {
|
|
uint64_t micros;
|
|
uint64_t cpu_micros;
|
|
uint64_t bytes_read_non_output_levels;
|
|
uint64_t bytes_read_output_level;
|
|
uint64_t bytes_written;
|
|
uint64_t bytes_moved;
|
|
int num_input_files_in_non_output_levels;
|
|
int num_input_files_in_output_level;
|
|
int num_output_files;
|
|
uint64_t num_input_records;
|
|
uint64_t num_dropped_records;
|
|
int count;
|
|
|
|
explicit CompactionStats() {}
|
|
|
|
explicit CompactionStats(CompactionReason /*reason*/, int /*c*/) {}
|
|
|
|
explicit CompactionStats(const CompactionStats& /*c*/) {}
|
|
|
|
void Add(const CompactionStats& /*c*/) {}
|
|
|
|
void Subtract(const CompactionStats& /*c*/) {}
|
|
};
|
|
|
|
void AddCompactionStats(int /*level*/, Env::Priority /*thread_pri*/,
|
|
const CompactionStats& /*stats*/) {}
|
|
|
|
void IncBytesMoved(int /*level*/, uint64_t /*amount*/) {}
|
|
|
|
void AddCFStats(InternalCFStatsType /*type*/, uint64_t /*value*/) {}
|
|
|
|
void AddDBStats(InternalDBStatsType /*type*/, uint64_t /*value*/,
|
|
bool /*concurrent */ = false) {}
|
|
|
|
HistogramImpl* GetFileReadHist(int /*level*/) { return nullptr; }
|
|
|
|
uint64_t GetBackgroundErrorCount() const { return 0; }
|
|
|
|
uint64_t BumpAndGetBackgroundErrorCount() { return 0; }
|
|
|
|
bool GetStringProperty(const DBPropertyInfo& /*property_info*/,
|
|
const Slice& /*property*/, std::string* /*value*/) {
|
|
return false;
|
|
}
|
|
|
|
bool GetMapProperty(const DBPropertyInfo& /*property_info*/,
|
|
const Slice& /*property*/,
|
|
std::map<std::string, std::string>* /*value*/) {
|
|
return false;
|
|
}
|
|
|
|
bool GetIntProperty(const DBPropertyInfo& /*property_info*/, uint64_t* /*value*/,
|
|
DBImpl* /*db*/) const {
|
|
return false;
|
|
}
|
|
|
|
bool GetIntPropertyOutOfMutex(const DBPropertyInfo& /*property_info*/,
|
|
Version* /*version*/, uint64_t* /*value*/) const {
|
|
return false;
|
|
}
|
|
};
|
|
#endif // !ROCKSDB_LITE
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|