3dff28cf9b
Summary: For performance purposes, the lower level routines were changed to use a SystemClock* instead of a std::shared_ptr<SystemClock>. The shared ptr has some performance degradation on certain hardware classes. For most of the system, there is no risk of the pointer being deleted/invalid because the shared_ptr will be stored elsewhere. For example, the ImmutableDBOptions stores the Env which has a std::shared_ptr<SystemClock> in it. The SystemClock* within the ImmutableDBOptions is essentially a "short cut" to gain access to this constant resource. There were a few classes (PeriodicWorkScheduler?) where the "short cut" property did not hold. In those cases, the shared pointer was preserved. Using db_bench readrandom perf_level=3 on my EC2 box, this change performed as well or better than 6.17: 6.17: readrandom : 28.046 micros/op 854902 ops/sec; 61.3 MB/s (355999 of 355999 found) 6.18: readrandom : 32.615 micros/op 735306 ops/sec; 52.7 MB/s (290999 of 290999 found) PR: readrandom : 27.500 micros/op 871909 ops/sec; 62.5 MB/s (367999 of 367999 found) (Note that the times for 6.18 are prior to revert of the SystemClock). Pull Request resolved: https://github.com/facebook/rocksdb/pull/8033 Reviewed By: pdillinger Differential Revision: D27014563 Pulled By: mrambacher fbshipit-source-id: ad0459eba03182e454391b5926bf5cdd45657b67
2310 lines
93 KiB
C++
2310 lines
93 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
#ifdef GFLAGS
|
|
#include "tools/block_cache_analyzer/block_cache_trace_analyzer.h"
|
|
|
|
#include <algorithm>
|
|
#include <cinttypes>
|
|
#include <cstdio>
|
|
#include <cstdlib>
|
|
#include <fstream>
|
|
#include <iomanip>
|
|
#include <iostream>
|
|
#include <memory>
|
|
#include <random>
|
|
#include <sstream>
|
|
|
|
#include "monitoring/histogram.h"
|
|
#include "rocksdb/system_clock.h"
|
|
#include "util/gflags_compat.h"
|
|
#include "util/string_util.h"
|
|
|
|
using GFLAGS_NAMESPACE::ParseCommandLineFlags;
|
|
|
|
DEFINE_string(block_cache_trace_path, "", "The trace file path.");
|
|
DEFINE_bool(is_block_cache_human_readable_trace, false,
|
|
"Is the trace file provided for analysis generated by running "
|
|
"block_cache_trace_analyzer with "
|
|
"FLAGS_human_readable_trace_file_path is specified.");
|
|
DEFINE_string(
|
|
block_cache_sim_config_path, "",
|
|
"The config file path. One cache configuration per line. The format of a "
|
|
"cache configuration is "
|
|
"cache_name,num_shard_bits,ghost_capacity,cache_capacity_1,...,cache_"
|
|
"capacity_N. Supported cache names are lru, lru_priority, lru_hybrid, and "
|
|
"lru_hybrid_no_insert_on_row_miss. User may also add a prefix 'ghost_' to "
|
|
"a cache_name to add a ghost cache in front of the real cache. "
|
|
"ghost_capacity and cache_capacity can be xK, xM or xG where x is a "
|
|
"positive number.");
|
|
DEFINE_int32(block_cache_trace_downsample_ratio, 1,
|
|
"The trace collected accesses on one in every "
|
|
"block_cache_trace_downsample_ratio blocks. We scale "
|
|
"down the simulated cache size by this ratio.");
|
|
DEFINE_bool(print_block_size_stats, false,
|
|
"Print block size distribution and the distribution break down by "
|
|
"block type and column family.");
|
|
DEFINE_bool(print_access_count_stats, false,
|
|
"Print access count distribution and the distribution break down "
|
|
"by block type and column family.");
|
|
DEFINE_bool(print_data_block_access_count_stats, false,
|
|
"Print data block accesses by user Get and Multi-Get.");
|
|
DEFINE_int32(cache_sim_warmup_seconds, 0,
|
|
"The number of seconds to warmup simulated caches. The hit/miss "
|
|
"counters are reset after the warmup completes.");
|
|
DEFINE_int32(analyze_bottom_k_access_count_blocks, 0,
|
|
"Print out detailed access information for blocks with their "
|
|
"number of accesses are the bottom k among all blocks.");
|
|
DEFINE_int32(analyze_top_k_access_count_blocks, 0,
|
|
"Print out detailed access information for blocks with their "
|
|
"number of accesses are the top k among all blocks.");
|
|
DEFINE_string(block_cache_analysis_result_dir, "",
|
|
"The directory that saves block cache analysis results.");
|
|
DEFINE_string(
|
|
timeline_labels, "",
|
|
"Group the number of accesses per block per second using these labels. "
|
|
"Possible labels are a combination of the following: cf (column family), "
|
|
"sst, level, bt (block type), caller, block. For example, label \"cf_bt\" "
|
|
"means the number of access per second is grouped by unique pairs of "
|
|
"\"cf_bt\". A label \"all\" contains the aggregated number of accesses per "
|
|
"second across all possible labels.");
|
|
DEFINE_string(reuse_distance_labels, "",
|
|
"Group the reuse distance of a block using these labels. Reuse "
|
|
"distance is defined as the cumulated size of unique blocks read "
|
|
"between two consecutive accesses on the same block.");
|
|
DEFINE_string(
|
|
reuse_distance_buckets, "",
|
|
"Group blocks by their reuse distances given these buckets. For "
|
|
"example, if 'reuse_distance_buckets' is '1K,1M,1G', we will "
|
|
"create four buckets. The first three buckets contain the number of "
|
|
"blocks with reuse distance less than 1KB, between 1K and 1M, between 1M "
|
|
"and 1G, respectively. The last bucket contains the number of blocks with "
|
|
"reuse distance larger than 1G. ");
|
|
DEFINE_string(
|
|
reuse_interval_labels, "",
|
|
"Group the reuse interval of a block using these labels. Reuse "
|
|
"interval is defined as the time between two consecutive accesses "
|
|
"on the same block.");
|
|
DEFINE_string(
|
|
reuse_interval_buckets, "",
|
|
"Group blocks by their reuse interval given these buckets. For "
|
|
"example, if 'reuse_distance_buckets' is '1,10,100', we will "
|
|
"create four buckets. The first three buckets contain the number of "
|
|
"blocks with reuse interval less than 1 second, between 1 second and 10 "
|
|
"seconds, between 10 seconds and 100 seconds, respectively. The last "
|
|
"bucket contains the number of blocks with reuse interval longer than 100 "
|
|
"seconds.");
|
|
DEFINE_string(
|
|
reuse_lifetime_labels, "",
|
|
"Group the reuse lifetime of a block using these labels. Reuse "
|
|
"lifetime is defined as the time interval between the first access on a "
|
|
"block and the last access on the same block. For blocks that are only "
|
|
"accessed once, its lifetime is set to kMaxUint64.");
|
|
DEFINE_string(
|
|
reuse_lifetime_buckets, "",
|
|
"Group blocks by their reuse lifetime given these buckets. For "
|
|
"example, if 'reuse_lifetime_buckets' is '1,10,100', we will "
|
|
"create four buckets. The first three buckets contain the number of "
|
|
"blocks with reuse lifetime less than 1 second, between 1 second and 10 "
|
|
"seconds, between 10 seconds and 100 seconds, respectively. The last "
|
|
"bucket contains the number of blocks with reuse lifetime longer than 100 "
|
|
"seconds.");
|
|
DEFINE_string(
|
|
analyze_callers, "",
|
|
"The list of callers to perform a detailed analysis on. If speicfied, the "
|
|
"analyzer will output a detailed percentage of accesses for each caller "
|
|
"break down by column family, level, and block type. A list of available "
|
|
"callers are: Get, MultiGet, Iterator, ApproximateSize, VerifyChecksum, "
|
|
"SSTDumpTool, ExternalSSTIngestion, Repair, Prefetch, Compaction, "
|
|
"CompactionRefill, Flush, SSTFileReader, Uncategorized.");
|
|
DEFINE_string(access_count_buckets, "",
|
|
"Group number of blocks by their access count given these "
|
|
"buckets. If specified, the analyzer will output a detailed "
|
|
"analysis on the number of blocks grouped by their access count "
|
|
"break down by block type and column family.");
|
|
DEFINE_int32(analyze_blocks_reuse_k_reuse_window, 0,
|
|
"Analyze the percentage of blocks that are accessed in the "
|
|
"[k, 2*k] seconds are accessed again in the next [2*k, 3*k], "
|
|
"[3*k, 4*k],...,[k*(n-1), k*n] seconds. ");
|
|
DEFINE_string(analyze_get_spatial_locality_labels, "",
|
|
"Group data blocks using these labels.");
|
|
DEFINE_string(analyze_get_spatial_locality_buckets, "",
|
|
"Group data blocks by their statistics using these buckets.");
|
|
DEFINE_string(skew_labels, "",
|
|
"Group the access count of a block using these labels.");
|
|
DEFINE_string(skew_buckets, "", "Group the skew labels using these buckets.");
|
|
DEFINE_bool(mrc_only, false,
|
|
"Evaluate alternative cache policies only. When this flag is true, "
|
|
"the analyzer does NOT maintain states of each block in memory for "
|
|
"analysis. It only feeds the accesses into the cache simulators.");
|
|
DEFINE_string(
|
|
analyze_correlation_coefficients_labels, "",
|
|
"Analyze the correlation coefficients of features such as number of past "
|
|
"accesses with regard to the number of accesses till the next access.");
|
|
DEFINE_int32(analyze_correlation_coefficients_max_number_of_values, 1000000,
|
|
"The maximum number of values for a feature. If the number of "
|
|
"values for a feature is larger than this max, it randomly "
|
|
"selects 'max' number of values.");
|
|
DEFINE_string(human_readable_trace_file_path, "",
|
|
"The filt path that saves human readable access records.");
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
namespace {
|
|
|
|
const std::string kMissRatioCurveFileName = "mrc";
|
|
const std::string kGroupbyBlock = "block";
|
|
const std::string kGroupbyTable = "table";
|
|
const std::string kGroupbyColumnFamily = "cf";
|
|
const std::string kGroupbySSTFile = "sst";
|
|
const std::string kGroupbyBlockType = "bt";
|
|
const std::string kGroupbyCaller = "caller";
|
|
const std::string kGroupbyLevel = "level";
|
|
const std::string kGroupbyAll = "all";
|
|
const std::set<std::string> kGroupbyLabels{
|
|
kGroupbyBlock, kGroupbyColumnFamily, kGroupbySSTFile, kGroupbyLevel,
|
|
kGroupbyBlockType, kGroupbyCaller, kGroupbyAll};
|
|
const std::string kSupportedCacheNames =
|
|
" lru ghost_lru lru_priority ghost_lru_priority lru_hybrid "
|
|
"ghost_lru_hybrid lru_hybrid_no_insert_on_row_miss "
|
|
"ghost_lru_hybrid_no_insert_on_row_miss ";
|
|
|
|
// The suffix for the generated csv files.
|
|
const std::string kFileNameSuffixMissRatioTimeline = "miss_ratio_timeline";
|
|
const std::string kFileNameSuffixMissTimeline = "miss_timeline";
|
|
const std::string kFileNameSuffixSkew = "skewness";
|
|
const std::string kFileNameSuffixAccessTimeline = "access_timeline";
|
|
const std::string kFileNameSuffixCorrelation = "correlation_input";
|
|
const std::string kFileNameSuffixAvgReuseIntervalNaccesses =
|
|
"avg_reuse_interval_naccesses";
|
|
const std::string kFileNameSuffixAvgReuseInterval = "avg_reuse_interval";
|
|
const std::string kFileNameSuffixReuseInterval = "access_reuse_interval";
|
|
const std::string kFileNameSuffixReuseLifetime = "reuse_lifetime";
|
|
const std::string kFileNameSuffixAccessReuseBlocksTimeline =
|
|
"reuse_blocks_timeline";
|
|
const std::string kFileNameSuffixPercentOfAccessSummary =
|
|
"percentage_of_accesses_summary";
|
|
const std::string kFileNameSuffixPercentRefKeys = "percent_ref_keys";
|
|
const std::string kFileNameSuffixPercentDataSizeOnRefKeys =
|
|
"percent_data_size_on_ref_keys";
|
|
const std::string kFileNameSuffixPercentAccessesOnRefKeys =
|
|
"percent_accesses_on_ref_keys";
|
|
const std::string kFileNameSuffixAccessCountSummary = "access_count_summary";
|
|
|
|
std::string block_type_to_string(TraceType type) {
|
|
switch (type) {
|
|
case kBlockTraceFilterBlock:
|
|
return "Filter";
|
|
case kBlockTraceDataBlock:
|
|
return "Data";
|
|
case kBlockTraceIndexBlock:
|
|
return "Index";
|
|
case kBlockTraceRangeDeletionBlock:
|
|
return "RangeDeletion";
|
|
case kBlockTraceUncompressionDictBlock:
|
|
return "UncompressionDict";
|
|
default:
|
|
break;
|
|
}
|
|
// This cannot happen.
|
|
return "InvalidType";
|
|
}
|
|
|
|
std::string caller_to_string(TableReaderCaller caller) {
|
|
switch (caller) {
|
|
case kUserGet:
|
|
return "Get";
|
|
case kUserMultiGet:
|
|
return "MultiGet";
|
|
case kUserIterator:
|
|
return "Iterator";
|
|
case kUserApproximateSize:
|
|
return "ApproximateSize";
|
|
case kUserVerifyChecksum:
|
|
return "VerifyChecksum";
|
|
case kSSTDumpTool:
|
|
return "SSTDumpTool";
|
|
case kExternalSSTIngestion:
|
|
return "ExternalSSTIngestion";
|
|
case kRepair:
|
|
return "Repair";
|
|
case kPrefetch:
|
|
return "Prefetch";
|
|
case kCompaction:
|
|
return "Compaction";
|
|
case kCompactionRefill:
|
|
return "CompactionRefill";
|
|
case kFlush:
|
|
return "Flush";
|
|
case kSSTFileReader:
|
|
return "SSTFileReader";
|
|
case kUncategorized:
|
|
return "Uncategorized";
|
|
default:
|
|
break;
|
|
}
|
|
// This cannot happen.
|
|
return "InvalidCaller";
|
|
}
|
|
|
|
TableReaderCaller string_to_caller(std::string caller_str) {
|
|
if (caller_str == "Get") {
|
|
return kUserGet;
|
|
} else if (caller_str == "MultiGet") {
|
|
return kUserMultiGet;
|
|
} else if (caller_str == "Iterator") {
|
|
return kUserIterator;
|
|
} else if (caller_str == "ApproximateSize") {
|
|
return kUserApproximateSize;
|
|
} else if (caller_str == "VerifyChecksum") {
|
|
return kUserVerifyChecksum;
|
|
} else if (caller_str == "SSTDumpTool") {
|
|
return kSSTDumpTool;
|
|
} else if (caller_str == "ExternalSSTIngestion") {
|
|
return kExternalSSTIngestion;
|
|
} else if (caller_str == "Repair") {
|
|
return kRepair;
|
|
} else if (caller_str == "Prefetch") {
|
|
return kPrefetch;
|
|
} else if (caller_str == "Compaction") {
|
|
return kCompaction;
|
|
} else if (caller_str == "CompactionRefill") {
|
|
return kCompactionRefill;
|
|
} else if (caller_str == "Flush") {
|
|
return kFlush;
|
|
} else if (caller_str == "SSTFileReader") {
|
|
return kSSTFileReader;
|
|
} else if (caller_str == "Uncategorized") {
|
|
return kUncategorized;
|
|
}
|
|
return TableReaderCaller::kMaxBlockCacheLookupCaller;
|
|
}
|
|
|
|
bool is_user_access(TableReaderCaller caller) {
|
|
switch (caller) {
|
|
case kUserGet:
|
|
case kUserMultiGet:
|
|
case kUserIterator:
|
|
case kUserApproximateSize:
|
|
case kUserVerifyChecksum:
|
|
return true;
|
|
default:
|
|
break;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
const char kBreakLine[] =
|
|
"***************************************************************\n";
|
|
|
|
void print_break_lines(uint32_t num_break_lines) {
|
|
for (uint32_t i = 0; i < num_break_lines; i++) {
|
|
fprintf(stdout, kBreakLine);
|
|
}
|
|
}
|
|
|
|
double percent(uint64_t numerator, uint64_t denomenator) {
|
|
if (denomenator == 0) {
|
|
return -1;
|
|
}
|
|
return static_cast<double>(numerator * 100.0 / denomenator);
|
|
}
|
|
|
|
std::map<uint64_t, uint64_t> adjust_time_unit(
|
|
const std::map<uint64_t, uint64_t>& time_stats, uint64_t time_unit) {
|
|
if (time_unit == 1) {
|
|
return time_stats;
|
|
}
|
|
std::map<uint64_t, uint64_t> adjusted_time_stats;
|
|
for (auto const& time : time_stats) {
|
|
adjusted_time_stats[static_cast<uint64_t>(time.first / time_unit)] +=
|
|
time.second;
|
|
}
|
|
return adjusted_time_stats;
|
|
}
|
|
} // namespace
|
|
|
|
void BlockCacheTraceAnalyzer::WriteMissRatioCurves() const {
|
|
if (!cache_simulator_) {
|
|
return;
|
|
}
|
|
if (output_dir_.empty()) {
|
|
return;
|
|
}
|
|
uint64_t trace_duration =
|
|
trace_end_timestamp_in_seconds_ - trace_start_timestamp_in_seconds_;
|
|
uint64_t total_accesses = access_sequence_number_;
|
|
const std::string output_miss_ratio_curve_path =
|
|
output_dir_ + "/" + std::to_string(trace_duration) + "_" +
|
|
std::to_string(total_accesses) + "_" + kMissRatioCurveFileName;
|
|
std::ofstream out(output_miss_ratio_curve_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
// Write header.
|
|
const std::string header =
|
|
"cache_name,num_shard_bits,ghost_capacity,capacity,miss_ratio,total_"
|
|
"accesses";
|
|
out << header << std::endl;
|
|
for (auto const& config_caches : cache_simulator_->sim_caches()) {
|
|
const CacheConfiguration& config = config_caches.first;
|
|
for (uint32_t i = 0; i < config.cache_capacities.size(); i++) {
|
|
double miss_ratio =
|
|
config_caches.second[i]->miss_ratio_stats().miss_ratio();
|
|
// Write the body.
|
|
out << config.cache_name;
|
|
out << ",";
|
|
out << config.num_shard_bits;
|
|
out << ",";
|
|
out << config.ghost_cache_capacity;
|
|
out << ",";
|
|
out << config.cache_capacities[i];
|
|
out << ",";
|
|
out << std::fixed << std::setprecision(4) << miss_ratio;
|
|
out << ",";
|
|
out << config_caches.second[i]->miss_ratio_stats().total_accesses();
|
|
out << std::endl;
|
|
}
|
|
}
|
|
out.close();
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::UpdateFeatureVectors(
|
|
const std::vector<uint64_t>& access_sequence_number_timeline,
|
|
const std::vector<uint64_t>& access_timeline, const std::string& label,
|
|
std::map<std::string, Features>* label_features,
|
|
std::map<std::string, Predictions>* label_predictions) const {
|
|
if (access_sequence_number_timeline.empty() || access_timeline.empty()) {
|
|
return;
|
|
}
|
|
assert(access_timeline.size() == access_sequence_number_timeline.size());
|
|
uint64_t prev_access_sequence_number = access_sequence_number_timeline[0];
|
|
uint64_t prev_access_timestamp = access_timeline[0];
|
|
for (uint32_t i = 0; i < access_sequence_number_timeline.size(); i++) {
|
|
uint64_t num_accesses_since_last_access =
|
|
access_sequence_number_timeline[i] - prev_access_sequence_number;
|
|
uint64_t elapsed_time_since_last_access =
|
|
access_timeline[i] - prev_access_timestamp;
|
|
prev_access_sequence_number = access_sequence_number_timeline[i];
|
|
prev_access_timestamp = access_timeline[i];
|
|
if (i < access_sequence_number_timeline.size() - 1) {
|
|
(*label_features)[label].num_accesses_since_last_access.push_back(
|
|
num_accesses_since_last_access);
|
|
(*label_features)[label].num_past_accesses.push_back(i);
|
|
(*label_features)[label].elapsed_time_since_last_access.push_back(
|
|
elapsed_time_since_last_access);
|
|
}
|
|
if (i >= 1) {
|
|
(*label_predictions)[label].num_accesses_till_next_access.push_back(
|
|
num_accesses_since_last_access);
|
|
(*label_predictions)[label].elapsed_time_till_next_access.push_back(
|
|
elapsed_time_since_last_access);
|
|
}
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteMissRatioTimeline(uint64_t time_unit) const {
|
|
if (!cache_simulator_ || output_dir_.empty()) {
|
|
return;
|
|
}
|
|
std::map<uint64_t, std::map<std::string, std::map<uint64_t, double>>>
|
|
cs_name_timeline;
|
|
uint64_t start_time = port::kMaxUint64;
|
|
uint64_t end_time = 0;
|
|
const std::map<uint64_t, uint64_t>& trace_num_misses =
|
|
adjust_time_unit(miss_ratio_stats_.num_misses_timeline(), time_unit);
|
|
const std::map<uint64_t, uint64_t>& trace_num_accesses =
|
|
adjust_time_unit(miss_ratio_stats_.num_accesses_timeline(), time_unit);
|
|
assert(trace_num_misses.size() == trace_num_accesses.size());
|
|
for (auto const& num_miss : trace_num_misses) {
|
|
uint64_t time = num_miss.first;
|
|
start_time = std::min(start_time, time);
|
|
end_time = std::max(end_time, time);
|
|
uint64_t miss = num_miss.second;
|
|
auto it = trace_num_accesses.find(time);
|
|
assert(it != trace_num_accesses.end());
|
|
uint64_t access = it->second;
|
|
cs_name_timeline[port::kMaxUint64]["trace"][time] = percent(miss, access);
|
|
}
|
|
for (auto const& config_caches : cache_simulator_->sim_caches()) {
|
|
const CacheConfiguration& config = config_caches.first;
|
|
std::string cache_label = config.cache_name + "-" +
|
|
std::to_string(config.num_shard_bits) + "-" +
|
|
std::to_string(config.ghost_cache_capacity);
|
|
for (uint32_t i = 0; i < config.cache_capacities.size(); i++) {
|
|
const std::map<uint64_t, uint64_t>& num_misses = adjust_time_unit(
|
|
config_caches.second[i]->miss_ratio_stats().num_misses_timeline(),
|
|
time_unit);
|
|
const std::map<uint64_t, uint64_t>& num_accesses = adjust_time_unit(
|
|
config_caches.second[i]->miss_ratio_stats().num_accesses_timeline(),
|
|
time_unit);
|
|
assert(num_misses.size() == num_accesses.size());
|
|
for (auto const& num_miss : num_misses) {
|
|
uint64_t time = num_miss.first;
|
|
start_time = std::min(start_time, time);
|
|
end_time = std::max(end_time, time);
|
|
uint64_t miss = num_miss.second;
|
|
auto it = num_accesses.find(time);
|
|
assert(it != num_accesses.end());
|
|
uint64_t access = it->second;
|
|
cs_name_timeline[config.cache_capacities[i]][cache_label][time] =
|
|
percent(miss, access);
|
|
}
|
|
}
|
|
}
|
|
for (auto const& it : cs_name_timeline) {
|
|
const std::string output_miss_ratio_timeline_path =
|
|
output_dir_ + "/" + std::to_string(it.first) + "_" +
|
|
std::to_string(time_unit) + "_" + kFileNameSuffixMissRatioTimeline;
|
|
std::ofstream out(output_miss_ratio_timeline_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("time");
|
|
for (uint64_t now = start_time; now <= end_time; now++) {
|
|
header += ",";
|
|
header += std::to_string(now);
|
|
}
|
|
out << header << std::endl;
|
|
for (auto const& label : it.second) {
|
|
std::string row(label.first);
|
|
for (uint64_t now = start_time; now <= end_time; now++) {
|
|
auto misses = label.second.find(now);
|
|
row += ",";
|
|
if (misses != label.second.end()) {
|
|
row += std::to_string(misses->second);
|
|
} else {
|
|
row += "0";
|
|
}
|
|
}
|
|
out << row << std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteMissTimeline(uint64_t time_unit) const {
|
|
if (!cache_simulator_ || output_dir_.empty()) {
|
|
return;
|
|
}
|
|
std::map<uint64_t, std::map<std::string, std::map<uint64_t, uint64_t>>>
|
|
cs_name_timeline;
|
|
uint64_t start_time = port::kMaxUint64;
|
|
uint64_t end_time = 0;
|
|
const std::map<uint64_t, uint64_t>& trace_num_misses =
|
|
adjust_time_unit(miss_ratio_stats_.num_misses_timeline(), time_unit);
|
|
for (auto const& num_miss : trace_num_misses) {
|
|
uint64_t time = num_miss.first;
|
|
start_time = std::min(start_time, time);
|
|
end_time = std::max(end_time, time);
|
|
uint64_t miss = num_miss.second;
|
|
cs_name_timeline[port::kMaxUint64]["trace"][time] = miss;
|
|
}
|
|
for (auto const& config_caches : cache_simulator_->sim_caches()) {
|
|
const CacheConfiguration& config = config_caches.first;
|
|
std::string cache_label = config.cache_name + "-" +
|
|
std::to_string(config.num_shard_bits) + "-" +
|
|
std::to_string(config.ghost_cache_capacity);
|
|
for (uint32_t i = 0; i < config.cache_capacities.size(); i++) {
|
|
const std::map<uint64_t, uint64_t>& num_misses = adjust_time_unit(
|
|
config_caches.second[i]->miss_ratio_stats().num_misses_timeline(),
|
|
time_unit);
|
|
for (auto const& num_miss : num_misses) {
|
|
uint64_t time = num_miss.first;
|
|
start_time = std::min(start_time, time);
|
|
end_time = std::max(end_time, time);
|
|
uint64_t miss = num_miss.second;
|
|
cs_name_timeline[config.cache_capacities[i]][cache_label][time] = miss;
|
|
}
|
|
}
|
|
}
|
|
for (auto const& it : cs_name_timeline) {
|
|
const std::string output_miss_ratio_timeline_path =
|
|
output_dir_ + "/" + std::to_string(it.first) + "_" +
|
|
std::to_string(time_unit) + "_" + kFileNameSuffixMissTimeline;
|
|
std::ofstream out(output_miss_ratio_timeline_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("time");
|
|
for (uint64_t now = start_time; now <= end_time; now++) {
|
|
header += ",";
|
|
header += std::to_string(now);
|
|
}
|
|
out << header << std::endl;
|
|
for (auto const& label : it.second) {
|
|
std::string row(label.first);
|
|
for (uint64_t now = start_time; now <= end_time; now++) {
|
|
auto misses = label.second.find(now);
|
|
row += ",";
|
|
if (misses != label.second.end()) {
|
|
row += std::to_string(misses->second);
|
|
} else {
|
|
row += "0";
|
|
}
|
|
}
|
|
out << row << std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteSkewness(
|
|
const std::string& label_str, const std::vector<uint64_t>& percent_buckets,
|
|
TraceType target_block_type) const {
|
|
std::set<std::string> labels = ParseLabelStr(label_str);
|
|
std::map<std::string, uint64_t> label_naccesses;
|
|
uint64_t total_naccesses = 0;
|
|
auto block_callback = [&](const std::string& cf_name, uint64_t fd,
|
|
uint32_t level, TraceType type,
|
|
const std::string& /*block_key*/, uint64_t block_id,
|
|
const BlockAccessInfo& block) {
|
|
if (target_block_type != TraceType::kTraceMax &&
|
|
target_block_type != type) {
|
|
return;
|
|
}
|
|
const std::string label = BuildLabel(
|
|
labels, cf_name, fd, level, type,
|
|
TableReaderCaller::kMaxBlockCacheLookupCaller, block_id, block);
|
|
label_naccesses[label] += block.num_accesses;
|
|
total_naccesses += block.num_accesses;
|
|
};
|
|
TraverseBlocks(block_callback, &labels);
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_bucket_naccesses;
|
|
std::vector<std::pair<std::string, uint64_t>> pairs;
|
|
for (auto const& itr : label_naccesses) {
|
|
pairs.push_back(itr);
|
|
}
|
|
// Sort in descending order.
|
|
sort(pairs.begin(), pairs.end(),
|
|
[](const std::pair<std::string, uint64_t>& a,
|
|
const std::pair<std::string, uint64_t>& b) {
|
|
return b.second < a.second;
|
|
});
|
|
|
|
size_t prev_start_index = 0;
|
|
for (auto const& percent : percent_buckets) {
|
|
label_bucket_naccesses[label_str][percent] = 0;
|
|
size_t end_index = 0;
|
|
if (percent == port::kMaxUint64) {
|
|
end_index = label_naccesses.size();
|
|
} else {
|
|
end_index = percent * label_naccesses.size() / 100;
|
|
}
|
|
for (size_t i = prev_start_index; i < end_index; i++) {
|
|
label_bucket_naccesses[label_str][percent] += pairs[i].second;
|
|
}
|
|
prev_start_index = end_index;
|
|
}
|
|
std::string filename_suffix;
|
|
if (target_block_type != TraceType::kTraceMax) {
|
|
filename_suffix = block_type_to_string(target_block_type);
|
|
filename_suffix += "_";
|
|
}
|
|
filename_suffix += kFileNameSuffixSkew;
|
|
WriteStatsToFile(label_str, percent_buckets, filename_suffix,
|
|
label_bucket_naccesses, total_naccesses);
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteCorrelationFeatures(
|
|
const std::string& label_str, uint32_t max_number_of_values) const {
|
|
std::set<std::string> labels = ParseLabelStr(label_str);
|
|
std::map<std::string, Features> label_features;
|
|
std::map<std::string, Predictions> label_predictions;
|
|
auto block_callback =
|
|
[&](const std::string& cf_name, uint64_t fd, uint32_t level,
|
|
TraceType block_type, const std::string& /*block_key*/,
|
|
uint64_t /*block_key_id*/, const BlockAccessInfo& block) {
|
|
if (block.table_id == 0 && labels.find(kGroupbyTable) != labels.end()) {
|
|
// We only know table id information for get requests.
|
|
return;
|
|
}
|
|
if (labels.find(kGroupbyCaller) != labels.end()) {
|
|
// Group by caller.
|
|
for (auto const& caller_map : block.caller_access_timeline) {
|
|
const std::string label =
|
|
BuildLabel(labels, cf_name, fd, level, block_type,
|
|
caller_map.first, /*block_id=*/0, block);
|
|
auto it = block.caller_access_sequence__number_timeline.find(
|
|
caller_map.first);
|
|
assert(it != block.caller_access_sequence__number_timeline.end());
|
|
UpdateFeatureVectors(it->second, caller_map.second, label,
|
|
&label_features, &label_predictions);
|
|
}
|
|
return;
|
|
}
|
|
const std::string label =
|
|
BuildLabel(labels, cf_name, fd, level, block_type,
|
|
TableReaderCaller::kMaxBlockCacheLookupCaller,
|
|
/*block_id=*/0, block);
|
|
UpdateFeatureVectors(block.access_sequence_number_timeline,
|
|
block.access_timeline, label, &label_features,
|
|
&label_predictions);
|
|
};
|
|
TraverseBlocks(block_callback, &labels);
|
|
WriteCorrelationFeaturesToFile(label_str, label_features, label_predictions,
|
|
max_number_of_values);
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteCorrelationFeaturesToFile(
|
|
const std::string& label,
|
|
const std::map<std::string, Features>& label_features,
|
|
const std::map<std::string, Predictions>& label_predictions,
|
|
uint32_t max_number_of_values) const {
|
|
for (auto const& label_feature_vectors : label_features) {
|
|
const Features& past = label_feature_vectors.second;
|
|
auto it = label_predictions.find(label_feature_vectors.first);
|
|
assert(it != label_predictions.end());
|
|
const Predictions& future = it->second;
|
|
const std::string output_path = output_dir_ + "/" + label + "_" +
|
|
label_feature_vectors.first + "_" +
|
|
kFileNameSuffixCorrelation;
|
|
std::ofstream out(output_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header(
|
|
"num_accesses_since_last_access,elapsed_time_since_last_access,num_"
|
|
"past_accesses,num_accesses_till_next_access,elapsed_time_till_next_"
|
|
"access");
|
|
out << header << std::endl;
|
|
std::vector<uint32_t> indexes;
|
|
for (uint32_t i = 0; i < past.num_accesses_since_last_access.size(); i++) {
|
|
indexes.push_back(i);
|
|
}
|
|
RandomShuffle(indexes.begin(), indexes.end());
|
|
for (uint32_t i = 0; i < max_number_of_values && i < indexes.size(); i++) {
|
|
uint32_t rand_index = indexes[i];
|
|
out << std::to_string(past.num_accesses_since_last_access[rand_index])
|
|
<< ",";
|
|
out << std::to_string(past.elapsed_time_since_last_access[rand_index])
|
|
<< ",";
|
|
out << std::to_string(past.num_past_accesses[rand_index]) << ",";
|
|
out << std::to_string(future.num_accesses_till_next_access[rand_index])
|
|
<< ",";
|
|
out << std::to_string(future.elapsed_time_till_next_access[rand_index])
|
|
<< std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteCorrelationFeaturesForGet(
|
|
uint32_t max_number_of_values) const {
|
|
std::string label = "GetKeyInfo";
|
|
std::map<std::string, Features> label_features;
|
|
std::map<std::string, Predictions> label_predictions;
|
|
for (auto const& get_info : get_key_info_map_) {
|
|
const GetKeyInfo& info = get_info.second;
|
|
UpdateFeatureVectors(info.access_sequence_number_timeline,
|
|
info.access_timeline, label, &label_features,
|
|
&label_predictions);
|
|
}
|
|
WriteCorrelationFeaturesToFile(label, label_features, label_predictions,
|
|
max_number_of_values);
|
|
}
|
|
|
|
std::set<std::string> BlockCacheTraceAnalyzer::ParseLabelStr(
|
|
const std::string& label_str) const {
|
|
std::stringstream ss(label_str);
|
|
std::set<std::string> labels;
|
|
// label_str is in the form of "label1_label2_label3", e.g., cf_bt.
|
|
while (ss.good()) {
|
|
std::string label_name;
|
|
getline(ss, label_name, '_');
|
|
if (kGroupbyLabels.find(label_name) == kGroupbyLabels.end()) {
|
|
// Unknown label name.
|
|
fprintf(stderr, "Unknown label name %s, label string %s\n",
|
|
label_name.c_str(), label_str.c_str());
|
|
return {};
|
|
}
|
|
labels.insert(label_name);
|
|
}
|
|
return labels;
|
|
}
|
|
|
|
std::string BlockCacheTraceAnalyzer::BuildLabel(
|
|
const std::set<std::string>& labels, const std::string& cf_name,
|
|
uint64_t fd, uint32_t level, TraceType type, TableReaderCaller caller,
|
|
uint64_t block_key, const BlockAccessInfo& block) const {
|
|
std::map<std::string, std::string> label_value_map;
|
|
label_value_map[kGroupbyAll] = kGroupbyAll;
|
|
label_value_map[kGroupbyLevel] = std::to_string(level);
|
|
label_value_map[kGroupbyCaller] = caller_to_string(caller);
|
|
label_value_map[kGroupbySSTFile] = std::to_string(fd);
|
|
label_value_map[kGroupbyBlockType] = block_type_to_string(type);
|
|
label_value_map[kGroupbyColumnFamily] = cf_name;
|
|
label_value_map[kGroupbyBlock] = std::to_string(block_key);
|
|
label_value_map[kGroupbyTable] = std::to_string(block.table_id);
|
|
// Concatenate the label values.
|
|
std::string label;
|
|
for (auto const& l : labels) {
|
|
label += label_value_map[l];
|
|
label += "-";
|
|
}
|
|
if (!label.empty()) {
|
|
label.pop_back();
|
|
}
|
|
return label;
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::TraverseBlocks(
|
|
std::function<void(const std::string& /*cf_name*/, uint64_t /*fd*/,
|
|
uint32_t /*level*/, TraceType /*block_type*/,
|
|
const std::string& /*block_key*/,
|
|
uint64_t /*block_key_id*/,
|
|
const BlockAccessInfo& /*block_access_info*/)>
|
|
block_callback,
|
|
std::set<std::string>* labels) const {
|
|
for (auto const& cf_aggregates : cf_aggregates_map_) {
|
|
// Stats per column family.
|
|
const std::string& cf_name = cf_aggregates.first;
|
|
for (auto const& file_aggregates : cf_aggregates.second.fd_aggregates_map) {
|
|
// Stats per SST file.
|
|
const uint64_t fd = file_aggregates.first;
|
|
const uint32_t level = file_aggregates.second.level;
|
|
for (auto const& block_type_aggregates :
|
|
file_aggregates.second.block_type_aggregates_map) {
|
|
// Stats per block type.
|
|
const TraceType type = block_type_aggregates.first;
|
|
for (auto const& block_access_info :
|
|
block_type_aggregates.second.block_access_info_map) {
|
|
// Stats per block.
|
|
if (labels && block_access_info.second.table_id == 0 &&
|
|
labels->find(kGroupbyTable) != labels->end()) {
|
|
// We only know table id information for get requests.
|
|
return;
|
|
}
|
|
block_callback(cf_name, fd, level, type, block_access_info.first,
|
|
block_access_info.second.block_id,
|
|
block_access_info.second);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteGetSpatialLocality(
|
|
const std::string& label_str,
|
|
const std::vector<uint64_t>& percent_buckets) const {
|
|
std::set<std::string> labels = ParseLabelStr(label_str);
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_pnrefkeys_nblocks;
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_pnrefs_nblocks;
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_pndatasize_nblocks;
|
|
uint64_t nblocks = 0;
|
|
auto block_callback = [&](const std::string& cf_name, uint64_t fd,
|
|
uint32_t level, TraceType /*block_type*/,
|
|
const std::string& /*block_key*/,
|
|
uint64_t /*block_key_id*/,
|
|
const BlockAccessInfo& block) {
|
|
if (block.num_keys == 0) {
|
|
return;
|
|
}
|
|
uint64_t naccesses = 0;
|
|
for (auto const& key_access : block.key_num_access_map) {
|
|
for (auto const& caller_access : key_access.second) {
|
|
if (caller_access.first == TableReaderCaller::kUserGet) {
|
|
naccesses += caller_access.second;
|
|
}
|
|
}
|
|
}
|
|
const std::string label =
|
|
BuildLabel(labels, cf_name, fd, level, TraceType::kBlockTraceDataBlock,
|
|
TableReaderCaller::kUserGet, /*block_id=*/0, block);
|
|
|
|
const uint64_t percent_referenced_for_existing_keys =
|
|
static_cast<uint64_t>(std::max(
|
|
percent(block.key_num_access_map.size(), block.num_keys), 0.0));
|
|
const uint64_t percent_accesses_for_existing_keys =
|
|
static_cast<uint64_t>(std::max(
|
|
percent(block.num_referenced_key_exist_in_block, naccesses), 0.0));
|
|
const uint64_t percent_referenced_data_size = static_cast<uint64_t>(
|
|
std::max(percent(block.referenced_data_size, block.block_size), 0.0));
|
|
if (label_pnrefkeys_nblocks.find(label) == label_pnrefkeys_nblocks.end()) {
|
|
for (auto const& percent_bucket : percent_buckets) {
|
|
label_pnrefkeys_nblocks[label][percent_bucket] = 0;
|
|
label_pnrefs_nblocks[label][percent_bucket] = 0;
|
|
label_pndatasize_nblocks[label][percent_bucket] = 0;
|
|
}
|
|
}
|
|
label_pnrefkeys_nblocks[label]
|
|
.upper_bound(percent_referenced_for_existing_keys)
|
|
->second += 1;
|
|
label_pnrefs_nblocks[label]
|
|
.upper_bound(percent_accesses_for_existing_keys)
|
|
->second += 1;
|
|
label_pndatasize_nblocks[label]
|
|
.upper_bound(percent_referenced_data_size)
|
|
->second += 1;
|
|
nblocks += 1;
|
|
};
|
|
TraverseBlocks(block_callback, &labels);
|
|
WriteStatsToFile(label_str, percent_buckets, kFileNameSuffixPercentRefKeys,
|
|
label_pnrefkeys_nblocks, nblocks);
|
|
WriteStatsToFile(label_str, percent_buckets,
|
|
kFileNameSuffixPercentAccessesOnRefKeys,
|
|
label_pnrefs_nblocks, nblocks);
|
|
WriteStatsToFile(label_str, percent_buckets,
|
|
kFileNameSuffixPercentDataSizeOnRefKeys,
|
|
label_pndatasize_nblocks, nblocks);
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteAccessTimeline(const std::string& label_str,
|
|
uint64_t time_unit,
|
|
bool user_access_only) const {
|
|
std::set<std::string> labels = ParseLabelStr(label_str);
|
|
uint64_t start_time = port::kMaxUint64;
|
|
uint64_t end_time = 0;
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_access_timeline;
|
|
std::map<uint64_t, std::vector<std::string>> access_count_block_id_map;
|
|
|
|
auto block_callback = [&](const std::string& cf_name, uint64_t fd,
|
|
uint32_t level, TraceType type,
|
|
const std::string& /*block_key*/, uint64_t block_id,
|
|
const BlockAccessInfo& block) {
|
|
uint64_t naccesses = 0;
|
|
for (auto const& timeline : block.caller_num_accesses_timeline) {
|
|
const TableReaderCaller caller = timeline.first;
|
|
if (user_access_only && !is_user_access(caller)) {
|
|
continue;
|
|
}
|
|
const std::string label =
|
|
BuildLabel(labels, cf_name, fd, level, type, caller, block_id, block);
|
|
for (auto const& naccess : timeline.second) {
|
|
const uint64_t timestamp = naccess.first / time_unit;
|
|
const uint64_t num = naccess.second;
|
|
label_access_timeline[label][timestamp] += num;
|
|
start_time = std::min(start_time, timestamp);
|
|
end_time = std::max(end_time, timestamp);
|
|
naccesses += num;
|
|
}
|
|
}
|
|
if (naccesses > 0) {
|
|
access_count_block_id_map[naccesses].push_back(std::to_string(block_id));
|
|
}
|
|
};
|
|
TraverseBlocks(block_callback, &labels);
|
|
|
|
// We have label_access_timeline now. Write them into a file.
|
|
const std::string user_access_prefix =
|
|
user_access_only ? "user_access_only_" : "all_access_";
|
|
const std::string output_path = output_dir_ + "/" + user_access_prefix +
|
|
label_str + "_" + std::to_string(time_unit) +
|
|
"_" + kFileNameSuffixAccessTimeline;
|
|
std::ofstream out(output_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("time");
|
|
if (labels.find("block") != labels.end()) {
|
|
for (uint64_t now = start_time; now <= end_time; now++) {
|
|
header += ",";
|
|
header += std::to_string(now);
|
|
}
|
|
out << header << std::endl;
|
|
// Write the most frequently accessed blocks first.
|
|
for (auto naccess_it = access_count_block_id_map.rbegin();
|
|
naccess_it != access_count_block_id_map.rend(); naccess_it++) {
|
|
for (auto& block_id_it : naccess_it->second) {
|
|
std::string row(block_id_it);
|
|
for (uint64_t now = start_time; now <= end_time; now++) {
|
|
auto it = label_access_timeline[block_id_it].find(now);
|
|
row += ",";
|
|
if (it != label_access_timeline[block_id_it].end()) {
|
|
row += std::to_string(it->second);
|
|
} else {
|
|
row += "0";
|
|
}
|
|
}
|
|
out << row << std::endl;
|
|
}
|
|
}
|
|
out.close();
|
|
return;
|
|
}
|
|
for (uint64_t now = start_time; now <= end_time; now++) {
|
|
header += ",";
|
|
header += std::to_string(now);
|
|
}
|
|
out << header << std::endl;
|
|
for (auto const& label : label_access_timeline) {
|
|
std::string row(label.first);
|
|
for (uint64_t now = start_time; now <= end_time; now++) {
|
|
auto it = label.second.find(now);
|
|
row += ",";
|
|
if (it != label.second.end()) {
|
|
row += std::to_string(it->second);
|
|
} else {
|
|
row += "0";
|
|
}
|
|
}
|
|
out << row << std::endl;
|
|
}
|
|
|
|
out.close();
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteReuseDistance(
|
|
const std::string& label_str,
|
|
const std::vector<uint64_t>& distance_buckets) const {
|
|
std::set<std::string> labels = ParseLabelStr(label_str);
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_distance_num_reuses;
|
|
uint64_t total_num_reuses = 0;
|
|
auto block_callback = [&](const std::string& cf_name, uint64_t fd,
|
|
uint32_t level, TraceType type,
|
|
const std::string& /*block_key*/, uint64_t block_id,
|
|
const BlockAccessInfo& block) {
|
|
const std::string label = BuildLabel(
|
|
labels, cf_name, fd, level, type,
|
|
TableReaderCaller::kMaxBlockCacheLookupCaller, block_id, block);
|
|
if (label_distance_num_reuses.find(label) ==
|
|
label_distance_num_reuses.end()) {
|
|
// The first time we encounter this label.
|
|
for (auto const& distance_bucket : distance_buckets) {
|
|
label_distance_num_reuses[label][distance_bucket] = 0;
|
|
}
|
|
}
|
|
for (auto const& reuse_distance : block.reuse_distance_count) {
|
|
label_distance_num_reuses[label]
|
|
.upper_bound(reuse_distance.first)
|
|
->second += reuse_distance.second;
|
|
total_num_reuses += reuse_distance.second;
|
|
}
|
|
};
|
|
TraverseBlocks(block_callback, &labels);
|
|
// We have label_naccesses and label_distance_num_reuses now. Write them into
|
|
// a file.
|
|
const std::string output_path =
|
|
output_dir_ + "/" + label_str + "_reuse_distance";
|
|
std::ofstream out(output_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("bucket");
|
|
for (auto const& label_it : label_distance_num_reuses) {
|
|
header += ",";
|
|
header += label_it.first;
|
|
}
|
|
out << header << std::endl;
|
|
for (auto const& bucket : distance_buckets) {
|
|
std::string row(std::to_string(bucket));
|
|
for (auto const& label_it : label_distance_num_reuses) {
|
|
auto const& it = label_it.second.find(bucket);
|
|
assert(it != label_it.second.end());
|
|
row += ",";
|
|
row += std::to_string(percent(it->second, total_num_reuses));
|
|
}
|
|
out << row << std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::UpdateReuseIntervalStats(
|
|
const std::string& label, const std::vector<uint64_t>& time_buckets,
|
|
const std::map<uint64_t, uint64_t> timeline,
|
|
std::map<std::string, std::map<uint64_t, uint64_t>>* label_time_num_reuses,
|
|
uint64_t* total_num_reuses) const {
|
|
assert(label_time_num_reuses);
|
|
assert(total_num_reuses);
|
|
if (label_time_num_reuses->find(label) == label_time_num_reuses->end()) {
|
|
// The first time we encounter this label.
|
|
for (auto const& time_bucket : time_buckets) {
|
|
(*label_time_num_reuses)[label][time_bucket] = 0;
|
|
}
|
|
}
|
|
auto it = timeline.begin();
|
|
uint64_t prev_timestamp = it->first;
|
|
const uint64_t prev_num = it->second;
|
|
it++;
|
|
// Reused within one second.
|
|
if (prev_num > 1) {
|
|
(*label_time_num_reuses)[label].upper_bound(0)->second += prev_num - 1;
|
|
*total_num_reuses += prev_num - 1;
|
|
}
|
|
while (it != timeline.end()) {
|
|
const uint64_t timestamp = it->first;
|
|
const uint64_t num = it->second;
|
|
const uint64_t reuse_interval = timestamp - prev_timestamp;
|
|
(*label_time_num_reuses)[label].upper_bound(reuse_interval)->second += 1;
|
|
if (num > 1) {
|
|
(*label_time_num_reuses)[label].upper_bound(0)->second += num - 1;
|
|
}
|
|
prev_timestamp = timestamp;
|
|
*total_num_reuses += num;
|
|
it++;
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteStatsToFile(
|
|
const std::string& label_str, const std::vector<uint64_t>& time_buckets,
|
|
const std::string& filename_suffix,
|
|
const std::map<std::string, std::map<uint64_t, uint64_t>>& label_data,
|
|
uint64_t ntotal) const {
|
|
const std::string output_path =
|
|
output_dir_ + "/" + label_str + "_" + filename_suffix;
|
|
std::ofstream out(output_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("bucket");
|
|
for (auto const& label_it : label_data) {
|
|
header += ",";
|
|
header += label_it.first;
|
|
}
|
|
out << header << std::endl;
|
|
for (auto const& bucket : time_buckets) {
|
|
std::string row(std::to_string(bucket));
|
|
for (auto const& label_it : label_data) {
|
|
auto const& it = label_it.second.find(bucket);
|
|
assert(it != label_it.second.end());
|
|
row += ",";
|
|
row += std::to_string(percent(it->second, ntotal));
|
|
}
|
|
out << row << std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteReuseInterval(
|
|
const std::string& label_str,
|
|
const std::vector<uint64_t>& time_buckets) const {
|
|
std::set<std::string> labels = ParseLabelStr(label_str);
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_time_num_reuses;
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_avg_reuse_nblocks;
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_avg_reuse_naccesses;
|
|
|
|
uint64_t total_num_reuses = 0;
|
|
uint64_t total_nblocks = 0;
|
|
uint64_t total_accesses = 0;
|
|
auto block_callback = [&](const std::string& cf_name, uint64_t fd,
|
|
uint32_t level, TraceType type,
|
|
const std::string& /*block_key*/, uint64_t block_id,
|
|
const BlockAccessInfo& block) {
|
|
total_nblocks++;
|
|
total_accesses += block.num_accesses;
|
|
uint64_t avg_reuse_interval = 0;
|
|
if (block.num_accesses > 1) {
|
|
avg_reuse_interval = ((block.last_access_time - block.first_access_time) /
|
|
kMicrosInSecond) /
|
|
block.num_accesses;
|
|
} else {
|
|
avg_reuse_interval = port::kMaxUint64 - 1;
|
|
}
|
|
if (labels.find(kGroupbyCaller) != labels.end()) {
|
|
for (auto const& timeline : block.caller_num_accesses_timeline) {
|
|
const TableReaderCaller caller = timeline.first;
|
|
const std::string label = BuildLabel(labels, cf_name, fd, level, type,
|
|
caller, block_id, block);
|
|
UpdateReuseIntervalStats(label, time_buckets, timeline.second,
|
|
&label_time_num_reuses, &total_num_reuses);
|
|
}
|
|
return;
|
|
}
|
|
// Does not group by caller so we need to flatten the access timeline.
|
|
const std::string label = BuildLabel(
|
|
labels, cf_name, fd, level, type,
|
|
TableReaderCaller::kMaxBlockCacheLookupCaller, block_id, block);
|
|
std::map<uint64_t, uint64_t> timeline;
|
|
for (auto const& caller_timeline : block.caller_num_accesses_timeline) {
|
|
for (auto const& time_naccess : caller_timeline.second) {
|
|
timeline[time_naccess.first] += time_naccess.second;
|
|
}
|
|
}
|
|
UpdateReuseIntervalStats(label, time_buckets, timeline,
|
|
&label_time_num_reuses, &total_num_reuses);
|
|
if (label_avg_reuse_nblocks.find(label) == label_avg_reuse_nblocks.end()) {
|
|
for (auto const& time_bucket : time_buckets) {
|
|
label_avg_reuse_nblocks[label][time_bucket] = 0;
|
|
label_avg_reuse_naccesses[label][time_bucket] = 0;
|
|
}
|
|
}
|
|
label_avg_reuse_nblocks[label].upper_bound(avg_reuse_interval)->second += 1;
|
|
label_avg_reuse_naccesses[label].upper_bound(avg_reuse_interval)->second +=
|
|
block.num_accesses;
|
|
};
|
|
TraverseBlocks(block_callback, &labels);
|
|
|
|
// Write the stats into files.
|
|
WriteStatsToFile(label_str, time_buckets, kFileNameSuffixReuseInterval,
|
|
label_time_num_reuses, total_num_reuses);
|
|
WriteStatsToFile(label_str, time_buckets, kFileNameSuffixAvgReuseInterval,
|
|
label_avg_reuse_nblocks, total_nblocks);
|
|
WriteStatsToFile(label_str, time_buckets,
|
|
kFileNameSuffixAvgReuseIntervalNaccesses,
|
|
label_avg_reuse_naccesses, total_accesses);
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteReuseLifetime(
|
|
const std::string& label_str,
|
|
const std::vector<uint64_t>& time_buckets) const {
|
|
std::set<std::string> labels = ParseLabelStr(label_str);
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> label_lifetime_nblocks;
|
|
uint64_t total_nblocks = 0;
|
|
auto block_callback = [&](const std::string& cf_name, uint64_t fd,
|
|
uint32_t level, TraceType type,
|
|
const std::string& /*block_key*/, uint64_t block_id,
|
|
const BlockAccessInfo& block) {
|
|
uint64_t lifetime = 0;
|
|
if (block.num_accesses > 1) {
|
|
lifetime =
|
|
(block.last_access_time - block.first_access_time) / kMicrosInSecond;
|
|
} else {
|
|
lifetime = port::kMaxUint64 - 1;
|
|
}
|
|
const std::string label = BuildLabel(
|
|
labels, cf_name, fd, level, type,
|
|
TableReaderCaller::kMaxBlockCacheLookupCaller, block_id, block);
|
|
|
|
if (label_lifetime_nblocks.find(label) == label_lifetime_nblocks.end()) {
|
|
// The first time we encounter this label.
|
|
for (auto const& time_bucket : time_buckets) {
|
|
label_lifetime_nblocks[label][time_bucket] = 0;
|
|
}
|
|
}
|
|
label_lifetime_nblocks[label].upper_bound(lifetime)->second += 1;
|
|
total_nblocks += 1;
|
|
};
|
|
TraverseBlocks(block_callback, &labels);
|
|
WriteStatsToFile(label_str, time_buckets, kFileNameSuffixReuseLifetime,
|
|
label_lifetime_nblocks, total_nblocks);
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteBlockReuseTimeline(
|
|
const uint64_t reuse_window, bool user_access_only, TraceType block_type) const {
|
|
// A map from block key to an array of bools that states whether a block is
|
|
// accessed in a time window.
|
|
std::map<uint64_t, std::vector<bool>> block_accessed;
|
|
const uint64_t trace_duration =
|
|
trace_end_timestamp_in_seconds_ - trace_start_timestamp_in_seconds_;
|
|
const uint64_t reuse_vector_size = (trace_duration / reuse_window);
|
|
if (reuse_vector_size < 2) {
|
|
// The reuse window is less than 2. We cannot calculate the reused
|
|
// percentage of blocks.
|
|
return;
|
|
}
|
|
auto block_callback = [&](const std::string& /*cf_name*/, uint64_t /*fd*/,
|
|
uint32_t /*level*/, TraceType /*type*/,
|
|
const std::string& /*block_key*/, uint64_t block_id,
|
|
const BlockAccessInfo& block) {
|
|
if (block_accessed.find(block_id) == block_accessed.end()) {
|
|
block_accessed[block_id].resize(reuse_vector_size);
|
|
for (uint64_t i = 0; i < reuse_vector_size; i++) {
|
|
block_accessed[block_id][i] = false;
|
|
}
|
|
}
|
|
for (auto const& caller_num : block.caller_num_accesses_timeline) {
|
|
const TableReaderCaller caller = caller_num.first;
|
|
for (auto const& timeline : caller_num.second) {
|
|
const uint64_t timestamp = timeline.first;
|
|
const uint64_t elapsed_time =
|
|
timestamp - trace_start_timestamp_in_seconds_;
|
|
if (!user_access_only || is_user_access(caller)) {
|
|
uint64_t index =
|
|
std::min(elapsed_time / reuse_window, reuse_vector_size - 1);
|
|
block_accessed[block_id][index] = true;
|
|
}
|
|
}
|
|
}
|
|
};
|
|
TraverseBlocks(block_callback);
|
|
|
|
// A cell is the number of blocks accessed in a reuse window.
|
|
std::unique_ptr<uint64_t[]> reuse_table(new uint64_t[reuse_vector_size * reuse_vector_size]);
|
|
for (uint64_t start_time = 0; start_time < reuse_vector_size; start_time++) {
|
|
// Initialize the reuse_table.
|
|
for (uint64_t i = 0; i < reuse_vector_size; i++) {
|
|
reuse_table[start_time * reuse_vector_size + i] = 0;
|
|
}
|
|
// Examine all blocks.
|
|
for (auto const& block : block_accessed) {
|
|
for (uint64_t i = start_time; i < reuse_vector_size; i++) {
|
|
if (block.second[start_time] && block.second[i]) {
|
|
// This block is accessed at start time and at the current time. We
|
|
// increment reuse_table[start_time][i] since it is reused at the ith
|
|
// window.
|
|
reuse_table[start_time * reuse_vector_size + i]++;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
const std::string user_access_prefix =
|
|
user_access_only ? "_user_access_only_" : "_all_access_";
|
|
const std::string output_path =
|
|
output_dir_ + "/" + block_type_to_string(block_type) +
|
|
user_access_prefix + std::to_string(reuse_window) + "_" +
|
|
kFileNameSuffixAccessReuseBlocksTimeline;
|
|
std::ofstream out(output_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("start_time");
|
|
for (uint64_t start_time = 0; start_time < reuse_vector_size; start_time++) {
|
|
header += ",";
|
|
header += std::to_string(start_time);
|
|
}
|
|
out << header << std::endl;
|
|
for (uint64_t start_time = 0; start_time < reuse_vector_size; start_time++) {
|
|
std::string row(std::to_string(start_time * reuse_window));
|
|
for (uint64_t j = 0; j < reuse_vector_size; j++) {
|
|
row += ",";
|
|
if (j < start_time) {
|
|
row += "100.0";
|
|
} else {
|
|
row += std::to_string(percent(reuse_table[start_time * reuse_vector_size + j],
|
|
reuse_table[start_time * reuse_vector_size + start_time]));
|
|
}
|
|
}
|
|
out << row << std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
|
|
std::string BlockCacheTraceAnalyzer::OutputPercentAccessStats(
|
|
uint64_t total_accesses,
|
|
const std::map<std::string, uint64_t>& cf_access_count) const {
|
|
std::string row;
|
|
for (auto const& cf_aggregates : cf_aggregates_map_) {
|
|
const std::string& cf_name = cf_aggregates.first;
|
|
const auto& naccess = cf_access_count.find(cf_name);
|
|
row += ",";
|
|
if (naccess != cf_access_count.end()) {
|
|
row += std::to_string(percent(naccess->second, total_accesses));
|
|
} else {
|
|
row += "0";
|
|
}
|
|
}
|
|
return row;
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WritePercentAccessSummaryStats() const {
|
|
std::map<TableReaderCaller, std::map<std::string, uint64_t>>
|
|
caller_cf_accesses;
|
|
uint64_t total_accesses = 0;
|
|
auto block_callback =
|
|
[&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/,
|
|
TraceType /*type*/, const std::string& /*block_key*/,
|
|
uint64_t /*block_id*/, const BlockAccessInfo& block) {
|
|
for (auto const& caller_num : block.caller_num_access_map) {
|
|
const TableReaderCaller caller = caller_num.first;
|
|
const uint64_t naccess = caller_num.second;
|
|
caller_cf_accesses[caller][cf_name] += naccess;
|
|
total_accesses += naccess;
|
|
}
|
|
};
|
|
TraverseBlocks(block_callback);
|
|
|
|
const std::string output_path =
|
|
output_dir_ + "/" + kFileNameSuffixPercentOfAccessSummary;
|
|
std::ofstream out(output_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("caller");
|
|
for (auto const& cf_name : cf_aggregates_map_) {
|
|
header += ",";
|
|
header += cf_name.first;
|
|
}
|
|
out << header << std::endl;
|
|
for (auto const& cf_naccess_it : caller_cf_accesses) {
|
|
const TableReaderCaller caller = cf_naccess_it.first;
|
|
std::string row;
|
|
row += caller_to_string(caller);
|
|
row += OutputPercentAccessStats(total_accesses, cf_naccess_it.second);
|
|
out << row << std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteDetailedPercentAccessSummaryStats(
|
|
TableReaderCaller analyzing_caller) const {
|
|
std::map<uint32_t, std::map<std::string, uint64_t>> level_cf_accesses;
|
|
std::map<TraceType, std::map<std::string, uint64_t>> bt_cf_accesses;
|
|
uint64_t total_accesses = 0;
|
|
auto block_callback =
|
|
[&](const std::string& cf_name, uint64_t /*fd*/, uint32_t level,
|
|
TraceType type, const std::string& /*block_key*/,
|
|
uint64_t /*block_id*/, const BlockAccessInfo& block) {
|
|
for (auto const& caller_num : block.caller_num_access_map) {
|
|
const TableReaderCaller caller = caller_num.first;
|
|
if (caller == analyzing_caller) {
|
|
const uint64_t naccess = caller_num.second;
|
|
level_cf_accesses[level][cf_name] += naccess;
|
|
bt_cf_accesses[type][cf_name] += naccess;
|
|
total_accesses += naccess;
|
|
}
|
|
}
|
|
};
|
|
TraverseBlocks(block_callback);
|
|
{
|
|
const std::string output_path =
|
|
output_dir_ + "/" + caller_to_string(analyzing_caller) + "_level_" +
|
|
kFileNameSuffixPercentOfAccessSummary;
|
|
std::ofstream out(output_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("level");
|
|
for (auto const& cf_name : cf_aggregates_map_) {
|
|
header += ",";
|
|
header += cf_name.first;
|
|
}
|
|
out << header << std::endl;
|
|
for (auto const& level_naccess_it : level_cf_accesses) {
|
|
const uint32_t level = level_naccess_it.first;
|
|
std::string row;
|
|
row += std::to_string(level);
|
|
row += OutputPercentAccessStats(total_accesses, level_naccess_it.second);
|
|
out << row << std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
{
|
|
const std::string output_path =
|
|
output_dir_ + "/" + caller_to_string(analyzing_caller) + "_bt_" +
|
|
kFileNameSuffixPercentOfAccessSummary;
|
|
std::ofstream out(output_path);
|
|
if (!out.is_open()) {
|
|
return;
|
|
}
|
|
std::string header("bt");
|
|
for (auto const& cf_name : cf_aggregates_map_) {
|
|
header += ",";
|
|
header += cf_name.first;
|
|
}
|
|
out << header << std::endl;
|
|
for (auto const& bt_naccess_it : bt_cf_accesses) {
|
|
const TraceType bt = bt_naccess_it.first;
|
|
std::string row;
|
|
row += block_type_to_string(bt);
|
|
row += OutputPercentAccessStats(total_accesses, bt_naccess_it.second);
|
|
out << row << std::endl;
|
|
}
|
|
out.close();
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::WriteAccessCountSummaryStats(
|
|
const std::vector<uint64_t>& access_count_buckets,
|
|
bool user_access_only) const {
|
|
// x: buckets.
|
|
// y: # of accesses.
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> bt_access_nblocks;
|
|
std::map<std::string, std::map<uint64_t, uint64_t>> cf_access_nblocks;
|
|
uint64_t total_nblocks = 0;
|
|
auto block_callback =
|
|
[&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/,
|
|
TraceType type, const std::string& /*block_key*/,
|
|
uint64_t /*block_id*/, const BlockAccessInfo& block) {
|
|
const std::string type_str = block_type_to_string(type);
|
|
if (cf_access_nblocks.find(cf_name) == cf_access_nblocks.end()) {
|
|
// initialize.
|
|
for (auto& access : access_count_buckets) {
|
|
cf_access_nblocks[cf_name][access] = 0;
|
|
}
|
|
}
|
|
if (bt_access_nblocks.find(type_str) == bt_access_nblocks.end()) {
|
|
// initialize.
|
|
for (auto& access : access_count_buckets) {
|
|
bt_access_nblocks[type_str][access] = 0;
|
|
}
|
|
}
|
|
uint64_t naccesses = 0;
|
|
for (auto const& caller_access : block.caller_num_access_map) {
|
|
if (!user_access_only || is_user_access(caller_access.first)) {
|
|
naccesses += caller_access.second;
|
|
}
|
|
}
|
|
if (naccesses == 0) {
|
|
return;
|
|
}
|
|
total_nblocks += 1;
|
|
bt_access_nblocks[type_str].upper_bound(naccesses)->second += 1;
|
|
cf_access_nblocks[cf_name].upper_bound(naccesses)->second += 1;
|
|
};
|
|
TraverseBlocks(block_callback);
|
|
const std::string user_access_prefix =
|
|
user_access_only ? "user_access_only_" : "all_access_";
|
|
WriteStatsToFile("cf", access_count_buckets,
|
|
user_access_prefix + kFileNameSuffixAccessCountSummary,
|
|
cf_access_nblocks, total_nblocks);
|
|
WriteStatsToFile("bt", access_count_buckets,
|
|
user_access_prefix + kFileNameSuffixAccessCountSummary,
|
|
bt_access_nblocks, total_nblocks);
|
|
}
|
|
|
|
BlockCacheTraceAnalyzer::BlockCacheTraceAnalyzer(
|
|
const std::string& trace_file_path, const std::string& output_dir,
|
|
const std::string& human_readable_trace_file_path,
|
|
bool compute_reuse_distance, bool mrc_only,
|
|
bool is_human_readable_trace_file,
|
|
std::unique_ptr<BlockCacheTraceSimulator>&& cache_simulator)
|
|
: env_(ROCKSDB_NAMESPACE::Env::Default()),
|
|
trace_file_path_(trace_file_path),
|
|
output_dir_(output_dir),
|
|
human_readable_trace_file_path_(human_readable_trace_file_path),
|
|
compute_reuse_distance_(compute_reuse_distance),
|
|
mrc_only_(mrc_only),
|
|
is_human_readable_trace_file_(is_human_readable_trace_file),
|
|
cache_simulator_(std::move(cache_simulator)) {}
|
|
|
|
void BlockCacheTraceAnalyzer::ComputeReuseDistance(
|
|
BlockAccessInfo* info) const {
|
|
assert(info);
|
|
if (info->num_accesses == 0) {
|
|
return;
|
|
}
|
|
uint64_t reuse_distance = 0;
|
|
for (auto const& block_key : info->unique_blocks_since_last_access) {
|
|
auto const& it = block_info_map_.find(block_key);
|
|
// This block must exist.
|
|
assert(it != block_info_map_.end());
|
|
reuse_distance += it->second->block_size;
|
|
}
|
|
info->reuse_distance_count[reuse_distance] += 1;
|
|
// We clear this hash set since this is the second access on this block.
|
|
info->unique_blocks_since_last_access.clear();
|
|
}
|
|
|
|
Status BlockCacheTraceAnalyzer::RecordAccess(
|
|
const BlockCacheTraceRecord& access) {
|
|
ColumnFamilyAccessInfoAggregate& cf_aggr = cf_aggregates_map_[access.cf_name];
|
|
SSTFileAccessInfoAggregate& file_aggr =
|
|
cf_aggr.fd_aggregates_map[access.sst_fd_number];
|
|
file_aggr.level = access.level;
|
|
BlockTypeAccessInfoAggregate& block_type_aggr =
|
|
file_aggr.block_type_aggregates_map[access.block_type];
|
|
if (block_type_aggr.block_access_info_map.find(access.block_key) ==
|
|
block_type_aggr.block_access_info_map.end()) {
|
|
block_type_aggr.block_access_info_map[access.block_key].block_id =
|
|
unique_block_id_;
|
|
unique_block_id_++;
|
|
}
|
|
BlockAccessInfo& block_access_info =
|
|
block_type_aggr.block_access_info_map[access.block_key];
|
|
if (compute_reuse_distance_) {
|
|
ComputeReuseDistance(&block_access_info);
|
|
}
|
|
block_access_info.AddAccess(access, access_sequence_number_);
|
|
block_info_map_[access.block_key] = &block_access_info;
|
|
uint64_t get_key_id = 0;
|
|
if (access.caller == TableReaderCaller::kUserGet &&
|
|
access.get_id != BlockCacheTraceHelper::kReservedGetId) {
|
|
std::string user_key = ExtractUserKey(access.referenced_key).ToString();
|
|
if (get_key_info_map_.find(user_key) == get_key_info_map_.end()) {
|
|
get_key_info_map_[user_key].key_id = unique_get_key_id_;
|
|
unique_get_key_id_++;
|
|
}
|
|
get_key_id = get_key_info_map_[user_key].key_id;
|
|
get_key_info_map_[user_key].AddAccess(access, access_sequence_number_);
|
|
}
|
|
|
|
if (compute_reuse_distance_) {
|
|
// Add this block to all existing blocks.
|
|
for (auto& cf_aggregates : cf_aggregates_map_) {
|
|
for (auto& file_aggregates : cf_aggregates.second.fd_aggregates_map) {
|
|
for (auto& block_type_aggregates :
|
|
file_aggregates.second.block_type_aggregates_map) {
|
|
for (auto& existing_block :
|
|
block_type_aggregates.second.block_access_info_map) {
|
|
existing_block.second.unique_blocks_since_last_access.insert(
|
|
access.block_key);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return human_readable_trace_writer_.WriteHumanReadableTraceRecord(
|
|
access, block_access_info.block_id, get_key_id);
|
|
}
|
|
|
|
Status BlockCacheTraceAnalyzer::Analyze() {
|
|
SystemClock* clock = env_->GetSystemClock().get();
|
|
std::unique_ptr<BlockCacheTraceReader> reader;
|
|
Status s = Status::OK();
|
|
if (is_human_readable_trace_file_) {
|
|
reader.reset(new BlockCacheHumanReadableTraceReader(trace_file_path_));
|
|
} else {
|
|
std::unique_ptr<TraceReader> trace_reader;
|
|
s = NewFileTraceReader(env_, EnvOptions(), trace_file_path_, &trace_reader);
|
|
if (!s.ok()) {
|
|
return s;
|
|
}
|
|
reader.reset(new BlockCacheTraceReader(std::move(trace_reader)));
|
|
s = reader->ReadHeader(&header_);
|
|
if (!s.ok()) {
|
|
return s;
|
|
}
|
|
}
|
|
if (!human_readable_trace_file_path_.empty()) {
|
|
s = human_readable_trace_writer_.NewWritableFile(
|
|
human_readable_trace_file_path_, env_);
|
|
if (!s.ok()) {
|
|
return s;
|
|
}
|
|
}
|
|
uint64_t start = clock->NowMicros();
|
|
uint64_t time_interval = 0;
|
|
while (s.ok()) {
|
|
BlockCacheTraceRecord access;
|
|
s = reader->ReadAccess(&access);
|
|
if (!s.ok()) {
|
|
break;
|
|
}
|
|
if (!mrc_only_) {
|
|
s = RecordAccess(access);
|
|
if (!s.ok()) {
|
|
break;
|
|
}
|
|
}
|
|
if (trace_start_timestamp_in_seconds_ == 0) {
|
|
trace_start_timestamp_in_seconds_ =
|
|
access.access_timestamp / kMicrosInSecond;
|
|
}
|
|
trace_end_timestamp_in_seconds_ = access.access_timestamp / kMicrosInSecond;
|
|
miss_ratio_stats_.UpdateMetrics(access.access_timestamp,
|
|
is_user_access(access.caller),
|
|
access.is_cache_hit == Boolean::kFalse);
|
|
if (cache_simulator_) {
|
|
cache_simulator_->Access(access);
|
|
}
|
|
access_sequence_number_++;
|
|
uint64_t now = clock->NowMicros();
|
|
uint64_t duration = (now - start) / kMicrosInSecond;
|
|
if (duration > 10 * time_interval) {
|
|
uint64_t trace_duration =
|
|
trace_end_timestamp_in_seconds_ - trace_start_timestamp_in_seconds_;
|
|
fprintf(stdout,
|
|
"Running for %" PRIu64 " seconds: Processed %" PRIu64
|
|
" records/second. Trace duration %" PRIu64
|
|
" seconds. Observed miss ratio %.2f\n",
|
|
duration, duration > 0 ? access_sequence_number_ / duration : 0,
|
|
trace_duration, miss_ratio_stats_.miss_ratio());
|
|
time_interval++;
|
|
}
|
|
}
|
|
uint64_t now = clock->NowMicros();
|
|
uint64_t duration = (now - start) / kMicrosInSecond;
|
|
uint64_t trace_duration =
|
|
trace_end_timestamp_in_seconds_ - trace_start_timestamp_in_seconds_;
|
|
fprintf(stdout,
|
|
"Running for %" PRIu64 " seconds: Processed %" PRIu64
|
|
" records/second. Trace duration %" PRIu64
|
|
" seconds. Observed miss ratio %.2f\n",
|
|
duration, duration > 0 ? access_sequence_number_ / duration : 0,
|
|
trace_duration, miss_ratio_stats_.miss_ratio());
|
|
return s;
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::PrintBlockSizeStats() const {
|
|
HistogramStat bs_stats;
|
|
std::map<TraceType, HistogramStat> bt_stats_map;
|
|
std::map<std::string, std::map<TraceType, HistogramStat>> cf_bt_stats_map;
|
|
auto block_callback =
|
|
[&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/,
|
|
TraceType type, const std::string& /*block_key*/,
|
|
uint64_t /*block_id*/, const BlockAccessInfo& block) {
|
|
if (block.block_size == 0) {
|
|
// Block size may be 0 when 1) compaction observes a cache miss and
|
|
// does not insert the missing block into the cache again. 2)
|
|
// fetching filter blocks in SST files at the last level.
|
|
return;
|
|
}
|
|
bs_stats.Add(block.block_size);
|
|
bt_stats_map[type].Add(block.block_size);
|
|
cf_bt_stats_map[cf_name][type].Add(block.block_size);
|
|
};
|
|
TraverseBlocks(block_callback);
|
|
fprintf(stdout, "Block size stats: \n%s", bs_stats.ToString().c_str());
|
|
for (auto const& bt_stats : bt_stats_map) {
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout, "Block size stats for block type %s: \n%s",
|
|
block_type_to_string(bt_stats.first).c_str(),
|
|
bt_stats.second.ToString().c_str());
|
|
}
|
|
for (auto const& cf_bt_stats : cf_bt_stats_map) {
|
|
const std::string& cf_name = cf_bt_stats.first;
|
|
for (auto const& bt_stats : cf_bt_stats.second) {
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout,
|
|
"Block size stats for column family %s and block type %s: \n%s",
|
|
cf_name.c_str(), block_type_to_string(bt_stats.first).c_str(),
|
|
bt_stats.second.ToString().c_str());
|
|
}
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::PrintAccessCountStats(bool user_access_only,
|
|
uint32_t bottom_k,
|
|
uint32_t top_k) const {
|
|
HistogramStat access_stats;
|
|
std::map<TraceType, HistogramStat> bt_stats_map;
|
|
std::map<std::string, std::map<TraceType, HistogramStat>> cf_bt_stats_map;
|
|
std::map<uint64_t, std::vector<std::string>> access_count_blocks;
|
|
auto block_callback = [&](const std::string& cf_name, uint64_t /*fd*/,
|
|
uint32_t /*level*/, TraceType type,
|
|
const std::string& block_key, uint64_t /*block_id*/,
|
|
const BlockAccessInfo& block) {
|
|
uint64_t naccesses = 0;
|
|
for (auto const& caller_access : block.caller_num_access_map) {
|
|
if (!user_access_only || is_user_access(caller_access.first)) {
|
|
naccesses += caller_access.second;
|
|
}
|
|
}
|
|
if (naccesses == 0) {
|
|
return;
|
|
}
|
|
if (type == TraceType::kBlockTraceDataBlock) {
|
|
access_count_blocks[naccesses].push_back(block_key);
|
|
}
|
|
access_stats.Add(naccesses);
|
|
bt_stats_map[type].Add(naccesses);
|
|
cf_bt_stats_map[cf_name][type].Add(naccesses);
|
|
};
|
|
TraverseBlocks(block_callback);
|
|
fprintf(stdout,
|
|
"Block access count stats: The number of accesses per block. %s\n%s",
|
|
user_access_only ? "User accesses only" : "All accesses",
|
|
access_stats.ToString().c_str());
|
|
uint32_t bottom_k_index = 0;
|
|
for (auto naccess_it = access_count_blocks.begin();
|
|
naccess_it != access_count_blocks.end(); naccess_it++) {
|
|
bottom_k_index++;
|
|
if (bottom_k_index >= bottom_k) {
|
|
break;
|
|
}
|
|
std::map<TableReaderCaller, uint64_t> caller_naccesses;
|
|
uint64_t naccesses = 0;
|
|
for (auto const& block_id : naccess_it->second) {
|
|
BlockAccessInfo* block = block_info_map_.find(block_id)->second;
|
|
for (auto const& caller_access : block->caller_num_access_map) {
|
|
if (!user_access_only || is_user_access(caller_access.first)) {
|
|
caller_naccesses[caller_access.first] += caller_access.second;
|
|
naccesses += caller_access.second;
|
|
}
|
|
}
|
|
}
|
|
std::string statistics("Caller:");
|
|
for (auto const& caller_naccessess_it : caller_naccesses) {
|
|
statistics += caller_to_string(caller_naccessess_it.first);
|
|
statistics += ":";
|
|
statistics +=
|
|
std::to_string(percent(caller_naccessess_it.second, naccesses));
|
|
statistics += ",";
|
|
}
|
|
fprintf(stdout,
|
|
"Bottom %" PRIu32 " access count. Access count=%" PRIu64
|
|
" nblocks=%" ROCKSDB_PRIszt " %s\n",
|
|
bottom_k, naccess_it->first, naccess_it->second.size(),
|
|
statistics.c_str());
|
|
}
|
|
|
|
uint32_t top_k_index = 0;
|
|
for (auto naccess_it = access_count_blocks.rbegin();
|
|
naccess_it != access_count_blocks.rend(); naccess_it++) {
|
|
top_k_index++;
|
|
if (top_k_index >= top_k) {
|
|
break;
|
|
}
|
|
for (auto const& block_id : naccess_it->second) {
|
|
BlockAccessInfo* block = block_info_map_.find(block_id)->second;
|
|
std::string statistics("Caller:");
|
|
uint64_t naccesses = 0;
|
|
for (auto const& caller_access : block->caller_num_access_map) {
|
|
if (!user_access_only || is_user_access(caller_access.first)) {
|
|
naccesses += caller_access.second;
|
|
}
|
|
}
|
|
assert(naccesses > 0);
|
|
for (auto const& caller_access : block->caller_num_access_map) {
|
|
if (!user_access_only || is_user_access(caller_access.first)) {
|
|
statistics += ",";
|
|
statistics += caller_to_string(caller_access.first);
|
|
statistics += ":";
|
|
statistics +=
|
|
std::to_string(percent(caller_access.second, naccesses));
|
|
}
|
|
}
|
|
uint64_t ref_keys_accesses = 0;
|
|
uint64_t ref_keys_does_not_exist_accesses = 0;
|
|
for (auto const& ref_key_caller_access : block->key_num_access_map) {
|
|
for (auto const& caller_access : ref_key_caller_access.second) {
|
|
if (!user_access_only || is_user_access(caller_access.first)) {
|
|
ref_keys_accesses += caller_access.second;
|
|
}
|
|
}
|
|
}
|
|
for (auto const& ref_key_caller_access :
|
|
block->non_exist_key_num_access_map) {
|
|
for (auto const& caller_access : ref_key_caller_access.second) {
|
|
if (!user_access_only || is_user_access(caller_access.first)) {
|
|
ref_keys_does_not_exist_accesses += caller_access.second;
|
|
}
|
|
}
|
|
}
|
|
statistics += ",nkeys=";
|
|
statistics += std::to_string(block->num_keys);
|
|
statistics += ",block_size=";
|
|
statistics += std::to_string(block->block_size);
|
|
statistics += ",num_ref_keys=";
|
|
statistics += std::to_string(block->key_num_access_map.size());
|
|
statistics += ",percent_access_ref_keys=";
|
|
statistics += std::to_string(percent(ref_keys_accesses, naccesses));
|
|
statistics += ",num_ref_keys_does_not_exist=";
|
|
statistics += std::to_string(block->non_exist_key_num_access_map.size());
|
|
statistics += ",percent_access_ref_keys_does_not_exist=";
|
|
statistics +=
|
|
std::to_string(percent(ref_keys_does_not_exist_accesses, naccesses));
|
|
statistics += ",ref_data_size=";
|
|
statistics += std::to_string(block->referenced_data_size);
|
|
fprintf(stdout,
|
|
"Top %" PRIu32 " access count blocks access_count=%" PRIu64
|
|
" %s\n",
|
|
top_k, naccess_it->first, statistics.c_str());
|
|
}
|
|
}
|
|
|
|
for (auto const& bt_stats : bt_stats_map) {
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout, "Break down by block type %s: \n%s",
|
|
block_type_to_string(bt_stats.first).c_str(),
|
|
bt_stats.second.ToString().c_str());
|
|
}
|
|
for (auto const& cf_bt_stats : cf_bt_stats_map) {
|
|
const std::string& cf_name = cf_bt_stats.first;
|
|
for (auto const& bt_stats : cf_bt_stats.second) {
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout,
|
|
"Break down by column family %s and block type "
|
|
"%s: \n%s",
|
|
cf_name.c_str(), block_type_to_string(bt_stats.first).c_str(),
|
|
bt_stats.second.ToString().c_str());
|
|
}
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::PrintDataBlockAccessStats() const {
|
|
HistogramStat existing_keys_stats;
|
|
std::map<std::string, HistogramStat> cf_existing_keys_stats_map;
|
|
HistogramStat non_existing_keys_stats;
|
|
std::map<std::string, HistogramStat> cf_non_existing_keys_stats_map;
|
|
HistogramStat block_access_stats;
|
|
std::map<std::string, HistogramStat> cf_block_access_info;
|
|
HistogramStat percent_referenced_bytes;
|
|
std::map<std::string, HistogramStat> cf_percent_referenced_bytes;
|
|
// Total number of accesses in a data block / number of keys in a data block.
|
|
HistogramStat avg_naccesses_per_key_in_a_data_block;
|
|
std::map<std::string, HistogramStat> cf_avg_naccesses_per_key_in_a_data_block;
|
|
// The standard deviation on the number of accesses of a key in a data block.
|
|
HistogramStat stdev_naccesses_per_key_in_a_data_block;
|
|
std::map<std::string, HistogramStat>
|
|
cf_stdev_naccesses_per_key_in_a_data_block;
|
|
auto block_callback =
|
|
[&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/,
|
|
TraceType /*type*/, const std::string& /*block_key*/,
|
|
uint64_t /*block_id*/, const BlockAccessInfo& block) {
|
|
if (block.num_keys == 0) {
|
|
return;
|
|
}
|
|
// Use four decimal points.
|
|
uint64_t percent_referenced_for_existing_keys = (uint64_t)(
|
|
((double)block.key_num_access_map.size() / (double)block.num_keys) *
|
|
10000.0);
|
|
uint64_t percent_referenced_for_non_existing_keys =
|
|
(uint64_t)(((double)block.non_exist_key_num_access_map.size() /
|
|
(double)block.num_keys) *
|
|
10000.0);
|
|
uint64_t percent_accesses_for_existing_keys =
|
|
(uint64_t)(((double)block.num_referenced_key_exist_in_block /
|
|
(double)block.num_accesses) *
|
|
10000.0);
|
|
|
|
HistogramStat hist_naccess_per_key;
|
|
for (auto const& key_access : block.key_num_access_map) {
|
|
for (auto const& caller_access : key_access.second) {
|
|
hist_naccess_per_key.Add(caller_access.second);
|
|
}
|
|
}
|
|
uint64_t avg_accesses =
|
|
static_cast<uint64_t>(hist_naccess_per_key.Average());
|
|
uint64_t stdev_accesses =
|
|
static_cast<uint64_t>(hist_naccess_per_key.StandardDeviation());
|
|
avg_naccesses_per_key_in_a_data_block.Add(avg_accesses);
|
|
cf_avg_naccesses_per_key_in_a_data_block[cf_name].Add(avg_accesses);
|
|
stdev_naccesses_per_key_in_a_data_block.Add(stdev_accesses);
|
|
cf_stdev_naccesses_per_key_in_a_data_block[cf_name].Add(stdev_accesses);
|
|
|
|
existing_keys_stats.Add(percent_referenced_for_existing_keys);
|
|
cf_existing_keys_stats_map[cf_name].Add(
|
|
percent_referenced_for_existing_keys);
|
|
non_existing_keys_stats.Add(percent_referenced_for_non_existing_keys);
|
|
cf_non_existing_keys_stats_map[cf_name].Add(
|
|
percent_referenced_for_non_existing_keys);
|
|
block_access_stats.Add(percent_accesses_for_existing_keys);
|
|
cf_block_access_info[cf_name].Add(percent_accesses_for_existing_keys);
|
|
};
|
|
TraverseBlocks(block_callback);
|
|
fprintf(stdout,
|
|
"Histogram on the number of referenced keys existing in a block over "
|
|
"the total number of keys in a block: \n%s",
|
|
existing_keys_stats.ToString().c_str());
|
|
for (auto const& cf_stats : cf_existing_keys_stats_map) {
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout, "Break down by column family %s: \n%s",
|
|
cf_stats.first.c_str(), cf_stats.second.ToString().c_str());
|
|
}
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(
|
|
stdout,
|
|
"Histogram on the number of referenced keys DO NOT exist in a block over "
|
|
"the total number of keys in a block: \n%s",
|
|
non_existing_keys_stats.ToString().c_str());
|
|
for (auto const& cf_stats : cf_non_existing_keys_stats_map) {
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout, "Break down by column family %s: \n%s",
|
|
cf_stats.first.c_str(), cf_stats.second.ToString().c_str());
|
|
}
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout,
|
|
"Histogram on the number of accesses on keys exist in a block over "
|
|
"the total number of accesses in a block: \n%s",
|
|
block_access_stats.ToString().c_str());
|
|
for (auto const& cf_stats : cf_block_access_info) {
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout, "Break down by column family %s: \n%s",
|
|
cf_stats.first.c_str(), cf_stats.second.ToString().c_str());
|
|
}
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(
|
|
stdout,
|
|
"Histogram on the average number of accesses per key in a block: \n%s",
|
|
avg_naccesses_per_key_in_a_data_block.ToString().c_str());
|
|
for (auto const& cf_stats : cf_avg_naccesses_per_key_in_a_data_block) {
|
|
fprintf(stdout, "Break down by column family %s: \n%s",
|
|
cf_stats.first.c_str(), cf_stats.second.ToString().c_str());
|
|
}
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout,
|
|
"Histogram on the standard deviation of the number of accesses per "
|
|
"key in a block: \n%s",
|
|
stdev_naccesses_per_key_in_a_data_block.ToString().c_str());
|
|
for (auto const& cf_stats : cf_stdev_naccesses_per_key_in_a_data_block) {
|
|
fprintf(stdout, "Break down by column family %s: \n%s",
|
|
cf_stats.first.c_str(), cf_stats.second.ToString().c_str());
|
|
}
|
|
}
|
|
|
|
void BlockCacheTraceAnalyzer::PrintStatsSummary() const {
|
|
uint64_t total_num_files = 0;
|
|
uint64_t total_num_blocks = 0;
|
|
uint64_t total_num_accesses = 0;
|
|
std::map<TraceType, uint64_t> bt_num_blocks_map;
|
|
std::map<TableReaderCaller, uint64_t> caller_num_access_map;
|
|
std::map<TableReaderCaller, std::map<TraceType, uint64_t>>
|
|
caller_bt_num_access_map;
|
|
std::map<TableReaderCaller, std::map<uint32_t, uint64_t>>
|
|
caller_level_num_access_map;
|
|
for (auto const& cf_aggregates : cf_aggregates_map_) {
|
|
// Stats per column family.
|
|
const std::string& cf_name = cf_aggregates.first;
|
|
uint64_t cf_num_files = 0;
|
|
uint64_t cf_num_blocks = 0;
|
|
std::map<TraceType, uint64_t> cf_bt_blocks;
|
|
uint64_t cf_num_accesses = 0;
|
|
std::map<TableReaderCaller, uint64_t> cf_caller_num_accesses_map;
|
|
std::map<TableReaderCaller, std::map<uint64_t, uint64_t>>
|
|
cf_caller_level_num_accesses_map;
|
|
std::map<TableReaderCaller, std::map<uint64_t, uint64_t>>
|
|
cf_caller_file_num_accesses_map;
|
|
std::map<TableReaderCaller, std::map<TraceType, uint64_t>>
|
|
cf_caller_bt_num_accesses_map;
|
|
total_num_files += cf_aggregates.second.fd_aggregates_map.size();
|
|
for (auto const& file_aggregates : cf_aggregates.second.fd_aggregates_map) {
|
|
// Stats per SST file.
|
|
const uint64_t fd = file_aggregates.first;
|
|
const uint32_t level = file_aggregates.second.level;
|
|
cf_num_files++;
|
|
for (auto const& block_type_aggregates :
|
|
file_aggregates.second.block_type_aggregates_map) {
|
|
// Stats per block type.
|
|
const TraceType type = block_type_aggregates.first;
|
|
cf_bt_blocks[type] +=
|
|
block_type_aggregates.second.block_access_info_map.size();
|
|
total_num_blocks +=
|
|
block_type_aggregates.second.block_access_info_map.size();
|
|
bt_num_blocks_map[type] +=
|
|
block_type_aggregates.second.block_access_info_map.size();
|
|
for (auto const& block_access_info :
|
|
block_type_aggregates.second.block_access_info_map) {
|
|
// Stats per block.
|
|
cf_num_blocks++;
|
|
for (auto const& stats :
|
|
block_access_info.second.caller_num_access_map) {
|
|
// Stats per caller.
|
|
const TableReaderCaller caller = stats.first;
|
|
const uint64_t num_accesses = stats.second;
|
|
// Overall stats.
|
|
total_num_accesses += num_accesses;
|
|
caller_num_access_map[caller] += num_accesses;
|
|
caller_bt_num_access_map[caller][type] += num_accesses;
|
|
caller_level_num_access_map[caller][level] += num_accesses;
|
|
// Column Family stats.
|
|
cf_num_accesses += num_accesses;
|
|
cf_caller_num_accesses_map[caller] += num_accesses;
|
|
cf_caller_level_num_accesses_map[caller][level] += num_accesses;
|
|
cf_caller_file_num_accesses_map[caller][fd] += num_accesses;
|
|
cf_caller_bt_num_accesses_map[caller][type] += num_accesses;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Print stats.
|
|
print_break_lines(/*num_break_lines=*/3);
|
|
fprintf(stdout, "Statistics for column family %s:\n", cf_name.c_str());
|
|
fprintf(stdout,
|
|
" Number of files:%" PRIu64 " Number of blocks: %" PRIu64
|
|
" Number of accesses: %" PRIu64 "\n",
|
|
cf_num_files, cf_num_blocks, cf_num_accesses);
|
|
for (auto block_type : cf_bt_blocks) {
|
|
fprintf(stdout, "Number of %s blocks: %" PRIu64 " Percent: %.2f\n",
|
|
block_type_to_string(block_type.first).c_str(), block_type.second,
|
|
percent(block_type.second, cf_num_blocks));
|
|
}
|
|
for (auto caller : cf_caller_num_accesses_map) {
|
|
const uint64_t naccesses = caller.second;
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
fprintf(stdout,
|
|
"Caller %s: Number of accesses %" PRIu64 " Percent: %.2f\n",
|
|
caller_to_string(caller.first).c_str(), naccesses,
|
|
percent(naccesses, cf_num_accesses));
|
|
fprintf(stdout, "Caller %s: Number of accesses per level break down\n",
|
|
caller_to_string(caller.first).c_str());
|
|
for (auto naccess_level :
|
|
cf_caller_level_num_accesses_map[caller.first]) {
|
|
fprintf(stdout,
|
|
"\t Level %" PRIu64 ": Number of accesses: %" PRIu64
|
|
" Percent: %.2f\n",
|
|
naccess_level.first, naccess_level.second,
|
|
percent(naccess_level.second, naccesses));
|
|
}
|
|
fprintf(stdout, "Caller %s: Number of accesses per file break down\n",
|
|
caller_to_string(caller.first).c_str());
|
|
for (auto naccess_file : cf_caller_file_num_accesses_map[caller.first]) {
|
|
fprintf(stdout,
|
|
"\t File %" PRIu64 ": Number of accesses: %" PRIu64
|
|
" Percent: %.2f\n",
|
|
naccess_file.first, naccess_file.second,
|
|
percent(naccess_file.second, naccesses));
|
|
}
|
|
fprintf(stdout,
|
|
"Caller %s: Number of accesses per block type break down\n",
|
|
caller_to_string(caller.first).c_str());
|
|
for (auto naccess_type : cf_caller_bt_num_accesses_map[caller.first]) {
|
|
fprintf(stdout,
|
|
"\t Block Type %s: Number of accesses: %" PRIu64
|
|
" Percent: %.2f\n",
|
|
block_type_to_string(naccess_type.first).c_str(),
|
|
naccess_type.second, percent(naccess_type.second, naccesses));
|
|
}
|
|
}
|
|
}
|
|
print_break_lines(/*num_break_lines=*/3);
|
|
fprintf(stdout, "Overall statistics:\n");
|
|
fprintf(stdout,
|
|
"Number of files: %" PRIu64 " Number of blocks: %" PRIu64
|
|
" Number of accesses: %" PRIu64 "\n",
|
|
total_num_files, total_num_blocks, total_num_accesses);
|
|
for (auto block_type : bt_num_blocks_map) {
|
|
fprintf(stdout, "Number of %s blocks: %" PRIu64 " Percent: %.2f\n",
|
|
block_type_to_string(block_type.first).c_str(), block_type.second,
|
|
percent(block_type.second, total_num_blocks));
|
|
}
|
|
for (auto caller : caller_num_access_map) {
|
|
print_break_lines(/*num_break_lines=*/1);
|
|
uint64_t naccesses = caller.second;
|
|
fprintf(stdout, "Caller %s: Number of accesses %" PRIu64 " Percent: %.2f\n",
|
|
caller_to_string(caller.first).c_str(), naccesses,
|
|
percent(naccesses, total_num_accesses));
|
|
fprintf(stdout, "Caller %s: Number of accesses per level break down\n",
|
|
caller_to_string(caller.first).c_str());
|
|
for (auto naccess_level : caller_level_num_access_map[caller.first]) {
|
|
fprintf(stdout,
|
|
"\t Level %d: Number of accesses: %" PRIu64 " Percent: %.2f\n",
|
|
naccess_level.first, naccess_level.second,
|
|
percent(naccess_level.second, naccesses));
|
|
}
|
|
fprintf(stdout, "Caller %s: Number of accesses per block type break down\n",
|
|
caller_to_string(caller.first).c_str());
|
|
for (auto naccess_type : caller_bt_num_access_map[caller.first]) {
|
|
fprintf(stdout,
|
|
"\t Block Type %s: Number of accesses: %" PRIu64
|
|
" Percent: %.2f\n",
|
|
block_type_to_string(naccess_type.first).c_str(),
|
|
naccess_type.second, percent(naccess_type.second, naccesses));
|
|
}
|
|
}
|
|
}
|
|
|
|
std::vector<CacheConfiguration> parse_cache_config_file(
|
|
const std::string& config_path) {
|
|
std::ifstream file(config_path);
|
|
if (!file.is_open()) {
|
|
return {};
|
|
}
|
|
std::vector<CacheConfiguration> configs;
|
|
std::string line;
|
|
while (getline(file, line)) {
|
|
CacheConfiguration cache_config;
|
|
std::stringstream ss(line);
|
|
std::vector<std::string> config_strs;
|
|
while (ss.good()) {
|
|
std::string substr;
|
|
getline(ss, substr, ',');
|
|
config_strs.push_back(substr);
|
|
}
|
|
// Sanity checks.
|
|
if (config_strs.size() < 4) {
|
|
fprintf(stderr, "Invalid cache simulator configuration %s\n",
|
|
line.c_str());
|
|
exit(1);
|
|
}
|
|
if (kSupportedCacheNames.find(" " + config_strs[0] + " ") ==
|
|
std::string::npos) {
|
|
fprintf(stderr, "Invalid cache name %s. Supported cache names are %s\n",
|
|
line.c_str(), kSupportedCacheNames.c_str());
|
|
exit(1);
|
|
}
|
|
cache_config.cache_name = config_strs[0];
|
|
cache_config.num_shard_bits = ParseUint32(config_strs[1]);
|
|
cache_config.ghost_cache_capacity = ParseUint64(config_strs[2]);
|
|
for (uint32_t i = 3; i < config_strs.size(); i++) {
|
|
uint64_t capacity = ParseUint64(config_strs[i]);
|
|
if (capacity == 0) {
|
|
fprintf(stderr, "Invalid cache capacity %s, %s\n",
|
|
config_strs[i].c_str(), line.c_str());
|
|
exit(1);
|
|
}
|
|
cache_config.cache_capacities.push_back(capacity);
|
|
}
|
|
configs.push_back(cache_config);
|
|
}
|
|
file.close();
|
|
return configs;
|
|
}
|
|
|
|
std::vector<uint64_t> parse_buckets(const std::string& bucket_str) {
|
|
std::vector<uint64_t> buckets;
|
|
std::stringstream ss(bucket_str);
|
|
while (ss.good()) {
|
|
std::string bucket;
|
|
getline(ss, bucket, ',');
|
|
buckets.push_back(ParseUint64(bucket));
|
|
}
|
|
buckets.push_back(port::kMaxUint64);
|
|
return buckets;
|
|
}
|
|
|
|
int block_cache_trace_analyzer_tool(int argc, char** argv) {
|
|
ParseCommandLineFlags(&argc, &argv, true);
|
|
if (FLAGS_block_cache_trace_path.empty()) {
|
|
fprintf(stderr, "block cache trace path is empty\n");
|
|
exit(1);
|
|
}
|
|
uint64_t warmup_seconds =
|
|
FLAGS_cache_sim_warmup_seconds > 0 ? FLAGS_cache_sim_warmup_seconds : 0;
|
|
uint32_t downsample_ratio = FLAGS_block_cache_trace_downsample_ratio > 0
|
|
? FLAGS_block_cache_trace_downsample_ratio
|
|
: 0;
|
|
std::vector<CacheConfiguration> cache_configs =
|
|
parse_cache_config_file(FLAGS_block_cache_sim_config_path);
|
|
std::unique_ptr<BlockCacheTraceSimulator> cache_simulator;
|
|
if (!cache_configs.empty()) {
|
|
cache_simulator.reset(new BlockCacheTraceSimulator(
|
|
warmup_seconds, downsample_ratio, cache_configs));
|
|
Status s = cache_simulator->InitializeCaches();
|
|
if (!s.ok()) {
|
|
fprintf(stderr, "Cannot initialize cache simulators %s\n",
|
|
s.ToString().c_str());
|
|
exit(1);
|
|
}
|
|
}
|
|
BlockCacheTraceAnalyzer analyzer(
|
|
FLAGS_block_cache_trace_path, FLAGS_block_cache_analysis_result_dir,
|
|
FLAGS_human_readable_trace_file_path,
|
|
!FLAGS_reuse_distance_labels.empty(), FLAGS_mrc_only,
|
|
FLAGS_is_block_cache_human_readable_trace, std::move(cache_simulator));
|
|
Status s = analyzer.Analyze();
|
|
if (!s.IsIncomplete() && !s.ok()) {
|
|
// Read all traces.
|
|
fprintf(stderr, "Cannot process the trace %s\n", s.ToString().c_str());
|
|
exit(1);
|
|
}
|
|
fprintf(stdout, "Status: %s\n", s.ToString().c_str());
|
|
analyzer.WriteMissRatioCurves();
|
|
analyzer.WriteMissRatioTimeline(1);
|
|
analyzer.WriteMissRatioTimeline(kSecondInMinute);
|
|
analyzer.WriteMissRatioTimeline(kSecondInHour);
|
|
analyzer.WriteMissTimeline(1);
|
|
analyzer.WriteMissTimeline(kSecondInMinute);
|
|
analyzer.WriteMissTimeline(kSecondInHour);
|
|
|
|
if (FLAGS_mrc_only) {
|
|
fprintf(stdout,
|
|
"Skipping the analysis statistics since the user wants to compute "
|
|
"MRC only");
|
|
return 0;
|
|
}
|
|
|
|
analyzer.PrintStatsSummary();
|
|
if (FLAGS_print_access_count_stats) {
|
|
print_break_lines(/*num_break_lines=*/3);
|
|
analyzer.PrintAccessCountStats(
|
|
/*user_access_only=*/false, FLAGS_analyze_bottom_k_access_count_blocks,
|
|
FLAGS_analyze_top_k_access_count_blocks);
|
|
print_break_lines(/*num_break_lines=*/3);
|
|
analyzer.PrintAccessCountStats(
|
|
/*user_access_only=*/true, FLAGS_analyze_bottom_k_access_count_blocks,
|
|
FLAGS_analyze_top_k_access_count_blocks);
|
|
}
|
|
if (FLAGS_print_block_size_stats) {
|
|
print_break_lines(/*num_break_lines=*/3);
|
|
analyzer.PrintBlockSizeStats();
|
|
}
|
|
if (FLAGS_print_data_block_access_count_stats) {
|
|
print_break_lines(/*num_break_lines=*/3);
|
|
analyzer.PrintDataBlockAccessStats();
|
|
}
|
|
print_break_lines(/*num_break_lines=*/3);
|
|
|
|
if (!FLAGS_timeline_labels.empty()) {
|
|
std::stringstream ss(FLAGS_timeline_labels);
|
|
while (ss.good()) {
|
|
std::string label;
|
|
getline(ss, label, ',');
|
|
if (label.find("block") != std::string::npos) {
|
|
analyzer.WriteAccessTimeline(label, kSecondInMinute, true);
|
|
analyzer.WriteAccessTimeline(label, kSecondInMinute, false);
|
|
analyzer.WriteAccessTimeline(label, kSecondInHour, true);
|
|
analyzer.WriteAccessTimeline(label, kSecondInHour, false);
|
|
} else {
|
|
analyzer.WriteAccessTimeline(label, kSecondInMinute, false);
|
|
analyzer.WriteAccessTimeline(label, kSecondInHour, false);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (!FLAGS_analyze_callers.empty()) {
|
|
analyzer.WritePercentAccessSummaryStats();
|
|
std::stringstream ss(FLAGS_analyze_callers);
|
|
while (ss.good()) {
|
|
std::string caller;
|
|
getline(ss, caller, ',');
|
|
analyzer.WriteDetailedPercentAccessSummaryStats(string_to_caller(caller));
|
|
}
|
|
}
|
|
|
|
if (!FLAGS_access_count_buckets.empty()) {
|
|
std::vector<uint64_t> buckets = parse_buckets(FLAGS_access_count_buckets);
|
|
analyzer.WriteAccessCountSummaryStats(buckets, /*user_access_only=*/true);
|
|
analyzer.WriteAccessCountSummaryStats(buckets, /*user_access_only=*/false);
|
|
}
|
|
|
|
if (!FLAGS_reuse_distance_labels.empty() &&
|
|
!FLAGS_reuse_distance_buckets.empty()) {
|
|
std::vector<uint64_t> buckets = parse_buckets(FLAGS_reuse_distance_buckets);
|
|
std::stringstream ss(FLAGS_reuse_distance_labels);
|
|
while (ss.good()) {
|
|
std::string label;
|
|
getline(ss, label, ',');
|
|
analyzer.WriteReuseDistance(label, buckets);
|
|
}
|
|
}
|
|
|
|
if (!FLAGS_reuse_interval_labels.empty() &&
|
|
!FLAGS_reuse_interval_buckets.empty()) {
|
|
std::vector<uint64_t> buckets = parse_buckets(FLAGS_reuse_interval_buckets);
|
|
std::stringstream ss(FLAGS_reuse_interval_labels);
|
|
while (ss.good()) {
|
|
std::string label;
|
|
getline(ss, label, ',');
|
|
analyzer.WriteReuseInterval(label, buckets);
|
|
}
|
|
}
|
|
|
|
if (!FLAGS_reuse_lifetime_labels.empty() &&
|
|
!FLAGS_reuse_lifetime_buckets.empty()) {
|
|
std::vector<uint64_t> buckets = parse_buckets(FLAGS_reuse_lifetime_buckets);
|
|
std::stringstream ss(FLAGS_reuse_lifetime_labels);
|
|
while (ss.good()) {
|
|
std::string label;
|
|
getline(ss, label, ',');
|
|
analyzer.WriteReuseLifetime(label, buckets);
|
|
}
|
|
}
|
|
|
|
if (FLAGS_analyze_blocks_reuse_k_reuse_window != 0) {
|
|
std::vector<TraceType> block_types{TraceType::kBlockTraceIndexBlock,
|
|
TraceType::kBlockTraceDataBlock,
|
|
TraceType::kBlockTraceFilterBlock};
|
|
for (auto block_type : block_types) {
|
|
analyzer.WriteBlockReuseTimeline(
|
|
FLAGS_analyze_blocks_reuse_k_reuse_window,
|
|
/*user_access_only=*/true, block_type);
|
|
analyzer.WriteBlockReuseTimeline(
|
|
FLAGS_analyze_blocks_reuse_k_reuse_window,
|
|
/*user_access_only=*/false, block_type);
|
|
}
|
|
}
|
|
|
|
if (!FLAGS_analyze_get_spatial_locality_labels.empty() &&
|
|
!FLAGS_analyze_get_spatial_locality_buckets.empty()) {
|
|
std::vector<uint64_t> buckets =
|
|
parse_buckets(FLAGS_analyze_get_spatial_locality_buckets);
|
|
std::stringstream ss(FLAGS_analyze_get_spatial_locality_labels);
|
|
while (ss.good()) {
|
|
std::string label;
|
|
getline(ss, label, ',');
|
|
analyzer.WriteGetSpatialLocality(label, buckets);
|
|
}
|
|
}
|
|
|
|
if (!FLAGS_analyze_correlation_coefficients_labels.empty()) {
|
|
std::stringstream ss(FLAGS_analyze_correlation_coefficients_labels);
|
|
while (ss.good()) {
|
|
std::string label;
|
|
getline(ss, label, ',');
|
|
analyzer.WriteCorrelationFeatures(
|
|
label, FLAGS_analyze_correlation_coefficients_max_number_of_values);
|
|
}
|
|
analyzer.WriteCorrelationFeaturesForGet(
|
|
FLAGS_analyze_correlation_coefficients_max_number_of_values);
|
|
}
|
|
|
|
if (!FLAGS_skew_labels.empty() && !FLAGS_skew_buckets.empty()) {
|
|
std::vector<uint64_t> buckets = parse_buckets(FLAGS_skew_buckets);
|
|
std::stringstream ss(FLAGS_skew_labels);
|
|
while (ss.good()) {
|
|
std::string label;
|
|
getline(ss, label, ',');
|
|
if (label.find("block") != std::string::npos) {
|
|
analyzer.WriteSkewness(label, buckets,
|
|
TraceType::kBlockTraceIndexBlock);
|
|
analyzer.WriteSkewness(label, buckets,
|
|
TraceType::kBlockTraceFilterBlock);
|
|
analyzer.WriteSkewness(label, buckets, TraceType::kBlockTraceDataBlock);
|
|
analyzer.WriteSkewness(label, buckets, TraceType::kTraceMax);
|
|
} else {
|
|
analyzer.WriteSkewness(label, buckets, TraceType::kTraceMax);
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
|
|
#endif // GFLAGS
|
|
#endif // ROCKSDB_LITE
|