2019-06-11 21:18:37 +02:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
|
|
|
|
#pragma once
|
|
|
|
|
|
|
|
#include <map>
|
2019-06-25 05:38:20 +02:00
|
|
|
#include <set>
|
2019-06-11 21:18:37 +02:00
|
|
|
#include <vector>
|
|
|
|
|
2019-07-13 01:52:15 +02:00
|
|
|
#include "db/dbformat.h"
|
2019-06-11 21:18:37 +02:00
|
|
|
#include "rocksdb/env.h"
|
2021-08-12 04:31:44 +02:00
|
|
|
#include "rocksdb/trace_record.h"
|
Support computing miss ratio curves using sim_cache. (#5449)
Summary:
This PR adds a BlockCacheTraceSimulator that reports the miss ratios given different cache configurations. A cache configuration contains "cache_name,num_shard_bits,cache_capacities". For example, "lru, 1, 1K, 2K, 4M, 4G".
When we replay the trace, we also perform lookups and inserts on the simulated caches.
In the end, it reports the miss ratio for each tuple <cache_name, num_shard_bits, cache_capacity> in a output file.
This PR also adds a main source block_cache_trace_analyzer so that we can run the analyzer in command line.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5449
Test Plan:
Added tests for block_cache_trace_analyzer.
COMPILE_WITH_ASAN=1 make check -j32.
Differential Revision: D15797073
Pulled By: HaoyuHuang
fbshipit-source-id: aef0c5c2e7938f3e8b6a10d4a6a50e6928ecf408
2019-06-18 01:33:40 +02:00
|
|
|
#include "rocksdb/utilities/sim_cache.h"
|
2019-06-11 21:18:37 +02:00
|
|
|
#include "trace_replay/block_cache_tracer.h"
|
2019-07-01 21:43:14 +02:00
|
|
|
#include "utilities/simulator_cache/cache_simulator.h"
|
2019-06-11 21:18:37 +02:00
|
|
|
|
2020-02-20 21:07:53 +01:00
|
|
|
namespace ROCKSDB_NAMESPACE {
|
2019-07-23 02:47:54 +02:00
|
|
|
|
|
|
|
// Statistics of a key refereneced by a Get.
|
|
|
|
struct GetKeyInfo {
|
|
|
|
uint64_t key_id = 0;
|
|
|
|
std::vector<uint64_t> access_sequence_number_timeline;
|
|
|
|
std::vector<uint64_t> access_timeline;
|
|
|
|
|
|
|
|
void AddAccess(const BlockCacheTraceRecord& access,
|
|
|
|
uint64_t access_sequnce_number) {
|
|
|
|
access_sequence_number_timeline.push_back(access_sequnce_number);
|
|
|
|
access_timeline.push_back(access.access_timestamp);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2019-06-11 21:18:37 +02:00
|
|
|
// Statistics of a block.
|
|
|
|
struct BlockAccessInfo {
|
2019-07-23 02:47:54 +02:00
|
|
|
uint64_t block_id = 0;
|
Pysim more algorithms (#5644)
Summary:
This PR adds four more eviction policies.
- OPT [1]
- Hyperbolic caching [2]
- ARC [3]
- GreedyDualSize [4]
[1] L. A. Belady. 1966. A Study of Replacement Algorithms for a Virtual-storage Computer. IBM Syst. J. 5, 2 (June 1966), 78-101. DOI=http://dx.doi.org/10.1147/sj.52.0078
[2] Aaron Blankstein, Siddhartha Sen, and Michael J. Freedman. 2017. Hyperbolic caching: flexible caching for web applications. In Proceedings of the 2017 USENIX Conference on Usenix Annual Technical Conference (USENIX ATC '17). USENIX Association, Berkeley, CA, USA, 499-511.
[3] Nimrod Megiddo and Dharmendra S. Modha. 2003. ARC: A Self-Tuning, Low Overhead Replacement Cache. In Proceedings of the 2nd USENIX Conference on File and Storage Technologies (FAST '03). USENIX Association, Berkeley, CA, USA, 115-130.
[4] N. Young. The k-server dual and loose competitiveness for paging. Algorithmica, June 1994, vol. 11,(no.6):525-41. Rewritten version of ''On-line caching as cache size varies'', in The 2nd Annual ACM-SIAM Symposium on Discrete Algorithms, 241-250, 1991.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5644
Differential Revision: D16548817
Pulled By: HaoyuHuang
fbshipit-source-id: 838f76db9179f07911abaab46c97e1c929cfcd63
2019-08-07 03:47:39 +02:00
|
|
|
uint64_t table_id = 0;
|
|
|
|
uint64_t block_offset = 0;
|
2019-06-11 21:18:37 +02:00
|
|
|
uint64_t num_accesses = 0;
|
|
|
|
uint64_t block_size = 0;
|
|
|
|
uint64_t first_access_time = 0;
|
|
|
|
uint64_t last_access_time = 0;
|
|
|
|
uint64_t num_keys = 0;
|
2019-07-13 01:52:15 +02:00
|
|
|
std::map<std::string, std::map<TableReaderCaller, uint64_t>>
|
2019-06-11 21:18:37 +02:00
|
|
|
key_num_access_map; // for keys exist in this block.
|
2019-07-13 01:52:15 +02:00
|
|
|
std::map<std::string, std::map<TableReaderCaller, uint64_t>>
|
2019-06-11 21:18:37 +02:00
|
|
|
non_exist_key_num_access_map; // for keys do not exist in this block.
|
|
|
|
uint64_t num_referenced_key_exist_in_block = 0;
|
2019-07-13 01:52:15 +02:00
|
|
|
uint64_t referenced_data_size = 0;
|
2019-06-20 23:28:22 +02:00
|
|
|
std::map<TableReaderCaller, uint64_t> caller_num_access_map;
|
2019-06-25 05:38:20 +02:00
|
|
|
// caller:timestamp:number_of_accesses. The granularity of the timestamp is
|
|
|
|
// seconds.
|
|
|
|
std::map<TableReaderCaller, std::map<uint64_t, uint64_t>>
|
|
|
|
caller_num_accesses_timeline;
|
|
|
|
// Unique blocks since the last access.
|
|
|
|
std::set<std::string> unique_blocks_since_last_access;
|
|
|
|
// Number of reuses grouped by reuse distance.
|
|
|
|
std::map<uint64_t, uint64_t> reuse_distance_count;
|
2019-06-11 21:18:37 +02:00
|
|
|
|
2019-07-23 02:47:54 +02:00
|
|
|
// The access sequence numbers of this block.
|
|
|
|
std::vector<uint64_t> access_sequence_number_timeline;
|
|
|
|
std::map<TableReaderCaller, std::vector<uint64_t>>
|
|
|
|
caller_access_sequence__number_timeline;
|
|
|
|
// The access timestamp in microseconds of this block.
|
|
|
|
std::vector<uint64_t> access_timeline;
|
|
|
|
std::map<TableReaderCaller, std::vector<uint64_t>> caller_access_timeline;
|
|
|
|
|
|
|
|
void AddAccess(const BlockCacheTraceRecord& access,
|
|
|
|
uint64_t access_sequnce_number) {
|
2019-07-13 01:52:15 +02:00
|
|
|
if (block_size != 0 && access.block_size != 0) {
|
|
|
|
assert(block_size == access.block_size);
|
|
|
|
}
|
|
|
|
if (num_keys != 0 && access.num_keys_in_block != 0) {
|
|
|
|
assert(num_keys == access.num_keys_in_block);
|
|
|
|
}
|
2019-06-11 21:18:37 +02:00
|
|
|
if (first_access_time == 0) {
|
|
|
|
first_access_time = access.access_timestamp;
|
|
|
|
}
|
Pysim more algorithms (#5644)
Summary:
This PR adds four more eviction policies.
- OPT [1]
- Hyperbolic caching [2]
- ARC [3]
- GreedyDualSize [4]
[1] L. A. Belady. 1966. A Study of Replacement Algorithms for a Virtual-storage Computer. IBM Syst. J. 5, 2 (June 1966), 78-101. DOI=http://dx.doi.org/10.1147/sj.52.0078
[2] Aaron Blankstein, Siddhartha Sen, and Michael J. Freedman. 2017. Hyperbolic caching: flexible caching for web applications. In Proceedings of the 2017 USENIX Conference on Usenix Annual Technical Conference (USENIX ATC '17). USENIX Association, Berkeley, CA, USA, 499-511.
[3] Nimrod Megiddo and Dharmendra S. Modha. 2003. ARC: A Self-Tuning, Low Overhead Replacement Cache. In Proceedings of the 2nd USENIX Conference on File and Storage Technologies (FAST '03). USENIX Association, Berkeley, CA, USA, 115-130.
[4] N. Young. The k-server dual and loose competitiveness for paging. Algorithmica, June 1994, vol. 11,(no.6):525-41. Rewritten version of ''On-line caching as cache size varies'', in The 2nd Annual ACM-SIAM Symposium on Discrete Algorithms, 241-250, 1991.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5644
Differential Revision: D16548817
Pulled By: HaoyuHuang
fbshipit-source-id: 838f76db9179f07911abaab46c97e1c929cfcd63
2019-08-07 03:47:39 +02:00
|
|
|
table_id = BlockCacheTraceHelper::GetTableId(access);
|
|
|
|
block_offset = BlockCacheTraceHelper::GetBlockOffsetInFile(access);
|
2019-06-11 21:18:37 +02:00
|
|
|
last_access_time = access.access_timestamp;
|
|
|
|
block_size = access.block_size;
|
|
|
|
caller_num_access_map[access.caller]++;
|
|
|
|
num_accesses++;
|
2019-06-25 05:38:20 +02:00
|
|
|
// access.access_timestamp is in microsecond.
|
|
|
|
const uint64_t timestamp_in_seconds =
|
|
|
|
access.access_timestamp / kMicrosInSecond;
|
|
|
|
caller_num_accesses_timeline[access.caller][timestamp_in_seconds] += 1;
|
2019-07-23 02:47:54 +02:00
|
|
|
// Populate the feature vectors.
|
|
|
|
access_sequence_number_timeline.push_back(access_sequnce_number);
|
|
|
|
caller_access_sequence__number_timeline[access.caller].push_back(
|
|
|
|
access_sequnce_number);
|
|
|
|
access_timeline.push_back(access.access_timestamp);
|
|
|
|
caller_access_timeline[access.caller].push_back(access.access_timestamp);
|
2019-07-17 22:02:00 +02:00
|
|
|
if (BlockCacheTraceHelper::IsGetOrMultiGetOnDataBlock(access.block_type,
|
|
|
|
access.caller)) {
|
2019-06-11 21:18:37 +02:00
|
|
|
num_keys = access.num_keys_in_block;
|
2019-06-15 02:37:24 +02:00
|
|
|
if (access.referenced_key_exist_in_block == Boolean::kTrue) {
|
2019-07-13 01:52:15 +02:00
|
|
|
if (key_num_access_map.find(access.referenced_key) ==
|
|
|
|
key_num_access_map.end()) {
|
|
|
|
referenced_data_size += access.referenced_data_size;
|
|
|
|
}
|
|
|
|
key_num_access_map[access.referenced_key][access.caller]++;
|
2019-06-11 21:18:37 +02:00
|
|
|
num_referenced_key_exist_in_block++;
|
2019-07-13 01:52:15 +02:00
|
|
|
if (referenced_data_size > block_size && block_size != 0) {
|
|
|
|
ParsedInternalKey internal_key;
|
2020-10-28 18:11:13 +01:00
|
|
|
Status s = ParseInternalKey(access.referenced_key, &internal_key,
|
|
|
|
false /* log_err_key */); // TODO
|
2020-10-01 04:15:42 +02:00
|
|
|
assert(s.ok()); // TODO
|
2019-07-13 01:52:15 +02:00
|
|
|
}
|
2019-06-11 21:18:37 +02:00
|
|
|
} else {
|
2019-07-13 01:52:15 +02:00
|
|
|
non_exist_key_num_access_map[access.referenced_key][access.caller]++;
|
2019-06-11 21:18:37 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// Aggregates stats of a block given a block type.
|
|
|
|
struct BlockTypeAccessInfoAggregate {
|
|
|
|
std::map<std::string, BlockAccessInfo> block_access_info_map;
|
|
|
|
};
|
|
|
|
|
|
|
|
// Aggregates BlockTypeAggregate given a SST file.
|
|
|
|
struct SSTFileAccessInfoAggregate {
|
|
|
|
uint32_t level;
|
|
|
|
std::map<TraceType, BlockTypeAccessInfoAggregate> block_type_aggregates_map;
|
|
|
|
};
|
|
|
|
|
|
|
|
// Aggregates SSTFileAggregate given a column family.
|
|
|
|
struct ColumnFamilyAccessInfoAggregate {
|
|
|
|
std::map<uint64_t, SSTFileAccessInfoAggregate> fd_aggregates_map;
|
|
|
|
};
|
|
|
|
|
2019-07-23 02:47:54 +02:00
|
|
|
struct Features {
|
|
|
|
std::vector<uint64_t> elapsed_time_since_last_access;
|
|
|
|
std::vector<uint64_t> num_accesses_since_last_access;
|
|
|
|
std::vector<uint64_t> num_past_accesses;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct Predictions {
|
|
|
|
std::vector<uint64_t> elapsed_time_till_next_access;
|
|
|
|
std::vector<uint64_t> num_accesses_till_next_access;
|
|
|
|
};
|
|
|
|
|
2019-06-11 21:18:37 +02:00
|
|
|
class BlockCacheTraceAnalyzer {
|
|
|
|
public:
|
Support computing miss ratio curves using sim_cache. (#5449)
Summary:
This PR adds a BlockCacheTraceSimulator that reports the miss ratios given different cache configurations. A cache configuration contains "cache_name,num_shard_bits,cache_capacities". For example, "lru, 1, 1K, 2K, 4M, 4G".
When we replay the trace, we also perform lookups and inserts on the simulated caches.
In the end, it reports the miss ratio for each tuple <cache_name, num_shard_bits, cache_capacity> in a output file.
This PR also adds a main source block_cache_trace_analyzer so that we can run the analyzer in command line.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5449
Test Plan:
Added tests for block_cache_trace_analyzer.
COMPILE_WITH_ASAN=1 make check -j32.
Differential Revision: D15797073
Pulled By: HaoyuHuang
fbshipit-source-id: aef0c5c2e7938f3e8b6a10d4a6a50e6928ecf408
2019-06-18 01:33:40 +02:00
|
|
|
BlockCacheTraceAnalyzer(
|
2019-06-25 05:38:20 +02:00
|
|
|
const std::string& trace_file_path, const std::string& output_dir,
|
2019-07-23 02:47:54 +02:00
|
|
|
const std::string& human_readable_trace_file_path,
|
|
|
|
bool compute_reuse_distance, bool mrc_only,
|
2019-08-09 22:09:04 +02:00
|
|
|
bool is_human_readable_trace_file,
|
Support computing miss ratio curves using sim_cache. (#5449)
Summary:
This PR adds a BlockCacheTraceSimulator that reports the miss ratios given different cache configurations. A cache configuration contains "cache_name,num_shard_bits,cache_capacities". For example, "lru, 1, 1K, 2K, 4M, 4G".
When we replay the trace, we also perform lookups and inserts on the simulated caches.
In the end, it reports the miss ratio for each tuple <cache_name, num_shard_bits, cache_capacity> in a output file.
This PR also adds a main source block_cache_trace_analyzer so that we can run the analyzer in command line.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5449
Test Plan:
Added tests for block_cache_trace_analyzer.
COMPILE_WITH_ASAN=1 make check -j32.
Differential Revision: D15797073
Pulled By: HaoyuHuang
fbshipit-source-id: aef0c5c2e7938f3e8b6a10d4a6a50e6928ecf408
2019-06-18 01:33:40 +02:00
|
|
|
std::unique_ptr<BlockCacheTraceSimulator>&& cache_simulator);
|
2019-06-11 21:18:37 +02:00
|
|
|
~BlockCacheTraceAnalyzer() = default;
|
|
|
|
// No copy and move.
|
|
|
|
BlockCacheTraceAnalyzer(const BlockCacheTraceAnalyzer&) = delete;
|
|
|
|
BlockCacheTraceAnalyzer& operator=(const BlockCacheTraceAnalyzer&) = delete;
|
|
|
|
BlockCacheTraceAnalyzer(BlockCacheTraceAnalyzer&&) = delete;
|
|
|
|
BlockCacheTraceAnalyzer& operator=(BlockCacheTraceAnalyzer&&) = delete;
|
|
|
|
|
|
|
|
// Read all access records in the given trace_file, maintains the stats of
|
|
|
|
// a block, and aggregates the information by block type, sst file, and column
|
|
|
|
// family. Subsequently, the caller may call Print* functions to print
|
|
|
|
// statistics.
|
|
|
|
Status Analyze();
|
|
|
|
|
|
|
|
// Print a summary of statistics of the trace, e.g.,
|
|
|
|
// Number of files: 2 Number of blocks: 50 Number of accesses: 50
|
|
|
|
// Number of Index blocks: 10
|
|
|
|
// Number of Filter blocks: 10
|
|
|
|
// Number of Data blocks: 10
|
|
|
|
// Number of UncompressionDict blocks: 10
|
|
|
|
// Number of RangeDeletion blocks: 10
|
|
|
|
// ***************************************************************
|
|
|
|
// Caller Get: Number of accesses 10
|
|
|
|
// Caller Get: Number of accesses per level break down
|
|
|
|
// Level 0: Number of accesses: 10
|
|
|
|
// Caller Get: Number of accesses per block type break down
|
|
|
|
// Block Type Index: Number of accesses: 2
|
|
|
|
// Block Type Filter: Number of accesses: 2
|
|
|
|
// Block Type Data: Number of accesses: 2
|
|
|
|
// Block Type UncompressionDict: Number of accesses: 2
|
|
|
|
// Block Type RangeDeletion: Number of accesses: 2
|
|
|
|
void PrintStatsSummary() const;
|
|
|
|
|
|
|
|
// Print block size distribution and the distribution break down by block type
|
|
|
|
// and column family.
|
|
|
|
void PrintBlockSizeStats() const;
|
|
|
|
|
|
|
|
// Print access count distribution and the distribution break down by block
|
|
|
|
// type and column family.
|
2019-07-13 01:52:15 +02:00
|
|
|
void PrintAccessCountStats(bool user_access_only, uint32_t bottom_k,
|
|
|
|
uint32_t top_k) const;
|
2019-06-11 21:18:37 +02:00
|
|
|
|
|
|
|
// Print data block accesses by user Get and Multi-Get.
|
|
|
|
// It prints out 1) A histogram on the percentage of keys accessed in a data
|
|
|
|
// block break down by if a referenced key exists in the data block andthe
|
|
|
|
// histogram break down by column family. 2) A histogram on the percentage of
|
|
|
|
// accesses on keys exist in a data block and its break down by column family.
|
|
|
|
void PrintDataBlockAccessStats() const;
|
|
|
|
|
2019-07-13 01:52:15 +02:00
|
|
|
// Write the percentage of accesses break down by column family into a csv
|
|
|
|
// file saved in 'output_dir'.
|
|
|
|
//
|
|
|
|
// The file is named "percentage_of_accesses_summary". The file format is
|
|
|
|
// caller,cf_0,cf_1,...,cf_n where the cf_i is the column family name found in
|
|
|
|
// the trace.
|
|
|
|
void WritePercentAccessSummaryStats() const;
|
|
|
|
|
|
|
|
// Write the percentage of accesses for the given caller break down by column
|
|
|
|
// family, level, and block type into a csv file saved in 'output_dir'.
|
|
|
|
//
|
|
|
|
// It generates two files: 1) caller_level_percentage_of_accesses_summary and
|
|
|
|
// 2) caller_bt_percentage_of_accesses_summary which break down by the level
|
|
|
|
// and block type, respectively. The file format is
|
|
|
|
// level/bt,cf_0,cf_1,...,cf_n where cf_i is the column family name found in
|
|
|
|
// the trace.
|
|
|
|
void WriteDetailedPercentAccessSummaryStats(TableReaderCaller caller) const;
|
|
|
|
|
|
|
|
// Write the access count summary into a csv file saved in 'output_dir'.
|
|
|
|
// It groups blocks by their access count.
|
|
|
|
//
|
|
|
|
// It generates two files: 1) cf_access_count_summary and 2)
|
|
|
|
// bt_access_count_summary which break down the access count by column family
|
|
|
|
// and block type, respectively. The file format is
|
|
|
|
// cf/bt,bucket_0,bucket_1,...,bucket_N.
|
|
|
|
void WriteAccessCountSummaryStats(
|
|
|
|
const std::vector<uint64_t>& access_count_buckets,
|
|
|
|
bool user_access_only) const;
|
|
|
|
|
2019-06-25 05:38:20 +02:00
|
|
|
// Write miss ratio curves of simulated cache configurations into a csv file
|
2019-07-13 01:52:15 +02:00
|
|
|
// named "mrc" saved in 'output_dir'.
|
|
|
|
//
|
|
|
|
// The file format is
|
|
|
|
// "cache_name,num_shard_bits,capacity,miss_ratio,total_accesses".
|
2019-06-25 05:38:20 +02:00
|
|
|
void WriteMissRatioCurves() const;
|
|
|
|
|
2019-07-23 02:47:54 +02:00
|
|
|
// Write miss ratio timeline of simulated cache configurations into several
|
|
|
|
// csv files, one per cache capacity saved in 'output_dir'.
|
|
|
|
//
|
|
|
|
// The file format is
|
|
|
|
// "time,label_1_access_per_second,label_2_access_per_second,...,label_N_access_per_second"
|
|
|
|
// where N is the number of unique cache names
|
|
|
|
// (cache_name+num_shard_bits+ghost_capacity).
|
|
|
|
void WriteMissRatioTimeline(uint64_t time_unit) const;
|
|
|
|
|
|
|
|
// Write misses timeline of simulated cache configurations into several
|
|
|
|
// csv files, one per cache capacity saved in 'output_dir'.
|
|
|
|
//
|
|
|
|
// The file format is
|
|
|
|
// "time,label_1_access_per_second,label_2_access_per_second,...,label_N_access_per_second"
|
|
|
|
// where N is the number of unique cache names
|
|
|
|
// (cache_name+num_shard_bits+ghost_capacity).
|
|
|
|
void WriteMissTimeline(uint64_t time_unit) const;
|
|
|
|
|
2019-06-25 05:38:20 +02:00
|
|
|
// Write the access timeline into a csv file saved in 'output_dir'.
|
2019-07-13 01:52:15 +02:00
|
|
|
//
|
|
|
|
// The file is named "label_access_timeline".The file format is
|
|
|
|
// "time,label_1_access_per_second,label_2_access_per_second,...,label_N_access_per_second"
|
|
|
|
// where N is the number of unique labels found in the trace.
|
|
|
|
void WriteAccessTimeline(const std::string& label, uint64_t time_unit,
|
|
|
|
bool user_access_only) const;
|
2019-06-25 05:38:20 +02:00
|
|
|
|
|
|
|
// Write the reuse distance into a csv file saved in 'output_dir'. Reuse
|
|
|
|
// distance is defined as the cumulated size of unique blocks read between two
|
|
|
|
// consective accesses on the same block.
|
2019-07-13 01:52:15 +02:00
|
|
|
//
|
|
|
|
// The file is named "label_reuse_distance". The file format is
|
|
|
|
// bucket,label_1,label_2,...,label_N.
|
2019-06-25 05:38:20 +02:00
|
|
|
void WriteReuseDistance(const std::string& label_str,
|
2019-07-13 01:52:15 +02:00
|
|
|
const std::vector<uint64_t>& distance_buckets) const;
|
2019-06-25 05:38:20 +02:00
|
|
|
|
|
|
|
// Write the reuse interval into a csv file saved in 'output_dir'. Reuse
|
|
|
|
// interval is defined as the time between two consecutive accesses on the
|
2019-07-13 01:52:15 +02:00
|
|
|
// same block.
|
|
|
|
//
|
|
|
|
// The file is named "label_reuse_interval". The file format is
|
|
|
|
// bucket,label_1,label_2,...,label_N.
|
2019-06-25 05:38:20 +02:00
|
|
|
void WriteReuseInterval(const std::string& label_str,
|
2019-07-13 01:52:15 +02:00
|
|
|
const std::vector<uint64_t>& time_buckets) const;
|
|
|
|
|
|
|
|
// Write the reuse lifetime into a csv file saved in 'output_dir'. Reuse
|
|
|
|
// lifetime is defined as the time interval between the first access of a
|
|
|
|
// block and its last access.
|
|
|
|
//
|
|
|
|
// The file is named "label_reuse_lifetime". The file format is
|
|
|
|
// bucket,label_1,label_2,...,label_N.
|
|
|
|
void WriteReuseLifetime(const std::string& label_str,
|
|
|
|
const std::vector<uint64_t>& time_buckets) const;
|
|
|
|
|
|
|
|
// Write the reuse timeline into a csv file saved in 'output_dir'.
|
|
|
|
//
|
|
|
|
// The file is named
|
|
|
|
// "block_type_user_access_only_reuse_window_reuse_timeline". The file format
|
|
|
|
// is start_time,0,1,...,N where N equals trace_duration / reuse_window.
|
2019-11-27 01:55:46 +01:00
|
|
|
void WriteBlockReuseTimeline(const uint64_t reuse_window, bool user_access_only,
|
2019-07-13 01:52:15 +02:00
|
|
|
TraceType block_type) const;
|
|
|
|
|
|
|
|
// Write the Get spatical locality into csv files saved in 'output_dir'.
|
|
|
|
//
|
|
|
|
// It generates three csv files. label_percent_ref_keys,
|
|
|
|
// label_percent_accesses_on_ref_keys, and
|
|
|
|
// label_percent_data_size_on_ref_keys.
|
|
|
|
void WriteGetSpatialLocality(
|
|
|
|
const std::string& label_str,
|
|
|
|
const std::vector<uint64_t>& percent_buckets) const;
|
Support computing miss ratio curves using sim_cache. (#5449)
Summary:
This PR adds a BlockCacheTraceSimulator that reports the miss ratios given different cache configurations. A cache configuration contains "cache_name,num_shard_bits,cache_capacities". For example, "lru, 1, 1K, 2K, 4M, 4G".
When we replay the trace, we also perform lookups and inserts on the simulated caches.
In the end, it reports the miss ratio for each tuple <cache_name, num_shard_bits, cache_capacity> in a output file.
This PR also adds a main source block_cache_trace_analyzer so that we can run the analyzer in command line.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5449
Test Plan:
Added tests for block_cache_trace_analyzer.
COMPILE_WITH_ASAN=1 make check -j32.
Differential Revision: D15797073
Pulled By: HaoyuHuang
fbshipit-source-id: aef0c5c2e7938f3e8b6a10d4a6a50e6928ecf408
2019-06-18 01:33:40 +02:00
|
|
|
|
2019-07-23 02:47:54 +02:00
|
|
|
void WriteCorrelationFeatures(const std::string& label_str,
|
|
|
|
uint32_t max_number_of_values) const;
|
|
|
|
|
|
|
|
void WriteCorrelationFeaturesForGet(uint32_t max_number_of_values) const;
|
|
|
|
|
Pysim more algorithms (#5644)
Summary:
This PR adds four more eviction policies.
- OPT [1]
- Hyperbolic caching [2]
- ARC [3]
- GreedyDualSize [4]
[1] L. A. Belady. 1966. A Study of Replacement Algorithms for a Virtual-storage Computer. IBM Syst. J. 5, 2 (June 1966), 78-101. DOI=http://dx.doi.org/10.1147/sj.52.0078
[2] Aaron Blankstein, Siddhartha Sen, and Michael J. Freedman. 2017. Hyperbolic caching: flexible caching for web applications. In Proceedings of the 2017 USENIX Conference on Usenix Annual Technical Conference (USENIX ATC '17). USENIX Association, Berkeley, CA, USA, 499-511.
[3] Nimrod Megiddo and Dharmendra S. Modha. 2003. ARC: A Self-Tuning, Low Overhead Replacement Cache. In Proceedings of the 2nd USENIX Conference on File and Storage Technologies (FAST '03). USENIX Association, Berkeley, CA, USA, 115-130.
[4] N. Young. The k-server dual and loose competitiveness for paging. Algorithmica, June 1994, vol. 11,(no.6):525-41. Rewritten version of ''On-line caching as cache size varies'', in The 2nd Annual ACM-SIAM Symposium on Discrete Algorithms, 241-250, 1991.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5644
Differential Revision: D16548817
Pulled By: HaoyuHuang
fbshipit-source-id: 838f76db9179f07911abaab46c97e1c929cfcd63
2019-08-07 03:47:39 +02:00
|
|
|
void WriteSkewness(const std::string& label_str,
|
|
|
|
const std::vector<uint64_t>& percent_buckets,
|
|
|
|
TraceType target_block_type) const;
|
|
|
|
|
2019-06-11 21:18:37 +02:00
|
|
|
const std::map<std::string, ColumnFamilyAccessInfoAggregate>&
|
|
|
|
TEST_cf_aggregates_map() const {
|
|
|
|
return cf_aggregates_map_;
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
2019-06-25 05:38:20 +02:00
|
|
|
std::set<std::string> ParseLabelStr(const std::string& label_str) const;
|
|
|
|
|
|
|
|
std::string BuildLabel(const std::set<std::string>& labels,
|
|
|
|
const std::string& cf_name, uint64_t fd,
|
|
|
|
uint32_t level, TraceType type,
|
Pysim more algorithms (#5644)
Summary:
This PR adds four more eviction policies.
- OPT [1]
- Hyperbolic caching [2]
- ARC [3]
- GreedyDualSize [4]
[1] L. A. Belady. 1966. A Study of Replacement Algorithms for a Virtual-storage Computer. IBM Syst. J. 5, 2 (June 1966), 78-101. DOI=http://dx.doi.org/10.1147/sj.52.0078
[2] Aaron Blankstein, Siddhartha Sen, and Michael J. Freedman. 2017. Hyperbolic caching: flexible caching for web applications. In Proceedings of the 2017 USENIX Conference on Usenix Annual Technical Conference (USENIX ATC '17). USENIX Association, Berkeley, CA, USA, 499-511.
[3] Nimrod Megiddo and Dharmendra S. Modha. 2003. ARC: A Self-Tuning, Low Overhead Replacement Cache. In Proceedings of the 2nd USENIX Conference on File and Storage Technologies (FAST '03). USENIX Association, Berkeley, CA, USA, 115-130.
[4] N. Young. The k-server dual and loose competitiveness for paging. Algorithmica, June 1994, vol. 11,(no.6):525-41. Rewritten version of ''On-line caching as cache size varies'', in The 2nd Annual ACM-SIAM Symposium on Discrete Algorithms, 241-250, 1991.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5644
Differential Revision: D16548817
Pulled By: HaoyuHuang
fbshipit-source-id: 838f76db9179f07911abaab46c97e1c929cfcd63
2019-08-07 03:47:39 +02:00
|
|
|
TableReaderCaller caller, uint64_t block_key,
|
|
|
|
const BlockAccessInfo& block) const;
|
2019-06-25 05:38:20 +02:00
|
|
|
|
|
|
|
void ComputeReuseDistance(BlockAccessInfo* info) const;
|
|
|
|
|
2019-07-23 02:47:54 +02:00
|
|
|
Status RecordAccess(const BlockCacheTraceRecord& access);
|
2019-06-11 21:18:37 +02:00
|
|
|
|
2019-06-25 05:38:20 +02:00
|
|
|
void UpdateReuseIntervalStats(
|
2019-07-13 01:52:15 +02:00
|
|
|
const std::string& label, const std::vector<uint64_t>& time_buckets,
|
2019-06-25 05:38:20 +02:00
|
|
|
const std::map<uint64_t, uint64_t> timeline,
|
|
|
|
std::map<std::string, std::map<uint64_t, uint64_t>>*
|
|
|
|
label_time_num_reuses,
|
|
|
|
uint64_t* total_num_reuses) const;
|
|
|
|
|
2019-07-13 01:52:15 +02:00
|
|
|
std::string OutputPercentAccessStats(
|
|
|
|
uint64_t total_accesses,
|
|
|
|
const std::map<std::string, uint64_t>& cf_access_count) const;
|
|
|
|
|
|
|
|
void WriteStatsToFile(
|
|
|
|
const std::string& label_str, const std::vector<uint64_t>& time_buckets,
|
|
|
|
const std::string& filename_suffix,
|
|
|
|
const std::map<std::string, std::map<uint64_t, uint64_t>>& label_data,
|
|
|
|
uint64_t ntotal) const;
|
|
|
|
|
|
|
|
void TraverseBlocks(
|
|
|
|
std::function<void(const std::string& /*cf_name*/, uint64_t /*fd*/,
|
|
|
|
uint32_t /*level*/, TraceType /*block_type*/,
|
|
|
|
const std::string& /*block_key*/,
|
|
|
|
uint64_t /*block_key_id*/,
|
|
|
|
const BlockAccessInfo& /*block_access_info*/)>
|
Pysim more algorithms (#5644)
Summary:
This PR adds four more eviction policies.
- OPT [1]
- Hyperbolic caching [2]
- ARC [3]
- GreedyDualSize [4]
[1] L. A. Belady. 1966. A Study of Replacement Algorithms for a Virtual-storage Computer. IBM Syst. J. 5, 2 (June 1966), 78-101. DOI=http://dx.doi.org/10.1147/sj.52.0078
[2] Aaron Blankstein, Siddhartha Sen, and Michael J. Freedman. 2017. Hyperbolic caching: flexible caching for web applications. In Proceedings of the 2017 USENIX Conference on Usenix Annual Technical Conference (USENIX ATC '17). USENIX Association, Berkeley, CA, USA, 499-511.
[3] Nimrod Megiddo and Dharmendra S. Modha. 2003. ARC: A Self-Tuning, Low Overhead Replacement Cache. In Proceedings of the 2nd USENIX Conference on File and Storage Technologies (FAST '03). USENIX Association, Berkeley, CA, USA, 115-130.
[4] N. Young. The k-server dual and loose competitiveness for paging. Algorithmica, June 1994, vol. 11,(no.6):525-41. Rewritten version of ''On-line caching as cache size varies'', in The 2nd Annual ACM-SIAM Symposium on Discrete Algorithms, 241-250, 1991.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5644
Differential Revision: D16548817
Pulled By: HaoyuHuang
fbshipit-source-id: 838f76db9179f07911abaab46c97e1c929cfcd63
2019-08-07 03:47:39 +02:00
|
|
|
block_callback,
|
|
|
|
std::set<std::string>* labels = nullptr) const;
|
2019-07-13 01:52:15 +02:00
|
|
|
|
2019-07-23 02:47:54 +02:00
|
|
|
void UpdateFeatureVectors(
|
|
|
|
const std::vector<uint64_t>& access_sequence_number_timeline,
|
|
|
|
const std::vector<uint64_t>& access_timeline, const std::string& label,
|
|
|
|
std::map<std::string, Features>* label_features,
|
|
|
|
std::map<std::string, Predictions>* label_predictions) const;
|
|
|
|
|
|
|
|
void WriteCorrelationFeaturesToFile(
|
|
|
|
const std::string& label,
|
|
|
|
const std::map<std::string, Features>& label_features,
|
|
|
|
const std::map<std::string, Predictions>& label_predictions,
|
|
|
|
uint32_t max_number_of_values) const;
|
|
|
|
|
2020-02-20 21:07:53 +01:00
|
|
|
ROCKSDB_NAMESPACE::Env* env_;
|
Support computing miss ratio curves using sim_cache. (#5449)
Summary:
This PR adds a BlockCacheTraceSimulator that reports the miss ratios given different cache configurations. A cache configuration contains "cache_name,num_shard_bits,cache_capacities". For example, "lru, 1, 1K, 2K, 4M, 4G".
When we replay the trace, we also perform lookups and inserts on the simulated caches.
In the end, it reports the miss ratio for each tuple <cache_name, num_shard_bits, cache_capacity> in a output file.
This PR also adds a main source block_cache_trace_analyzer so that we can run the analyzer in command line.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5449
Test Plan:
Added tests for block_cache_trace_analyzer.
COMPILE_WITH_ASAN=1 make check -j32.
Differential Revision: D15797073
Pulled By: HaoyuHuang
fbshipit-source-id: aef0c5c2e7938f3e8b6a10d4a6a50e6928ecf408
2019-06-18 01:33:40 +02:00
|
|
|
const std::string trace_file_path_;
|
2019-06-25 05:38:20 +02:00
|
|
|
const std::string output_dir_;
|
2019-07-23 02:47:54 +02:00
|
|
|
std::string human_readable_trace_file_path_;
|
2019-07-13 01:52:15 +02:00
|
|
|
const bool compute_reuse_distance_;
|
2019-07-23 02:47:54 +02:00
|
|
|
const bool mrc_only_;
|
2019-08-09 22:09:04 +02:00
|
|
|
const bool is_human_readable_trace_file_;
|
Support computing miss ratio curves using sim_cache. (#5449)
Summary:
This PR adds a BlockCacheTraceSimulator that reports the miss ratios given different cache configurations. A cache configuration contains "cache_name,num_shard_bits,cache_capacities". For example, "lru, 1, 1K, 2K, 4M, 4G".
When we replay the trace, we also perform lookups and inserts on the simulated caches.
In the end, it reports the miss ratio for each tuple <cache_name, num_shard_bits, cache_capacity> in a output file.
This PR also adds a main source block_cache_trace_analyzer so that we can run the analyzer in command line.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5449
Test Plan:
Added tests for block_cache_trace_analyzer.
COMPILE_WITH_ASAN=1 make check -j32.
Differential Revision: D15797073
Pulled By: HaoyuHuang
fbshipit-source-id: aef0c5c2e7938f3e8b6a10d4a6a50e6928ecf408
2019-06-18 01:33:40 +02:00
|
|
|
|
2019-06-11 21:18:37 +02:00
|
|
|
BlockCacheTraceHeader header_;
|
Support computing miss ratio curves using sim_cache. (#5449)
Summary:
This PR adds a BlockCacheTraceSimulator that reports the miss ratios given different cache configurations. A cache configuration contains "cache_name,num_shard_bits,cache_capacities". For example, "lru, 1, 1K, 2K, 4M, 4G".
When we replay the trace, we also perform lookups and inserts on the simulated caches.
In the end, it reports the miss ratio for each tuple <cache_name, num_shard_bits, cache_capacity> in a output file.
This PR also adds a main source block_cache_trace_analyzer so that we can run the analyzer in command line.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5449
Test Plan:
Added tests for block_cache_trace_analyzer.
COMPILE_WITH_ASAN=1 make check -j32.
Differential Revision: D15797073
Pulled By: HaoyuHuang
fbshipit-source-id: aef0c5c2e7938f3e8b6a10d4a6a50e6928ecf408
2019-06-18 01:33:40 +02:00
|
|
|
std::unique_ptr<BlockCacheTraceSimulator> cache_simulator_;
|
2019-06-11 21:18:37 +02:00
|
|
|
std::map<std::string, ColumnFamilyAccessInfoAggregate> cf_aggregates_map_;
|
2019-06-25 05:38:20 +02:00
|
|
|
std::map<std::string, BlockAccessInfo*> block_info_map_;
|
2019-07-23 02:47:54 +02:00
|
|
|
std::unordered_map<std::string, GetKeyInfo> get_key_info_map_;
|
|
|
|
uint64_t access_sequence_number_ = 0;
|
2019-07-13 01:52:15 +02:00
|
|
|
uint64_t trace_start_timestamp_in_seconds_ = 0;
|
|
|
|
uint64_t trace_end_timestamp_in_seconds_ = 0;
|
2019-07-23 02:47:54 +02:00
|
|
|
MissRatioStats miss_ratio_stats_;
|
|
|
|
uint64_t unique_block_id_ = 1;
|
|
|
|
uint64_t unique_get_key_id_ = 1;
|
2019-08-09 22:09:04 +02:00
|
|
|
BlockCacheHumanReadableTraceWriter human_readable_trace_writer_;
|
2019-06-11 21:18:37 +02:00
|
|
|
};
|
|
|
|
|
Support computing miss ratio curves using sim_cache. (#5449)
Summary:
This PR adds a BlockCacheTraceSimulator that reports the miss ratios given different cache configurations. A cache configuration contains "cache_name,num_shard_bits,cache_capacities". For example, "lru, 1, 1K, 2K, 4M, 4G".
When we replay the trace, we also perform lookups and inserts on the simulated caches.
In the end, it reports the miss ratio for each tuple <cache_name, num_shard_bits, cache_capacity> in a output file.
This PR also adds a main source block_cache_trace_analyzer so that we can run the analyzer in command line.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5449
Test Plan:
Added tests for block_cache_trace_analyzer.
COMPILE_WITH_ASAN=1 make check -j32.
Differential Revision: D15797073
Pulled By: HaoyuHuang
fbshipit-source-id: aef0c5c2e7938f3e8b6a10d4a6a50e6928ecf408
2019-06-18 01:33:40 +02:00
|
|
|
int block_cache_trace_analyzer_tool(int argc, char** argv);
|
|
|
|
|
2020-02-20 21:07:53 +01:00
|
|
|
} // namespace ROCKSDB_NAMESPACE
|