d5a46c40e5
Summary: If the block Cache is full with strict_capacity_limit=false, then our CacheEntryStatsCollector could be immediately evicted on release, so iterating through column families with shared block cache could trigger re-scan for each CF. This change fixes that problem by pinning the CacheEntryStatsCollector from InternalStats so that it's not evicted. I had originally thought that this object could participate in LRU like everything else, but even though a re-load+re-scan only touches memory, it can be orders of magnitude more expensive than other cache misses. One service in Facebook has scans that take ~20s over 100GB block cache that is mostly 4KB entries. (The up-side of this bug and https://github.com/facebook/rocksdb/issues/8369 is that we had a natural experiment on the effect on some service metrics even with block cache scans running continuously in the background--a kind of worst case scenario. Metrics like latency were not affected enough to trigger warnings.) Other smaller fixes: 20s is already a sizable portion of 600s stats dump period, or 180s default max age to force re-scan, so added logic to ensure that (for each block cache) we don't spend more than 0.2% of our background thread time scanning it. Nevertheless, "foreground" requests for cache entry stats (calls to `db->GetMapProperty(DB::Properties::kBlockCacheEntryStats)`) are permitted to consume more CPU. Renamed field to cache_entry_stats_ to match code style. This change is intended for patching in 6.21 release. Pull Request resolved: https://github.com/facebook/rocksdb/pull/8385 Test Plan: unit test expanded to cover new logic (detect regression), some manual testing with db_bench Reviewed By: ajkr Differential Revision: D29042759 Pulled By: pdillinger fbshipit-source-id: 236faa902397f50038c618f50fbc8cf3f277308c
166 lines
6.4 KiB
C++
166 lines
6.4 KiB
C++
// Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
#pragma once
|
|
|
|
#include <array>
|
|
#include <cstdint>
|
|
#include <memory>
|
|
#include <mutex>
|
|
|
|
#include "cache/cache_helpers.h"
|
|
#include "port/lang.h"
|
|
#include "rocksdb/cache.h"
|
|
#include "rocksdb/status.h"
|
|
#include "rocksdb/system_clock.h"
|
|
#include "util/coding_lean.h"
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
// A generic helper object for gathering stats about cache entries by
|
|
// iterating over them with ApplyToAllEntries. This class essentially
|
|
// solves the problem of slowing down a Cache with too many stats
|
|
// collectors that could be sharing stat results, such as from multiple
|
|
// column families or multiple DBs sharing a Cache. We employ a few
|
|
// mitigations:
|
|
// * Only one collector for a particular kind of Stats is alive
|
|
// for each Cache. This is guaranteed using the Cache itself to hold
|
|
// the collector.
|
|
// * A mutex ensures only one thread is gathering stats for this
|
|
// collector.
|
|
// * The most recent gathered stats are saved and simply copied to
|
|
// satisfy requests within a time window (default: 3 minutes) of
|
|
// completion of the most recent stat gathering.
|
|
//
|
|
// Template parameter Stats must be copyable and trivially constructable,
|
|
// as well as...
|
|
// concept Stats {
|
|
// // Notification before applying callback to all entries
|
|
// void BeginCollection(Cache*, SystemClock*, uint64_t start_time_micros);
|
|
// // Get the callback to apply to all entries. `callback`
|
|
// // type must be compatible with Cache::ApplyToAllEntries
|
|
// callback GetEntryCallback();
|
|
// // Notification after applying callback to all entries
|
|
// void EndCollection(Cache*, SystemClock*, uint64_t end_time_micros);
|
|
// // Notification that a collection was skipped because of
|
|
// // sufficiently recent saved results.
|
|
// void SkippedCollection();
|
|
// }
|
|
template <class Stats>
|
|
class CacheEntryStatsCollector {
|
|
public:
|
|
// Gathers stats and saves results into `stats`
|
|
//
|
|
// Maximum allowed age for a "hit" on saved results is determined by the
|
|
// two interval parameters. Both set to 0 forces a re-scan. For example
|
|
// with min_interval_seconds=300 and min_interval_factor=100, if the last
|
|
// scan took 10s, we would only rescan ("miss") if the age in seconds of
|
|
// the saved results is > max(300, 100*10).
|
|
// Justification: scans can vary wildly in duration, e.g. from 0.02 sec
|
|
// to as much as 20 seconds, so we want to be able to cap the absolute
|
|
// and relative frequency of scans.
|
|
void GetStats(Stats *stats, int min_interval_seconds,
|
|
int min_interval_factor) {
|
|
// Waits for any pending reader or writer (collector)
|
|
std::lock_guard<std::mutex> lock(mutex_);
|
|
|
|
uint64_t max_age_micros =
|
|
static_cast<uint64_t>(std::max(min_interval_seconds, 0)) * 1000000U;
|
|
|
|
if (last_end_time_micros_ > last_start_time_micros_ &&
|
|
min_interval_factor > 0) {
|
|
max_age_micros = std::max(
|
|
max_age_micros, min_interval_factor * (last_end_time_micros_ -
|
|
last_start_time_micros_));
|
|
}
|
|
|
|
uint64_t start_time_micros = clock_->NowMicros();
|
|
if ((start_time_micros - last_end_time_micros_) > max_age_micros) {
|
|
last_start_time_micros_ = start_time_micros;
|
|
saved_stats_.BeginCollection(cache_, clock_, start_time_micros);
|
|
|
|
cache_->ApplyToAllEntries(saved_stats_.GetEntryCallback(), {});
|
|
TEST_SYNC_POINT_CALLBACK(
|
|
"CacheEntryStatsCollector::GetStats:AfterApplyToAllEntries", nullptr);
|
|
|
|
uint64_t end_time_micros = clock_->NowMicros();
|
|
last_end_time_micros_ = end_time_micros;
|
|
saved_stats_.EndCollection(cache_, clock_, end_time_micros);
|
|
} else {
|
|
saved_stats_.SkippedCollection();
|
|
}
|
|
// Copy to caller
|
|
*stats = saved_stats_;
|
|
}
|
|
|
|
Cache *GetCache() const { return cache_; }
|
|
|
|
// Gets or creates a shared instance of CacheEntryStatsCollector in the
|
|
// cache itself, and saves into `ptr`. This shared_ptr will hold the
|
|
// entry in cache until all refs are destroyed.
|
|
static Status GetShared(Cache *cache, SystemClock *clock,
|
|
std::shared_ptr<CacheEntryStatsCollector> *ptr) {
|
|
std::array<uint64_t, 3> cache_key_data{
|
|
{// First 16 bytes == md5 of class name
|
|
0x7eba5a8fb5437c90U, 0x8ca68c9b11655855U,
|
|
// Last 8 bytes based on a function pointer to make unique for each
|
|
// template instantiation
|
|
reinterpret_cast<uint64_t>(&CacheEntryStatsCollector::GetShared)}};
|
|
Slice cache_key = GetSlice(&cache_key_data);
|
|
|
|
Cache::Handle *h = cache->Lookup(cache_key);
|
|
if (h == nullptr) {
|
|
// Not yet in cache, but Cache doesn't provide a built-in way to
|
|
// avoid racing insert. So we double-check under a shared mutex,
|
|
// inspired by TableCache.
|
|
STATIC_AVOID_DESTRUCTION(std::mutex, static_mutex);
|
|
std::lock_guard<std::mutex> lock(static_mutex);
|
|
|
|
h = cache->Lookup(cache_key);
|
|
if (h == nullptr) {
|
|
auto new_ptr = new CacheEntryStatsCollector(cache, clock);
|
|
// TODO: non-zero charge causes some tests that count block cache
|
|
// usage to go flaky. Fix the problem somehow so we can use an
|
|
// accurate charge.
|
|
size_t charge = 0;
|
|
Status s = cache->Insert(cache_key, new_ptr, charge, Deleter, &h,
|
|
Cache::Priority::HIGH);
|
|
if (!s.ok()) {
|
|
assert(h == nullptr);
|
|
return s;
|
|
}
|
|
}
|
|
}
|
|
// If we reach here, shared entry is in cache with handle `h`.
|
|
assert(cache->GetDeleter(h) == Deleter);
|
|
|
|
// Build an aliasing shared_ptr that keeps `ptr` in cache while there
|
|
// are references.
|
|
*ptr = MakeSharedCacheHandleGuard<CacheEntryStatsCollector>(cache, h);
|
|
return Status::OK();
|
|
}
|
|
|
|
private:
|
|
explicit CacheEntryStatsCollector(Cache *cache, SystemClock *clock)
|
|
: saved_stats_(),
|
|
last_start_time_micros_(0),
|
|
last_end_time_micros_(/*pessimistic*/ 10000000),
|
|
cache_(cache),
|
|
clock_(clock) {}
|
|
|
|
static void Deleter(const Slice &, void *value) {
|
|
delete static_cast<CacheEntryStatsCollector *>(value);
|
|
}
|
|
|
|
std::mutex mutex_;
|
|
Stats saved_stats_;
|
|
uint64_t last_start_time_micros_;
|
|
uint64_t last_end_time_micros_;
|
|
Cache *const cache_;
|
|
SystemClock *const clock_;
|
|
};
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|