d755c62f92
Summary: This provides provides an implementation of PersistentCacheTier that is specialized for RAM. This tier does not persist data though. Why do we need this tier ? This is ideal as tier 0. This tier can host data that is too hot. Why can't we use Cache variants ? Yes you can use them instead. This tier can potentially outperform BlockCache in RAW mode by virtue of compression and compressed cache in block cache doesn't seem very popular. Potentially this tier can be modified to under stand the disadvantage of the tier below and retain data that the tier below is bad at handling (for example index and bloom data that is huge in size) Test Plan: Run unit tests added Subscribers: andrewkr, dhruba, leveldb Differential Revision: https://reviews.facebook.net/D57069
167 lines
4.8 KiB
C++
167 lines
4.8 KiB
C++
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under the BSD-style license found in the
|
|
// LICENSE file in the root directory of this source tree. An additional grant
|
|
// of patent rights can be found in the PATENTS file in the same directory.
|
|
//
|
|
#pragma once
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
#include "util/random.h"
|
|
#include "utilities/persistent_cache/hash_table.h"
|
|
#include "utilities/persistent_cache/lrulist.h"
|
|
|
|
namespace rocksdb {
|
|
|
|
// Evictable Hash Table
|
|
//
|
|
// Hash table index where least accessed (or one of the least accessed) elements
|
|
// can be evicted.
|
|
//
|
|
// Please note EvictableHashTable can only be created for pointer type objects
|
|
template <class T, class Hash, class Equal>
|
|
class EvictableHashTable : private HashTable<T*, Hash, Equal> {
|
|
public:
|
|
typedef HashTable<T*, Hash, Equal> hash_table;
|
|
|
|
explicit EvictableHashTable(const size_t capacity = 1024 * 1024,
|
|
const float load_factor = 2.0,
|
|
const uint32_t nlocks = 256)
|
|
: HashTable<T*, Hash, Equal>(capacity, load_factor, nlocks),
|
|
lru_lists_(new LRUList<T>[hash_table::nlocks_]) {
|
|
assert(lru_lists_);
|
|
}
|
|
|
|
virtual ~EvictableHashTable() { AssertEmptyLRU(); }
|
|
|
|
//
|
|
// Insert given record to hash table (and LRU list)
|
|
//
|
|
bool Insert(T* t) {
|
|
const uint64_t h = Hash()(t);
|
|
typename hash_table::Bucket& bucket = GetBucket(h);
|
|
LRUListType& lru = GetLRUList(h);
|
|
port::RWMutex& lock = GetMutex(h);
|
|
|
|
WriteLock _(&lock);
|
|
if (hash_table::Insert(&bucket, t)) {
|
|
lru.Push(t);
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
//
|
|
// Lookup hash table
|
|
//
|
|
// Please note that read lock should be held by the caller. This is because
|
|
// the caller owns the data, and should hold the read lock as long as he
|
|
// operates on the data.
|
|
bool Find(T* t, T** ret) {
|
|
const uint64_t h = Hash()(t);
|
|
typename hash_table::Bucket& bucket = GetBucket(h);
|
|
LRUListType& lru = GetLRUList(h);
|
|
port::RWMutex& lock = GetMutex(h);
|
|
|
|
ReadLock _(&lock);
|
|
if (hash_table::Find(&bucket, t, ret)) {
|
|
++(*ret)->refs_;
|
|
lru.Touch(*ret);
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
//
|
|
// Evict one of the least recently used object
|
|
//
|
|
T* Evict(const std::function<void(T*)>& fn = nullptr) {
|
|
const size_t start_idx = rand_.Next() % hash_table::nlocks_;
|
|
T* t = nullptr;
|
|
|
|
// iterate from start_idx .. 0 .. start_idx
|
|
for (size_t i = 0; !t && i < hash_table::nlocks_; ++i) {
|
|
const size_t idx = (start_idx + i) % hash_table::nlocks_;
|
|
|
|
WriteLock _(&hash_table::locks_[idx]);
|
|
LRUListType& lru = lru_lists_[idx];
|
|
if (!lru.IsEmpty() && (t = lru.Pop())) {
|
|
assert(!t->refs_);
|
|
// We got an item to evict, erase from the bucket
|
|
const uint64_t h = Hash()(t);
|
|
typename hash_table::Bucket& bucket = GetBucket(h);
|
|
T* tmp = nullptr;
|
|
bool status = hash_table::Erase(&bucket, t, &tmp);
|
|
assert(t == tmp);
|
|
(void)status;
|
|
assert(status);
|
|
if (fn) {
|
|
fn(t);
|
|
}
|
|
break;
|
|
}
|
|
assert(!t);
|
|
}
|
|
return t;
|
|
}
|
|
|
|
void Clear(void (*fn)(T*)) {
|
|
for (uint32_t i = 0; i < hash_table::nbuckets_; ++i) {
|
|
const uint32_t lock_idx = i % hash_table::nlocks_;
|
|
WriteLock _(&hash_table::locks_[lock_idx]);
|
|
auto& lru_list = lru_lists_[lock_idx];
|
|
auto& bucket = hash_table::buckets_[i];
|
|
for (auto* t : bucket.list_) {
|
|
lru_list.Unlink(t);
|
|
(*fn)(t);
|
|
}
|
|
bucket.list_.clear();
|
|
}
|
|
// make sure that all LRU lists are emptied
|
|
AssertEmptyLRU();
|
|
}
|
|
|
|
void AssertEmptyLRU() {
|
|
#ifndef NDEBUG
|
|
for (uint32_t i = 0; i < hash_table::nlocks_; ++i) {
|
|
WriteLock _(&hash_table::locks_[i]);
|
|
auto& lru_list = lru_lists_[i];
|
|
assert(lru_list.IsEmpty());
|
|
}
|
|
#endif
|
|
}
|
|
|
|
//
|
|
// Fetch the mutex associated with a key
|
|
// This call is used to hold the lock for a given data for extended period of
|
|
// time.
|
|
port::RWMutex* GetMutex(T* t) { return hash_table::GetMutex(t); }
|
|
|
|
private:
|
|
typedef LRUList<T> LRUListType;
|
|
|
|
typename hash_table::Bucket& GetBucket(const uint64_t h) {
|
|
const uint32_t bucket_idx = h % hash_table::nbuckets_;
|
|
return hash_table::buckets_[bucket_idx];
|
|
}
|
|
|
|
LRUListType& GetLRUList(const uint64_t h) {
|
|
const uint32_t bucket_idx = h % hash_table::nbuckets_;
|
|
const uint32_t lock_idx = bucket_idx % hash_table::nlocks_;
|
|
return lru_lists_[lock_idx];
|
|
}
|
|
|
|
port::RWMutex& GetMutex(const uint64_t h) {
|
|
const uint32_t bucket_idx = h % hash_table::nbuckets_;
|
|
const uint32_t lock_idx = bucket_idx % hash_table::nlocks_;
|
|
return hash_table::locks_[lock_idx];
|
|
}
|
|
|
|
Random64 rand_{static_cast<uint64_t>(time(nullptr))};
|
|
std::unique_ptr<LRUListType[]> lru_lists_;
|
|
};
|
|
|
|
} // namespace rocksdb
|
|
|
|
#endif
|