7259e28d91
Summary: Enhancement to MultiGet batching to read data blocks required for keys in a batch in parallel from disk. It uses Env::MultiRead() API to read multiple blocks and reduce latency. Pull Request resolved: https://github.com/facebook/rocksdb/pull/5464 Test Plan: 1. make check 2. make asan_check 3. make asan_crash Differential Revision: D15911771 Pulled By: anand1976 fbshipit-source-id: 605036b9af0f90ca0020dc87c3a86b4da6e83394
401 lines
14 KiB
C++
401 lines
14 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
#pragma once
|
|
#include <atomic>
|
|
#include <sstream>
|
|
#include <string>
|
|
#include "port/port.h"
|
|
#include "rocksdb/env.h"
|
|
#include "rocksdb/listener.h"
|
|
#include "rocksdb/rate_limiter.h"
|
|
#include "test_util/sync_point.h"
|
|
#include "util/aligned_buffer.h"
|
|
|
|
namespace rocksdb {
|
|
|
|
class Statistics;
|
|
class HistogramImpl;
|
|
|
|
// This file provides the following main abstractions:
|
|
// SequentialFileReader : wrapper over Env::SequentialFile
|
|
// RandomAccessFileReader : wrapper over Env::RandomAccessFile
|
|
// WritableFileWriter : wrapper over Env::WritableFile
|
|
// In addition, it also exposed NewReadaheadRandomAccessFile, NewWritableFile,
|
|
// and ReadOneLine primitives.
|
|
|
|
// NewReadaheadRandomAccessFile provides a wrapper over RandomAccessFile to
|
|
// always prefetch additional data with every read. This is mainly used in
|
|
// Compaction Table Readers.
|
|
std::unique_ptr<RandomAccessFile> NewReadaheadRandomAccessFile(
|
|
std::unique_ptr<RandomAccessFile>&& file, size_t readahead_size);
|
|
|
|
// SequentialFileReader is a wrapper on top of Env::SequentialFile. It handles
|
|
// Buffered (i.e when page cache is enabled) and Direct (with O_DIRECT / page
|
|
// cache disabled) reads appropriately, and also updates the IO stats.
|
|
class SequentialFileReader {
|
|
private:
|
|
std::unique_ptr<SequentialFile> file_;
|
|
std::string file_name_;
|
|
std::atomic<size_t> offset_; // read offset
|
|
|
|
public:
|
|
explicit SequentialFileReader(std::unique_ptr<SequentialFile>&& _file,
|
|
const std::string& _file_name)
|
|
: file_(std::move(_file)), file_name_(_file_name), offset_(0) {}
|
|
|
|
SequentialFileReader(SequentialFileReader&& o) ROCKSDB_NOEXCEPT {
|
|
*this = std::move(o);
|
|
}
|
|
|
|
SequentialFileReader& operator=(SequentialFileReader&& o) ROCKSDB_NOEXCEPT {
|
|
file_ = std::move(o.file_);
|
|
return *this;
|
|
}
|
|
|
|
SequentialFileReader(const SequentialFileReader&) = delete;
|
|
SequentialFileReader& operator=(const SequentialFileReader&) = delete;
|
|
|
|
Status Read(size_t n, Slice* result, char* scratch);
|
|
|
|
Status Skip(uint64_t n);
|
|
|
|
void Rewind();
|
|
|
|
SequentialFile* file() { return file_.get(); }
|
|
|
|
std::string file_name() { return file_name_; }
|
|
|
|
bool use_direct_io() const { return file_->use_direct_io(); }
|
|
};
|
|
|
|
// RandomAccessFileReader is a wrapper on top of Env::RnadomAccessFile. It is
|
|
// responsible for:
|
|
// - Handling Buffered and Direct reads appropriately.
|
|
// - Rate limiting compaction reads.
|
|
// - Notifying any interested listeners on the completion of a read.
|
|
// - Updating IO stats.
|
|
class RandomAccessFileReader {
|
|
private:
|
|
#ifndef ROCKSDB_LITE
|
|
void NotifyOnFileReadFinish(uint64_t offset, size_t length,
|
|
const FileOperationInfo::TimePoint& start_ts,
|
|
const FileOperationInfo::TimePoint& finish_ts,
|
|
const Status& status) const {
|
|
FileOperationInfo info(file_name_, start_ts, finish_ts);
|
|
info.offset = offset;
|
|
info.length = length;
|
|
info.status = status;
|
|
|
|
for (auto& listener : listeners_) {
|
|
listener->OnFileReadFinish(info);
|
|
}
|
|
}
|
|
#endif // ROCKSDB_LITE
|
|
|
|
bool ShouldNotifyListeners() const { return !listeners_.empty(); }
|
|
|
|
std::unique_ptr<RandomAccessFile> file_;
|
|
std::string file_name_;
|
|
Env* env_;
|
|
Statistics* stats_;
|
|
uint32_t hist_type_;
|
|
HistogramImpl* file_read_hist_;
|
|
RateLimiter* rate_limiter_;
|
|
bool for_compaction_;
|
|
std::vector<std::shared_ptr<EventListener>> listeners_;
|
|
|
|
public:
|
|
explicit RandomAccessFileReader(
|
|
std::unique_ptr<RandomAccessFile>&& raf, std::string _file_name,
|
|
Env* env = nullptr, Statistics* stats = nullptr, uint32_t hist_type = 0,
|
|
HistogramImpl* file_read_hist = nullptr,
|
|
RateLimiter* rate_limiter = nullptr, bool for_compaction = false,
|
|
const std::vector<std::shared_ptr<EventListener>>& listeners = {})
|
|
: file_(std::move(raf)),
|
|
file_name_(std::move(_file_name)),
|
|
env_(env),
|
|
stats_(stats),
|
|
hist_type_(hist_type),
|
|
file_read_hist_(file_read_hist),
|
|
rate_limiter_(rate_limiter),
|
|
for_compaction_(for_compaction),
|
|
listeners_() {
|
|
#ifndef ROCKSDB_LITE
|
|
std::for_each(listeners.begin(), listeners.end(),
|
|
[this](const std::shared_ptr<EventListener>& e) {
|
|
if (e->ShouldBeNotifiedOnFileIO()) {
|
|
listeners_.emplace_back(e);
|
|
}
|
|
});
|
|
#else // !ROCKSDB_LITE
|
|
(void)listeners;
|
|
#endif
|
|
}
|
|
|
|
RandomAccessFileReader(RandomAccessFileReader&& o) ROCKSDB_NOEXCEPT {
|
|
*this = std::move(o);
|
|
}
|
|
|
|
RandomAccessFileReader& operator=(RandomAccessFileReader&& o)
|
|
ROCKSDB_NOEXCEPT {
|
|
file_ = std::move(o.file_);
|
|
env_ = std::move(o.env_);
|
|
stats_ = std::move(o.stats_);
|
|
hist_type_ = std::move(o.hist_type_);
|
|
file_read_hist_ = std::move(o.file_read_hist_);
|
|
rate_limiter_ = std::move(o.rate_limiter_);
|
|
for_compaction_ = std::move(o.for_compaction_);
|
|
return *this;
|
|
}
|
|
|
|
RandomAccessFileReader(const RandomAccessFileReader&) = delete;
|
|
RandomAccessFileReader& operator=(const RandomAccessFileReader&) = delete;
|
|
|
|
Status Read(uint64_t offset, size_t n, Slice* result, char* scratch,
|
|
bool for_compaction = false) const;
|
|
|
|
Status MultiRead(ReadRequest* reqs, size_t num_reqs) const;
|
|
|
|
Status Prefetch(uint64_t offset, size_t n) const {
|
|
return file_->Prefetch(offset, n);
|
|
}
|
|
|
|
RandomAccessFile* file() { return file_.get(); }
|
|
|
|
std::string file_name() const { return file_name_; }
|
|
|
|
bool use_direct_io() const { return file_->use_direct_io(); }
|
|
};
|
|
|
|
// WritableFileWriter is a wrapper on top of Env::WritableFile. It provides
|
|
// facilities to:
|
|
// - Handle Buffered and Direct writes.
|
|
// - Rate limit writes.
|
|
// - Flush and Sync the data to the underlying filesystem.
|
|
// - Notify any interested listeners on the completion of a write.
|
|
// - Update IO stats.
|
|
class WritableFileWriter {
|
|
private:
|
|
#ifndef ROCKSDB_LITE
|
|
void NotifyOnFileWriteFinish(uint64_t offset, size_t length,
|
|
const FileOperationInfo::TimePoint& start_ts,
|
|
const FileOperationInfo::TimePoint& finish_ts,
|
|
const Status& status) {
|
|
FileOperationInfo info(file_name_, start_ts, finish_ts);
|
|
info.offset = offset;
|
|
info.length = length;
|
|
info.status = status;
|
|
|
|
for (auto& listener : listeners_) {
|
|
listener->OnFileWriteFinish(info);
|
|
}
|
|
}
|
|
#endif // ROCKSDB_LITE
|
|
|
|
bool ShouldNotifyListeners() const { return !listeners_.empty(); }
|
|
|
|
std::unique_ptr<WritableFile> writable_file_;
|
|
std::string file_name_;
|
|
Env* env_;
|
|
AlignedBuffer buf_;
|
|
size_t max_buffer_size_;
|
|
// Actually written data size can be used for truncate
|
|
// not counting padding data
|
|
uint64_t filesize_;
|
|
#ifndef ROCKSDB_LITE
|
|
// This is necessary when we use unbuffered access
|
|
// and writes must happen on aligned offsets
|
|
// so we need to go back and write that page again
|
|
uint64_t next_write_offset_;
|
|
#endif // ROCKSDB_LITE
|
|
bool pending_sync_;
|
|
uint64_t last_sync_size_;
|
|
uint64_t bytes_per_sync_;
|
|
RateLimiter* rate_limiter_;
|
|
Statistics* stats_;
|
|
std::vector<std::shared_ptr<EventListener>> listeners_;
|
|
|
|
public:
|
|
WritableFileWriter(
|
|
std::unique_ptr<WritableFile>&& file, const std::string& _file_name,
|
|
const EnvOptions& options, Env* env = nullptr,
|
|
Statistics* stats = nullptr,
|
|
const std::vector<std::shared_ptr<EventListener>>& listeners = {})
|
|
: writable_file_(std::move(file)),
|
|
file_name_(_file_name),
|
|
env_(env),
|
|
buf_(),
|
|
max_buffer_size_(options.writable_file_max_buffer_size),
|
|
filesize_(0),
|
|
#ifndef ROCKSDB_LITE
|
|
next_write_offset_(0),
|
|
#endif // ROCKSDB_LITE
|
|
pending_sync_(false),
|
|
last_sync_size_(0),
|
|
bytes_per_sync_(options.bytes_per_sync),
|
|
rate_limiter_(options.rate_limiter),
|
|
stats_(stats),
|
|
listeners_() {
|
|
TEST_SYNC_POINT_CALLBACK("WritableFileWriter::WritableFileWriter:0",
|
|
reinterpret_cast<void*>(max_buffer_size_));
|
|
buf_.Alignment(writable_file_->GetRequiredBufferAlignment());
|
|
buf_.AllocateNewBuffer(std::min((size_t)65536, max_buffer_size_));
|
|
#ifndef ROCKSDB_LITE
|
|
std::for_each(listeners.begin(), listeners.end(),
|
|
[this](const std::shared_ptr<EventListener>& e) {
|
|
if (e->ShouldBeNotifiedOnFileIO()) {
|
|
listeners_.emplace_back(e);
|
|
}
|
|
});
|
|
#else // !ROCKSDB_LITE
|
|
(void)listeners;
|
|
#endif
|
|
}
|
|
|
|
WritableFileWriter(const WritableFileWriter&) = delete;
|
|
|
|
WritableFileWriter& operator=(const WritableFileWriter&) = delete;
|
|
|
|
~WritableFileWriter() { Close(); }
|
|
|
|
std::string file_name() const { return file_name_; }
|
|
|
|
Status Append(const Slice& data);
|
|
|
|
Status Pad(const size_t pad_bytes);
|
|
|
|
Status Flush();
|
|
|
|
Status Close();
|
|
|
|
Status Sync(bool use_fsync);
|
|
|
|
// Sync only the data that was already Flush()ed. Safe to call concurrently
|
|
// with Append() and Flush(). If !writable_file_->IsSyncThreadSafe(),
|
|
// returns NotSupported status.
|
|
Status SyncWithoutFlush(bool use_fsync);
|
|
|
|
uint64_t GetFileSize() { return filesize_; }
|
|
|
|
Status InvalidateCache(size_t offset, size_t length) {
|
|
return writable_file_->InvalidateCache(offset, length);
|
|
}
|
|
|
|
WritableFile* writable_file() const { return writable_file_.get(); }
|
|
|
|
bool use_direct_io() { return writable_file_->use_direct_io(); }
|
|
|
|
bool TEST_BufferIsEmpty() { return buf_.CurrentSize() == 0; }
|
|
|
|
private:
|
|
// Used when os buffering is OFF and we are writing
|
|
// DMA such as in Direct I/O mode
|
|
#ifndef ROCKSDB_LITE
|
|
Status WriteDirect();
|
|
#endif // !ROCKSDB_LITE
|
|
// Normal write
|
|
Status WriteBuffered(const char* data, size_t size);
|
|
Status RangeSync(uint64_t offset, uint64_t nbytes);
|
|
Status SyncInternal(bool use_fsync);
|
|
};
|
|
|
|
// FilePrefetchBuffer is a smart buffer to store and read data from a file.
|
|
class FilePrefetchBuffer {
|
|
public:
|
|
// Constructor.
|
|
//
|
|
// All arguments are optional.
|
|
// file_reader : the file reader to use. Can be a nullptr.
|
|
// readahead_size : the initial readahead size.
|
|
// max_readahead_size : the maximum readahead size.
|
|
// If max_readahead_size > readahead_size, the readahead size will be
|
|
// doubled on every IO until max_readahead_size is hit.
|
|
// Typically this is set as a multiple of readahead_size.
|
|
// max_readahead_size should be greater than equal to readahead_size.
|
|
// enable : controls whether reading from the buffer is enabled.
|
|
// If false, TryReadFromCache() always return false, and we only take stats
|
|
// for the minimum offset if track_min_offset = true.
|
|
// track_min_offset : Track the minimum offset ever read and collect stats on
|
|
// it. Used for adaptable readahead of the file footer/metadata.
|
|
//
|
|
// Automatic readhead is enabled for a file if file_reader, readahead_size,
|
|
// and max_readahead_size are passed in.
|
|
// If file_reader is a nullptr, setting readadhead_size and max_readahead_size
|
|
// does not make any sense. So it does nothing.
|
|
// A user can construct a FilePrefetchBuffer without any arguments, but use
|
|
// `Prefetch` to load data into the buffer.
|
|
FilePrefetchBuffer(RandomAccessFileReader* file_reader = nullptr,
|
|
size_t readadhead_size = 0, size_t max_readahead_size = 0,
|
|
bool enable = true, bool track_min_offset = false)
|
|
: buffer_offset_(0),
|
|
file_reader_(file_reader),
|
|
readahead_size_(readadhead_size),
|
|
max_readahead_size_(max_readahead_size),
|
|
min_offset_read_(port::kMaxSizet),
|
|
enable_(enable),
|
|
track_min_offset_(track_min_offset) {}
|
|
|
|
// Load data into the buffer from a file.
|
|
// reader : the file reader.
|
|
// offset : the file offset to start reading from.
|
|
// n : the number of bytes to read.
|
|
// for_compaction : if prefetch is done for compaction read.
|
|
Status Prefetch(RandomAccessFileReader* reader, uint64_t offset, size_t n,
|
|
bool for_compaction = false);
|
|
|
|
// Tries returning the data for a file raed from this buffer, if that data is
|
|
// in the buffer.
|
|
// It handles tracking the minimum read offset if track_min_offset = true.
|
|
// It also does the exponential readahead when readadhead_size is set as part
|
|
// of the constructor.
|
|
//
|
|
// offset : the file offset.
|
|
// n : the number of bytes.
|
|
// result : output buffer to put the data into.
|
|
// for_compaction : if cache read is done for compaction read.
|
|
bool TryReadFromCache(uint64_t offset, size_t n, Slice* result,
|
|
bool for_compaction = false);
|
|
|
|
// The minimum `offset` ever passed to TryReadFromCache(). This will nly be
|
|
// tracked if track_min_offset = true.
|
|
size_t min_offset_read() const { return min_offset_read_; }
|
|
|
|
private:
|
|
AlignedBuffer buffer_;
|
|
uint64_t buffer_offset_;
|
|
RandomAccessFileReader* file_reader_;
|
|
size_t readahead_size_;
|
|
size_t max_readahead_size_;
|
|
// The minimum `offset` ever passed to TryReadFromCache().
|
|
size_t min_offset_read_;
|
|
// if false, TryReadFromCache() always return false, and we only take stats
|
|
// for track_min_offset_ if track_min_offset_ = true
|
|
bool enable_;
|
|
// If true, track minimum `offset` ever passed to TryReadFromCache(), which
|
|
// can be fetched from min_offset_read().
|
|
bool track_min_offset_;
|
|
};
|
|
|
|
// Returns a WritableFile.
|
|
//
|
|
// env : the Env.
|
|
// fname : the file name.
|
|
// result : output arg. A WritableFile based on `fname` returned.
|
|
// options : the Env Options.
|
|
extern Status NewWritableFile(Env* env, const std::string& fname,
|
|
std::unique_ptr<WritableFile>* result,
|
|
const EnvOptions& options);
|
|
|
|
// Read a single line from a file.
|
|
bool ReadOneLine(std::istringstream* iss, SequentialFile* seq_file,
|
|
std::string* output, bool* has_data, Status* result);
|
|
|
|
} // namespace rocksdb
|