2013-10-16 23:59:46 +02:00
|
|
|
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
|
|
|
// This source code is licensed under the BSD-style license found in the
|
|
|
|
// LICENSE file in the root directory of this source tree. An additional grant
|
|
|
|
// of patent rights can be found in the PATENTS file in the same directory.
|
|
|
|
//
|
2011-03-18 23:37:00 +01:00
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
#include "table/block_based_table_builder.h"
|
2011-03-18 23:37:00 +01:00
|
|
|
|
|
|
|
#include <assert.h>
|
2013-11-20 01:29:42 +01:00
|
|
|
#include <inttypes.h>
|
|
|
|
#include <stdio.h>
|
2013-10-10 20:43:24 +02:00
|
|
|
|
2014-03-01 03:19:07 +01:00
|
|
|
#include <map>
|
|
|
|
#include <memory>
|
2014-05-15 23:09:03 +02:00
|
|
|
#include <string>
|
|
|
|
#include <unordered_map>
|
2014-03-01 03:19:07 +01:00
|
|
|
|
|
|
|
#include "db/dbformat.h"
|
|
|
|
|
2013-09-02 08:23:40 +02:00
|
|
|
#include "rocksdb/cache.h"
|
2013-08-23 17:38:13 +02:00
|
|
|
#include "rocksdb/comparator.h"
|
|
|
|
#include "rocksdb/env.h"
|
|
|
|
#include "rocksdb/filter_policy.h"
|
2014-03-01 03:19:07 +01:00
|
|
|
#include "rocksdb/flush_block_policy.h"
|
2013-08-23 17:38:13 +02:00
|
|
|
#include "rocksdb/options.h"
|
2014-03-01 03:19:07 +01:00
|
|
|
#include "rocksdb/table.h"
|
|
|
|
|
2013-09-02 08:23:40 +02:00
|
|
|
#include "table/block.h"
|
2014-03-01 03:19:07 +01:00
|
|
|
#include "table/block_based_table_reader.h"
|
2011-03-18 23:37:00 +01:00
|
|
|
#include "table/block_builder.h"
|
2012-04-17 17:36:46 +02:00
|
|
|
#include "table/filter_block.h"
|
2011-03-18 23:37:00 +01:00
|
|
|
#include "table/format.h"
|
2013-12-05 22:09:13 +01:00
|
|
|
#include "table/meta_blocks.h"
|
2014-03-01 03:19:07 +01:00
|
|
|
#include "table/table_builder.h"
|
|
|
|
|
2011-03-18 23:37:00 +01:00
|
|
|
#include "util/coding.h"
|
|
|
|
#include "util/crc32c.h"
|
2013-06-17 19:11:10 +02:00
|
|
|
#include "util/stop_watch.h"
|
2014-05-01 20:09:32 +02:00
|
|
|
#include "util/xxhash.h"
|
2011-03-18 23:37:00 +01:00
|
|
|
|
2013-10-04 06:49:15 +02:00
|
|
|
namespace rocksdb {
|
2011-03-18 23:37:00 +01:00
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
extern const std::string kHashIndexPrefixesBlock;
|
|
|
|
extern const std::string kHashIndexPrefixesMetadataBlock;
|
2013-10-10 20:43:24 +02:00
|
|
|
|
2014-03-01 03:19:07 +01:00
|
|
|
typedef BlockBasedTableOptions::IndexType IndexType;
|
|
|
|
|
|
|
|
// The interface for building index.
|
|
|
|
// Instruction for adding a new concrete IndexBuilder:
|
|
|
|
// 1. Create a subclass instantiated from IndexBuilder.
|
|
|
|
// 2. Add a new entry associated with that subclass in TableOptions::IndexType.
|
|
|
|
// 3. Add a create function for the new subclass in CreateIndexBuilder.
|
|
|
|
// Note: we can devise more advanced design to simplify the process for adding
|
|
|
|
// new subclass, which will, on the other hand, increase the code complexity and
|
|
|
|
// catch unwanted attention from readers. Given that we won't add/change
|
|
|
|
// indexes frequently, it makes sense to just embrace a more straightforward
|
|
|
|
// design that just works.
|
|
|
|
class IndexBuilder {
|
|
|
|
public:
|
2014-05-15 23:09:03 +02:00
|
|
|
// Index builder will construct a set of blocks which contain:
|
|
|
|
// 1. One primary index block.
|
|
|
|
// 2. (Optional) a set of metablocks that contains the metadata of the
|
|
|
|
// primary index.
|
|
|
|
struct IndexBlocks {
|
|
|
|
Slice index_block_contents;
|
|
|
|
std::unordered_map<std::string, Slice> meta_blocks;
|
|
|
|
};
|
2014-03-01 03:19:07 +01:00
|
|
|
explicit IndexBuilder(const Comparator* comparator)
|
|
|
|
: comparator_(comparator) {}
|
|
|
|
|
|
|
|
virtual ~IndexBuilder() {}
|
|
|
|
|
|
|
|
// Add a new index entry to index block.
|
|
|
|
// To allow further optimization, we provide `last_key_in_current_block` and
|
|
|
|
// `first_key_in_next_block`, based on which the specific implementation can
|
|
|
|
// determine the best index key to be used for the index block.
|
|
|
|
// @last_key_in_current_block: this parameter maybe overridden with the value
|
|
|
|
// "substitute key".
|
|
|
|
// @first_key_in_next_block: it will be nullptr if the entry being added is
|
|
|
|
// the last one in the table
|
|
|
|
//
|
|
|
|
// REQUIRES: Finish() has not yet been called.
|
2014-05-15 23:09:03 +02:00
|
|
|
virtual void AddIndexEntry(std::string* last_key_in_current_block,
|
|
|
|
const Slice* first_key_in_next_block,
|
|
|
|
const BlockHandle& block_handle) = 0;
|
|
|
|
|
|
|
|
// This method will be called whenever a key is added. The subclasses may
|
|
|
|
// override OnKeyAdded() if they need to collect additional information.
|
|
|
|
virtual void OnKeyAdded(const Slice& key) {}
|
2014-03-01 03:19:07 +01:00
|
|
|
|
|
|
|
// Inform the index builder that all entries has been written. Block builder
|
|
|
|
// may therefore perform any operation required for block finalization.
|
|
|
|
//
|
|
|
|
// REQUIRES: Finish() has not yet been called.
|
2014-05-15 23:09:03 +02:00
|
|
|
virtual Status Finish(IndexBlocks* index_blocks) = 0;
|
2014-03-01 03:19:07 +01:00
|
|
|
|
|
|
|
// Get the estimated size for index block.
|
|
|
|
virtual size_t EstimatedSize() const = 0;
|
|
|
|
|
|
|
|
protected:
|
|
|
|
const Comparator* comparator_;
|
|
|
|
};
|
|
|
|
|
2014-04-11 02:23:49 +02:00
|
|
|
// This index builder builds space-efficient index block.
|
2014-03-01 03:19:07 +01:00
|
|
|
//
|
|
|
|
// Optimizations:
|
|
|
|
// 1. Made block's `block_restart_interval` to be 1, which will avoid linear
|
|
|
|
// search when doing index lookup.
|
|
|
|
// 2. Shorten the key length for index block. Other than honestly using the
|
|
|
|
// last key in the data block as the index key, we instead find a shortest
|
|
|
|
// substitute key that serves the same function.
|
2014-04-10 23:19:43 +02:00
|
|
|
class ShortenedIndexBuilder : public IndexBuilder {
|
2014-03-01 03:19:07 +01:00
|
|
|
public:
|
2014-04-10 23:19:43 +02:00
|
|
|
explicit ShortenedIndexBuilder(const Comparator* comparator)
|
2014-03-01 03:19:07 +01:00
|
|
|
: IndexBuilder(comparator),
|
2014-09-02 20:49:38 +02:00
|
|
|
index_block_builder_(1 /* block_restart_interval == 1 */) {}
|
2014-03-01 03:19:07 +01:00
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
virtual void AddIndexEntry(std::string* last_key_in_current_block,
|
|
|
|
const Slice* first_key_in_next_block,
|
|
|
|
const BlockHandle& block_handle) override {
|
2014-03-01 03:19:07 +01:00
|
|
|
if (first_key_in_next_block != nullptr) {
|
|
|
|
comparator_->FindShortestSeparator(last_key_in_current_block,
|
|
|
|
*first_key_in_next_block);
|
|
|
|
} else {
|
|
|
|
comparator_->FindShortSuccessor(last_key_in_current_block);
|
|
|
|
}
|
|
|
|
|
|
|
|
std::string handle_encoding;
|
|
|
|
block_handle.EncodeTo(&handle_encoding);
|
|
|
|
index_block_builder_.Add(*last_key_in_current_block, handle_encoding);
|
|
|
|
}
|
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
virtual Status Finish(IndexBlocks* index_blocks) {
|
|
|
|
index_blocks->index_block_contents = index_block_builder_.Finish();
|
|
|
|
return Status::OK();
|
|
|
|
}
|
2014-03-01 03:19:07 +01:00
|
|
|
|
|
|
|
virtual size_t EstimatedSize() const {
|
|
|
|
return index_block_builder_.CurrentSizeEstimate();
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
BlockBuilder index_block_builder_;
|
|
|
|
};
|
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
// HashIndexBuilder contains a binary-searchable primary index and the
|
|
|
|
// metadata for secondary hash index construction.
|
|
|
|
// The metadata for hash index consists two parts:
|
|
|
|
// - a metablock that compactly contains a sequence of prefixes. All prefixes
|
|
|
|
// are stored consectively without any metadata (like, prefix sizes) being
|
|
|
|
// stored, which is kept in the other metablock.
|
|
|
|
// - a metablock contains the metadata of the prefixes, including prefix size,
|
|
|
|
// restart index and number of block it spans. The format looks like:
|
|
|
|
//
|
|
|
|
// +-----------------+---------------------------+---------------------+ <=prefix 1
|
|
|
|
// | length: 4 bytes | restart interval: 4 bytes | num-blocks: 4 bytes |
|
|
|
|
// +-----------------+---------------------------+---------------------+ <=prefix 2
|
|
|
|
// | length: 4 bytes | restart interval: 4 bytes | num-blocks: 4 bytes |
|
|
|
|
// +-----------------+---------------------------+---------------------+
|
|
|
|
// | |
|
|
|
|
// | .... |
|
|
|
|
// | |
|
|
|
|
// +-----------------+---------------------------+---------------------+ <=prefix n
|
|
|
|
// | length: 4 bytes | restart interval: 4 bytes | num-blocks: 4 bytes |
|
|
|
|
// +-----------------+---------------------------+---------------------+
|
|
|
|
//
|
|
|
|
// The reason of separating these two metablocks is to enable the efficiently
|
|
|
|
// reuse the first metablock during hash index construction without unnecessary
|
|
|
|
// data copy or small heap allocations for prefixes.
|
|
|
|
class HashIndexBuilder : public IndexBuilder {
|
2014-04-10 23:19:43 +02:00
|
|
|
public:
|
2014-05-15 23:09:03 +02:00
|
|
|
explicit HashIndexBuilder(const Comparator* comparator,
|
|
|
|
const SliceTransform* hash_key_extractor)
|
2014-04-10 23:19:43 +02:00
|
|
|
: IndexBuilder(comparator),
|
2014-05-15 23:09:03 +02:00
|
|
|
primary_index_builder(comparator),
|
|
|
|
hash_key_extractor_(hash_key_extractor) {}
|
|
|
|
|
|
|
|
virtual void AddIndexEntry(std::string* last_key_in_current_block,
|
|
|
|
const Slice* first_key_in_next_block,
|
|
|
|
const BlockHandle& block_handle) override {
|
|
|
|
++current_restart_index_;
|
|
|
|
primary_index_builder.AddIndexEntry(last_key_in_current_block,
|
|
|
|
first_key_in_next_block, block_handle);
|
|
|
|
}
|
2014-04-10 23:19:43 +02:00
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
virtual void OnKeyAdded(const Slice& key) override {
|
|
|
|
auto key_prefix = hash_key_extractor_->Transform(key);
|
|
|
|
bool is_first_entry = pending_block_num_ == 0;
|
|
|
|
|
|
|
|
// Keys may share the prefix
|
|
|
|
if (is_first_entry || pending_entry_prefix_ != key_prefix) {
|
|
|
|
if (!is_first_entry) {
|
|
|
|
FlushPendingPrefix();
|
|
|
|
}
|
|
|
|
|
|
|
|
// need a hard copy otherwise the underlying data changes all the time.
|
|
|
|
// TODO(kailiu) ToString() is expensive. We may speed up can avoid data
|
|
|
|
// copy.
|
|
|
|
pending_entry_prefix_ = key_prefix.ToString();
|
|
|
|
pending_block_num_ = 1;
|
|
|
|
pending_entry_index_ = current_restart_index_;
|
|
|
|
} else {
|
|
|
|
// entry number increments when keys share the prefix reside in
|
|
|
|
// differnt data blocks.
|
|
|
|
auto last_restart_index = pending_entry_index_ + pending_block_num_ - 1;
|
|
|
|
assert(last_restart_index <= current_restart_index_);
|
|
|
|
if (last_restart_index != current_restart_index_) {
|
|
|
|
++pending_block_num_;
|
|
|
|
}
|
|
|
|
}
|
2014-04-10 23:19:43 +02:00
|
|
|
}
|
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
virtual Status Finish(IndexBlocks* index_blocks) {
|
|
|
|
FlushPendingPrefix();
|
|
|
|
primary_index_builder.Finish(index_blocks);
|
|
|
|
index_blocks->meta_blocks.insert(
|
|
|
|
{kHashIndexPrefixesBlock.c_str(), prefix_block_});
|
|
|
|
index_blocks->meta_blocks.insert(
|
|
|
|
{kHashIndexPrefixesMetadataBlock.c_str(), prefix_meta_block_});
|
|
|
|
return Status::OK();
|
|
|
|
}
|
2014-04-10 23:19:43 +02:00
|
|
|
|
|
|
|
virtual size_t EstimatedSize() const {
|
2014-05-15 23:09:03 +02:00
|
|
|
return primary_index_builder.EstimatedSize() + prefix_block_.size() +
|
|
|
|
prefix_meta_block_.size();
|
2014-04-10 23:19:43 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
2014-05-15 23:09:03 +02:00
|
|
|
void FlushPendingPrefix() {
|
|
|
|
prefix_block_.append(pending_entry_prefix_.data(),
|
|
|
|
pending_entry_prefix_.size());
|
|
|
|
PutVarint32(&prefix_meta_block_, pending_entry_prefix_.size());
|
|
|
|
PutVarint32(&prefix_meta_block_, pending_entry_index_);
|
|
|
|
PutVarint32(&prefix_meta_block_, pending_block_num_);
|
|
|
|
}
|
|
|
|
|
|
|
|
ShortenedIndexBuilder primary_index_builder;
|
|
|
|
const SliceTransform* hash_key_extractor_;
|
|
|
|
|
|
|
|
// stores a sequence of prefixes
|
|
|
|
std::string prefix_block_;
|
|
|
|
// stores the metadata of prefixes
|
|
|
|
std::string prefix_meta_block_;
|
|
|
|
|
|
|
|
// The following 3 variables keeps unflushed prefix and its metadata.
|
|
|
|
// The details of block_num and entry_index can be found in
|
|
|
|
// "block_hash_index.{h,cc}"
|
|
|
|
uint32_t pending_block_num_ = 0;
|
|
|
|
uint32_t pending_entry_index_ = 0;
|
|
|
|
std::string pending_entry_prefix_;
|
|
|
|
|
|
|
|
uint64_t current_restart_index_ = 0;
|
2014-04-10 23:19:43 +02:00
|
|
|
};
|
|
|
|
|
2014-03-01 03:19:07 +01:00
|
|
|
// Create a index builder based on its type.
|
2014-05-15 23:09:03 +02:00
|
|
|
IndexBuilder* CreateIndexBuilder(IndexType type, const Comparator* comparator,
|
|
|
|
const SliceTransform* prefix_extractor) {
|
2014-03-01 03:19:07 +01:00
|
|
|
switch (type) {
|
|
|
|
case BlockBasedTableOptions::kBinarySearch: {
|
2014-04-10 23:19:43 +02:00
|
|
|
return new ShortenedIndexBuilder(comparator);
|
|
|
|
}
|
2014-05-15 23:09:03 +02:00
|
|
|
case BlockBasedTableOptions::kHashSearch: {
|
|
|
|
return new HashIndexBuilder(comparator, prefix_extractor);
|
|
|
|
}
|
2014-03-01 03:19:07 +01:00
|
|
|
default: {
|
|
|
|
assert(!"Do not recognize the index type ");
|
|
|
|
return nullptr;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// impossible.
|
|
|
|
assert(false);
|
|
|
|
return nullptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool GoodCompressionRatio(size_t compressed_size, size_t raw_size) {
|
2013-10-10 20:43:24 +02:00
|
|
|
// Check to see if compressed less than 12.5%
|
|
|
|
return compressed_size < raw_size - (raw_size / 8u);
|
|
|
|
}
|
|
|
|
|
2014-03-01 03:19:07 +01:00
|
|
|
Slice CompressBlock(const Slice& raw,
|
|
|
|
const CompressionOptions& compression_options,
|
|
|
|
CompressionType* type, std::string* compressed_output) {
|
|
|
|
if (*type == kNoCompression) {
|
|
|
|
return raw;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Will return compressed block contents if (1) the compression method is
|
|
|
|
// supported in this platform and (2) the compression rate is "good enough".
|
|
|
|
switch (*type) {
|
|
|
|
case kSnappyCompression:
|
|
|
|
if (port::Snappy_Compress(compression_options, raw.data(), raw.size(),
|
|
|
|
compressed_output) &&
|
|
|
|
GoodCompressionRatio(compressed_output->size(), raw.size())) {
|
|
|
|
return *compressed_output;
|
|
|
|
}
|
|
|
|
break; // fall back to no compression.
|
|
|
|
case kZlibCompression:
|
|
|
|
if (port::Zlib_Compress(compression_options, raw.data(), raw.size(),
|
|
|
|
compressed_output) &&
|
|
|
|
GoodCompressionRatio(compressed_output->size(), raw.size())) {
|
|
|
|
return *compressed_output;
|
|
|
|
}
|
|
|
|
break; // fall back to no compression.
|
|
|
|
case kBZip2Compression:
|
|
|
|
if (port::BZip2_Compress(compression_options, raw.data(), raw.size(),
|
|
|
|
compressed_output) &&
|
|
|
|
GoodCompressionRatio(compressed_output->size(), raw.size())) {
|
|
|
|
return *compressed_output;
|
|
|
|
}
|
|
|
|
break; // fall back to no compression.
|
|
|
|
case kLZ4Compression:
|
|
|
|
if (port::LZ4_Compress(compression_options, raw.data(), raw.size(),
|
|
|
|
compressed_output) &&
|
|
|
|
GoodCompressionRatio(compressed_output->size(), raw.size())) {
|
|
|
|
return *compressed_output;
|
|
|
|
}
|
|
|
|
break; // fall back to no compression.
|
|
|
|
case kLZ4HCCompression:
|
|
|
|
if (port::LZ4HC_Compress(compression_options, raw.data(), raw.size(),
|
|
|
|
compressed_output) &&
|
|
|
|
GoodCompressionRatio(compressed_output->size(), raw.size())) {
|
|
|
|
return *compressed_output;
|
|
|
|
}
|
|
|
|
break; // fall back to no compression.
|
|
|
|
default: {} // Do not recognize this compression type
|
|
|
|
}
|
|
|
|
|
|
|
|
// Compression method is not supported, or not good compression ratio, so just
|
|
|
|
// fall back to uncompressed form.
|
|
|
|
*type = kNoCompression;
|
|
|
|
return raw;
|
|
|
|
}
|
|
|
|
|
2013-12-05 01:35:48 +01:00
|
|
|
// kBlockBasedTableMagicNumber was picked by running
|
2014-05-01 20:09:32 +02:00
|
|
|
// echo rocksdb.table.block_based | sha1sum
|
2013-12-05 00:09:41 +01:00
|
|
|
// and taking the leading 64 bits.
|
2013-12-05 01:35:48 +01:00
|
|
|
// Please note that kBlockBasedTableMagicNumber may also be accessed by
|
2013-12-05 00:09:41 +01:00
|
|
|
// other .cc files so it have to be explicitly declared with "extern".
|
2014-05-01 20:09:32 +02:00
|
|
|
extern const uint64_t kBlockBasedTableMagicNumber = 0x88e241b785f4cff7ull;
|
|
|
|
// We also support reading and writing legacy block based table format (for
|
|
|
|
// backwards compatibility)
|
|
|
|
extern const uint64_t kLegacyBlockBasedTableMagicNumber = 0xdb4775248b80fb57ull;
|
2013-12-05 00:09:41 +01:00
|
|
|
|
2014-03-01 03:19:07 +01:00
|
|
|
// A collector that collects properties of interest to block-based table.
|
|
|
|
// For now this class looks heavy-weight since we only write one additional
|
|
|
|
// property.
|
|
|
|
// But in the forseeable future, we will add more and more properties that are
|
|
|
|
// specific to block-based table.
|
|
|
|
class BlockBasedTableBuilder::BlockBasedTablePropertiesCollector
|
|
|
|
: public TablePropertiesCollector {
|
|
|
|
public:
|
2014-05-15 23:09:03 +02:00
|
|
|
explicit BlockBasedTablePropertiesCollector(
|
2014-03-01 03:19:07 +01:00
|
|
|
BlockBasedTableOptions::IndexType index_type)
|
|
|
|
: index_type_(index_type) {}
|
|
|
|
|
|
|
|
virtual Status Add(const Slice& key, const Slice& value) {
|
|
|
|
// Intentionally left blank. Have no interest in collecting stats for
|
|
|
|
// individual key/value pairs.
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
|
|
|
virtual Status Finish(UserCollectedProperties* properties) {
|
|
|
|
std::string val;
|
|
|
|
PutFixed32(&val, static_cast<uint32_t>(index_type_));
|
|
|
|
properties->insert({BlockBasedTablePropertyNames::kIndexType, val});
|
|
|
|
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
|
|
|
// The name of the properties collector can be used for debugging purpose.
|
|
|
|
virtual const char* Name() const {
|
|
|
|
return "BlockBasedTablePropertiesCollector";
|
|
|
|
}
|
|
|
|
|
|
|
|
virtual UserCollectedProperties GetReadableProperties() const {
|
|
|
|
// Intentionally left blank.
|
|
|
|
return UserCollectedProperties();
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
BlockBasedTableOptions::IndexType index_type_;
|
|
|
|
};
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
struct BlockBasedTableBuilder::Rep {
|
2014-08-25 23:22:05 +02:00
|
|
|
const Options options;
|
|
|
|
const BlockBasedTableOptions table_options;
|
2014-01-27 22:53:22 +01:00
|
|
|
const InternalKeyComparator& internal_comparator;
|
2011-03-18 23:37:00 +01:00
|
|
|
WritableFile* file;
|
2013-10-10 20:43:24 +02:00
|
|
|
uint64_t offset = 0;
|
2011-03-18 23:37:00 +01:00
|
|
|
Status status;
|
|
|
|
BlockBuilder data_block;
|
2014-05-15 23:09:03 +02:00
|
|
|
|
|
|
|
InternalKeySliceTransform internal_prefix_transform;
|
2014-03-01 03:19:07 +01:00
|
|
|
std::unique_ptr<IndexBuilder> index_builder;
|
|
|
|
|
2011-03-18 23:37:00 +01:00
|
|
|
std::string last_key;
|
2013-10-30 18:52:33 +01:00
|
|
|
CompressionType compression_type;
|
2013-11-20 01:29:42 +01:00
|
|
|
TableProperties props;
|
2013-10-10 20:43:24 +02:00
|
|
|
|
|
|
|
bool closed = false; // Either Finish() or Abandon() has been called.
|
2012-04-17 17:36:46 +02:00
|
|
|
FilterBlockBuilder* filter_block;
|
2013-09-02 08:23:40 +02:00
|
|
|
char compressed_cache_key_prefix[BlockBasedTable::kMaxCacheKeyPrefixSize];
|
|
|
|
size_t compressed_cache_key_prefix_size;
|
2011-03-18 23:37:00 +01:00
|
|
|
|
|
|
|
BlockHandle pending_handle; // Handle to add to index block
|
|
|
|
|
|
|
|
std::string compressed_output;
|
2013-11-08 06:27:21 +01:00
|
|
|
std::unique_ptr<FlushBlockPolicy> flush_block_policy;
|
2011-03-18 23:37:00 +01:00
|
|
|
|
TablePropertiesCollectorFactory
Summary:
This diff addresses task #4296714 and rethinks how users provide us with TablePropertiesCollectors as part of Options.
Here's description of task #4296714:
I'm debugging #4295529 and noticed that our count of user properties kDeletedKeys is wrong. We're sharing one single InternalKeyPropertiesCollector with all Table Builders. In LOG Files, we're outputting number of kDeletedKeys as connected with a single table, while it's actually the total count of deleted keys since creation of the DB.
For example, this table has 3155 entries and 1391828 deleted keys.
The problem with current approach that we call methods on a single TablePropertiesCollector for all the tables we create. Even worse, we could do it from multiple threads at the same time and TablePropertiesCollector has no way of knowing which table we're calling it for.
Good part: Looks like nobody inside Facebook is using Options::table_properties_collectors. This means we should be able to painfully change the API.
In this change, I introduce TablePropertiesCollectorFactory. For every table we create, we call `CreateTablePropertiesCollector`, which creates a TablePropertiesCollector for a single table. We then use it sequentially from a single thread, which means it doesn't have to be thread-safe.
Test Plan:
Added a test in table_properties_collector_test that fails on master (build two tables, assert that kDeletedKeys count is correct for the second one).
Also, all other tests
Reviewers: sdong, dhruba, haobo, kailiu
Reviewed By: kailiu
CC: leveldb
Differential Revision: https://reviews.facebook.net/D18579
2014-05-13 21:30:55 +02:00
|
|
|
std::vector<std::unique_ptr<TablePropertiesCollector>>
|
|
|
|
table_properties_collectors;
|
|
|
|
|
2014-08-25 23:22:05 +02:00
|
|
|
Rep(const Options& opt, const BlockBasedTableOptions& table_opt,
|
|
|
|
const InternalKeyComparator& icomparator,
|
|
|
|
WritableFile* f, CompressionType compression_type)
|
2011-03-18 23:37:00 +01:00
|
|
|
: options(opt),
|
2014-08-25 23:22:05 +02:00
|
|
|
table_options(table_opt),
|
2014-01-27 22:53:22 +01:00
|
|
|
internal_comparator(icomparator),
|
2011-03-18 23:37:00 +01:00
|
|
|
file(f),
|
2014-09-02 20:49:38 +02:00
|
|
|
data_block(table_options.block_restart_interval),
|
2014-05-15 23:09:03 +02:00
|
|
|
internal_prefix_transform(options.prefix_extractor.get()),
|
2014-08-25 23:22:05 +02:00
|
|
|
index_builder(CreateIndexBuilder(
|
|
|
|
table_options.index_type, &internal_comparator,
|
|
|
|
&this->internal_prefix_transform)),
|
2013-10-30 18:52:33 +01:00
|
|
|
compression_type(compression_type),
|
2014-08-25 23:22:05 +02:00
|
|
|
filter_block(table_options.filter_policy == nullptr ?
|
|
|
|
nullptr :
|
|
|
|
new FilterBlockBuilder(opt, table_options, &internal_comparator)),
|
|
|
|
flush_block_policy(
|
|
|
|
table_options.flush_block_policy_factory->NewFlushBlockPolicy(
|
|
|
|
table_options, data_block)) {
|
TablePropertiesCollectorFactory
Summary:
This diff addresses task #4296714 and rethinks how users provide us with TablePropertiesCollectors as part of Options.
Here's description of task #4296714:
I'm debugging #4295529 and noticed that our count of user properties kDeletedKeys is wrong. We're sharing one single InternalKeyPropertiesCollector with all Table Builders. In LOG Files, we're outputting number of kDeletedKeys as connected with a single table, while it's actually the total count of deleted keys since creation of the DB.
For example, this table has 3155 entries and 1391828 deleted keys.
The problem with current approach that we call methods on a single TablePropertiesCollector for all the tables we create. Even worse, we could do it from multiple threads at the same time and TablePropertiesCollector has no way of knowing which table we're calling it for.
Good part: Looks like nobody inside Facebook is using Options::table_properties_collectors. This means we should be able to painfully change the API.
In this change, I introduce TablePropertiesCollectorFactory. For every table we create, we call `CreateTablePropertiesCollector`, which creates a TablePropertiesCollector for a single table. We then use it sequentially from a single thread, which means it doesn't have to be thread-safe.
Test Plan:
Added a test in table_properties_collector_test that fails on master (build two tables, assert that kDeletedKeys count is correct for the second one).
Also, all other tests
Reviewers: sdong, dhruba, haobo, kailiu
Reviewed By: kailiu
CC: leveldb
Differential Revision: https://reviews.facebook.net/D18579
2014-05-13 21:30:55 +02:00
|
|
|
for (auto& collector_factories :
|
|
|
|
options.table_properties_collector_factories) {
|
|
|
|
table_properties_collectors.emplace_back(
|
|
|
|
collector_factories->CreateTablePropertiesCollector());
|
|
|
|
}
|
|
|
|
table_properties_collectors.emplace_back(
|
2014-08-25 23:22:05 +02:00
|
|
|
new BlockBasedTablePropertiesCollector(table_options.index_type));
|
2014-03-01 03:19:07 +01:00
|
|
|
}
|
2011-03-18 23:37:00 +01:00
|
|
|
};
|
|
|
|
|
2013-11-20 07:00:48 +01:00
|
|
|
BlockBasedTableBuilder::BlockBasedTableBuilder(
|
2014-03-01 03:19:07 +01:00
|
|
|
const Options& options, const BlockBasedTableOptions& table_options,
|
|
|
|
const InternalKeyComparator& internal_comparator, WritableFile* file,
|
2013-11-20 07:00:48 +01:00
|
|
|
CompressionType compression_type)
|
2014-08-25 23:22:05 +02:00
|
|
|
: rep_(new Rep(options, table_options, internal_comparator,
|
|
|
|
file, compression_type)) {
|
2013-03-01 03:04:58 +01:00
|
|
|
if (rep_->filter_block != nullptr) {
|
2012-04-17 17:36:46 +02:00
|
|
|
rep_->filter_block->StartBlock(0);
|
|
|
|
}
|
2014-08-25 23:22:05 +02:00
|
|
|
if (table_options.block_cache_compressed.get() != nullptr) {
|
2013-12-03 20:17:58 +01:00
|
|
|
BlockBasedTable::GenerateCachePrefix(
|
2014-08-25 23:22:05 +02:00
|
|
|
table_options.block_cache_compressed.get(), file,
|
2013-12-03 20:17:58 +01:00
|
|
|
&rep_->compressed_cache_key_prefix[0],
|
|
|
|
&rep_->compressed_cache_key_prefix_size);
|
2013-09-02 08:23:40 +02:00
|
|
|
}
|
2011-03-18 23:37:00 +01:00
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
BlockBasedTableBuilder::~BlockBasedTableBuilder() {
|
2011-03-18 23:37:00 +01:00
|
|
|
assert(rep_->closed); // Catch errors where caller forgot to call Finish()
|
2012-04-17 17:36:46 +02:00
|
|
|
delete rep_->filter_block;
|
2011-03-18 23:37:00 +01:00
|
|
|
delete rep_;
|
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
void BlockBasedTableBuilder::Add(const Slice& key, const Slice& value) {
|
2011-03-18 23:37:00 +01:00
|
|
|
Rep* r = rep_;
|
|
|
|
assert(!r->closed);
|
|
|
|
if (!ok()) return;
|
2013-11-20 01:29:42 +01:00
|
|
|
if (r->props.num_entries > 0) {
|
2014-01-27 22:53:22 +01:00
|
|
|
assert(r->internal_comparator.Compare(key, Slice(r->last_key)) > 0);
|
2011-03-18 23:37:00 +01:00
|
|
|
}
|
2014-06-13 04:03:22 +02:00
|
|
|
|
2013-11-08 06:27:21 +01:00
|
|
|
auto should_flush = r->flush_block_policy->Update(key, value);
|
|
|
|
if (should_flush) {
|
|
|
|
assert(!r->data_block.empty());
|
2013-05-15 19:34:02 +02:00
|
|
|
Flush();
|
|
|
|
|
2013-11-08 06:27:21 +01:00
|
|
|
// Add item to index block.
|
|
|
|
// We do not emit the index entry for a block until we have seen the
|
|
|
|
// first key for the next data block. This allows us to use shorter
|
|
|
|
// keys in the index block. For example, consider a block boundary
|
|
|
|
// between the keys "the quick brown fox" and "the who". We can use
|
|
|
|
// "the r" as the key for the index block entry since it is >= all
|
|
|
|
// entries in the first block and < all entries in subsequent
|
|
|
|
// blocks.
|
|
|
|
if (ok()) {
|
2014-05-15 23:09:03 +02:00
|
|
|
r->index_builder->AddIndexEntry(&r->last_key, &key, r->pending_handle);
|
2013-11-08 06:27:21 +01:00
|
|
|
}
|
2011-03-18 23:37:00 +01:00
|
|
|
}
|
|
|
|
|
2013-03-01 03:04:58 +01:00
|
|
|
if (r->filter_block != nullptr) {
|
2014-08-29 02:06:29 +02:00
|
|
|
r->filter_block->AddKey(ExtractUserKey(key));
|
2012-04-17 17:36:46 +02:00
|
|
|
}
|
|
|
|
|
2011-03-18 23:37:00 +01:00
|
|
|
r->last_key.assign(key.data(), key.size());
|
|
|
|
r->data_block.Add(key, value);
|
2013-11-20 01:29:42 +01:00
|
|
|
r->props.num_entries++;
|
|
|
|
r->props.raw_key_size += key.size();
|
|
|
|
r->props.raw_value_size += value.size();
|
2013-10-16 20:50:50 +02:00
|
|
|
|
2014-06-13 04:03:22 +02:00
|
|
|
r->index_builder->OnKeyAdded(key);
|
TablePropertiesCollectorFactory
Summary:
This diff addresses task #4296714 and rethinks how users provide us with TablePropertiesCollectors as part of Options.
Here's description of task #4296714:
I'm debugging #4295529 and noticed that our count of user properties kDeletedKeys is wrong. We're sharing one single InternalKeyPropertiesCollector with all Table Builders. In LOG Files, we're outputting number of kDeletedKeys as connected with a single table, while it's actually the total count of deleted keys since creation of the DB.
For example, this table has 3155 entries and 1391828 deleted keys.
The problem with current approach that we call methods on a single TablePropertiesCollector for all the tables we create. Even worse, we could do it from multiple threads at the same time and TablePropertiesCollector has no way of knowing which table we're calling it for.
Good part: Looks like nobody inside Facebook is using Options::table_properties_collectors. This means we should be able to painfully change the API.
In this change, I introduce TablePropertiesCollectorFactory. For every table we create, we call `CreateTablePropertiesCollector`, which creates a TablePropertiesCollector for a single table. We then use it sequentially from a single thread, which means it doesn't have to be thread-safe.
Test Plan:
Added a test in table_properties_collector_test that fails on master (build two tables, assert that kDeletedKeys count is correct for the second one).
Also, all other tests
Reviewers: sdong, dhruba, haobo, kailiu
Reviewed By: kailiu
CC: leveldb
Differential Revision: https://reviews.facebook.net/D18579
2014-05-13 21:30:55 +02:00
|
|
|
NotifyCollectTableCollectorsOnAdd(key, value, r->table_properties_collectors,
|
|
|
|
r->options.info_log.get());
|
2011-03-18 23:37:00 +01:00
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
void BlockBasedTableBuilder::Flush() {
|
2011-03-18 23:37:00 +01:00
|
|
|
Rep* r = rep_;
|
|
|
|
assert(!r->closed);
|
|
|
|
if (!ok()) return;
|
|
|
|
if (r->data_block.empty()) return;
|
|
|
|
WriteBlock(&r->data_block, &r->pending_handle);
|
|
|
|
if (ok()) {
|
|
|
|
r->status = r->file->Flush();
|
|
|
|
}
|
2013-03-01 03:04:58 +01:00
|
|
|
if (r->filter_block != nullptr) {
|
2012-04-17 17:36:46 +02:00
|
|
|
r->filter_block->StartBlock(r->offset);
|
|
|
|
}
|
2013-11-20 01:29:42 +01:00
|
|
|
r->props.data_size = r->offset;
|
|
|
|
++r->props.num_data_blocks;
|
2012-06-28 08:41:33 +02:00
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
void BlockBasedTableBuilder::WriteBlock(BlockBuilder* block,
|
|
|
|
BlockHandle* handle) {
|
2014-03-01 03:19:07 +01:00
|
|
|
WriteBlock(block->Finish(), handle);
|
|
|
|
block->Reset();
|
|
|
|
}
|
|
|
|
|
|
|
|
void BlockBasedTableBuilder::WriteBlock(const Slice& raw_block_contents,
|
|
|
|
BlockHandle* handle) {
|
2011-03-18 23:37:00 +01:00
|
|
|
// File format contains a sequence of blocks where each block has:
|
|
|
|
// block_data: uint8[n]
|
|
|
|
// type: uint8
|
|
|
|
// crc: uint32
|
|
|
|
assert(ok());
|
|
|
|
Rep* r = rep_;
|
|
|
|
|
2014-03-01 03:19:07 +01:00
|
|
|
auto type = r->compression_type;
|
2014-06-09 21:26:09 +02:00
|
|
|
Slice block_contents;
|
|
|
|
if (raw_block_contents.size() < kCompressionSizeLimit) {
|
|
|
|
block_contents =
|
|
|
|
CompressBlock(raw_block_contents, r->options.compression_opts, &type,
|
|
|
|
&r->compressed_output);
|
|
|
|
} else {
|
|
|
|
RecordTick(r->options.statistics.get(), NUMBER_BLOCK_NOT_COMPRESSED);
|
|
|
|
type = kNoCompression;
|
|
|
|
block_contents = raw_block_contents;
|
|
|
|
}
|
2012-04-17 17:36:46 +02:00
|
|
|
WriteRawBlock(block_contents, type, handle);
|
|
|
|
r->compressed_output.clear();
|
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
void BlockBasedTableBuilder::WriteRawBlock(const Slice& block_contents,
|
|
|
|
CompressionType type,
|
|
|
|
BlockHandle* handle) {
|
2012-04-17 17:36:46 +02:00
|
|
|
Rep* r = rep_;
|
2013-11-22 23:14:05 +01:00
|
|
|
StopWatch sw(r->options.env, r->options.statistics.get(),
|
|
|
|
WRITE_RAW_BLOCK_MICROS);
|
2011-03-18 23:37:00 +01:00
|
|
|
handle->set_offset(r->offset);
|
|
|
|
handle->set_size(block_contents.size());
|
|
|
|
r->status = r->file->Append(block_contents);
|
|
|
|
if (r->status.ok()) {
|
|
|
|
char trailer[kBlockTrailerSize];
|
|
|
|
trailer[0] = type;
|
2014-05-01 20:09:32 +02:00
|
|
|
char* trailer_without_type = trailer + 1;
|
2014-08-25 23:22:05 +02:00
|
|
|
switch (r->table_options.checksum) {
|
2014-05-01 20:09:32 +02:00
|
|
|
case kNoChecksum:
|
|
|
|
// we don't support no checksum yet
|
|
|
|
assert(false);
|
|
|
|
// intentional fallthrough in release binary
|
|
|
|
case kCRC32c: {
|
|
|
|
auto crc = crc32c::Value(block_contents.data(), block_contents.size());
|
|
|
|
crc = crc32c::Extend(crc, trailer, 1); // Extend to cover block type
|
|
|
|
EncodeFixed32(trailer_without_type, crc32c::Mask(crc));
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case kxxHash: {
|
|
|
|
void* xxh = XXH32_init(0);
|
|
|
|
XXH32_update(xxh, block_contents.data(), block_contents.size());
|
|
|
|
XXH32_update(xxh, trailer, 1); // Extend to cover block type
|
|
|
|
EncodeFixed32(trailer_without_type, XXH32_digest(xxh));
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-03-18 23:37:00 +01:00
|
|
|
r->status = r->file->Append(Slice(trailer, kBlockTrailerSize));
|
2013-09-02 08:23:40 +02:00
|
|
|
if (r->status.ok()) {
|
|
|
|
r->status = InsertBlockInCache(block_contents, type, handle);
|
|
|
|
}
|
2011-03-18 23:37:00 +01:00
|
|
|
if (r->status.ok()) {
|
|
|
|
r->offset += block_contents.size() + kBlockTrailerSize;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
Status BlockBasedTableBuilder::status() const {
|
2011-03-18 23:37:00 +01:00
|
|
|
return rep_->status;
|
|
|
|
}
|
|
|
|
|
2013-09-02 08:23:40 +02:00
|
|
|
static void DeleteCachedBlock(const Slice& key, void* value) {
|
|
|
|
Block* block = reinterpret_cast<Block*>(value);
|
|
|
|
delete block;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Make a copy of the block contents and insert into compressed block cache
|
|
|
|
//
|
|
|
|
Status BlockBasedTableBuilder::InsertBlockInCache(const Slice& block_contents,
|
2014-07-16 15:45:49 +02:00
|
|
|
const CompressionType type,
|
|
|
|
const BlockHandle* handle) {
|
2013-09-02 08:23:40 +02:00
|
|
|
Rep* r = rep_;
|
2014-08-25 23:22:05 +02:00
|
|
|
Cache* block_cache_compressed = r->table_options.block_cache_compressed.get();
|
2013-09-02 08:23:40 +02:00
|
|
|
|
|
|
|
if (type != kNoCompression && block_cache_compressed != nullptr) {
|
|
|
|
|
|
|
|
Cache::Handle* cache_handle = nullptr;
|
|
|
|
size_t size = block_contents.size();
|
|
|
|
|
2014-07-16 15:45:49 +02:00
|
|
|
char* ubuf = new char[size + 1]; // make a new copy
|
2013-09-02 08:23:40 +02:00
|
|
|
memcpy(ubuf, block_contents.data(), size);
|
2014-07-16 15:45:49 +02:00
|
|
|
ubuf[size] = type;
|
2013-09-02 08:23:40 +02:00
|
|
|
|
|
|
|
BlockContents results;
|
|
|
|
Slice sl(ubuf, size);
|
|
|
|
results.data = sl;
|
|
|
|
results.cachable = true; // XXX
|
|
|
|
results.heap_allocated = true;
|
|
|
|
results.compression_type = type;
|
|
|
|
|
|
|
|
Block* block = new Block(results);
|
|
|
|
|
|
|
|
// make cache key by appending the file offset to the cache prefix id
|
|
|
|
char* end = EncodeVarint64(
|
|
|
|
r->compressed_cache_key_prefix +
|
|
|
|
r->compressed_cache_key_prefix_size,
|
|
|
|
handle->offset());
|
|
|
|
Slice key(r->compressed_cache_key_prefix, static_cast<size_t>
|
|
|
|
(end - r->compressed_cache_key_prefix));
|
|
|
|
|
|
|
|
// Insert into compressed block cache.
|
|
|
|
cache_handle = block_cache_compressed->Insert(key, block, block->size(),
|
|
|
|
&DeleteCachedBlock);
|
|
|
|
block_cache_compressed->Release(cache_handle);
|
|
|
|
|
|
|
|
// Invalidate OS cache.
|
|
|
|
r->file->InvalidateCache(r->offset, size);
|
|
|
|
}
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
Status BlockBasedTableBuilder::Finish() {
|
2011-03-18 23:37:00 +01:00
|
|
|
Rep* r = rep_;
|
2013-11-08 06:27:21 +01:00
|
|
|
bool empty_data_block = r->data_block.empty();
|
2011-03-18 23:37:00 +01:00
|
|
|
Flush();
|
|
|
|
assert(!r->closed);
|
|
|
|
r->closed = true;
|
2012-04-17 17:36:46 +02:00
|
|
|
|
2013-11-08 06:27:21 +01:00
|
|
|
BlockHandle filter_block_handle,
|
|
|
|
metaindex_block_handle,
|
|
|
|
index_block_handle;
|
2012-04-17 17:36:46 +02:00
|
|
|
|
|
|
|
// Write filter block
|
2013-03-01 03:04:58 +01:00
|
|
|
if (ok() && r->filter_block != nullptr) {
|
2013-11-20 01:29:42 +01:00
|
|
|
auto filter_contents = r->filter_block->Finish();
|
|
|
|
r->props.filter_size = filter_contents.size();
|
|
|
|
WriteRawBlock(filter_contents, kNoCompression, &filter_block_handle);
|
2012-04-17 17:36:46 +02:00
|
|
|
}
|
|
|
|
|
2013-11-20 01:29:42 +01:00
|
|
|
// To make sure properties block is able to keep the accurate size of index
|
2013-10-10 20:43:24 +02:00
|
|
|
// block, we will finish writing all index entries here and flush them
|
|
|
|
// to storage after metaindex block is written.
|
2013-11-08 06:27:21 +01:00
|
|
|
if (ok() && !empty_data_block) {
|
2014-05-15 23:09:03 +02:00
|
|
|
r->index_builder->AddIndexEntry(
|
|
|
|
&r->last_key, nullptr /* no next data block */, r->pending_handle);
|
|
|
|
}
|
|
|
|
|
|
|
|
IndexBuilder::IndexBlocks index_blocks;
|
|
|
|
auto s = r->index_builder->Finish(&index_blocks);
|
|
|
|
if (!s.ok()) {
|
|
|
|
return s;
|
2013-10-10 20:43:24 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Write meta blocks and metaindex block with the following order.
|
|
|
|
// 1. [meta block: filter]
|
2014-05-15 23:09:03 +02:00
|
|
|
// 2. [other meta blocks]
|
|
|
|
// 3. [meta block: properties]
|
|
|
|
// 4. [metaindex block]
|
|
|
|
// write meta blocks
|
|
|
|
MetaIndexBuilder meta_index_builder;
|
|
|
|
for (const auto& item : index_blocks.meta_blocks) {
|
|
|
|
BlockHandle block_handle;
|
|
|
|
WriteBlock(item.second, &block_handle);
|
|
|
|
meta_index_builder.Add(item.first, block_handle);
|
|
|
|
}
|
2013-10-10 20:43:24 +02:00
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
if (ok()) {
|
2013-03-01 03:04:58 +01:00
|
|
|
if (r->filter_block != nullptr) {
|
2013-10-10 20:43:24 +02:00
|
|
|
// Add mapping from "<filter_block_prefix>.Name" to location
|
|
|
|
// of filter data.
|
2013-10-29 01:54:09 +01:00
|
|
|
std::string key = BlockBasedTable::kFilterBlockPrefix;
|
2014-08-25 23:22:05 +02:00
|
|
|
key.append(r->table_options.filter_policy->Name());
|
2014-05-15 23:09:03 +02:00
|
|
|
meta_index_builder.Add(key, filter_block_handle);
|
2013-10-10 20:43:24 +02:00
|
|
|
}
|
|
|
|
|
2013-11-20 01:29:42 +01:00
|
|
|
// Write properties block.
|
2013-10-10 20:43:24 +02:00
|
|
|
{
|
2013-12-05 22:09:13 +01:00
|
|
|
PropertyBlockBuilder property_block_builder;
|
|
|
|
std::vector<std::string> failed_user_prop_collectors;
|
2014-08-25 23:22:05 +02:00
|
|
|
r->props.filter_policy_name = r->table_options.filter_policy != nullptr ?
|
|
|
|
r->table_options.filter_policy->Name() : "";
|
2013-11-20 01:29:42 +01:00
|
|
|
r->props.index_size =
|
2014-03-01 03:19:07 +01:00
|
|
|
r->index_builder->EstimatedSize() + kBlockTrailerSize;
|
2013-10-10 20:43:24 +02:00
|
|
|
|
2013-12-05 22:09:13 +01:00
|
|
|
// Add basic properties
|
|
|
|
property_block_builder.AddTableProperty(r->props);
|
2013-10-16 20:50:50 +02:00
|
|
|
|
2014-03-01 03:19:07 +01:00
|
|
|
// Add use collected properties
|
TablePropertiesCollectorFactory
Summary:
This diff addresses task #4296714 and rethinks how users provide us with TablePropertiesCollectors as part of Options.
Here's description of task #4296714:
I'm debugging #4295529 and noticed that our count of user properties kDeletedKeys is wrong. We're sharing one single InternalKeyPropertiesCollector with all Table Builders. In LOG Files, we're outputting number of kDeletedKeys as connected with a single table, while it's actually the total count of deleted keys since creation of the DB.
For example, this table has 3155 entries and 1391828 deleted keys.
The problem with current approach that we call methods on a single TablePropertiesCollector for all the tables we create. Even worse, we could do it from multiple threads at the same time and TablePropertiesCollector has no way of knowing which table we're calling it for.
Good part: Looks like nobody inside Facebook is using Options::table_properties_collectors. This means we should be able to painfully change the API.
In this change, I introduce TablePropertiesCollectorFactory. For every table we create, we call `CreateTablePropertiesCollector`, which creates a TablePropertiesCollector for a single table. We then use it sequentially from a single thread, which means it doesn't have to be thread-safe.
Test Plan:
Added a test in table_properties_collector_test that fails on master (build two tables, assert that kDeletedKeys count is correct for the second one).
Also, all other tests
Reviewers: sdong, dhruba, haobo, kailiu
Reviewed By: kailiu
CC: leveldb
Differential Revision: https://reviews.facebook.net/D18579
2014-05-13 21:30:55 +02:00
|
|
|
NotifyCollectTableCollectorsOnFinish(r->table_properties_collectors,
|
|
|
|
r->options.info_log.get(),
|
|
|
|
&property_block_builder);
|
2013-10-10 20:43:24 +02:00
|
|
|
|
2013-11-20 01:29:42 +01:00
|
|
|
BlockHandle properties_block_handle;
|
2013-12-05 22:09:13 +01:00
|
|
|
WriteRawBlock(
|
|
|
|
property_block_builder.Finish(),
|
|
|
|
kNoCompression,
|
|
|
|
&properties_block_handle
|
2013-10-10 20:43:24 +02:00
|
|
|
);
|
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
meta_index_builder.Add(kPropertiesBlock, properties_block_handle);
|
2013-12-05 22:09:13 +01:00
|
|
|
} // end of properties block writing
|
2014-05-15 23:09:03 +02:00
|
|
|
} // meta blocks
|
2012-04-17 17:36:46 +02:00
|
|
|
|
|
|
|
// Write index block
|
2011-03-18 23:37:00 +01:00
|
|
|
if (ok()) {
|
2014-05-15 23:09:03 +02:00
|
|
|
// flush the meta index block
|
|
|
|
WriteRawBlock(meta_index_builder.Finish(), kNoCompression,
|
|
|
|
&metaindex_block_handle);
|
|
|
|
WriteBlock(index_blocks.index_block_contents, &index_block_handle);
|
2011-03-18 23:37:00 +01:00
|
|
|
}
|
2012-04-17 17:36:46 +02:00
|
|
|
|
|
|
|
// Write footer
|
2011-03-18 23:37:00 +01:00
|
|
|
if (ok()) {
|
2014-05-01 20:09:32 +02:00
|
|
|
// No need to write out new footer if we're using default checksum.
|
|
|
|
// We're writing legacy magic number because we want old versions of RocksDB
|
|
|
|
// be able to read files generated with new release (just in case if
|
|
|
|
// somebody wants to roll back after an upgrade)
|
|
|
|
// TODO(icanadi) at some point in the future, when we're absolutely sure
|
|
|
|
// nobody will roll back to RocksDB 2.x versions, retire the legacy magic
|
|
|
|
// number and always write new table files with new magic number
|
2014-08-25 23:22:05 +02:00
|
|
|
bool legacy = (r->table_options.checksum == kCRC32c);
|
2014-05-01 20:09:32 +02:00
|
|
|
Footer footer(legacy ? kLegacyBlockBasedTableMagicNumber
|
|
|
|
: kBlockBasedTableMagicNumber);
|
2011-03-18 23:37:00 +01:00
|
|
|
footer.set_metaindex_handle(metaindex_block_handle);
|
|
|
|
footer.set_index_handle(index_block_handle);
|
2014-08-25 23:22:05 +02:00
|
|
|
footer.set_checksum(r->table_options.checksum);
|
2011-03-18 23:37:00 +01:00
|
|
|
std::string footer_encoding;
|
|
|
|
footer.EncodeTo(&footer_encoding);
|
|
|
|
r->status = r->file->Append(footer_encoding);
|
|
|
|
if (r->status.ok()) {
|
|
|
|
r->offset += footer_encoding.size();
|
|
|
|
}
|
|
|
|
}
|
2013-11-20 01:29:42 +01:00
|
|
|
|
|
|
|
// Print out the table stats
|
|
|
|
if (ok()) {
|
|
|
|
// user collected properties
|
|
|
|
std::string user_collected;
|
|
|
|
user_collected.reserve(1024);
|
TablePropertiesCollectorFactory
Summary:
This diff addresses task #4296714 and rethinks how users provide us with TablePropertiesCollectors as part of Options.
Here's description of task #4296714:
I'm debugging #4295529 and noticed that our count of user properties kDeletedKeys is wrong. We're sharing one single InternalKeyPropertiesCollector with all Table Builders. In LOG Files, we're outputting number of kDeletedKeys as connected with a single table, while it's actually the total count of deleted keys since creation of the DB.
For example, this table has 3155 entries and 1391828 deleted keys.
The problem with current approach that we call methods on a single TablePropertiesCollector for all the tables we create. Even worse, we could do it from multiple threads at the same time and TablePropertiesCollector has no way of knowing which table we're calling it for.
Good part: Looks like nobody inside Facebook is using Options::table_properties_collectors. This means we should be able to painfully change the API.
In this change, I introduce TablePropertiesCollectorFactory. For every table we create, we call `CreateTablePropertiesCollector`, which creates a TablePropertiesCollector for a single table. We then use it sequentially from a single thread, which means it doesn't have to be thread-safe.
Test Plan:
Added a test in table_properties_collector_test that fails on master (build two tables, assert that kDeletedKeys count is correct for the second one).
Also, all other tests
Reviewers: sdong, dhruba, haobo, kailiu
Reviewed By: kailiu
CC: leveldb
Differential Revision: https://reviews.facebook.net/D18579
2014-05-13 21:30:55 +02:00
|
|
|
for (const auto& collector : r->table_properties_collectors) {
|
2013-11-20 01:29:42 +01:00
|
|
|
for (const auto& prop : collector->GetReadableProperties()) {
|
|
|
|
user_collected.append(prop.first);
|
|
|
|
user_collected.append("=");
|
|
|
|
user_collected.append(prop.second);
|
|
|
|
user_collected.append("; ");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Log(
|
|
|
|
r->options.info_log,
|
|
|
|
"Table was constructed:\n"
|
2013-11-22 02:54:23 +01:00
|
|
|
" [basic properties]: %s\n"
|
|
|
|
" [user collected properties]: %s",
|
2013-11-20 01:29:42 +01:00
|
|
|
r->props.ToString().c_str(),
|
|
|
|
user_collected.c_str()
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
2011-03-18 23:37:00 +01:00
|
|
|
return r->status;
|
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
void BlockBasedTableBuilder::Abandon() {
|
2011-03-18 23:37:00 +01:00
|
|
|
Rep* r = rep_;
|
|
|
|
assert(!r->closed);
|
|
|
|
r->closed = true;
|
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
uint64_t BlockBasedTableBuilder::NumEntries() const {
|
2013-11-20 01:29:42 +01:00
|
|
|
return rep_->props.num_entries;
|
2011-03-18 23:37:00 +01:00
|
|
|
}
|
|
|
|
|
2013-10-29 01:54:09 +01:00
|
|
|
uint64_t BlockBasedTableBuilder::FileSize() const {
|
2011-03-18 23:37:00 +01:00
|
|
|
return rep_->offset;
|
|
|
|
}
|
|
|
|
|
2014-05-15 23:09:03 +02:00
|
|
|
const std::string BlockBasedTable::kFilterBlockPrefix = "filter.";
|
2013-12-05 22:09:13 +01:00
|
|
|
|
2013-10-04 06:49:15 +02:00
|
|
|
} // namespace rocksdb
|