2016-02-10 00:12:00 +01:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2017-07-16 01:03:42 +02:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2014-11-21 00:54:47 +01:00
|
|
|
//
|
|
|
|
// Copyright (c) 2012 Facebook.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file.
|
|
|
|
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
|
2017-04-24 23:57:27 +02:00
|
|
|
#include "utilities/checkpoint/checkpoint_impl.h"
|
2014-11-21 00:54:47 +01:00
|
|
|
|
|
|
|
#ifndef __STDC_FORMAT_MACROS
|
|
|
|
#define __STDC_FORMAT_MACROS
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#include <inttypes.h>
|
|
|
|
#include <algorithm>
|
|
|
|
#include <string>
|
2017-04-24 23:57:27 +02:00
|
|
|
#include <vector>
|
|
|
|
|
2015-06-20 01:08:31 +02:00
|
|
|
#include "db/wal_manager.h"
|
2017-04-04 03:27:24 +02:00
|
|
|
#include "port/port.h"
|
2014-11-21 00:54:47 +01:00
|
|
|
#include "rocksdb/db.h"
|
|
|
|
#include "rocksdb/env.h"
|
2015-06-20 01:08:31 +02:00
|
|
|
#include "rocksdb/transaction_log.h"
|
2017-04-24 23:57:27 +02:00
|
|
|
#include "rocksdb/utilities/checkpoint.h"
|
2014-11-21 00:54:47 +01:00
|
|
|
#include "util/file_util.h"
|
2017-04-04 03:27:24 +02:00
|
|
|
#include "util/filename.h"
|
2016-03-17 18:07:21 +01:00
|
|
|
#include "util/sync_point.h"
|
2014-11-21 00:54:47 +01:00
|
|
|
|
|
|
|
namespace rocksdb {
|
|
|
|
|
|
|
|
Status Checkpoint::Create(DB* db, Checkpoint** checkpoint_ptr) {
|
|
|
|
*checkpoint_ptr = new CheckpointImpl(db);
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
2017-07-22 03:13:59 +02:00
|
|
|
Status Checkpoint::CreateCheckpoint(const std::string& checkpoint_dir,
|
|
|
|
uint64_t log_size_for_flush) {
|
2014-11-21 00:54:47 +01:00
|
|
|
return Status::NotSupported("");
|
|
|
|
}
|
|
|
|
|
|
|
|
// Builds an openable snapshot of RocksDB
|
2017-03-22 01:53:21 +01:00
|
|
|
Status CheckpointImpl::CreateCheckpoint(const std::string& checkpoint_dir,
|
|
|
|
uint64_t log_size_for_flush) {
|
2016-12-29 03:38:20 +01:00
|
|
|
DBOptions db_options = db_->GetDBOptions();
|
2014-11-21 00:54:47 +01:00
|
|
|
|
2017-04-24 23:57:27 +02:00
|
|
|
Status s = db_->GetEnv()->FileExists(checkpoint_dir);
|
2015-07-21 02:20:40 +02:00
|
|
|
if (s.ok()) {
|
2014-11-21 00:54:47 +01:00
|
|
|
return Status::InvalidArgument("Directory exists");
|
2015-07-21 02:20:40 +02:00
|
|
|
} else if (!s.IsNotFound()) {
|
|
|
|
assert(s.IsIOError());
|
|
|
|
return s;
|
2014-11-21 00:54:47 +01:00
|
|
|
}
|
|
|
|
|
2017-04-24 23:57:27 +02:00
|
|
|
ROCKS_LOG_INFO(
|
|
|
|
db_options.info_log,
|
|
|
|
"Started the snapshot process -- creating snapshot in directory %s",
|
|
|
|
checkpoint_dir.c_str());
|
2018-01-30 05:59:59 +01:00
|
|
|
|
|
|
|
size_t final_nonslash_idx = checkpoint_dir.find_last_not_of('/');
|
|
|
|
if (final_nonslash_idx == std::string::npos) {
|
|
|
|
// npos means it's only slashes, which means it's the root directory, but it
|
|
|
|
// shouldn't be because we verified above the directory doesn't exist.
|
|
|
|
assert(false);
|
|
|
|
return Status::InvalidArgument("invalid checkpoint directory name");
|
|
|
|
}
|
|
|
|
|
|
|
|
std::string full_private_path =
|
|
|
|
checkpoint_dir.substr(0, final_nonslash_idx + 1) + ".tmp";
|
|
|
|
ROCKS_LOG_INFO(
|
|
|
|
db_options.info_log,
|
|
|
|
"Snapshot process -- using temporary directory %s",
|
|
|
|
full_private_path.c_str());
|
2017-04-24 23:57:27 +02:00
|
|
|
// create snapshot directory
|
|
|
|
s = db_->GetEnv()->CreateDir(full_private_path);
|
|
|
|
uint64_t sequence_number = 0;
|
|
|
|
if (s.ok()) {
|
|
|
|
db_->DisableFileDeletions();
|
|
|
|
s = CreateCustomCheckpoint(
|
|
|
|
db_options,
|
|
|
|
[&](const std::string& src_dirname, const std::string& fname,
|
|
|
|
FileType) {
|
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Hard Linking %s", fname.c_str());
|
|
|
|
return db_->GetEnv()->LinkFile(src_dirname + fname,
|
|
|
|
full_private_path + fname);
|
|
|
|
} /* link_file_cb */,
|
|
|
|
[&](const std::string& src_dirname, const std::string& fname,
|
|
|
|
uint64_t size_limit_bytes, FileType) {
|
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Copying %s", fname.c_str());
|
|
|
|
return CopyFile(db_->GetEnv(), src_dirname + fname,
|
|
|
|
full_private_path + fname, size_limit_bytes,
|
|
|
|
db_options.use_fsync);
|
|
|
|
} /* copy_file_cb */,
|
|
|
|
[&](const std::string& fname, const std::string& contents, FileType) {
|
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Creating %s", fname.c_str());
|
|
|
|
return CreateFile(db_->GetEnv(), full_private_path + fname, contents);
|
|
|
|
} /* create_file_cb */,
|
|
|
|
&sequence_number, log_size_for_flush);
|
|
|
|
// we copied all the files, enable file deletions
|
|
|
|
db_->EnableFileDeletions(false);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s.ok()) {
|
|
|
|
// move tmp private backup to real snapshot directory
|
|
|
|
s = db_->GetEnv()->RenameFile(full_private_path, checkpoint_dir);
|
|
|
|
}
|
|
|
|
if (s.ok()) {
|
|
|
|
unique_ptr<Directory> checkpoint_directory;
|
|
|
|
db_->GetEnv()->NewDirectory(checkpoint_dir, &checkpoint_directory);
|
|
|
|
if (checkpoint_directory != nullptr) {
|
|
|
|
s = checkpoint_directory->Fsync();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s.ok()) {
|
|
|
|
// here we know that we succeeded and installed the new snapshot
|
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Snapshot DONE. All is good");
|
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Snapshot sequence number: %" PRIu64,
|
|
|
|
sequence_number);
|
|
|
|
} else {
|
|
|
|
// clean all the files we might have created
|
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Snapshot failed -- %s",
|
|
|
|
s.ToString().c_str());
|
|
|
|
// we have to delete the dir and all its children
|
|
|
|
std::vector<std::string> subchildren;
|
|
|
|
db_->GetEnv()->GetChildren(full_private_path, &subchildren);
|
|
|
|
for (auto& subchild : subchildren) {
|
|
|
|
std::string subchild_path = full_private_path + "/" + subchild;
|
|
|
|
Status s1 = db_->GetEnv()->DeleteFile(subchild_path);
|
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Delete file %s -- %s",
|
|
|
|
subchild_path.c_str(), s1.ToString().c_str());
|
|
|
|
}
|
|
|
|
// finally delete the private dir
|
|
|
|
Status s1 = db_->GetEnv()->DeleteDir(full_private_path);
|
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Delete dir %s -- %s",
|
|
|
|
full_private_path.c_str(), s1.ToString().c_str());
|
|
|
|
}
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
Status CheckpointImpl::CreateCustomCheckpoint(
|
|
|
|
const DBOptions& db_options,
|
|
|
|
std::function<Status(const std::string& src_dirname,
|
|
|
|
const std::string& src_fname, FileType type)>
|
|
|
|
link_file_cb,
|
|
|
|
std::function<Status(const std::string& src_dirname,
|
|
|
|
const std::string& src_fname,
|
|
|
|
uint64_t size_limit_bytes, FileType type)>
|
|
|
|
copy_file_cb,
|
|
|
|
std::function<Status(const std::string& fname, const std::string& contents,
|
|
|
|
FileType type)>
|
|
|
|
create_file_cb,
|
|
|
|
uint64_t* sequence_number, uint64_t log_size_for_flush) {
|
|
|
|
Status s;
|
|
|
|
std::vector<std::string> live_files;
|
|
|
|
uint64_t manifest_file_size = 0;
|
|
|
|
uint64_t min_log_num = port::kMaxUint64;
|
|
|
|
*sequence_number = db_->GetLatestSequenceNumber();
|
|
|
|
bool same_fs = true;
|
|
|
|
VectorLogPtr live_wal_files;
|
|
|
|
|
2017-03-22 01:53:21 +01:00
|
|
|
bool flush_memtable = true;
|
2014-11-21 00:54:47 +01:00
|
|
|
if (s.ok()) {
|
2017-03-22 01:53:21 +01:00
|
|
|
if (!db_options.allow_2pc) {
|
2017-04-24 23:57:27 +02:00
|
|
|
if (log_size_for_flush == port::kMaxUint64) {
|
|
|
|
flush_memtable = false;
|
|
|
|
} else if (log_size_for_flush > 0) {
|
|
|
|
// If out standing log files are small, we skip the flush.
|
|
|
|
s = db_->GetSortedWalFiles(live_wal_files);
|
2017-03-22 01:53:21 +01:00
|
|
|
|
2017-04-24 23:57:27 +02:00
|
|
|
if (!s.ok()) {
|
|
|
|
return s;
|
|
|
|
}
|
2017-03-22 01:53:21 +01:00
|
|
|
|
2017-04-24 23:57:27 +02:00
|
|
|
// Don't flush column families if total log size is smaller than
|
|
|
|
// log_size_for_flush. We copy the log files instead.
|
|
|
|
// We may be able to cover 2PC case too.
|
|
|
|
uint64_t total_wal_size = 0;
|
|
|
|
for (auto& wal : live_wal_files) {
|
|
|
|
total_wal_size += wal->SizeFileBytes();
|
|
|
|
}
|
|
|
|
if (total_wal_size < log_size_for_flush) {
|
|
|
|
flush_memtable = false;
|
|
|
|
}
|
|
|
|
live_wal_files.clear();
|
2017-03-22 01:53:21 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-11-21 00:54:47 +01:00
|
|
|
// this will return live_files prefixed with "/"
|
2017-03-22 01:53:21 +01:00
|
|
|
s = db_->GetLiveFiles(live_files, &manifest_file_size, flush_memtable);
|
2016-12-28 20:53:29 +01:00
|
|
|
|
2016-12-29 03:38:20 +01:00
|
|
|
if (s.ok() && db_options.allow_2pc) {
|
2016-12-28 20:53:29 +01:00
|
|
|
// If 2PC is enabled, we need to get minimum log number after the flush.
|
|
|
|
// Need to refetch the live files to recapture the snapshot.
|
|
|
|
if (!db_->GetIntProperty(DB::Properties::kMinLogNumberToKeep,
|
|
|
|
&min_log_num)) {
|
|
|
|
return Status::InvalidArgument(
|
|
|
|
"2PC enabled but cannot fine the min log number to keep.");
|
|
|
|
}
|
|
|
|
// We need to refetch live files with flush to handle this case:
|
|
|
|
// A previous 000001.log contains the prepare record of transaction tnx1.
|
|
|
|
// The current log file is 000002.log, and sequence_number points to this
|
|
|
|
// file.
|
|
|
|
// After calling GetLiveFiles(), 000003.log is created.
|
|
|
|
// Then tnx1 is committed. The commit record is written to 000003.log.
|
|
|
|
// Now we fetch min_log_num, which will be 3.
|
|
|
|
// Then only 000002.log and 000003.log will be copied, and 000001.log will
|
|
|
|
// be skipped. 000003.log contains commit message of tnx1, but we don't
|
|
|
|
// have respective prepare record for it.
|
|
|
|
// In order to avoid this situation, we need to force flush to make sure
|
2017-04-24 23:57:27 +02:00
|
|
|
// all transactions committed before getting min_log_num will be flushed
|
2016-12-28 20:53:29 +01:00
|
|
|
// to SST files.
|
|
|
|
// We cannot get min_log_num before calling the GetLiveFiles() for the
|
|
|
|
// first time, because if we do that, all the logs files will be included,
|
|
|
|
// far more than needed.
|
2017-04-24 23:57:27 +02:00
|
|
|
s = db_->GetLiveFiles(live_files, &manifest_file_size, flush_memtable);
|
2016-12-28 20:53:29 +01:00
|
|
|
}
|
|
|
|
|
2016-03-17 18:07:21 +01:00
|
|
|
TEST_SYNC_POINT("CheckpointImpl::CreateCheckpoint:SavedLiveFiles1");
|
|
|
|
TEST_SYNC_POINT("CheckpointImpl::CreateCheckpoint:SavedLiveFiles2");
|
2017-06-24 23:06:43 +02:00
|
|
|
db_->FlushWAL(false /* sync */);
|
2014-11-21 00:54:47 +01:00
|
|
|
}
|
2015-06-20 01:08:31 +02:00
|
|
|
// if we have more than one column family, we need to also get WAL files
|
|
|
|
if (s.ok()) {
|
|
|
|
s = db_->GetSortedWalFiles(live_wal_files);
|
|
|
|
}
|
2014-11-21 00:54:47 +01:00
|
|
|
if (!s.ok()) {
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2015-06-20 01:08:31 +02:00
|
|
|
size_t wal_size = live_wal_files.size();
|
2014-11-21 00:54:47 +01:00
|
|
|
|
|
|
|
// copy/hard link live_files
|
2016-03-17 18:07:21 +01:00
|
|
|
std::string manifest_fname, current_fname;
|
2014-11-21 00:54:47 +01:00
|
|
|
for (size_t i = 0; s.ok() && i < live_files.size(); ++i) {
|
|
|
|
uint64_t number;
|
|
|
|
FileType type;
|
|
|
|
bool ok = ParseFileName(live_files[i], &number, &type);
|
|
|
|
if (!ok) {
|
|
|
|
s = Status::Corruption("Can't parse file name. This is very bad");
|
|
|
|
break;
|
|
|
|
}
|
2016-06-10 04:03:10 +02:00
|
|
|
// we should only get sst, options, manifest and current files here
|
2014-11-21 00:54:47 +01:00
|
|
|
assert(type == kTableFile || type == kDescriptorFile ||
|
2016-06-10 04:03:10 +02:00
|
|
|
type == kCurrentFile || type == kOptionsFile);
|
2014-11-21 00:54:47 +01:00
|
|
|
assert(live_files[i].size() > 0 && live_files[i][0] == '/');
|
2016-03-17 18:07:21 +01:00
|
|
|
if (type == kCurrentFile) {
|
|
|
|
// We will craft the current file manually to ensure it's consistent with
|
|
|
|
// the manifest number. This is necessary because current's file contents
|
|
|
|
// can change during checkpoint creation.
|
|
|
|
current_fname = live_files[i];
|
|
|
|
continue;
|
|
|
|
} else if (type == kDescriptorFile) {
|
|
|
|
manifest_fname = live_files[i];
|
|
|
|
}
|
2014-11-21 00:54:47 +01:00
|
|
|
std::string src_fname = live_files[i];
|
|
|
|
|
|
|
|
// rules:
|
|
|
|
// * if it's kTableFile, then it's shared
|
|
|
|
// * if it's kDescriptorFile, limit the size to manifest_file_size
|
|
|
|
// * always copy if cross-device link
|
|
|
|
if ((type == kTableFile) && same_fs) {
|
2017-04-24 23:57:27 +02:00
|
|
|
s = link_file_cb(db_->GetName(), src_fname, type);
|
2014-11-21 00:54:47 +01:00
|
|
|
if (s.IsNotSupported()) {
|
|
|
|
same_fs = false;
|
|
|
|
s = Status::OK();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if ((type != kTableFile) || (!same_fs)) {
|
2017-04-24 23:57:27 +02:00
|
|
|
s = copy_file_cb(db_->GetName(), src_fname,
|
|
|
|
(type == kDescriptorFile) ? manifest_file_size : 0,
|
|
|
|
type);
|
2014-11-21 00:54:47 +01:00
|
|
|
}
|
|
|
|
}
|
2016-03-17 18:07:21 +01:00
|
|
|
if (s.ok() && !current_fname.empty() && !manifest_fname.empty()) {
|
2017-04-24 23:57:27 +02:00
|
|
|
create_file_cb(current_fname, manifest_fname.substr(1) + "\n",
|
|
|
|
kCurrentFile);
|
2016-03-17 18:07:21 +01:00
|
|
|
}
|
2017-03-16 03:22:52 +01:00
|
|
|
ROCKS_LOG_INFO(db_options.info_log, "Number of log files %" ROCKSDB_PRIszt,
|
|
|
|
live_wal_files.size());
|
2015-06-20 01:08:31 +02:00
|
|
|
|
|
|
|
// Link WAL files. Copy exact size of last one because it is the only one
|
|
|
|
// that has changes after the last flush.
|
|
|
|
for (size_t i = 0; s.ok() && i < wal_size; ++i) {
|
|
|
|
if ((live_wal_files[i]->Type() == kAliveLogFile) &&
|
2017-03-22 01:53:21 +01:00
|
|
|
(!flush_memtable ||
|
2017-04-24 23:57:27 +02:00
|
|
|
live_wal_files[i]->StartSequence() >= *sequence_number ||
|
2016-12-28 20:53:29 +01:00
|
|
|
live_wal_files[i]->LogNumber() >= min_log_num)) {
|
2015-06-20 01:08:31 +02:00
|
|
|
if (i + 1 == wal_size) {
|
2017-04-24 23:57:27 +02:00
|
|
|
s = copy_file_cb(db_options.wal_dir, live_wal_files[i]->PathName(),
|
|
|
|
live_wal_files[i]->SizeFileBytes(), kLogFile);
|
2015-06-20 01:08:31 +02:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (same_fs) {
|
|
|
|
// we only care about live log files
|
2017-04-24 23:57:27 +02:00
|
|
|
s = link_file_cb(db_options.wal_dir, live_wal_files[i]->PathName(),
|
|
|
|
kLogFile);
|
2015-06-20 01:08:31 +02:00
|
|
|
if (s.IsNotSupported()) {
|
|
|
|
same_fs = false;
|
|
|
|
s = Status::OK();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (!same_fs) {
|
2017-04-24 23:57:27 +02:00
|
|
|
s = copy_file_cb(db_options.wal_dir, live_wal_files[i]->PathName(), 0,
|
|
|
|
kLogFile);
|
2015-06-20 01:08:31 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2014-11-21 00:54:47 +01:00
|
|
|
|
|
|
|
return s;
|
|
|
|
}
|
2017-04-24 23:57:27 +02:00
|
|
|
|
2014-11-21 00:54:47 +01:00
|
|
|
} // namespace rocksdb
|
|
|
|
|
|
|
|
#endif // ROCKSDB_LITE
|