fix some spelling typos (#6464)

Summary:
Found from Debian's "Lintian" program
Pull Request resolved: https://github.com/facebook/rocksdb/pull/6464

Differential Revision: D20162862

Pulled By: zhichao-cao

fbshipit-source-id: 06941ee2437b038b2b8045becbe9d2c6fbff3e12
This commit is contained in:
Michael R. Crusoe 2020-02-28 14:10:51 -08:00 committed by Facebook Github Bot
parent 41535d0218
commit 051696bf98
19 changed files with 39 additions and 39 deletions

View File

@ -216,7 +216,7 @@ missing_make_config_paths := $(shell \
done | sort | uniq) done | sort | uniq)
$(foreach path, $(missing_make_config_paths), \ $(foreach path, $(missing_make_config_paths), \
$(warning Warning: $(path) dont exist)) $(warning Warning: $(path) does not exist))
ifeq ($(PLATFORM), OS_AIX) ifeq ($(PLATFORM), OS_AIX)
# no debug info # no debug info

View File

@ -5801,7 +5801,7 @@ sub workdir {
. "-" . $self->seq(); . "-" . $self->seq();
} else { } else {
$workdir = $opt::workdir; $workdir = $opt::workdir;
# Rsync treats /./ special. We dont want that # Rsync treats /./ special. We don't want that
$workdir =~ s:/\./:/:g; # Remove /./ $workdir =~ s:/\./:/:g; # Remove /./
$workdir =~ s:/+$::; # Remove ending / if any $workdir =~ s:/+$::; # Remove ending / if any
$workdir =~ s:^\./::g; # Remove starting ./ if any $workdir =~ s:^\./::g; # Remove starting ./ if any

View File

@ -821,7 +821,7 @@ TEST_P(ColumnFamilyTest, BulkAddDrop) {
} }
TEST_P(ColumnFamilyTest, DropTest) { TEST_P(ColumnFamilyTest, DropTest) {
// first iteration - dont reopen DB before dropping // first iteration - don't reopen DB before dropping
// second iteration - reopen DB before dropping // second iteration - reopen DB before dropping
for (int iter = 0; iter < 2; ++iter) { for (int iter = 0; iter < 2; ++iter) {
Open({"default"}); Open({"default"});

View File

@ -2021,7 +2021,7 @@ class DBImpl : public DB {
// REQUIRES: mutex locked // REQUIRES: mutex locked
std::unique_ptr<ROCKSDB_NAMESPACE::RepeatableThread> thread_persist_stats_; std::unique_ptr<ROCKSDB_NAMESPACE::RepeatableThread> thread_persist_stats_;
// When set, we use a separate queue for writes that dont write to memtable. // When set, we use a separate queue for writes that don't write to memtable.
// In 2PC these are the writes at Prepare phase. // In 2PC these are the writes at Prepare phase.
const bool two_write_queues_; const bool two_write_queues_;
const bool manual_wal_flush_; const bool manual_wal_flush_;

View File

@ -1873,7 +1873,7 @@ TEST_P(DBIteratorTest, IterPrevKeyCrossingBlocksRandomized) {
DestroyAndReopen(options); DestroyAndReopen(options);
const int kNumKeys = 500; const int kNumKeys = 500;
// Small number of merge operands to make sure that DBIter::Prev() dont // Small number of merge operands to make sure that DBIter::Prev() don't
// fall back to Seek() // fall back to Seek()
const int kNumMergeOperands = 3; const int kNumMergeOperands = 3;
// Use value size that will make sure that every block contain 1 key // Use value size that will make sure that every block contain 1 key
@ -1908,7 +1908,7 @@ TEST_P(DBIteratorTest, IterPrevKeyCrossingBlocksRandomized) {
ASSERT_OK(Flush()); ASSERT_OK(Flush());
// Separate values and merge operands in different file so that we // Separate values and merge operands in different file so that we
// make sure that we dont merge them while flushing but actually // make sure that we don't merge them while flushing but actually
// merge them in the read path // merge them in the read path
for (int i = 0; i < kNumKeys; i++) { for (int i = 0; i < kNumKeys; i++) {
if (rnd.PercentTrue(kNoMergeOpPercentage)) { if (rnd.PercentTrue(kNoMergeOpPercentage)) {

View File

@ -46,7 +46,7 @@ Status ExternalSstFileIngestionJob::Prepare(
TablePropertiesCollectorFactory::Context::kUnknownColumnFamily && TablePropertiesCollectorFactory::Context::kUnknownColumnFamily &&
f.cf_id != cfd_->GetID()) { f.cf_id != cfd_->GetID()) {
return Status::InvalidArgument( return Status::InvalidArgument(
"External file column family id dont match"); "External file column family id don't match");
} }
} }
@ -55,7 +55,7 @@ Status ExternalSstFileIngestionJob::Prepare(
if (num_files == 0) { if (num_files == 0) {
return Status::InvalidArgument("The list of files is empty"); return Status::InvalidArgument("The list of files is empty");
} else if (num_files > 1) { } else if (num_files > 1) {
// Verify that passed files dont have overlapping ranges // Verify that passed files don't have overlapping ranges
autovector<const IngestedFileInfo*> sorted_files; autovector<const IngestedFileInfo*> sorted_files;
for (size_t i = 0; i < num_files; i++) { for (size_t i = 0; i < num_files; i++) {
sorted_files.push_back(&files_to_ingest_[i]); sorted_files.push_back(&files_to_ingest_[i]);
@ -212,7 +212,7 @@ Status ExternalSstFileIngestionJob::Run() {
if (ingestion_options_.snapshot_consistency && !db_snapshots_->empty()) { if (ingestion_options_.snapshot_consistency && !db_snapshots_->empty()) {
// We need to assign a global sequence number to all the files even // We need to assign a global sequence number to all the files even
// if the dont overlap with any ranges since we have snapshots // if the don't overlap with any ranges since we have snapshots
force_global_seqno = true; force_global_seqno = true;
} }
// It is safe to use this instead of LastAllocatedSequence since we are // It is safe to use this instead of LastAllocatedSequence since we are
@ -588,7 +588,7 @@ Status ExternalSstFileIngestionJob::AssignLevelAndSeqnoForIngestedFile(
continue; continue;
} }
// We dont overlap with any keys in this level, but we still need to check // We don't overlap with any keys in this level, but we still need to check
// if our file can fit in it // if our file can fit in it
if (IngestedFileFitInLevel(file_to_ingest, lvl)) { if (IngestedFileFitInLevel(file_to_ingest, lvl)) {
target_level = lvl; target_level = lvl;
@ -646,7 +646,7 @@ Status ExternalSstFileIngestionJob::AssignGlobalSeqnoForIngestedFile(
return Status::InvalidArgument("Global seqno is required, but disabled"); return Status::InvalidArgument("Global seqno is required, but disabled");
} else if (file_to_ingest->global_seqno_offset == 0) { } else if (file_to_ingest->global_seqno_offset == 0) {
return Status::InvalidArgument( return Status::InvalidArgument(
"Trying to set global seqno for a file that dont have a global seqno " "Trying to set global seqno for a file that don't have a global seqno "
"field"); "field");
} }

View File

@ -476,17 +476,17 @@ TEST_F(ExternalSSTFileTest, Basic) {
} }
ASSERT_NE(db_->GetLatestSequenceNumber(), 0U); ASSERT_NE(db_->GetLatestSequenceNumber(), 0U);
// Key range of file5 (400 => 499) dont overlap with any keys in DB // Key range of file5 (400 => 499) don't overlap with any keys in DB
ASSERT_OK(DeprecatedAddFile({file5})); ASSERT_OK(DeprecatedAddFile({file5}));
// This file has overlapping values with the existing data // This file has overlapping values with the existing data
s = DeprecatedAddFile({file6}); s = DeprecatedAddFile({file6});
ASSERT_FALSE(s.ok()) << s.ToString(); ASSERT_FALSE(s.ok()) << s.ToString();
// Key range of file7 (500 => 598) dont overlap with any keys in DB // Key range of file7 (500 => 598) don't overlap with any keys in DB
ASSERT_OK(DeprecatedAddFile({file7})); ASSERT_OK(DeprecatedAddFile({file7}));
// Key range of file7 (600 => 700) dont overlap with any keys in DB // Key range of file7 (600 => 700) don't overlap with any keys in DB
ASSERT_OK(DeprecatedAddFile({file8})); ASSERT_OK(DeprecatedAddFile({file8}));
// Make sure values are correct before and after flush/compaction // Make sure values are correct before and after flush/compaction
@ -1609,15 +1609,15 @@ TEST_F(ExternalSSTFileTest, AddExternalSstFileWithCustomCompartor) {
generated_files[7]}; generated_files[7]};
ASSERT_NOK(DeprecatedAddFile(in_files)); ASSERT_NOK(DeprecatedAddFile(in_files));
// These 2 files dont overlap with each other // These 2 files don't overlap with each other
in_files = {generated_files[0], generated_files[2]}; in_files = {generated_files[0], generated_files[2]};
ASSERT_OK(DeprecatedAddFile(in_files)); ASSERT_OK(DeprecatedAddFile(in_files));
// These 2 files dont overlap with each other but overlap with keys in DB // These 2 files don't overlap with each other but overlap with keys in DB
in_files = {generated_files[3], generated_files[7]}; in_files = {generated_files[3], generated_files[7]};
ASSERT_NOK(DeprecatedAddFile(in_files)); ASSERT_NOK(DeprecatedAddFile(in_files));
// Files dont overlap and dont overlap with DB key range // Files don't overlap and don't overlap with DB key range
in_files = {generated_files[4], generated_files[6], generated_files[8]}; in_files = {generated_files[4], generated_files[6], generated_files[8]};
ASSERT_OK(DeprecatedAddFile(in_files)); ASSERT_OK(DeprecatedAddFile(in_files));
@ -1797,7 +1797,7 @@ TEST_P(ExternalSSTFileTest, IngestFileWithGlobalSeqnoAssignedLevel) {
options, file_data, -1, true, write_global_seqno, options, file_data, -1, true, write_global_seqno,
verify_checksums_before_ingest, false, false, &true_data)); verify_checksums_before_ingest, false, false, &true_data));
// This file dont overlap with anything in the DB, will go to L4 // This file don't overlap with anything in the DB, will go to L4
ASSERT_EQ("0,0,0,0,1", FilesPerLevel()); ASSERT_EQ("0,0,0,0,1", FilesPerLevel());
// Insert 80 -> 130 using AddFile // Insert 80 -> 130 using AddFile
@ -1822,7 +1822,7 @@ TEST_P(ExternalSSTFileTest, IngestFileWithGlobalSeqnoAssignedLevel) {
options, file_data, -1, true, write_global_seqno, options, file_data, -1, true, write_global_seqno,
verify_checksums_before_ingest, false, false, &true_data)); verify_checksums_before_ingest, false, false, &true_data));
// This file dont overlap with anything in the DB and fit in L4 as well // This file don't overlap with anything in the DB and fit in L4 as well
ASSERT_EQ("2,0,0,0,2", FilesPerLevel()); ASSERT_EQ("2,0,0,0,2", FilesPerLevel());
// Insert 10 -> 40 using AddFile // Insert 10 -> 40 using AddFile
@ -2059,16 +2059,16 @@ TEST_F(ExternalSSTFileTest, FileWithCFInfo) {
IngestExternalFileOptions ifo; IngestExternalFileOptions ifo;
// SST CF dont match // SST CF don't match
ASSERT_NOK(db_->IngestExternalFile(handles_[0], {cf1_sst}, ifo)); ASSERT_NOK(db_->IngestExternalFile(handles_[0], {cf1_sst}, ifo));
// SST CF dont match // SST CF don't match
ASSERT_NOK(db_->IngestExternalFile(handles_[2], {cf1_sst}, ifo)); ASSERT_NOK(db_->IngestExternalFile(handles_[2], {cf1_sst}, ifo));
// SST CF match // SST CF match
ASSERT_OK(db_->IngestExternalFile(handles_[1], {cf1_sst}, ifo)); ASSERT_OK(db_->IngestExternalFile(handles_[1], {cf1_sst}, ifo));
// SST CF dont match // SST CF don't match
ASSERT_NOK(db_->IngestExternalFile(handles_[1], {cf_default_sst}, ifo)); ASSERT_NOK(db_->IngestExternalFile(handles_[1], {cf_default_sst}, ifo));
// SST CF dont match // SST CF don't match
ASSERT_NOK(db_->IngestExternalFile(handles_[2], {cf_default_sst}, ifo)); ASSERT_NOK(db_->IngestExternalFile(handles_[2], {cf_default_sst}, ifo));
// SST CF match // SST CF match
ASSERT_OK(db_->IngestExternalFile(handles_[0], {cf_default_sst}, ifo)); ASSERT_OK(db_->IngestExternalFile(handles_[0], {cf_default_sst}, ifo));

View File

@ -464,7 +464,7 @@ size_t WriteThread::EnterAsBatchGroupLeader(Writer* leader,
} }
if (w->callback != nullptr && !w->callback->AllowWriteBatching()) { if (w->callback != nullptr && !w->callback->AllowWriteBatching()) {
// dont batch writes that don't want to be batched // don't batch writes that don't want to be batched
break; break;
} }

2
env/io_posix.h vendored
View File

@ -27,7 +27,7 @@
#define POSIX_FADV_RANDOM 1 /* [MC1] expect random page refs */ #define POSIX_FADV_RANDOM 1 /* [MC1] expect random page refs */
#define POSIX_FADV_SEQUENTIAL 2 /* [MC1] expect sequential page refs */ #define POSIX_FADV_SEQUENTIAL 2 /* [MC1] expect sequential page refs */
#define POSIX_FADV_WILLNEED 3 /* [MC1] will need these pages */ #define POSIX_FADV_WILLNEED 3 /* [MC1] will need these pages */
#define POSIX_FADV_DONTNEED 4 /* [MC1] dont need these pages */ #define POSIX_FADV_DONTNEED 4 /* [MC1] don't need these pages */
#endif #endif
namespace ROCKSDB_NAMESPACE { namespace ROCKSDB_NAMESPACE {

View File

@ -216,7 +216,7 @@ void DeleteScheduler::BackgroundEmptyTrash() {
const FileAndDir& fad = queue_.front(); const FileAndDir& fad = queue_.front();
std::string path_in_trash = fad.fname; std::string path_in_trash = fad.fname;
// We dont need to hold the lock while deleting the file // We don't need to hold the lock while deleting the file
mu_.Unlock(); mu_.Unlock();
uint64_t deleted_bytes = 0; uint64_t deleted_bytes = 0;
bool is_complete = true; bool is_complete = true;

View File

@ -90,7 +90,7 @@ class DeleteSchedulerTest : public testing::Test {
} }
void NewDeleteScheduler() { void NewDeleteScheduler() {
// Tests in this file are for DeleteScheduler component and dont create any // Tests in this file are for DeleteScheduler component and don't create any
// DBs, so we need to set max_trash_db_ratio to 100% (instead of default // DBs, so we need to set max_trash_db_ratio to 100% (instead of default
// 25%) // 25%)
std::shared_ptr<FileSystem> std::shared_ptr<FileSystem>
@ -306,7 +306,7 @@ TEST_F(DeleteSchedulerTest, RateLimitingMultiThreaded) {
} }
// Disable rate limiting by setting rate_bytes_per_sec_ to 0 and make sure // Disable rate limiting by setting rate_bytes_per_sec_ to 0 and make sure
// that when DeleteScheduler delete a file it delete it immediately and dont // that when DeleteScheduler delete a file it delete it immediately and don't
// move it to trash // move it to trash
TEST_F(DeleteSchedulerTest, DisableRateLimiting) { TEST_F(DeleteSchedulerTest, DisableRateLimiting) {
int bg_delete_file = 0; int bg_delete_file = 0;

View File

@ -520,8 +520,8 @@ bool DataBlockIter::ParseNextDataKey(const char* limit) {
return false; return false;
} else { } else {
if (shared == 0) { if (shared == 0) {
// If this key dont share any bytes with prev key then we dont need // If this key doesn't share any bytes with prev key then we don't need
// to decode it and can use it's address in the block directly. // to decode it and can use its address in the block directly.
key_.SetKey(Slice(p, non_shared), false /* copy */); key_.SetKey(Slice(p, non_shared), false /* copy */);
key_pinned_ = true; key_pinned_ = true;
} else { } else {
@ -592,8 +592,8 @@ bool IndexBlockIter::ParseNextIndexKey() {
return false; return false;
} }
if (shared == 0) { if (shared == 0) {
// If this key dont share any bytes with prev key then we dont need // If this key doesn't share any bytes with prev key then we don't need
// to decode it and can use it's address in the block directly. // to decode it and can use its address in the block directly.
key_.SetKey(Slice(p, non_shared), false /* copy */); key_.SetKey(Slice(p, non_shared), false /* copy */);
key_pinned_ = true; key_pinned_ = true;
} else { } else {

View File

@ -122,7 +122,7 @@ class InternalIteratorBase : public Cleanable {
// iterate_upper_bound. // iterate_upper_bound.
virtual bool MayBeOutOfUpperBound() { return true; } virtual bool MayBeOutOfUpperBound() { return true; }
// Pass the PinnedIteratorsManager to the Iterator, most Iterators dont // Pass the PinnedIteratorsManager to the Iterator, most Iterators don't
// communicate with PinnedIteratorsManager so default implementation is no-op // communicate with PinnedIteratorsManager so default implementation is no-op
// but for Iterators that need to communicate with PinnedIteratorsManager // but for Iterators that need to communicate with PinnedIteratorsManager
// they will implement this function and use the passed pointer to communicate // they will implement this function and use the passed pointer to communicate

View File

@ -150,7 +150,7 @@ struct SstFileWriter::Rep {
if (bytes_since_last_fadvise > kFadviseTrigger || closing) { if (bytes_since_last_fadvise > kFadviseTrigger || closing) {
TEST_SYNC_POINT_CALLBACK("SstFileWriter::Rep::InvalidatePageCache", TEST_SYNC_POINT_CALLBACK("SstFileWriter::Rep::InvalidatePageCache",
&(bytes_since_last_fadvise)); &(bytes_since_last_fadvise));
// Tell the OS that we dont need this file in page cache // Tell the OS that we don't need this file in page cache
file_writer->InvalidateCache(0, 0); file_writer->InvalidateCache(0, 0);
last_fadvise_size = builder->FileSize(); last_fadvise_size = builder->FileSize();
} }

View File

@ -67,7 +67,7 @@ DEFINE_string(
"Group the number of accesses per block per second using these labels. " "Group the number of accesses per block per second using these labels. "
"Possible labels are a combination of the following: cf (column family), " "Possible labels are a combination of the following: cf (column family), "
"sst, level, bt (block type), caller, block. For example, label \"cf_bt\" " "sst, level, bt (block type), caller, block. For example, label \"cf_bt\" "
"means the number of acccess per second is grouped by unique pairs of " "means the number of access per second is grouped by unique pairs of "
"\"cf_bt\". A label \"all\" contains the aggregated number of accesses per " "\"cf_bt\". A label \"all\" contains the aggregated number of accesses per "
"second across all possible labels."); "second across all possible labels.");
DEFINE_string(reuse_distance_labels, "", DEFINE_string(reuse_distance_labels, "",

View File

@ -1061,7 +1061,7 @@ DEFINE_double(keyrange_dist_d, 0.0,
"f(x)=a*exp(b*x)+c*exp(d*x)"); "f(x)=a*exp(b*x)+c*exp(d*x)");
DEFINE_int64(keyrange_num, 1, DEFINE_int64(keyrange_num, 1,
"The number of key ranges that are in the same prefix " "The number of key ranges that are in the same prefix "
"group, each prefix range will have its key acccess " "group, each prefix range will have its key access "
"distribution"); "distribution");
DEFINE_double(key_dist_a, 0.0, DEFINE_double(key_dist_a, 0.0,
"The parameter 'a' of key access distribution model " "The parameter 'a' of key access distribution model "

View File

@ -32,7 +32,7 @@ Status BlobDBImpl::DisableFileDeletions() {
} }
ROCKS_LOG_INFO(db_options_.info_log, ROCKS_LOG_INFO(db_options_.info_log,
"Disalbed blob file deletions. count: %d", count); "Disabled blob file deletions. count: %d", count);
return Status::OK(); return Status::OK();
} }

View File

@ -5575,7 +5575,7 @@ class ThreeBytewiseComparator : public Comparator {
Slice nb = Slice(b.data(), b.size() < 3 ? b.size() : 3); Slice nb = Slice(b.data(), b.size() < 3 ? b.size() : 3);
return na == nb; return na == nb;
} }
// This methods below dont seem relevant to this test. Implement them if // These methods below don't seem relevant to this test. Implement them if
// proven othersize. // proven othersize.
void FindShortestSeparator(std::string* start, void FindShortestSeparator(std::string* start,
const Slice& limit) const override { const Slice& limit) const override {

View File

@ -194,7 +194,7 @@ class WritePreparedTxnDB : public PessimisticTransactionDB {
// happen after recovery, or it could be committed and evicted by another // happen after recovery, or it could be committed and evicted by another
// commit, or never committed. // commit, or never committed.
// At this point we dont know if it was committed or it is still prepared // At this point we don't know if it was committed or it is still prepared
max_evicted_seq_ub = max_evicted_seq_.load(std::memory_order_acquire); max_evicted_seq_ub = max_evicted_seq_.load(std::memory_order_acquire);
if (UNLIKELY(max_evicted_seq_lb != max_evicted_seq_ub)) { if (UNLIKELY(max_evicted_seq_lb != max_evicted_seq_ub)) {
continue; continue;