From 4bed00a44bdc0e64f92a57109d4c747d535e002f Mon Sep 17 00:00:00 2001 From: Poornima Chozhiyath Raman Date: Wed, 8 Jul 2015 15:21:10 -0700 Subject: [PATCH] Fix function name format according to google style Summary: Change the naming style of getter and setters according to Google C++ style in compaction.h file Test Plan: Compilation success Reviewers: sdong Reviewed By: sdong Subscribers: leveldb, dhruba Differential Revision: https://reviews.facebook.net/D41265 --- db/compaction.cc | 2 +- db/compaction.h | 14 +++++++------- db/compaction_job.cc | 18 +++++++++--------- db/db_impl.cc | 8 ++++---- db/db_test.cc | 6 +++--- 5 files changed, 24 insertions(+), 24 deletions(-) diff --git a/db/compaction.cc b/db/compaction.cc index a7f2a9742..8d4b5efda 100644 --- a/db/compaction.cc +++ b/db/compaction.cc @@ -174,7 +174,7 @@ bool Compaction::IsTrivialMove() const { } return (start_level_ != output_level_ && num_input_levels() == 1 && - input(0, 0)->fd.GetPathId() == GetOutputPathId() && + input(0, 0)->fd.GetPathId() == output_path_id() && InputCompressionMatchesOutput() && TotalFileSize(grandparents_) <= max_grandparent_overlap_bytes_); } diff --git a/db/compaction.h b/db/compaction.h index d40864f39..64cd3565a 100644 --- a/db/compaction.h +++ b/db/compaction.h @@ -109,20 +109,20 @@ class Compaction { } // Maximum size of files to build during this compaction. - uint64_t MaxOutputFileSize() const { return max_output_file_size_; } + uint64_t max_output_file_size() const { return max_output_file_size_; } // What compression for output - CompressionType OutputCompressionType() const { return output_compression_; } + CompressionType output_compression() const { return output_compression_; } // Whether need to write output file to second DB path. - uint32_t GetOutputPathId() const { return output_path_id_; } + uint32_t output_path_id() const { return output_path_id_; } // Is this a trivial compaction that can be implemented by just // moving a single input file to the next level (no merging or splitting) bool IsTrivialMove() const; // If true, then the compaction can be done by simply deleting input files. - bool IsDeletionCompaction() const { + bool deletion_compaction() const { return deletion_compaction_; } @@ -150,13 +150,13 @@ class Compaction { double score() const { return score_; } // Is this compaction creating a file in the bottom most level? - bool BottomMostLevel() { return bottommost_level_; } + bool bottommost_level() { return bottommost_level_; } // Does this compaction include all sst files? - bool IsFullCompaction() { return is_full_compaction_; } + bool is_full_compaction() { return is_full_compaction_; } // Was this compaction triggered manually by the client? - bool IsManualCompaction() { return is_manual_compaction_; } + bool is_manual_compaction() { return is_manual_compaction_; } // Used when allow_trivial_move option is set in // Universal compaction. If all the input files are diff --git a/db/compaction_job.cc b/db/compaction_job.cc index 9a4b7bc91..ac07851f2 100644 --- a/db/compaction_job.cc +++ b/db/compaction_job.cc @@ -238,12 +238,12 @@ void CompactionJob::ReportStartedCompaction( // In the current design, a CompactionJob is always created // for non-trivial compaction. assert(compaction->IsTrivialMove() == false || - compaction->IsManualCompaction() == true); + compaction->is_manual_compaction() == true); ThreadStatusUtil::SetThreadOperationProperty( ThreadStatus::COMPACTION_PROP_FLAGS, - compaction->IsManualCompaction() + - (compaction->IsDeletionCompaction() << 1)); + compaction->is_manual_compaction() + + (compaction->deletion_compaction() << 1)); ThreadStatusUtil::SetThreadOperationProperty( ThreadStatus::COMPACTION_TOTAL_INPUT_BYTES, @@ -263,7 +263,7 @@ void CompactionJob::ReportStartedCompaction( if (compaction_job_stats_) { compaction_job_stats_->is_manual_compaction = - compaction->IsManualCompaction(); + compaction->is_manual_compaction(); } } @@ -298,7 +298,7 @@ void CompactionJob::Prepare() { } // Is this compaction producing files at the bottommost level? - bottommost_level_ = compact_->compaction->BottomMostLevel(); + bottommost_level_ = compact_->compaction->bottommost_level(); } Status CompactionJob::Run() { @@ -864,7 +864,7 @@ Status CompactionJob::ProcessKeyValueCompaction(int64_t* imm_micros, // Close output file if it is big enough if (compact_->builder->FileSize() >= - compact_->compaction->MaxOutputFileSize()) { + compact_->compaction->max_output_file_size()) { status = FinishCompactionOutputFile(input); if (!status.ok()) { break; @@ -1160,7 +1160,7 @@ Status CompactionJob::OpenCompactionOutputFile() { uint64_t file_number = versions_->NewFileNumber(); // Make the output file std::string fname = TableFileName(db_options_.db_paths, file_number, - compact_->compaction->GetOutputPathId()); + compact_->compaction->output_path_id()); Status s = env_->NewWritableFile(fname, &compact_->outfile, env_options_); if (!s.ok()) { @@ -1174,7 +1174,7 @@ Status CompactionJob::OpenCompactionOutputFile() { } CompactionState::Output out; out.number = file_number; - out.path_id = compact_->compaction->GetOutputPathId(); + out.path_id = compact_->compaction->output_path_id(); out.smallest.Clear(); out.largest.Clear(); out.smallest_seqno = out.largest_seqno = 0; @@ -1198,7 +1198,7 @@ Status CompactionJob::OpenCompactionOutputFile() { compact_->builder.reset(NewTableBuilder( *cfd->ioptions(), cfd->internal_comparator(), cfd->int_tbl_prop_collector_factories(), compact_->outfile.get(), - compact_->compaction->OutputCompressionType(), + compact_->compaction->output_compression(), cfd->ioptions()->compression_opts, skip_filters)); LogFlush(db_options_.info_log); return s; diff --git a/db/db_impl.cc b/db/db_impl.cc index 75535c27d..030adc8e2 100644 --- a/db/db_impl.cc +++ b/db/db_impl.cc @@ -1617,7 +1617,7 @@ Status DBImpl::CompactFilesImpl( assert(c); c->SetInputVersion(version); // deletion compaction currently not allowed in CompactFiles. - assert(!c->IsDeletionCompaction()); + assert(!c->deletion_compaction()); auto yield_callback = [&]() { return CallFlushDuringCompaction( @@ -1628,7 +1628,7 @@ Status DBImpl::CompactFilesImpl( CompactionJob compaction_job( job_context->job_id, c.get(), db_options_, env_options_, versions_.get(), &shutting_down_, log_buffer, directories_.GetDbDir(), - directories_.GetDataDir(c->GetOutputPathId()), stats_, + directories_.GetDataDir(c->output_path_id()), stats_, snapshots_.GetAll(), table_cache_, std::move(yield_callback), &event_logger_, c->mutable_cf_options()->paranoid_file_checks, dbname_, nullptr); // Here we pass a nullptr for CompactionJobStats because @@ -2504,7 +2504,7 @@ Status DBImpl::BackgroundCompaction(bool* madeProgress, JobContext* job_context, if (!c) { // Nothing to do LogToBuffer(log_buffer, "Compaction nothing to do"); - } else if (c->IsDeletionCompaction()) { + } else if (c->deletion_compaction()) { // TODO(icanadi) Do we want to honor snapshots here? i.e. not delete old // file if there is alive snapshot pointing to it assert(c->num_input_files(1) == 0); @@ -2597,7 +2597,7 @@ Status DBImpl::BackgroundCompaction(bool* madeProgress, JobContext* job_context, CompactionJob compaction_job( job_context->job_id, c.get(), db_options_, env_options_, versions_.get(), &shutting_down_, log_buffer, directories_.GetDbDir(), - directories_.GetDataDir(c->GetOutputPathId()), stats_, + directories_.GetDataDir(c->output_path_id()), stats_, snapshots_.GetAll(), table_cache_, std::move(yield_callback), &event_logger_, c->mutable_cf_options()->paranoid_file_checks, dbname_, &compaction_job_stats); diff --git a/db/db_test.cc b/db/db_test.cc index a722d6eb8..42f4bf0cb 100644 --- a/db/db_test.cc +++ b/db/db_test.cc @@ -12292,7 +12292,7 @@ TEST_F(DBTest, DynamicLevelCompressionPerLevel2) { "LevelCompactionPicker::PickCompaction:Return", [&](void* arg) { Compaction* compaction = reinterpret_cast(arg); if (compaction->output_level() == 4) { - ASSERT_TRUE(compaction->OutputCompressionType() == kLZ4Compression); + ASSERT_TRUE(compaction->output_compression() == kLZ4Compression); num_lz4.fetch_add(1); } }); @@ -12327,10 +12327,10 @@ TEST_F(DBTest, DynamicLevelCompressionPerLevel2) { "LevelCompactionPicker::PickCompaction:Return", [&](void* arg) { Compaction* compaction = reinterpret_cast(arg); if (compaction->output_level() == 4 && compaction->start_level() == 3) { - ASSERT_TRUE(compaction->OutputCompressionType() == kZlibCompression); + ASSERT_TRUE(compaction->output_compression() == kZlibCompression); num_zlib.fetch_add(1); } else { - ASSERT_TRUE(compaction->OutputCompressionType() == kLZ4Compression); + ASSERT_TRUE(compaction->output_compression() == kLZ4Compression); num_lz4.fetch_add(1); } });