|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
//
|
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
|
|
|
|
#include "rocksdb/db_bench_tool.h"
|
|
|
|
|
|
|
|
#include "db/db_impl/db_impl.h"
|
|
|
|
#include "options/options_parser.h"
|
|
|
|
#include "rocksdb/utilities/options_util.h"
|
|
|
|
#include "test_util/testharness.h"
|
|
|
|
#include "test_util/testutil.h"
|
|
|
|
#include "util/random.h"
|
|
|
|
|
|
|
|
#ifdef GFLAGS
|
|
|
|
#include "util/gflags_compat.h"
|
|
|
|
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
namespace {
|
|
|
|
static const int kMaxArgCount = 100;
|
|
|
|
static const size_t kArgBufferSize = 100000;
|
|
|
|
} // namespace
|
|
|
|
|
|
|
|
class DBBenchTest : public testing::Test {
|
|
|
|
public:
|
|
|
|
DBBenchTest() : rnd_(0xFB) {
|
|
|
|
test_path_ = test::PerThreadDBPath("db_bench_test");
|
|
|
|
Env::Default()->CreateDir(test_path_);
|
|
|
|
db_path_ = test_path_ + "/db";
|
|
|
|
wal_path_ = test_path_ + "/wal";
|
|
|
|
}
|
|
|
|
|
|
|
|
~DBBenchTest() {
|
|
|
|
// DestroyDB(db_path_, Options());
|
|
|
|
}
|
|
|
|
|
|
|
|
void ResetArgs() {
|
|
|
|
argc_ = 0;
|
|
|
|
cursor_ = 0;
|
|
|
|
memset(arg_buffer_, 0, kArgBufferSize);
|
|
|
|
}
|
|
|
|
|
|
|
|
void AppendArgs(const std::vector<std::string>& args) {
|
|
|
|
for (const auto& arg : args) {
|
|
|
|
ASSERT_LE(cursor_ + arg.size() + 1, kArgBufferSize);
|
|
|
|
ASSERT_LE(argc_ + 1, kMaxArgCount);
|
|
|
|
snprintf(arg_buffer_ + cursor_, arg.size() + 1, "%s", arg.c_str());
|
|
|
|
|
|
|
|
argv_[argc_++] = arg_buffer_ + cursor_;
|
|
|
|
cursor_ += arg.size() + 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Gets the default options for this test/db_bench.
|
|
|
|
// Note that db_bench may change some of the default option values and that
|
|
|
|
// the database might as well. The options changed by db_bench are
|
|
|
|
// specified here; the ones by the DB are set via SanitizeOptions
|
|
|
|
Options GetDefaultOptions(CompactionStyle style = kCompactionStyleLevel,
|
|
|
|
int levels = 7) const {
|
|
|
|
Options opt;
|
|
|
|
|
|
|
|
opt.create_if_missing = true;
|
|
|
|
opt.max_open_files = 256;
|
|
|
|
opt.max_background_compactions = 10;
|
|
|
|
opt.dump_malloc_stats = true; // db_bench uses a different default
|
|
|
|
opt.compaction_style = style;
|
|
|
|
opt.num_levels = levels;
|
|
|
|
opt.compression = kNoCompression;
|
|
|
|
opt.arena_block_size = 8388608;
|
|
|
|
|
|
|
|
return SanitizeOptions(db_path_, opt);
|
|
|
|
}
|
|
|
|
|
|
|
|
void RunDbBench(const std::string& options_file_name) {
|
|
|
|
AppendArgs({"./db_bench", "--benchmarks=fillseq", "--use_existing_db=0",
|
|
|
|
"--num=1000", "--compression_type=none",
|
|
|
|
std::string(std::string("--db=") + db_path_).c_str(),
|
|
|
|
std::string(std::string("--wal_dir=") + wal_path_).c_str(),
|
|
|
|
std::string(std::string("--options_file=") + options_file_name)
|
|
|
|
.c_str()});
|
|
|
|
ASSERT_EQ(0, db_bench_tool(argc(), argv()));
|
|
|
|
}
|
|
|
|
|
|
|
|
void VerifyOptions(const Options& opt) {
|
|
|
|
DBOptions loaded_db_opts;
|
|
|
|
std::vector<ColumnFamilyDescriptor> cf_descs;
|
|
|
|
ASSERT_OK(LoadLatestOptions(db_path_, Env::Default(), &loaded_db_opts,
|
|
|
|
&cf_descs));
|
|
|
|
|
|
|
|
ConfigOptions exact;
|
|
|
|
exact.input_strings_escaped = false;
|
|
|
|
exact.sanity_level = ConfigOptions::kSanityLevelExactMatch;
|
|
|
|
ASSERT_OK(RocksDBOptionsParser::VerifyDBOptions(exact, DBOptions(opt),
|
|
|
|
loaded_db_opts));
|
|
|
|
ASSERT_OK(RocksDBOptionsParser::VerifyCFOptions(
|
|
|
|
exact, ColumnFamilyOptions(opt), cf_descs[0].options));
|
|
|
|
|
|
|
|
// check with the default rocksdb options and expect failure
|
|
|
|
ASSERT_NOK(RocksDBOptionsParser::VerifyDBOptions(exact, DBOptions(),
|
|
|
|
loaded_db_opts));
|
|
|
|
ASSERT_NOK(RocksDBOptionsParser::VerifyCFOptions(
|
|
|
|
exact, ColumnFamilyOptions(), cf_descs[0].options));
|
|
|
|
}
|
|
|
|
|
|
|
|
char** argv() { return argv_; }
|
|
|
|
|
|
|
|
int argc() { return argc_; }
|
|
|
|
|
|
|
|
std::string db_path_;
|
|
|
|
std::string test_path_;
|
|
|
|
std::string wal_path_;
|
|
|
|
|
|
|
|
char arg_buffer_[kArgBufferSize];
|
|
|
|
char* argv_[kMaxArgCount];
|
|
|
|
int argc_ = 0;
|
|
|
|
int cursor_ = 0;
|
|
|
|
Random rnd_;
|
|
|
|
};
|
|
|
|
|
|
|
|
namespace {} // namespace
|
|
|
|
|
|
|
|
TEST_F(DBBenchTest, OptionsFile) {
|
|
|
|
const std::string kOptionsFileName = test_path_ + "/OPTIONS_test";
|
|
|
|
Options opt = GetDefaultOptions();
|
|
|
|
ASSERT_OK(PersistRocksDBOptions(DBOptions(opt), {"default"},
|
|
|
|
{ColumnFamilyOptions(opt)}, kOptionsFileName,
|
|
|
|
opt.env->GetFileSystem().get()));
|
|
|
|
|
|
|
|
// override the following options as db_bench will not take these
|
|
|
|
// options from the options file
|
|
|
|
opt.wal_dir = wal_path_;
|
|
|
|
|
|
|
|
RunDbBench(kOptionsFileName);
|
|
|
|
opt.delayed_write_rate = 16 * 1024 * 1024; // Set by SanitizeOptions
|
|
|
|
|
|
|
|
VerifyOptions(opt);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBBenchTest, OptionsFileUniversal) {
|
|
|
|
const std::string kOptionsFileName = test_path_ + "/OPTIONS_test";
|
|
|
|
|
|
|
|
Options opt = GetDefaultOptions(kCompactionStyleUniversal, 1);
|
|
|
|
|
|
|
|
ASSERT_OK(PersistRocksDBOptions(DBOptions(opt), {"default"},
|
|
|
|
{ColumnFamilyOptions(opt)}, kOptionsFileName,
|
|
|
|
opt.env->GetFileSystem().get()));
|
|
|
|
|
|
|
|
// override the following options as db_bench will not take these
|
|
|
|
// options from the options file
|
|
|
|
opt.wal_dir = wal_path_;
|
|
|
|
RunDbBench(kOptionsFileName);
|
|
|
|
|
|
|
|
VerifyOptions(opt);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBBenchTest, OptionsFileMultiLevelUniversal) {
|
|
|
|
const std::string kOptionsFileName = test_path_ + "/OPTIONS_test";
|
|
|
|
|
|
|
|
Options opt = GetDefaultOptions(kCompactionStyleUniversal, 12);
|
|
|
|
|
|
|
|
ASSERT_OK(PersistRocksDBOptions(DBOptions(opt), {"default"},
|
|
|
|
{ColumnFamilyOptions(opt)}, kOptionsFileName,
|
|
|
|
opt.env->GetFileSystem().get()));
|
|
|
|
|
|
|
|
// override the following options as db_bench will not take these
|
|
|
|
// options from the options file
|
|
|
|
opt.wal_dir = wal_path_;
|
|
|
|
|
|
|
|
RunDbBench(kOptionsFileName);
|
|
|
|
VerifyOptions(opt);
|
|
|
|
}
|
|
|
|
|
|
|
|
const std::string options_file_content = R"OPTIONS_FILE(
|
|
|
|
[Version]
|
|
|
|
rocksdb_version=4.3.1
|
|
|
|
options_file_version=1.1
|
|
|
|
|
|
|
|
[DBOptions]
|
|
|
|
wal_bytes_per_sync=1048576
|
|
|
|
delete_obsolete_files_period_micros=0
|
|
|
|
WAL_ttl_seconds=0
|
|
|
|
WAL_size_limit_MB=0
|
|
|
|
db_write_buffer_size=0
|
|
|
|
max_subcompactions=1
|
|
|
|
table_cache_numshardbits=4
|
|
|
|
max_open_files=-1
|
|
|
|
max_file_opening_threads=10
|
|
|
|
max_background_compactions=5
|
|
|
|
use_fsync=false
|
|
|
|
use_adaptive_mutex=false
|
|
|
|
max_total_wal_size=18446744073709551615
|
|
|
|
compaction_readahead_size=0
|
|
|
|
keep_log_file_num=10
|
|
|
|
skip_stats_update_on_db_open=false
|
|
|
|
max_manifest_file_size=18446744073709551615
|
|
|
|
db_log_dir=
|
|
|
|
writable_file_max_buffer_size=1048576
|
|
|
|
paranoid_checks=true
|
|
|
|
is_fd_close_on_exec=true
|
|
|
|
bytes_per_sync=1048576
|
|
|
|
enable_thread_tracking=true
|
|
|
|
recycle_log_file_num=0
|
|
|
|
create_missing_column_families=false
|
|
|
|
log_file_time_to_roll=0
|
|
|
|
max_background_flushes=1
|
|
|
|
create_if_missing=true
|
|
|
|
error_if_exists=false
|
|
|
|
delayed_write_rate=1048576
|
|
|
|
manifest_preallocation_size=4194304
|
|
|
|
allow_mmap_reads=false
|
|
|
|
allow_mmap_writes=false
|
|
|
|
use_direct_reads=false
|
|
|
|
use_direct_io_for_flush_and_compaction=false
|
|
|
|
stats_dump_period_sec=600
|
|
|
|
allow_fallocate=true
|
|
|
|
max_log_file_size=83886080
|
|
|
|
random_access_max_buffer_size=1048576
|
|
|
|
advise_random_on_open=true
|
|
|
|
dump_malloc_stats=true
|
|
|
|
|
|
|
|
[CFOptions "default"]
|
|
|
|
compaction_filter_factory=nullptr
|
|
|
|
table_factory=BlockBasedTable
|
|
|
|
prefix_extractor=nullptr
|
|
|
|
comparator=leveldb.BytewiseComparator
|
|
|
|
compression_per_level=
|
|
|
|
max_bytes_for_level_base=104857600
|
|
|
|
bloom_locality=0
|
|
|
|
target_file_size_base=10485760
|
|
|
|
memtable_huge_page_size=0
|
|
|
|
max_successive_merges=1000
|
|
|
|
max_sequential_skip_in_iterations=8
|
|
|
|
arena_block_size=52428800
|
|
|
|
target_file_size_multiplier=1
|
|
|
|
source_compaction_factor=1
|
|
|
|
min_write_buffer_number_to_merge=1
|
|
|
|
max_write_buffer_number=2
|
|
|
|
write_buffer_size=419430400
|
|
|
|
max_grandparent_overlap_factor=10
|
|
|
|
max_bytes_for_level_multiplier=10
|
|
|
|
memtable_factory=SkipListFactory
|
|
|
|
compression=kNoCompression
|
|
|
|
min_partial_merge_operands=2
|
|
|
|
level0_stop_writes_trigger=100
|
|
|
|
num_levels=1
|
|
|
|
level0_slowdown_writes_trigger=50
|
|
|
|
level0_file_num_compaction_trigger=10
|
|
|
|
expanded_compaction_factor=25
|
|
|
|
max_write_buffer_number_to_maintain=0
|
Refactor trimming logic for immutable memtables (#5022)
Summary:
MyRocks currently sets `max_write_buffer_number_to_maintain` in order to maintain enough history for transaction conflict checking. The effectiveness of this approach depends on the size of memtables. When memtables are small, it may not keep enough history; when memtables are large, this may consume too much memory.
We are proposing a new way to configure memtable list history: by limiting the memory usage of immutable memtables. The new option is `max_write_buffer_size_to_maintain` and it will take precedence over the old `max_write_buffer_number_to_maintain` if they are both set to non-zero values. The new option accounts for the total memory usage of flushed immutable memtables and mutable memtable. When the total usage exceeds the limit, RocksDB may start dropping immutable memtables (which is also called trimming history), starting from the oldest one.
The semantics of the old option actually works both as an upper bound and lower bound. History trimming will start if number of immutable memtables exceeds the limit, but it will never go below (limit-1) due to history trimming.
In order the mimic the behavior with the new option, history trimming will stop if dropping the next immutable memtable causes the total memory usage go below the size limit. For example, assuming the size limit is set to 64MB, and there are 3 immutable memtables with sizes of 20, 30, 30. Although the total memory usage is 80MB > 64MB, dropping the oldest memtable will reduce the memory usage to 60MB < 64MB, so in this case no memtable will be dropped.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5022
Differential Revision: D14394062
Pulled By: miasantreble
fbshipit-source-id: 60457a509c6af89d0993f988c9b5c2aa9e45f5c5
5 years ago
|
|
|
max_write_buffer_size_to_maintain=0
|
|
|
|
verify_checksums_in_compaction=true
|
|
|
|
merge_operator=nullptr
|
|
|
|
memtable_prefix_bloom_bits=0
|
|
|
|
memtable_whole_key_filtering=true
|
|
|
|
paranoid_file_checks=false
|
|
|
|
inplace_update_num_locks=10000
|
|
|
|
optimize_filters_for_hits=false
|
|
|
|
level_compaction_dynamic_level_bytes=false
|
|
|
|
inplace_update_support=false
|
|
|
|
compaction_style=kCompactionStyleUniversal
|
|
|
|
memtable_prefix_bloom_probes=6
|
|
|
|
filter_deletes=false
|
|
|
|
hard_pending_compaction_bytes_limit=0
|
|
|
|
disable_auto_compactions=false
|
|
|
|
compaction_measure_io_stats=false
|
|
|
|
enable_blob_files=true
|
|
|
|
min_blob_size=16
|
|
|
|
blob_file_size=10485760
|
|
|
|
blob_compression_type=kNoCompression
|
|
|
|
enable_blob_garbage_collection=true
|
Make it possible to force the garbage collection of the oldest blob files (#8994)
Summary:
The current BlobDB garbage collection logic works by relocating the valid
blobs from the oldest blob files as they are encountered during compaction,
and cleaning up blob files once they contain nothing but garbage. However,
with sufficiently skewed workloads, it is theoretically possible to end up in a
situation when few or no compactions get scheduled for the SST files that contain
references to the oldest blob files, which can lead to increased space amp due
to the lack of GC.
In order to efficiently handle such workloads, the patch adds a new BlobDB
configuration option called `blob_garbage_collection_force_threshold`,
which signals to BlobDB to schedule targeted compactions for the SST files
that keep alive the oldest batch of blob files if the overall ratio of garbage in
the given blob files meets the threshold *and* all the given blob files are
eligible for GC based on `blob_garbage_collection_age_cutoff`. (For example,
if the new option is set to 0.9, targeted compactions will get scheduled if the
sum of garbage bytes meets or exceeds 90% of the sum of total bytes in the
oldest blob files, assuming all affected blob files are below the age-based cutoff.)
The net result of these targeted compactions is that the valid blobs in the oldest
blob files are relocated and the oldest blob files themselves cleaned up (since
*all* SST files that rely on them get compacted away).
These targeted compactions are similar to periodic compactions in the sense
that they force certain SST files that otherwise would not get picked up to undergo
compaction and also in the sense that instead of merging files from multiple levels,
they target a single file. (Note: such compactions might still include neighboring files
from the same level due to the need of having a "clean cut" boundary but they never
include any files from any other level.)
This functionality is currently only supported with the leveled compaction style
and is inactive by default (since the default value is set to 1.0, i.e. 100%).
Pull Request resolved: https://github.com/facebook/rocksdb/pull/8994
Test Plan: Ran `make check` and tested using `db_bench` and the stress/crash tests.
Reviewed By: riversand963
Differential Revision: D31489850
Pulled By: ltamasi
fbshipit-source-id: 44057d511726a0e2a03c5d9313d7511b3f0c4eab
3 years ago
|
|
|
blob_garbage_collection_age_cutoff=0.5
|
|
|
|
blob_garbage_collection_force_threshold=0.75
|
|
|
|
blob_compaction_readahead_size=262144
|
|
|
|
|
|
|
|
[TableOptions/BlockBasedTable "default"]
|
|
|
|
format_version=0
|
|
|
|
skip_table_builder_flush=false
|
|
|
|
cache_index_and_filter_blocks=false
|
|
|
|
flush_block_policy_factory=FlushBlockBySizePolicyFactory
|
|
|
|
index_type=kBinarySearch
|
|
|
|
whole_key_filtering=true
|
|
|
|
checksum=kCRC32c
|
|
|
|
no_block_cache=false
|
|
|
|
block_size=32768
|
|
|
|
block_size_deviation=10
|
|
|
|
block_restart_interval=16
|
|
|
|
filter_policy=rocksdb.BuiltinBloomFilter
|
|
|
|
)OPTIONS_FILE";
|
|
|
|
|
|
|
|
TEST_F(DBBenchTest, OptionsFileFromFile) {
|
|
|
|
const std::string kOptionsFileName = test_path_ + "/OPTIONS_flash";
|
|
|
|
std::unique_ptr<WritableFile> writable;
|
|
|
|
ASSERT_OK(Env::Default()->NewWritableFile(kOptionsFileName, &writable,
|
|
|
|
EnvOptions()));
|
|
|
|
ASSERT_OK(writable->Append(options_file_content));
|
|
|
|
ASSERT_OK(writable->Close());
|
|
|
|
|
|
|
|
DBOptions db_opt;
|
|
|
|
std::vector<ColumnFamilyDescriptor> cf_descs;
|
|
|
|
ASSERT_OK(LoadOptionsFromFile(kOptionsFileName, Env::Default(), &db_opt,
|
|
|
|
&cf_descs));
|
|
|
|
Options opt(db_opt, cf_descs[0].options);
|
|
|
|
opt.create_if_missing = true;
|
|
|
|
|
|
|
|
// override the following options as db_bench will not take these
|
|
|
|
// options from the options file
|
|
|
|
opt.wal_dir = wal_path_;
|
|
|
|
|
|
|
|
RunDbBench(kOptionsFileName);
|
|
|
|
|
|
|
|
VerifyOptions(SanitizeOptions(db_path_, opt));
|
|
|
|
}
|
|
|
|
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
|
|
GFLAGS_NAMESPACE::ParseCommandLineFlags(&argc, &argv, true);
|
|
|
|
return RUN_ALL_TESTS();
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
|
|
printf("Skip db_bench_tool_test as the required library GFLAG is missing.");
|
|
|
|
}
|
|
|
|
#endif // #ifdef GFLAGS
|