|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
//
|
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include <map>
|
|
|
|
#include <string>
|
|
|
|
#include <vector>
|
|
|
|
#include "db/db_impl/db_impl.h"
|
|
|
|
#include "db/version_set.h"
|
|
|
|
#include "db/write_batch_internal.h"
|
|
|
|
#include "file/filename.h"
|
|
|
|
#include "rocksdb/db.h"
|
|
|
|
#include "rocksdb/env.h"
|
|
|
|
#include "rocksdb/transaction_log.h"
|
|
|
|
#include "test_util/sync_point.h"
|
|
|
|
#include "test_util/testharness.h"
|
|
|
|
#include "test_util/testutil.h"
|
|
|
|
#include "util/string_util.h"
|
|
|
|
|
|
|
|
namespace rocksdb {
|
|
|
|
|
|
|
|
class DeleteFileTest : public testing::Test {
|
|
|
|
public:
|
|
|
|
std::string dbname_;
|
|
|
|
Options options_;
|
|
|
|
DB* db_;
|
|
|
|
Env* env_;
|
|
|
|
int numlevels_;
|
|
|
|
|
|
|
|
DeleteFileTest() {
|
|
|
|
db_ = nullptr;
|
|
|
|
env_ = Env::Default();
|
|
|
|
options_.delete_obsolete_files_period_micros = 0; // always do full purge
|
|
|
|
options_.enable_thread_tracking = true;
|
|
|
|
options_.write_buffer_size = 1024*1024*1000;
|
|
|
|
options_.target_file_size_base = 1024*1024*1000;
|
|
|
|
options_.max_bytes_for_level_base = 1024*1024*1000;
|
|
|
|
options_.WAL_ttl_seconds = 300; // Used to test log files
|
|
|
|
options_.WAL_size_limit_MB = 1024; // Used to test log files
|
|
|
|
dbname_ = test::PerThreadDBPath("deletefile_test");
|
|
|
|
options_.wal_dir = dbname_ + "/wal_files";
|
|
|
|
|
|
|
|
// clean up all the files that might have been there before
|
|
|
|
std::vector<std::string> old_files;
|
|
|
|
env_->GetChildren(dbname_, &old_files);
|
|
|
|
for (auto file : old_files) {
|
|
|
|
env_->DeleteFile(dbname_ + "/" + file);
|
|
|
|
}
|
|
|
|
env_->GetChildren(options_.wal_dir, &old_files);
|
|
|
|
for (auto file : old_files) {
|
|
|
|
env_->DeleteFile(options_.wal_dir + "/" + file);
|
|
|
|
}
|
|
|
|
|
|
|
|
DestroyDB(dbname_, options_);
|
|
|
|
numlevels_ = 7;
|
rocksdb: Replace ASSERT* with EXPECT* in functions that does not return void value
Summary:
gtest does not use exceptions to fail a unit test by design, and `ASSERT*`s are implemented using `return`. As a consequence we cannot use `ASSERT*` in a function that does not return `void` value ([[ https://code.google.com/p/googletest/wiki/AdvancedGuide#Assertion_Placement | 1]]), and have to fix our existing code. This diff does this in a generic way, with no manual changes.
In order to detect all existing `ASSERT*` that are used in functions that doesn't return void value, I change the code to generate compile errors for such cases.
In `util/testharness.h` I defined `EXPECT*` assertions, the same way as `ASSERT*`, and redefined `ASSERT*` to return `void`. Then executed:
```lang=bash
% USE_CLANG=1 make all -j55 -k 2> build.log
% perl -naF: -e 'print "-- -number=".$F[1]." ".$F[0]."\n" if /: error:/' \
build.log | xargs -L 1 perl -spi -e 's/ASSERT/EXPECT/g if $. == $number'
% make format
```
After that I reverted back change to `ASSERT*` in `util/testharness.h`. But preserved introduced `EXPECT*`, which is the same as `ASSERT*`. This will be deleted once switched to gtest.
This diff is independent and contains manual changes only in `util/testharness.h`.
Test Plan:
Make sure all tests are passing.
```lang=bash
% USE_CLANG=1 make check
```
Reviewers: igor, lgalanis, sdong, yufei.zhu, rven, meyering
Reviewed By: meyering
Subscribers: dhruba, leveldb
Differential Revision: https://reviews.facebook.net/D33333
10 years ago
|
|
|
EXPECT_OK(ReopenDB(true));
|
|
|
|
}
|
|
|
|
|
|
|
|
Status ReopenDB(bool create) {
|
|
|
|
delete db_;
|
|
|
|
if (create) {
|
|
|
|
DestroyDB(dbname_, options_);
|
|
|
|
}
|
|
|
|
db_ = nullptr;
|
|
|
|
options_.create_if_missing = create;
|
|
|
|
Status s = DB::Open(options_, dbname_, &db_);
|
|
|
|
assert(db_);
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
void CloseDB() {
|
|
|
|
delete db_;
|
|
|
|
db_ = nullptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
void AddKeys(int numkeys, int startkey = 0) {
|
|
|
|
WriteOptions options;
|
|
|
|
options.sync = false;
|
|
|
|
ReadOptions roptions;
|
|
|
|
for (int i = startkey; i < (numkeys + startkey) ; i++) {
|
|
|
|
std::string temp = ToString(i);
|
|
|
|
Slice key(temp);
|
|
|
|
Slice value(temp);
|
|
|
|
ASSERT_OK(db_->Put(options, key, value));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
int numKeysInLevels(
|
|
|
|
std::vector<LiveFileMetaData> &metadata,
|
|
|
|
std::vector<int> *keysperlevel = nullptr) {
|
|
|
|
|
|
|
|
if (keysperlevel != nullptr) {
|
|
|
|
keysperlevel->resize(numlevels_);
|
|
|
|
}
|
|
|
|
|
|
|
|
int numKeys = 0;
|
|
|
|
for (size_t i = 0; i < metadata.size(); i++) {
|
|
|
|
int startkey = atoi(metadata[i].smallestkey.c_str());
|
|
|
|
int endkey = atoi(metadata[i].largestkey.c_str());
|
|
|
|
int numkeysinfile = (endkey - startkey + 1);
|
|
|
|
numKeys += numkeysinfile;
|
|
|
|
if (keysperlevel != nullptr) {
|
|
|
|
(*keysperlevel)[(int)metadata[i].level] += numkeysinfile;
|
|
|
|
}
|
|
|
|
fprintf(stderr, "level %d name %s smallest %s largest %s\n",
|
|
|
|
metadata[i].level, metadata[i].name.c_str(),
|
|
|
|
metadata[i].smallestkey.c_str(),
|
|
|
|
metadata[i].largestkey.c_str());
|
|
|
|
}
|
|
|
|
return numKeys;
|
|
|
|
}
|
|
|
|
|
|
|
|
void CreateTwoLevels() {
|
|
|
|
AddKeys(50000, 10000);
|
|
|
|
DBImpl* dbi = reinterpret_cast<DBImpl*>(db_);
|
|
|
|
ASSERT_OK(dbi->TEST_FlushMemTable());
|
|
|
|
ASSERT_OK(dbi->TEST_WaitForFlushMemTable());
|
|
|
|
for (int i = 0; i < 2; ++i) {
|
|
|
|
ASSERT_OK(dbi->TEST_CompactRange(i, nullptr, nullptr));
|
|
|
|
}
|
|
|
|
|
|
|
|
AddKeys(50000, 10000);
|
|
|
|
ASSERT_OK(dbi->TEST_FlushMemTable());
|
|
|
|
ASSERT_OK(dbi->TEST_WaitForFlushMemTable());
|
|
|
|
ASSERT_OK(dbi->TEST_CompactRange(0, nullptr, nullptr));
|
|
|
|
}
|
|
|
|
|
|
|
|
void CheckFileTypeCounts(std::string& dir,
|
|
|
|
int required_log,
|
|
|
|
int required_sst,
|
|
|
|
int required_manifest) {
|
|
|
|
std::vector<std::string> filenames;
|
|
|
|
env_->GetChildren(dir, &filenames);
|
|
|
|
|
|
|
|
int log_cnt = 0, sst_cnt = 0, manifest_cnt = 0;
|
|
|
|
for (auto file : filenames) {
|
|
|
|
uint64_t number;
|
|
|
|
FileType type;
|
|
|
|
if (ParseFileName(file, &number, &type)) {
|
|
|
|
log_cnt += (type == kLogFile);
|
|
|
|
sst_cnt += (type == kTableFile);
|
|
|
|
manifest_cnt += (type == kDescriptorFile);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ASSERT_EQ(required_log, log_cnt);
|
|
|
|
ASSERT_EQ(required_sst, sst_cnt);
|
|
|
|
ASSERT_EQ(required_manifest, manifest_cnt);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void DoSleep(void* arg) {
|
|
|
|
auto test = reinterpret_cast<DeleteFileTest*>(arg);
|
|
|
|
test->env_->SleepForMicroseconds(2 * 1000 * 1000);
|
|
|
|
}
|
|
|
|
|
|
|
|
// An empty job to guard all jobs are processed
|
|
|
|
static void GuardFinish(void* /*arg*/) {
|
|
|
|
TEST_SYNC_POINT("DeleteFileTest::GuardFinish");
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_F(DeleteFileTest, AddKeysAndQueryLevels) {
|
|
|
|
CreateTwoLevels();
|
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
|
|
db_->GetLiveFilesMetaData(&metadata);
|
|
|
|
|
|
|
|
std::string level1file = "";
|
|
|
|
int level1keycount = 0;
|
|
|
|
std::string level2file = "";
|
|
|
|
int level2keycount = 0;
|
|
|
|
int level1index = 0;
|
|
|
|
int level2index = 1;
|
|
|
|
|
|
|
|
ASSERT_EQ((int)metadata.size(), 2);
|
|
|
|
if (metadata[0].level == 2) {
|
|
|
|
level1index = 1;
|
|
|
|
level2index = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
level1file = metadata[level1index].name;
|
|
|
|
int startkey = atoi(metadata[level1index].smallestkey.c_str());
|
|
|
|
int endkey = atoi(metadata[level1index].largestkey.c_str());
|
|
|
|
level1keycount = (endkey - startkey + 1);
|
|
|
|
level2file = metadata[level2index].name;
|
|
|
|
startkey = atoi(metadata[level2index].smallestkey.c_str());
|
|
|
|
endkey = atoi(metadata[level2index].largestkey.c_str());
|
|
|
|
level2keycount = (endkey - startkey + 1);
|
|
|
|
|
|
|
|
// COntrolled setup. Levels 1 and 2 should both have 50K files.
|
|
|
|
// This is a little fragile as it depends on the current
|
|
|
|
// compaction heuristics.
|
|
|
|
ASSERT_EQ(level1keycount, 50000);
|
|
|
|
ASSERT_EQ(level2keycount, 50000);
|
|
|
|
|
|
|
|
Status status = db_->DeleteFile("0.sst");
|
|
|
|
ASSERT_TRUE(status.IsInvalidArgument());
|
|
|
|
|
|
|
|
// intermediate level files cannot be deleted.
|
|
|
|
status = db_->DeleteFile(level1file);
|
|
|
|
ASSERT_TRUE(status.IsInvalidArgument());
|
|
|
|
|
|
|
|
// Lowest level file deletion should succeed.
|
|
|
|
ASSERT_OK(db_->DeleteFile(level2file));
|
|
|
|
|
|
|
|
CloseDB();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DeleteFileTest, PurgeObsoleteFilesTest) {
|
|
|
|
CreateTwoLevels();
|
|
|
|
// there should be only one (empty) log file because CreateTwoLevels()
|
|
|
|
// flushes the memtables to disk
|
|
|
|
CheckFileTypeCounts(options_.wal_dir, 1, 0, 0);
|
|
|
|
// 2 ssts, 1 manifest
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 2, 1);
|
|
|
|
std::string first("0"), last("999999");
|
|
|
|
CompactRangeOptions compact_options;
|
|
|
|
compact_options.change_level = true;
|
|
|
|
compact_options.target_level = 2;
|
|
|
|
Slice first_slice(first), last_slice(last);
|
|
|
|
db_->CompactRange(compact_options, &first_slice, &last_slice);
|
|
|
|
// 1 sst after compaction
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
// this time, we keep an iterator alive
|
|
|
|
ReopenDB(true);
|
|
|
|
Iterator *itr = nullptr;
|
|
|
|
CreateTwoLevels();
|
|
|
|
itr = db_->NewIterator(ReadOptions());
|
|
|
|
db_->CompactRange(compact_options, &first_slice, &last_slice);
|
|
|
|
// 3 sst after compaction with live iterator
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
|
|
delete itr;
|
|
|
|
// 1 sst after iterator deletion
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
CloseDB();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DeleteFileTest, BackgroundPurgeIteratorTest) {
|
|
|
|
std::string first("0"), last("999999");
|
|
|
|
CompactRangeOptions compact_options;
|
|
|
|
compact_options.change_level = true;
|
|
|
|
compact_options.target_level = 2;
|
|
|
|
Slice first_slice(first), last_slice(last);
|
|
|
|
|
|
|
|
// We keep an iterator alive
|
|
|
|
Iterator* itr = nullptr;
|
|
|
|
CreateTwoLevels();
|
|
|
|
ReadOptions options;
|
|
|
|
options.background_purge_on_iterator_cleanup = true;
|
|
|
|
itr = db_->NewIterator(options);
|
|
|
|
db_->CompactRange(compact_options, &first_slice, &last_slice);
|
|
|
|
// 3 sst after compaction with live iterator
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
|
|
test::SleepingBackgroundTask sleeping_task_before;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_before, Env::Priority::HIGH);
|
|
|
|
delete itr;
|
|
|
|
test::SleepingBackgroundTask sleeping_task_after;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_after, Env::Priority::HIGH);
|
|
|
|
|
|
|
|
// Make sure no purges are executed foreground
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
|
|
sleeping_task_before.WakeUp();
|
|
|
|
sleeping_task_before.WaitUntilDone();
|
|
|
|
|
|
|
|
// Make sure all background purges are executed
|
|
|
|
sleeping_task_after.WakeUp();
|
|
|
|
sleeping_task_after.WaitUntilDone();
|
|
|
|
// 1 sst after iterator deletion
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
CloseDB();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DeleteFileTest, BackgroundPurgeCFDropTest) {
|
|
|
|
auto do_test = [&](bool bg_purge) {
|
|
|
|
ColumnFamilyOptions co;
|
Refactor trimming logic for immutable memtables (#5022)
Summary:
MyRocks currently sets `max_write_buffer_number_to_maintain` in order to maintain enough history for transaction conflict checking. The effectiveness of this approach depends on the size of memtables. When memtables are small, it may not keep enough history; when memtables are large, this may consume too much memory.
We are proposing a new way to configure memtable list history: by limiting the memory usage of immutable memtables. The new option is `max_write_buffer_size_to_maintain` and it will take precedence over the old `max_write_buffer_number_to_maintain` if they are both set to non-zero values. The new option accounts for the total memory usage of flushed immutable memtables and mutable memtable. When the total usage exceeds the limit, RocksDB may start dropping immutable memtables (which is also called trimming history), starting from the oldest one.
The semantics of the old option actually works both as an upper bound and lower bound. History trimming will start if number of immutable memtables exceeds the limit, but it will never go below (limit-1) due to history trimming.
In order the mimic the behavior with the new option, history trimming will stop if dropping the next immutable memtable causes the total memory usage go below the size limit. For example, assuming the size limit is set to 64MB, and there are 3 immutable memtables with sizes of 20, 30, 30. Although the total memory usage is 80MB > 64MB, dropping the oldest memtable will reduce the memory usage to 60MB < 64MB, so in this case no memtable will be dropped.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5022
Differential Revision: D14394062
Pulled By: miasantreble
fbshipit-source-id: 60457a509c6af89d0993f988c9b5c2aa9e45f5c5
5 years ago
|
|
|
co.max_write_buffer_size_to_maintain =
|
|
|
|
static_cast<int64_t>(co.write_buffer_size);
|
|
|
|
WriteOptions wo;
|
|
|
|
FlushOptions fo;
|
|
|
|
ColumnFamilyHandle* cfh = nullptr;
|
|
|
|
|
|
|
|
ASSERT_OK(db_->CreateColumnFamily(co, "dropme", &cfh));
|
|
|
|
|
|
|
|
ASSERT_OK(db_->Put(wo, cfh, "pika", "chu"));
|
|
|
|
ASSERT_OK(db_->Flush(fo, cfh));
|
|
|
|
// Expect 1 sst file.
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
ASSERT_OK(db_->DropColumnFamily(cfh));
|
|
|
|
// Still 1 file, it won't be deleted while ColumnFamilyHandle is alive.
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
delete cfh;
|
|
|
|
test::SleepingBackgroundTask sleeping_task_after;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_after, Env::Priority::HIGH);
|
|
|
|
// If background purge is enabled, the file should still be there.
|
|
|
|
CheckFileTypeCounts(dbname_, 0, bg_purge ? 1 : 0, 1);
|
|
|
|
TEST_SYNC_POINT("DeleteFileTest::BackgroundPurgeCFDropTest:1");
|
|
|
|
|
|
|
|
// Execute background purges.
|
|
|
|
sleeping_task_after.WakeUp();
|
|
|
|
sleeping_task_after.WaitUntilDone();
|
|
|
|
// The file should have been deleted.
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 0, 1);
|
|
|
|
};
|
|
|
|
|
|
|
|
{
|
|
|
|
SCOPED_TRACE("avoid_unnecessary_blocking_io = false");
|
|
|
|
do_test(false);
|
|
|
|
}
|
|
|
|
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
SyncPoint::GetInstance()->LoadDependency(
|
|
|
|
{{"DeleteFileTest::BackgroundPurgeCFDropTest:1",
|
|
|
|
"DBImpl::BGWorkPurge:start"}});
|
|
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
options_.avoid_unnecessary_blocking_io = true;
|
|
|
|
ASSERT_OK(ReopenDB(false));
|
|
|
|
{
|
|
|
|
SCOPED_TRACE("avoid_unnecessary_blocking_io = true");
|
|
|
|
do_test(true);
|
|
|
|
}
|
|
|
|
|
|
|
|
CloseDB();
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
}
|
|
|
|
|
|
|
|
// This test is to reproduce a bug that read invalid ReadOption in iterator
|
|
|
|
// cleanup function
|
|
|
|
TEST_F(DeleteFileTest, BackgroundPurgeCopyOptions) {
|
|
|
|
std::string first("0"), last("999999");
|
|
|
|
CompactRangeOptions compact_options;
|
|
|
|
compact_options.change_level = true;
|
|
|
|
compact_options.target_level = 2;
|
|
|
|
Slice first_slice(first), last_slice(last);
|
|
|
|
|
|
|
|
// We keep an iterator alive
|
|
|
|
Iterator* itr = nullptr;
|
|
|
|
CreateTwoLevels();
|
|
|
|
ReadOptions* options = new ReadOptions();
|
|
|
|
options->background_purge_on_iterator_cleanup = true;
|
|
|
|
itr = db_->NewIterator(*options);
|
|
|
|
// ReadOptions is deleted, but iterator cleanup function should not be
|
|
|
|
// affected
|
|
|
|
delete options;
|
|
|
|
|
|
|
|
db_->CompactRange(compact_options, &first_slice, &last_slice);
|
|
|
|
// 3 sst after compaction with live iterator
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
|
|
delete itr;
|
|
|
|
|
|
|
|
test::SleepingBackgroundTask sleeping_task_after;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_after, Env::Priority::HIGH);
|
|
|
|
|
|
|
|
// Make sure all background purges are executed
|
|
|
|
sleeping_task_after.WakeUp();
|
|
|
|
sleeping_task_after.WaitUntilDone();
|
|
|
|
// 1 sst after iterator deletion
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
CloseDB();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DeleteFileTest, BackgroundPurgeTestMultipleJobs) {
|
|
|
|
std::string first("0"), last("999999");
|
|
|
|
CompactRangeOptions compact_options;
|
|
|
|
compact_options.change_level = true;
|
|
|
|
compact_options.target_level = 2;
|
|
|
|
Slice first_slice(first), last_slice(last);
|
|
|
|
|
|
|
|
// We keep an iterator alive
|
|
|
|
CreateTwoLevels();
|
|
|
|
ReadOptions options;
|
|
|
|
options.background_purge_on_iterator_cleanup = true;
|
|
|
|
Iterator* itr1 = db_->NewIterator(options);
|
|
|
|
CreateTwoLevels();
|
|
|
|
Iterator* itr2 = db_->NewIterator(options);
|
|
|
|
db_->CompactRange(compact_options, &first_slice, &last_slice);
|
|
|
|
// 5 sst files after 2 compactions with 2 live iterators
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 5, 1);
|
|
|
|
|
|
|
|
// ~DBImpl should wait until all BGWorkPurge are finished
|
|
|
|
rocksdb::SyncPoint::GetInstance()->LoadDependency(
|
|
|
|
{{"DBImpl::~DBImpl:WaitJob", "DBImpl::BGWorkPurge"},
|
|
|
|
{"DeleteFileTest::GuardFinish",
|
|
|
|
"DeleteFileTest::BackgroundPurgeTestMultipleJobs:DBClose"}});
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
|
|
|
delete itr1;
|
|
|
|
env_->Schedule(&DeleteFileTest::DoSleep, this, Env::Priority::HIGH);
|
|
|
|
delete itr2;
|
|
|
|
env_->Schedule(&DeleteFileTest::GuardFinish, nullptr, Env::Priority::HIGH);
|
|
|
|
CloseDB();
|
|
|
|
|
|
|
|
TEST_SYNC_POINT("DeleteFileTest::BackgroundPurgeTestMultipleJobs:DBClose");
|
|
|
|
// 1 sst after iterator deletion
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DeleteFileTest, DeleteFileWithIterator) {
|
|
|
|
CreateTwoLevels();
|
|
|
|
ReadOptions options;
|
|
|
|
Iterator* it = db_->NewIterator(options);
|
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
|
|
db_->GetLiveFilesMetaData(&metadata);
|
|
|
|
|
|
|
|
std::string level2file = "";
|
|
|
|
|
|
|
|
ASSERT_EQ((int)metadata.size(), 2);
|
|
|
|
if (metadata[0].level == 1) {
|
|
|
|
level2file = metadata[1].name;
|
|
|
|
} else {
|
|
|
|
level2file = metadata[0].name;
|
|
|
|
}
|
|
|
|
|
|
|
|
Status status = db_->DeleteFile(level2file);
|
|
|
|
fprintf(stdout, "Deletion status %s: %s\n",
|
|
|
|
level2file.c_str(), status.ToString().c_str());
|
|
|
|
ASSERT_TRUE(status.ok());
|
|
|
|
it->SeekToFirst();
|
|
|
|
int numKeysIterated = 0;
|
|
|
|
while(it->Valid()) {
|
|
|
|
numKeysIterated++;
|
|
|
|
it->Next();
|
|
|
|
}
|
|
|
|
ASSERT_EQ(numKeysIterated, 50000);
|
|
|
|
delete it;
|
|
|
|
CloseDB();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DeleteFileTest, DeleteLogFiles) {
|
|
|
|
AddKeys(10, 0);
|
|
|
|
VectorLogPtr logfiles;
|
|
|
|
db_->GetSortedWalFiles(logfiles);
|
|
|
|
ASSERT_GT(logfiles.size(), 0UL);
|
|
|
|
// Take the last log file which is expected to be alive and try to delete it
|
|
|
|
// Should not succeed because live logs are not allowed to be deleted
|
|
|
|
std::unique_ptr<LogFile> alive_log = std::move(logfiles.back());
|
|
|
|
ASSERT_EQ(alive_log->Type(), kAliveLogFile);
|
|
|
|
ASSERT_OK(env_->FileExists(options_.wal_dir + "/" + alive_log->PathName()));
|
|
|
|
fprintf(stdout, "Deleting alive log file %s\n",
|
|
|
|
alive_log->PathName().c_str());
|
|
|
|
ASSERT_TRUE(!db_->DeleteFile(alive_log->PathName()).ok());
|
|
|
|
ASSERT_OK(env_->FileExists(options_.wal_dir + "/" + alive_log->PathName()));
|
|
|
|
logfiles.clear();
|
|
|
|
|
|
|
|
// Call Flush to bring about a new working log file and add more keys
|
|
|
|
// Call Flush again to flush out memtable and move alive log to archived log
|
|
|
|
// and try to delete the archived log file
|
|
|
|
FlushOptions fopts;
|
|
|
|
db_->Flush(fopts);
|
|
|
|
AddKeys(10, 0);
|
|
|
|
db_->Flush(fopts);
|
|
|
|
db_->GetSortedWalFiles(logfiles);
|
|
|
|
ASSERT_GT(logfiles.size(), 0UL);
|
|
|
|
std::unique_ptr<LogFile> archived_log = std::move(logfiles.front());
|
|
|
|
ASSERT_EQ(archived_log->Type(), kArchivedLogFile);
|
|
|
|
ASSERT_OK(
|
|
|
|
env_->FileExists(options_.wal_dir + "/" + archived_log->PathName()));
|
|
|
|
fprintf(stdout, "Deleting archived log file %s\n",
|
|
|
|
archived_log->PathName().c_str());
|
|
|
|
ASSERT_OK(db_->DeleteFile(archived_log->PathName()));
|
|
|
|
ASSERT_EQ(Status::NotFound(), env_->FileExists(options_.wal_dir + "/" +
|
|
|
|
archived_log->PathName()));
|
|
|
|
CloseDB();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DeleteFileTest, DeleteNonDefaultColumnFamily) {
|
|
|
|
CloseDB();
|
|
|
|
DBOptions db_options;
|
|
|
|
db_options.create_if_missing = true;
|
|
|
|
db_options.create_missing_column_families = true;
|
|
|
|
std::vector<ColumnFamilyDescriptor> column_families;
|
|
|
|
column_families.emplace_back();
|
|
|
|
column_families.emplace_back("new_cf", ColumnFamilyOptions());
|
|
|
|
|
|
|
|
std::vector<rocksdb::ColumnFamilyHandle*> handles;
|
|
|
|
rocksdb::DB* db;
|
|
|
|
ASSERT_OK(DB::Open(db_options, dbname_, column_families, &handles, &db));
|
|
|
|
|
|
|
|
Random rnd(5);
|
|
|
|
for (int i = 0; i < 1000; ++i) {
|
|
|
|
ASSERT_OK(db->Put(WriteOptions(), handles[1], test::RandomKey(&rnd, 10),
|
|
|
|
test::RandomKey(&rnd, 10)));
|
|
|
|
}
|
|
|
|
ASSERT_OK(db->Flush(FlushOptions(), handles[1]));
|
|
|
|
for (int i = 0; i < 1000; ++i) {
|
|
|
|
ASSERT_OK(db->Put(WriteOptions(), handles[1], test::RandomKey(&rnd, 10),
|
|
|
|
test::RandomKey(&rnd, 10)));
|
|
|
|
}
|
|
|
|
ASSERT_OK(db->Flush(FlushOptions(), handles[1]));
|
|
|
|
|
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
|
|
db->GetLiveFilesMetaData(&metadata);
|
|
|
|
ASSERT_EQ(2U, metadata.size());
|
|
|
|
ASSERT_EQ("new_cf", metadata[0].column_family_name);
|
|
|
|
ASSERT_EQ("new_cf", metadata[1].column_family_name);
|
|
|
|
auto old_file = metadata[0].smallest_seqno < metadata[1].smallest_seqno
|
|
|
|
? metadata[0].name
|
|
|
|
: metadata[1].name;
|
|
|
|
auto new_file = metadata[0].smallest_seqno > metadata[1].smallest_seqno
|
|
|
|
? metadata[0].name
|
|
|
|
: metadata[1].name;
|
|
|
|
ASSERT_TRUE(db->DeleteFile(new_file).IsInvalidArgument());
|
|
|
|
ASSERT_OK(db->DeleteFile(old_file));
|
|
|
|
|
|
|
|
{
|
|
|
|
std::unique_ptr<Iterator> itr(db->NewIterator(ReadOptions(), handles[1]));
|
|
|
|
int count = 0;
|
|
|
|
for (itr->SeekToFirst(); itr->Valid(); itr->Next()) {
|
|
|
|
ASSERT_OK(itr->status());
|
|
|
|
++count;
|
|
|
|
}
|
|
|
|
ASSERT_EQ(count, 1000);
|
|
|
|
}
|
|
|
|
|
|
|
|
delete handles[0];
|
|
|
|
delete handles[1];
|
|
|
|
delete db;
|
|
|
|
|
|
|
|
ASSERT_OK(DB::Open(db_options, dbname_, column_families, &handles, &db));
|
|
|
|
{
|
|
|
|
std::unique_ptr<Iterator> itr(db->NewIterator(ReadOptions(), handles[1]));
|
|
|
|
int count = 0;
|
|
|
|
for (itr->SeekToFirst(); itr->Valid(); itr->Next()) {
|
|
|
|
ASSERT_OK(itr->status());
|
|
|
|
++count;
|
|
|
|
}
|
|
|
|
ASSERT_EQ(count, 1000);
|
|
|
|
}
|
|
|
|
|
|
|
|
delete handles[0];
|
|
|
|
delete handles[1];
|
|
|
|
delete db;
|
|
|
|
}
|
|
|
|
|
|
|
|
} //namespace rocksdb
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
|
|
return RUN_ALL_TESTS();
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
#include <stdio.h>
|
|
|
|
|
|
|
|
int main(int /*argc*/, char** /*argv*/) {
|
|
|
|
fprintf(stderr,
|
|
|
|
"SKIPPED as DBImpl::DeleteFile is not supported in ROCKSDB_LITE\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif // !ROCKSDB_LITE
|