fork of https://github.com/oxigraph/rocksdb and https://github.com/facebook/rocksdb for nextgraph and oxigraph
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
626 lines
19 KiB
626 lines
19 KiB
11 years ago
|
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
||
|
// This source code is licensed under the BSD-style license found in the
|
||
|
// LICENSE file in the root directory of this source tree. An additional grant
|
||
|
// of patent rights can be found in the PATENTS file in the same directory.
|
||
|
//
|
||
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
||
|
// Use of this source code is governed by a BSD-style license that can be
|
||
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
||
|
|
||
|
#include "rocksdb/transaction_log.h"
|
||
|
#include "utilities/utility_db.h"
|
||
|
#include "utilities/backupable_db.h"
|
||
|
#include "util/testharness.h"
|
||
|
#include "util/random.h"
|
||
|
#include "util/testutil.h"
|
||
|
#include "util/auto_roll_logger.h"
|
||
|
|
||
|
#include <string>
|
||
|
#include <algorithm>
|
||
|
|
||
|
namespace rocksdb {
|
||
|
|
||
|
namespace {
|
||
|
|
||
|
using std::unique_ptr;
|
||
|
|
||
|
class DummyDB : public StackableDB {
|
||
|
public:
|
||
|
/* implicit */
|
||
|
DummyDB(const Options& options, const std::string& dbname)
|
||
|
: StackableDB(nullptr), options_(options), dbname_(dbname),
|
||
|
deletions_enabled_(true) {}
|
||
|
|
||
|
virtual const std::string& GetName() const override {
|
||
|
return dbname_;
|
||
|
}
|
||
|
|
||
|
virtual Env* GetEnv() const override {
|
||
|
return options_.env;
|
||
|
}
|
||
|
|
||
|
virtual const Options& GetOptions() const override {
|
||
|
return options_;
|
||
|
}
|
||
|
|
||
|
virtual Status EnableFileDeletions() override {
|
||
|
ASSERT_TRUE(!deletions_enabled_);
|
||
|
deletions_enabled_ = true;
|
||
|
return Status::OK();
|
||
|
}
|
||
|
|
||
|
virtual Status DisableFileDeletions() override {
|
||
|
ASSERT_TRUE(deletions_enabled_);
|
||
|
deletions_enabled_ = false;
|
||
|
return Status::OK();
|
||
|
}
|
||
|
|
||
|
virtual Status GetLiveFiles(std::vector<std::string>& vec, uint64_t* mfs,
|
||
|
bool flush_memtable = true) override {
|
||
|
ASSERT_TRUE(!deletions_enabled_);
|
||
|
vec = live_files_;
|
||
|
*mfs = 100;
|
||
|
return Status::OK();
|
||
|
}
|
||
|
|
||
|
class DummyLogFile : public LogFile {
|
||
|
public:
|
||
|
/* implicit */
|
||
|
DummyLogFile(const std::string& path, bool alive = true)
|
||
|
: path_(path), alive_(alive) {}
|
||
|
|
||
|
virtual std::string PathName() const override {
|
||
|
return path_;
|
||
|
}
|
||
|
|
||
|
virtual uint64_t LogNumber() const {
|
||
|
// what business do you have calling this method?
|
||
|
ASSERT_TRUE(false);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
virtual WalFileType Type() const override {
|
||
|
return alive_ ? kAliveLogFile : kArchivedLogFile;
|
||
|
}
|
||
|
|
||
|
virtual SequenceNumber StartSequence() const {
|
||
|
// backupabledb should not need this method
|
||
|
ASSERT_TRUE(false);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
virtual uint64_t SizeFileBytes() const {
|
||
|
// backupabledb should not need this method
|
||
|
ASSERT_TRUE(false);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
private:
|
||
|
std::string path_;
|
||
|
bool alive_;
|
||
|
}; // DummyLogFile
|
||
|
|
||
|
virtual Status GetSortedWalFiles(VectorLogPtr& files) override {
|
||
|
ASSERT_TRUE(!deletions_enabled_);
|
||
|
files.resize(wal_files_.size());
|
||
|
for (size_t i = 0; i < files.size(); ++i) {
|
||
|
files[i].reset(
|
||
|
new DummyLogFile(wal_files_[i].first, wal_files_[i].second));
|
||
|
}
|
||
|
return Status::OK();
|
||
|
}
|
||
|
|
||
|
std::vector<std::string> live_files_;
|
||
|
// pair<filename, alive?>
|
||
|
std::vector<std::pair<std::string, bool>> wal_files_;
|
||
|
private:
|
||
|
Options options_;
|
||
|
std::string dbname_;
|
||
|
bool deletions_enabled_;
|
||
|
}; // DummyDB
|
||
|
|
||
|
class TestEnv : public EnvWrapper {
|
||
|
public:
|
||
|
explicit TestEnv(Env* t) : EnvWrapper(t) {}
|
||
|
|
||
|
class DummySequentialFile : public SequentialFile {
|
||
|
virtual Status Read(size_t n, Slice* result, char* scratch) {
|
||
|
size_t read_size = (n > size_left) ? size_left : n;
|
||
|
*result = Slice(scratch, read_size);
|
||
|
size_left -= read_size;
|
||
|
return Status::OK();
|
||
|
}
|
||
|
|
||
|
virtual Status Skip(uint64_t n) {
|
||
|
size_left = (n > size_left) ? size_left - n : 0;
|
||
|
return Status::OK();
|
||
|
}
|
||
|
private:
|
||
|
size_t size_left = 200;
|
||
|
};
|
||
|
|
||
|
Status NewSequentialFile(const std::string& f,
|
||
|
unique_ptr<SequentialFile>* r,
|
||
|
const EnvOptions& options) {
|
||
|
opened_files_.push_back(f);
|
||
|
if (dummy_sequential_file_) {
|
||
|
r->reset(new TestEnv::DummySequentialFile());
|
||
|
return Status::OK();
|
||
|
} else {
|
||
|
return EnvWrapper::NewSequentialFile(f, r, options);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
Status NewWritableFile(const std::string& f, unique_ptr<WritableFile>* r,
|
||
|
const EnvOptions& options) {
|
||
|
if (limit_written_files_ <= 0) {
|
||
|
return Status::IOError("Sorry, can't do this");
|
||
|
}
|
||
|
limit_written_files_--;
|
||
|
return EnvWrapper::NewWritableFile(f, r, options);
|
||
|
}
|
||
|
|
||
|
void AssertOpenedFiles(std::vector<std::string>& should_have_opened) {
|
||
|
sort(should_have_opened.begin(), should_have_opened.end());
|
||
|
sort(opened_files_.begin(), opened_files_.end());
|
||
|
ASSERT_TRUE(opened_files_ == should_have_opened);
|
||
|
}
|
||
|
|
||
|
void ClearOpenedFiles() {
|
||
|
opened_files_.clear();
|
||
|
}
|
||
|
|
||
|
void SetLimitWrittenFiles(uint64_t limit) {
|
||
|
limit_written_files_ = limit;
|
||
|
}
|
||
|
|
||
|
void SetDummySequentialFile(bool dummy_sequential_file) {
|
||
|
dummy_sequential_file_ = dummy_sequential_file;
|
||
|
}
|
||
|
|
||
|
private:
|
||
|
bool dummy_sequential_file_ = false;
|
||
|
std::vector<std::string> opened_files_;
|
||
|
uint64_t limit_written_files_ = 1000000;
|
||
|
}; // TestEnv
|
||
|
|
||
|
class FileManager : public EnvWrapper {
|
||
|
public:
|
||
|
explicit FileManager(Env* t) : EnvWrapper(t), rnd_(5) {}
|
||
|
|
||
|
Status DeleteRandomFileInDir(const std::string dir) {
|
||
|
std::vector<std::string> children;
|
||
|
GetChildren(dir, &children);
|
||
|
if (children.size() <= 2) { // . and ..
|
||
|
return Status::NotFound("");
|
||
|
}
|
||
|
while (true) {
|
||
|
int i = rnd_.Next() % children.size();
|
||
|
if (children[i] != "." && children[i] != "..") {
|
||
|
return DeleteFile(dir + "/" + children[i]);
|
||
|
}
|
||
|
}
|
||
|
// should never get here
|
||
|
assert(false);
|
||
|
return Status::NotFound("");
|
||
|
}
|
||
|
|
||
|
Status CorruptFile(const std::string& fname, uint64_t bytes_to_corrupt) {
|
||
|
uint64_t size;
|
||
|
Status s = GetFileSize(fname, &size);
|
||
|
if (!s.ok()) {
|
||
|
return s;
|
||
|
}
|
||
|
unique_ptr<RandomRWFile> file;
|
||
|
EnvOptions env_options;
|
||
|
env_options.use_mmap_writes = false;
|
||
|
s = NewRandomRWFile(fname, &file, env_options);
|
||
|
if (!s.ok()) {
|
||
|
return s;
|
||
|
}
|
||
|
|
||
|
for (uint64_t i = 0; s.ok() && i < bytes_to_corrupt; ++i) {
|
||
|
std::string tmp;
|
||
|
// write one random byte to a random position
|
||
|
s = file->Write(rnd_.Next() % size, test::RandomString(&rnd_, 1, &tmp));
|
||
|
}
|
||
|
return s;
|
||
|
}
|
||
|
|
||
|
Status WriteToFile(const std::string& fname, const std::string& data) {
|
||
|
unique_ptr<WritableFile> file;
|
||
|
EnvOptions env_options;
|
||
|
env_options.use_mmap_writes = false;
|
||
|
Status s = EnvWrapper::NewWritableFile(fname, &file, env_options);
|
||
|
if (!s.ok()) {
|
||
|
return s;
|
||
|
}
|
||
|
return file->Append(Slice(data));
|
||
|
}
|
||
|
private:
|
||
|
Random rnd_;
|
||
|
}; // FileManager
|
||
|
|
||
|
// utility functions
|
||
|
static void FillDB(DB* db, int from, int to) {
|
||
|
for (int i = from; i < to; ++i) {
|
||
|
std::string key = "testkey" + std::to_string(i);
|
||
|
std::string value = "testvalue" + std::to_string(i);
|
||
|
|
||
|
ASSERT_OK(db->Put(WriteOptions(), Slice(key), Slice(value)));
|
||
|
}
|
||
|
}
|
||
|
|
||
|
static void AssertExists(DB* db, int from, int to) {
|
||
|
for (int i = from; i < to; ++i) {
|
||
|
std::string key = "testkey" + std::to_string(i);
|
||
|
std::string value;
|
||
|
Status s = db->Get(ReadOptions(), Slice(key), &value);
|
||
|
ASSERT_EQ(value, "testvalue" + std::to_string(i));
|
||
|
}
|
||
|
}
|
||
|
|
||
|
static void AssertEmpty(DB* db, int from, int to) {
|
||
|
for (int i = from; i < to; ++i) {
|
||
|
std::string key = "testkey" + std::to_string(i);
|
||
|
std::string value = "testvalue" + std::to_string(i);
|
||
|
|
||
|
Status s = db->Get(ReadOptions(), Slice(key), &value);
|
||
|
ASSERT_TRUE(s.IsNotFound());
|
||
|
}
|
||
|
}
|
||
|
|
||
|
class BackupableDBTest {
|
||
|
public:
|
||
|
BackupableDBTest() {
|
||
|
// set up files
|
||
|
dbname_ = test::TmpDir() + "/backupable_db";
|
||
|
backupdir_ = test::TmpDir() + "/backupable_db_backup";
|
||
|
|
||
|
// set up envs
|
||
|
env_ = Env::Default();
|
||
|
test_db_env_.reset(new TestEnv(env_));
|
||
|
test_backup_env_.reset(new TestEnv(env_));
|
||
|
file_manager_.reset(new FileManager(env_));
|
||
|
|
||
|
// set up db options
|
||
|
options_.create_if_missing = true;
|
||
|
options_.paranoid_checks = true;
|
||
|
options_.write_buffer_size = 1 << 19; // 512KB
|
||
|
options_.env = test_db_env_.get();
|
||
|
options_.wal_dir = dbname_;
|
||
|
// set up backup db options
|
||
|
CreateLoggerFromOptions(dbname_, backupdir_, env_,
|
||
|
Options(), &logger);
|
||
|
backupable_options_.reset(new BackupableDBOptions(
|
||
|
backupdir_, test_backup_env_.get(), logger.get(), true));
|
||
|
|
||
|
// delete old files in db
|
||
|
DestroyDB(dbname_, Options());
|
||
|
}
|
||
|
|
||
|
void OpenBackupableDB(bool destroy_old_data = false, bool dummy = false) {
|
||
|
// reset all the defaults
|
||
|
test_backup_env_->SetLimitWrittenFiles(1000000);
|
||
|
test_db_env_->SetLimitWrittenFiles(1000000);
|
||
|
test_db_env_->SetDummySequentialFile(dummy);
|
||
|
|
||
|
DB* db;
|
||
|
if (dummy) {
|
||
|
dummy_db_ = new DummyDB(options_, dbname_);
|
||
|
db = dummy_db_;
|
||
|
} else {
|
||
|
ASSERT_OK(DB::Open(options_, dbname_, &db));
|
||
|
}
|
||
|
backupable_options_->destroy_old_data = destroy_old_data;
|
||
|
db_.reset(new BackupableDB(db, *backupable_options_));
|
||
|
}
|
||
|
|
||
|
void CloseBackupableDB() {
|
||
|
db_.reset(nullptr);
|
||
|
}
|
||
|
|
||
|
void OpenRestoreDB() {
|
||
|
backupable_options_->destroy_old_data = false;
|
||
|
restore_db_.reset(
|
||
|
new RestoreBackupableDB(test_db_env_.get(), *backupable_options_));
|
||
|
}
|
||
|
|
||
|
void CloseRestoreDB() {
|
||
|
restore_db_.reset(nullptr);
|
||
|
}
|
||
|
|
||
|
// restores backup backup_id and asserts the existence of
|
||
|
// [start_exist, end_exist> and not-existence of
|
||
|
// [end_exist, end>
|
||
|
//
|
||
|
// if backup_id == 0, it means restore from latest
|
||
|
// if end == 0, don't check AssertEmpty
|
||
|
void AssertBackupConsistency(BackupID backup_id, uint32_t start_exist,
|
||
|
uint32_t end_exist, uint32_t end = 0) {
|
||
|
bool opened_restore = false;
|
||
|
if (restore_db_.get() == nullptr) {
|
||
|
opened_restore = true;
|
||
|
OpenRestoreDB();
|
||
|
}
|
||
|
if (backup_id > 0) {
|
||
|
ASSERT_OK(restore_db_->RestoreDBFromBackup(backup_id, dbname_, dbname_));
|
||
|
} else {
|
||
|
ASSERT_OK(restore_db_->RestoreDBFromLatestBackup(dbname_, dbname_));
|
||
|
}
|
||
|
OpenBackupableDB();
|
||
|
AssertExists(db_.get(), start_exist, end_exist);
|
||
|
if (end != 0) {
|
||
|
AssertEmpty(db_.get(), end_exist, end);
|
||
|
}
|
||
|
CloseBackupableDB();
|
||
|
if (opened_restore) {
|
||
|
CloseRestoreDB();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// files
|
||
|
std::string dbname_;
|
||
|
std::string backupdir_;
|
||
|
|
||
|
// envs
|
||
|
Env* env_;
|
||
|
unique_ptr<TestEnv> test_db_env_;
|
||
|
unique_ptr<TestEnv> test_backup_env_;
|
||
|
unique_ptr<FileManager> file_manager_;
|
||
|
|
||
|
// all the dbs!
|
||
|
DummyDB* dummy_db_; // BackupableDB owns dummy_db_
|
||
|
unique_ptr<BackupableDB> db_;
|
||
|
unique_ptr<RestoreBackupableDB> restore_db_;
|
||
|
|
||
|
// options
|
||
|
Options options_;
|
||
|
unique_ptr<BackupableDBOptions> backupable_options_;
|
||
|
std::shared_ptr<Logger> logger;
|
||
|
}; // BackupableDBTest
|
||
|
|
||
|
void AppendPath(const std::string& path, std::vector<std::string>& v) {
|
||
|
for (auto& f : v) {
|
||
|
f = path + f;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// this will make sure that backup does not copy the same file twice
|
||
|
TEST(BackupableDBTest, NoDoubleCopy) {
|
||
|
OpenBackupableDB(true, true);
|
||
|
|
||
|
// should write 5 DB files + LATEST_BACKUP + one meta file
|
||
|
test_backup_env_->SetLimitWrittenFiles(7);
|
||
|
test_db_env_->ClearOpenedFiles();
|
||
|
test_db_env_->SetLimitWrittenFiles(0);
|
||
|
dummy_db_->live_files_ = { "/00010.sst", "/00011.sst",
|
||
|
"/CURRENT", "/MANIFEST-01" };
|
||
|
dummy_db_->wal_files_ = {{"/00011.log", true}, {"/00012.log", false}};
|
||
|
ASSERT_OK(db_->CreateNewBackup(false));
|
||
|
std::vector<std::string> should_have_openened = dummy_db_->live_files_;
|
||
|
should_have_openened.push_back("/00011.log");
|
||
|
AppendPath(dbname_, should_have_openened);
|
||
|
test_db_env_->AssertOpenedFiles(should_have_openened);
|
||
|
|
||
|
// should write 4 new DB files + LATEST_BACKUP + one meta file
|
||
|
// should not write/copy 00010.sst, since it's already there!
|
||
|
test_backup_env_->SetLimitWrittenFiles(6);
|
||
|
test_db_env_->ClearOpenedFiles();
|
||
|
dummy_db_->live_files_ = { "/00010.sst", "/00015.sst",
|
||
|
"/CURRENT", "/MANIFEST-01" };
|
||
|
dummy_db_->wal_files_ = {{"/00011.log", true}, {"/00012.log", false}};
|
||
|
ASSERT_OK(db_->CreateNewBackup(false));
|
||
|
// should not open 00010.sst - it's already there
|
||
|
should_have_openened = { "/00015.sst", "/CURRENT",
|
||
|
"/MANIFEST-01", "/00011.log" };
|
||
|
AppendPath(dbname_, should_have_openened);
|
||
|
test_db_env_->AssertOpenedFiles(should_have_openened);
|
||
|
|
||
|
ASSERT_OK(db_->DeleteBackup(1));
|
||
|
ASSERT_EQ(true,
|
||
|
test_backup_env_->FileExists(backupdir_ + "/shared/00010.sst"));
|
||
|
// 00011.sst was only in backup 1, should be deleted
|
||
|
ASSERT_EQ(false,
|
||
|
test_backup_env_->FileExists(backupdir_ + "/shared/00011.sst"));
|
||
|
ASSERT_EQ(true,
|
||
|
test_backup_env_->FileExists(backupdir_ + "/shared/00015.sst"));
|
||
|
|
||
|
// MANIFEST file size should be only 100
|
||
|
uint64_t size;
|
||
|
test_backup_env_->GetFileSize(backupdir_ + "/private/2/MANIFEST-01", &size);
|
||
|
ASSERT_EQ(100, size);
|
||
|
test_backup_env_->GetFileSize(backupdir_ + "/shared/00015.sst", &size);
|
||
|
ASSERT_EQ(200, size);
|
||
|
}
|
||
|
|
||
|
// test various kind of corruptions that may happen:
|
||
|
// 1. Not able to write a file for backup - that backup should fail,
|
||
|
// everything else should work
|
||
|
// 2. Corrupted/deleted LATEST_BACKUP - everything should work fine
|
||
|
// 3. Corrupted backup meta file or missing backuped file - we should
|
||
|
// not be able to open that backup, but all other backups should be
|
||
|
// fine
|
||
|
TEST(BackupableDBTest, CorruptionsTest) {
|
||
|
const int keys_iteration = 20000;
|
||
|
Random rnd(6);
|
||
|
Status s;
|
||
|
|
||
|
OpenBackupableDB(true);
|
||
|
// create five backups
|
||
|
for (int i = 0; i < 5; ++i) {
|
||
|
FillDB(db_.get(), keys_iteration * i, keys_iteration * (i + 1));
|
||
|
ASSERT_OK(db_->CreateNewBackup(!!(rnd.Next() % 2)));
|
||
|
}
|
||
|
|
||
|
// ---------- case 1. - fail a write -----------
|
||
|
// try creating backup 6, but fail a write
|
||
|
FillDB(db_.get(), keys_iteration * 5, keys_iteration * 6);
|
||
|
test_backup_env_->SetLimitWrittenFiles(2);
|
||
|
// should fail
|
||
|
s = db_->CreateNewBackup(!!(rnd.Next() % 2));
|
||
|
ASSERT_TRUE(!s.ok());
|
||
|
test_backup_env_->SetLimitWrittenFiles(1000000);
|
||
|
// latest backup should have all the keys
|
||
|
CloseBackupableDB();
|
||
|
AssertBackupConsistency(0, 0, keys_iteration * 5, keys_iteration * 6);
|
||
|
|
||
|
// ---------- case 2. - corrupt/delete latest backup -----------
|
||
|
ASSERT_OK(file_manager_->CorruptFile(backupdir_ + "/LATEST_BACKUP", 2));
|
||
|
AssertBackupConsistency(0, 0, keys_iteration * 5);
|
||
|
ASSERT_OK(file_manager_->DeleteFile(backupdir_ + "/LATEST_BACKUP"));
|
||
|
AssertBackupConsistency(0, 0, keys_iteration * 5);
|
||
|
// create backup 6, point LATEST_BACKUP to 5
|
||
|
OpenBackupableDB();
|
||
|
FillDB(db_.get(), keys_iteration * 5, keys_iteration * 6);
|
||
|
ASSERT_OK(db_->CreateNewBackup(false));
|
||
|
CloseBackupableDB();
|
||
|
ASSERT_OK(file_manager_->WriteToFile(backupdir_ + "/LATEST_BACKUP", "5"));
|
||
|
AssertBackupConsistency(0, 0, keys_iteration * 5, keys_iteration * 6);
|
||
|
// assert that all 6 data is gone!
|
||
|
ASSERT_TRUE(file_manager_->FileExists(backupdir_ + "/meta/6") == false);
|
||
|
ASSERT_TRUE(file_manager_->FileExists(backupdir_ + "/private/6") == false);
|
||
|
|
||
|
// --------- case 3. corrupted backup meta or missing backuped file ----
|
||
|
ASSERT_OK(file_manager_->CorruptFile(backupdir_ + "/meta/5", 3));
|
||
|
// since 5 meta is now corrupted, latest backup should be 4
|
||
|
AssertBackupConsistency(0, 0, keys_iteration * 4, keys_iteration * 5);
|
||
|
OpenRestoreDB();
|
||
|
s = restore_db_->RestoreDBFromBackup(5, dbname_, dbname_);
|
||
|
ASSERT_TRUE(!s.ok());
|
||
|
CloseRestoreDB();
|
||
|
ASSERT_OK(file_manager_->DeleteRandomFileInDir(backupdir_ + "/private/4"));
|
||
|
// 4 is corrupted, 3 is the latest backup now
|
||
|
AssertBackupConsistency(0, 0, keys_iteration * 3, keys_iteration * 5);
|
||
|
OpenRestoreDB();
|
||
|
s = restore_db_->RestoreDBFromBackup(4, dbname_, dbname_);
|
||
|
CloseRestoreDB();
|
||
|
ASSERT_TRUE(!s.ok());
|
||
|
|
||
|
// new backup should be 4!
|
||
|
OpenBackupableDB();
|
||
|
FillDB(db_.get(), keys_iteration * 3, keys_iteration * 4);
|
||
|
ASSERT_OK(db_->CreateNewBackup(!!(rnd.Next() % 2)));
|
||
|
CloseBackupableDB();
|
||
|
AssertBackupConsistency(4, 0, keys_iteration * 4, keys_iteration * 5);
|
||
|
}
|
||
|
|
||
|
// open DB, write, close DB, backup, restore, repeat
|
||
|
TEST(BackupableDBTest, OfflineIntegrationTest) {
|
||
|
// has to be a big number, so that it triggers the memtable flush
|
||
|
const int keys_iteration = 20000;
|
||
|
const int max_key = keys_iteration * 4 + 10;
|
||
|
// first iter -- flush before backup
|
||
|
// second iter -- don't flush before backup
|
||
|
for (int iter = 0; iter < 2; ++iter) {
|
||
|
// delete old data
|
||
|
DestroyDB(dbname_, Options());
|
||
|
bool destroy_data = true;
|
||
|
|
||
|
// every iteration --
|
||
|
// 1. insert new data in the DB
|
||
|
// 2. backup the DB
|
||
|
// 3. destroy the db
|
||
|
// 4. restore the db, check everything is still there
|
||
|
for (int i = 0; i < 5; ++i) {
|
||
|
// in last iteration, put smaller amount of data,
|
||
|
int fill_up_to = std::min(keys_iteration * (i + 1), max_key);
|
||
|
// ---- insert new data and back up ----
|
||
|
OpenBackupableDB(destroy_data);
|
||
|
destroy_data = false;
|
||
|
FillDB(db_.get(), keys_iteration * i, fill_up_to);
|
||
|
ASSERT_OK(db_->CreateNewBackup(iter == 0));
|
||
|
CloseBackupableDB();
|
||
|
DestroyDB(dbname_, Options());
|
||
|
|
||
|
// ---- make sure it's empty ----
|
||
|
OpenBackupableDB();
|
||
|
AssertEmpty(db_.get(), 0, fill_up_to);
|
||
|
CloseBackupableDB();
|
||
|
|
||
|
// ---- restore the DB ----
|
||
|
OpenRestoreDB();
|
||
|
if (i >= 3) { // test purge old backups
|
||
|
// when i == 4, purge to only 1 backup
|
||
|
// when i == 3, purge to 2 backups
|
||
|
ASSERT_OK(restore_db_->PurgeOldBackups(5 - i));
|
||
|
}
|
||
|
// ---- make sure the data is there ---
|
||
|
AssertBackupConsistency(0, 0, fill_up_to, max_key);
|
||
|
CloseRestoreDB();
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// open DB, write, backup, write, backup, close, restore
|
||
|
TEST(BackupableDBTest, OnlineIntegrationTest) {
|
||
|
// has to be a big number, so that it triggers the memtable flush
|
||
|
const int keys_iteration = 20000;
|
||
|
const int max_key = keys_iteration * 4 + 10;
|
||
|
Random rnd(7);
|
||
|
// delete old data
|
||
|
DestroyDB(dbname_, Options());
|
||
|
|
||
|
OpenBackupableDB(true);
|
||
|
// write some data, backup, repeat
|
||
|
for (int i = 0; i < 5; ++i) {
|
||
|
if (i == 4) {
|
||
|
// delete backup number 2, online delete!
|
||
|
OpenRestoreDB();
|
||
|
ASSERT_OK(restore_db_->DeleteBackup(2));
|
||
|
CloseRestoreDB();
|
||
|
}
|
||
|
// in last iteration, put smaller amount of data,
|
||
|
// so that backups can share sst files
|
||
|
int fill_up_to = std::min(keys_iteration * (i + 1), max_key);
|
||
|
FillDB(db_.get(), keys_iteration * i, fill_up_to);
|
||
|
// we should get consistent results with flush_before_backup
|
||
|
// set to both true and false
|
||
|
ASSERT_OK(db_->CreateNewBackup(!!(rnd.Next() % 2)));
|
||
|
}
|
||
|
// close and destroy
|
||
|
CloseBackupableDB();
|
||
|
DestroyDB(dbname_, Options());
|
||
|
|
||
|
// ---- make sure it's empty ----
|
||
|
OpenBackupableDB();
|
||
|
AssertEmpty(db_.get(), 0, max_key);
|
||
|
CloseBackupableDB();
|
||
|
|
||
|
// ---- restore every backup and verify all the data is there ----
|
||
|
OpenRestoreDB();
|
||
|
for (int i = 1; i <= 5; ++i) {
|
||
|
if (i == 2) {
|
||
|
// we deleted backup 2
|
||
|
Status s = restore_db_->RestoreDBFromBackup(2, dbname_, dbname_);
|
||
|
ASSERT_TRUE(!s.ok());
|
||
|
} else {
|
||
|
int fill_up_to = std::min(keys_iteration * i, max_key);
|
||
|
AssertBackupConsistency(i, 0, fill_up_to, max_key);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// delete some backups -- this should leave only backups 3 and 5 alive
|
||
|
ASSERT_OK(restore_db_->DeleteBackup(4));
|
||
|
ASSERT_OK(restore_db_->PurgeOldBackups(2));
|
||
|
|
||
|
std::vector<BackupInfo> backup_info;
|
||
|
restore_db_->GetBackupInfo(&backup_info);
|
||
|
ASSERT_EQ(2, backup_info.size());
|
||
|
|
||
|
// check backup 3
|
||
|
AssertBackupConsistency(3, 0, 3 * keys_iteration, max_key);
|
||
|
// check backup 5
|
||
|
AssertBackupConsistency(5, 0, max_key);
|
||
|
|
||
|
CloseRestoreDB();
|
||
|
}
|
||
|
|
||
|
} // anon namespace
|
||
|
|
||
|
} // namespace rocksdb
|
||
|
|
||
|
int main(int argc, char** argv) {
|
||
|
return rocksdb::test::RunAllTests();
|
||
|
}
|