|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
//
|
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
|
|
|
|
// Introduction of SyncPoint effectively disabled building and running this test
|
|
|
|
// in Release build.
|
|
|
|
// which is a pity, it is a good test
|
|
|
|
#if !defined(ROCKSDB_LITE)
|
|
|
|
|
|
|
|
#include "db/db_test_util.h"
|
|
|
|
#include "db/forward_iterator.h"
|
|
|
|
#include "port/stack_trace.h"
|
|
|
|
|
|
|
|
namespace rocksdb {
|
|
|
|
|
|
|
|
class DBTestTailingIterator : public DBTestBase {
|
|
|
|
public:
|
|
|
|
DBTestTailingIterator() : DBTestBase("/db_tailing_iterator_test") {}
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorSingle) {
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
|
|
|
|
// add a record and check that iter can see it
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "mirko", "fodor"));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "mirko");
|
|
|
|
|
|
|
|
iter->Next();
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorKeepAdding) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
std::string value(1024, 'a');
|
|
|
|
|
|
|
|
const int num_records = 10000;
|
|
|
|
for (int i = 0; i < num_records; ++i) {
|
|
|
|
char buf[32];
|
|
|
|
snprintf(buf, sizeof(buf), "%016d", i);
|
|
|
|
|
|
|
|
Slice key(buf, 16);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
|
|
|
|
iter->Seek(key);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().compare(key), 0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorSeekToNext) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
std::unique_ptr<Iterator> itern(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
std::string value(1024, 'a');
|
|
|
|
|
|
|
|
const int num_records = 1000;
|
|
|
|
for (int i = 1; i < num_records; ++i) {
|
|
|
|
char buf1[32];
|
|
|
|
char buf2[32];
|
|
|
|
snprintf(buf1, sizeof(buf1), "00a0%016d", i * 5);
|
|
|
|
|
|
|
|
Slice key(buf1, 20);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
|
|
|
|
if (i % 100 == 99) {
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
}
|
|
|
|
|
|
|
|
snprintf(buf2, sizeof(buf2), "00a0%016d", i * 5 - 2);
|
|
|
|
Slice target(buf2, 20);
|
|
|
|
iter->Seek(target);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().compare(key), 0);
|
|
|
|
if (i == 1) {
|
|
|
|
itern->SeekToFirst();
|
|
|
|
} else {
|
|
|
|
itern->Next();
|
|
|
|
}
|
|
|
|
ASSERT_TRUE(itern->Valid());
|
|
|
|
ASSERT_EQ(itern->key().compare(key), 0);
|
|
|
|
}
|
|
|
|
rocksdb::SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
for (int i = 2 * num_records; i > 0; --i) {
|
|
|
|
char buf1[32];
|
|
|
|
char buf2[32];
|
|
|
|
snprintf(buf1, sizeof(buf1), "00a0%016d", i * 5);
|
|
|
|
|
|
|
|
Slice key(buf1, 20);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
|
|
|
|
if (i % 100 == 99) {
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
}
|
|
|
|
|
|
|
|
snprintf(buf2, sizeof(buf2), "00a0%016d", i * 5 - 2);
|
|
|
|
Slice target(buf2, 20);
|
|
|
|
iter->Seek(target);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().compare(key), 0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorTrimSeekToNext) {
|
|
|
|
const uint64_t k150KB = 150 * 1024;
|
|
|
|
Options options;
|
|
|
|
options.write_buffer_size = k150KB;
|
|
|
|
options.max_write_buffer_number = 3;
|
|
|
|
options.min_write_buffer_number_to_merge = 2;
|
|
|
|
options.env = env_;
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
int num_iters, deleted_iters;
|
|
|
|
|
|
|
|
char bufe[32];
|
|
|
|
snprintf(bufe, sizeof(bufe), "00b0%016d", 0);
|
|
|
|
Slice keyu(bufe, 20);
|
|
|
|
read_options.iterate_upper_bound = &keyu;
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
std::unique_ptr<Iterator> itern(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
std::unique_ptr<Iterator> iterh(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
std::string value(1024, 'a');
|
|
|
|
bool file_iters_deleted = false;
|
Reuse file iterators in tailing iterator when memtable is flushed
Summary:
Under a tailing workload, there were increased block cache
misses when a memtable was flushed because we were rebuilding iterators
in that case since the version set changed. This was exacerbated in the
case of iterate_upper_bound, since file iterators which were over the
iterate_upper_bound would have been deleted and are now brought back as
part of the Rebuild, only to be deleted again. We now renew the iterators
and only build iterators for files which are added and delete file
iterators for files which are deleted.
Refer to https://reviews.facebook.net/D50463 for previous version
Test Plan: DBTestTailingIterator.TailingIteratorTrimSeekToNext
Reviewers: anthony, IslamAbdelRahman, igor, tnovak, yhchiang, sdong
Reviewed By: sdong
Subscribers: yhchiang, march, dhruba, leveldb, lovro
Differential Revision: https://reviews.facebook.net/D50679
9 years ago
|
|
|
bool file_iters_renewed_null = false;
|
|
|
|
bool file_iters_renewed_copy = false;
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"ForwardIterator::SeekInternal:Return", [&](void* arg) {
|
|
|
|
ForwardIterator* fiter = reinterpret_cast<ForwardIterator*>(arg);
|
|
|
|
ASSERT_TRUE(!file_iters_deleted ||
|
|
|
|
fiter->TEST_CheckDeletedIters(&deleted_iters, &num_iters));
|
|
|
|
});
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"ForwardIterator::Next:Return", [&](void* arg) {
|
|
|
|
ForwardIterator* fiter = reinterpret_cast<ForwardIterator*>(arg);
|
|
|
|
ASSERT_TRUE(!file_iters_deleted ||
|
|
|
|
fiter->TEST_CheckDeletedIters(&deleted_iters, &num_iters));
|
|
|
|
});
|
Reuse file iterators in tailing iterator when memtable is flushed
Summary:
Under a tailing workload, there were increased block cache
misses when a memtable was flushed because we were rebuilding iterators
in that case since the version set changed. This was exacerbated in the
case of iterate_upper_bound, since file iterators which were over the
iterate_upper_bound would have been deleted and are now brought back as
part of the Rebuild, only to be deleted again. We now renew the iterators
and only build iterators for files which are added and delete file
iterators for files which are deleted.
Refer to https://reviews.facebook.net/D50463 for previous version
Test Plan: DBTestTailingIterator.TailingIteratorTrimSeekToNext
Reviewers: anthony, IslamAbdelRahman, igor, tnovak, yhchiang, sdong
Reviewed By: sdong
Subscribers: yhchiang, march, dhruba, leveldb, lovro
Differential Revision: https://reviews.facebook.net/D50679
9 years ago
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"ForwardIterator::RenewIterators:Null",
|
|
|
|
[&](void* arg) { file_iters_renewed_null = true; });
|
Reuse file iterators in tailing iterator when memtable is flushed
Summary:
Under a tailing workload, there were increased block cache
misses when a memtable was flushed because we were rebuilding iterators
in that case since the version set changed. This was exacerbated in the
case of iterate_upper_bound, since file iterators which were over the
iterate_upper_bound would have been deleted and are now brought back as
part of the Rebuild, only to be deleted again. We now renew the iterators
and only build iterators for files which are added and delete file
iterators for files which are deleted.
Refer to https://reviews.facebook.net/D50463 for previous version
Test Plan: DBTestTailingIterator.TailingIteratorTrimSeekToNext
Reviewers: anthony, IslamAbdelRahman, igor, tnovak, yhchiang, sdong
Reviewed By: sdong
Subscribers: yhchiang, march, dhruba, leveldb, lovro
Differential Revision: https://reviews.facebook.net/D50679
9 years ago
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"ForwardIterator::RenewIterators:Copy",
|
|
|
|
[&](void* arg) { file_iters_renewed_copy = true; });
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
const int num_records = 1000;
|
|
|
|
for (int i = 1; i < num_records; ++i) {
|
|
|
|
char buf1[32];
|
|
|
|
char buf2[32];
|
|
|
|
char buf3[32];
|
|
|
|
char buf4[32];
|
|
|
|
snprintf(buf1, sizeof(buf1), "00a0%016d", i * 5);
|
|
|
|
snprintf(buf3, sizeof(buf3), "00b0%016d", i * 5);
|
|
|
|
|
|
|
|
Slice key(buf1, 20);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
Slice keyn(buf3, 20);
|
|
|
|
ASSERT_OK(Put(1, keyn, value));
|
|
|
|
|
|
|
|
if (i % 100 == 99) {
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
if (i == 299) {
|
|
|
|
file_iters_deleted = true;
|
|
|
|
}
|
|
|
|
snprintf(buf4, sizeof(buf4), "00a0%016d", i * 5 / 2);
|
|
|
|
Slice target(buf4, 20);
|
|
|
|
iterh->Seek(target);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
for (int j = (i + 1) * 5 / 2; j < i * 5; j += 5) {
|
|
|
|
iterh->Next();
|
|
|
|
ASSERT_TRUE(iterh->Valid());
|
|
|
|
}
|
|
|
|
if (i == 299) {
|
|
|
|
file_iters_deleted = false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
file_iters_deleted = true;
|
|
|
|
snprintf(buf2, sizeof(buf2), "00a0%016d", i * 5 - 2);
|
|
|
|
Slice target(buf2, 20);
|
|
|
|
iter->Seek(target);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().compare(key), 0);
|
|
|
|
ASSERT_LE(num_iters, 1);
|
|
|
|
if (i == 1) {
|
|
|
|
itern->SeekToFirst();
|
|
|
|
} else {
|
|
|
|
itern->Next();
|
|
|
|
}
|
|
|
|
ASSERT_TRUE(itern->Valid());
|
|
|
|
ASSERT_EQ(itern->key().compare(key), 0);
|
|
|
|
ASSERT_LE(num_iters, 1);
|
|
|
|
file_iters_deleted = false;
|
|
|
|
}
|
Reuse file iterators in tailing iterator when memtable is flushed
Summary:
Under a tailing workload, there were increased block cache
misses when a memtable was flushed because we were rebuilding iterators
in that case since the version set changed. This was exacerbated in the
case of iterate_upper_bound, since file iterators which were over the
iterate_upper_bound would have been deleted and are now brought back as
part of the Rebuild, only to be deleted again. We now renew the iterators
and only build iterators for files which are added and delete file
iterators for files which are deleted.
Refer to https://reviews.facebook.net/D50463 for previous version
Test Plan: DBTestTailingIterator.TailingIteratorTrimSeekToNext
Reviewers: anthony, IslamAbdelRahman, igor, tnovak, yhchiang, sdong
Reviewed By: sdong
Subscribers: yhchiang, march, dhruba, leveldb, lovro
Differential Revision: https://reviews.facebook.net/D50679
9 years ago
|
|
|
ASSERT_TRUE(file_iters_renewed_null);
|
|
|
|
ASSERT_TRUE(file_iters_renewed_copy);
|
|
|
|
iter = 0;
|
|
|
|
itern = 0;
|
|
|
|
iterh = 0;
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.no_block_cache = true;
|
|
|
|
table_options.block_cache_compressed = nullptr;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
ReopenWithColumnFamilies({"default", "pikachu"}, options);
|
|
|
|
read_options.read_tier = kBlockCacheTier;
|
|
|
|
std::unique_ptr<Iterator> iteri(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
char buf5[32];
|
|
|
|
snprintf(buf5, sizeof(buf5), "00a0%016d", (num_records / 2) * 5 - 2);
|
|
|
|
Slice target1(buf5, 20);
|
|
|
|
iteri->Seek(target1);
|
|
|
|
ASSERT_TRUE(iteri->status().IsIncomplete());
|
|
|
|
iteri = 0;
|
|
|
|
|
|
|
|
read_options.read_tier = kReadAllTier;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory());
|
|
|
|
ReopenWithColumnFamilies({"default", "pikachu"}, options);
|
|
|
|
iter.reset(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
for (int i = 2 * num_records; i > 0; --i) {
|
|
|
|
char buf1[32];
|
|
|
|
char buf2[32];
|
|
|
|
snprintf(buf1, sizeof(buf1), "00a0%016d", i * 5);
|
|
|
|
|
|
|
|
Slice key(buf1, 20);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
|
|
|
|
if (i % 100 == 99) {
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
}
|
|
|
|
|
|
|
|
snprintf(buf2, sizeof(buf2), "00a0%016d", i * 5 - 2);
|
|
|
|
Slice target(buf2, 20);
|
|
|
|
iter->Seek(target);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().compare(key), 0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorDeletes) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
|
|
|
|
// write a single record, read it using the iterator, then delete it
|
|
|
|
ASSERT_OK(Put(1, "0test", "test"));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "0test");
|
|
|
|
ASSERT_OK(Delete(1, "0test"));
|
|
|
|
|
|
|
|
// write many more records
|
|
|
|
const int num_records = 10000;
|
|
|
|
std::string value(1024, 'A');
|
|
|
|
|
|
|
|
for (int i = 0; i < num_records; ++i) {
|
|
|
|
char buf[32];
|
|
|
|
snprintf(buf, sizeof(buf), "1%015d", i);
|
|
|
|
|
|
|
|
Slice key(buf, 16);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
}
|
|
|
|
|
|
|
|
// force a flush to make sure that no records are read from memtable
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
|
|
|
|
// skip "0test"
|
|
|
|
iter->Next();
|
|
|
|
|
|
|
|
// make sure we can read all new records using the existing iterator
|
|
|
|
int count = 0;
|
|
|
|
for (; iter->Valid(); iter->Next(), ++count) ;
|
|
|
|
|
|
|
|
ASSERT_EQ(count, num_records);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorPrefixSeek) {
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
options.prefix_extractor.reset(NewFixedPrefixTransform(2));
|
|
|
|
options.memtable_factory.reset(NewHashSkipListRepFactory(16));
|
|
|
|
options.allow_concurrent_memtable_write = false;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
ASSERT_OK(Put(1, "0101", "test"));
|
|
|
|
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
|
|
|
|
ASSERT_OK(Put(1, "0202", "test"));
|
|
|
|
|
|
|
|
// Seek(0102) shouldn't find any records since 0202 has a different prefix
|
|
|
|
iter->Seek("0102");
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
|
|
|
|
iter->Seek("0202");
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "0202");
|
|
|
|
|
|
|
|
iter->Next();
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorIncomplete) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.read_tier = kBlockCacheTier;
|
|
|
|
|
|
|
|
std::string key("key");
|
|
|
|
std::string value("value");
|
|
|
|
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), key, value));
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
// we either see the entry or it's not in cache
|
|
|
|
ASSERT_TRUE(iter->Valid() || iter->status().IsIncomplete());
|
|
|
|
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
// should still be true after compaction
|
|
|
|
ASSERT_TRUE(iter->Valid() || iter->status().IsIncomplete());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorSeekToSame) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.compaction_style = kCompactionStyleUniversal;
|
|
|
|
options.write_buffer_size = 1000;
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
|
|
|
|
const int NROWS = 10000;
|
|
|
|
// Write rows with keys 00000, 00002, 00004 etc.
|
|
|
|
for (int i = 0; i < NROWS; ++i) {
|
|
|
|
char buf[100];
|
|
|
|
snprintf(buf, sizeof(buf), "%05d", 2*i);
|
|
|
|
std::string key(buf);
|
|
|
|
std::string value("value");
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), key, value));
|
|
|
|
}
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
// Seek to 00001. We expect to find 00002.
|
|
|
|
std::string start_key = "00001";
|
|
|
|
iter->Seek(start_key);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
|
|
|
|
std::string found = iter->key().ToString();
|
|
|
|
ASSERT_EQ("00002", found);
|
|
|
|
|
|
|
|
// Now seek to the same key. The iterator should remain in the same
|
|
|
|
// position.
|
|
|
|
iter->Seek(found);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(found, iter->key().ToString());
|
|
|
|
}
|
|
|
|
|
|
|
|
// Sets iterate_upper_bound and verifies that ForwardIterator doesn't call
|
|
|
|
// Seek() on immutable iterators when target key is >= prev_key and all
|
|
|
|
// iterators, including the memtable iterator, are over the upper bound.
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorUpperBound) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
|
|
|
|
const Slice upper_bound("20", 3);
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.iterate_upper_bound = &upper_bound;
|
|
|
|
|
|
|
|
ASSERT_OK(Put(1, "11", "11"));
|
|
|
|
ASSERT_OK(Put(1, "12", "12"));
|
|
|
|
ASSERT_OK(Put(1, "22", "22"));
|
|
|
|
ASSERT_OK(Flush(1)); // flush all those keys to an immutable SST file
|
|
|
|
|
|
|
|
// Add another key to the memtable.
|
|
|
|
ASSERT_OK(Put(1, "21", "21"));
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> it(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
it->Seek("12");
|
|
|
|
ASSERT_TRUE(it->Valid());
|
|
|
|
ASSERT_EQ("12", it->key().ToString());
|
|
|
|
|
|
|
|
it->Next();
|
|
|
|
// Not valid since "21" is over the upper bound.
|
|
|
|
ASSERT_FALSE(it->Valid());
|
|
|
|
|
|
|
|
// This keeps track of the number of times NeedToSeekImmutable() was true.
|
|
|
|
int immutable_seeks = 0;
|
|
|
|
rocksdb::SyncPoint::GetInstance()->SetCallBack(
|
|
|
|
"ForwardIterator::SeekInternal:Immutable",
|
|
|
|
[&](void* arg) { ++immutable_seeks; });
|
|
|
|
|
|
|
|
// Seek to 13. This should not require any immutable seeks.
|
|
|
|
rocksdb::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
it->Seek("13");
|
|
|
|
rocksdb::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
|
|
|
|
ASSERT_FALSE(it->Valid());
|
|
|
|
ASSERT_EQ(0, immutable_seeks);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, TailingIteratorGap) {
|
|
|
|
// level 1: [20, 25] [35, 40]
|
|
|
|
// level 2: [10 - 15] [45 - 50]
|
|
|
|
// level 3: [20, 30, 40]
|
|
|
|
// Previously there is a bug in tailing_iterator that if there is a gap in
|
|
|
|
// lower level, the key will be skipped if it is within the range between
|
|
|
|
// the largest key of index n file and the smallest key of index n+1 file
|
|
|
|
// if both file fit in that gap. In this example, 25 < key < 35
|
|
|
|
// https://github.com/facebook/rocksdb/issues/1372
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
|
|
|
|
ASSERT_OK(Put(1, "20", "20"));
|
|
|
|
ASSERT_OK(Put(1, "30", "30"));
|
|
|
|
ASSERT_OK(Put(1, "40", "40"));
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
MoveFilesToLevel(3, 1);
|
|
|
|
|
|
|
|
ASSERT_OK(Put(1, "10", "10"));
|
|
|
|
ASSERT_OK(Put(1, "15", "15"));
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
ASSERT_OK(Put(1, "45", "45"));
|
|
|
|
ASSERT_OK(Put(1, "50", "50"));
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
MoveFilesToLevel(2, 1);
|
|
|
|
|
|
|
|
ASSERT_OK(Put(1, "20", "20"));
|
|
|
|
ASSERT_OK(Put(1, "25", "25"));
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
ASSERT_OK(Put(1, "35", "35"));
|
|
|
|
ASSERT_OK(Put(1, "40", "40"));
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
MoveFilesToLevel(1, 1);
|
|
|
|
|
|
|
|
ColumnFamilyMetaData meta;
|
|
|
|
db_->GetColumnFamilyMetaData(handles_[1], &meta);
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> it(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
it->Seek("30");
|
|
|
|
ASSERT_TRUE(it->Valid());
|
|
|
|
ASSERT_EQ("30", it->key().ToString());
|
|
|
|
|
|
|
|
it->Next();
|
|
|
|
ASSERT_TRUE(it->Valid());
|
|
|
|
ASSERT_EQ("35", it->key().ToString());
|
|
|
|
|
|
|
|
it->Next();
|
|
|
|
ASSERT_TRUE(it->Valid());
|
|
|
|
ASSERT_EQ("40", it->key().ToString());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, ManagedTailingIteratorSingle) {
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.managed = true;
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
|
|
|
|
// add a record and check that iter can see it
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "mirko", "fodor"));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "mirko");
|
|
|
|
|
|
|
|
iter->Next();
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, ManagedTailingIteratorKeepAdding) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.managed = true;
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
std::string value(1024, 'a');
|
|
|
|
|
|
|
|
const int num_records = 10000;
|
|
|
|
for (int i = 0; i < num_records; ++i) {
|
|
|
|
char buf[32];
|
|
|
|
snprintf(buf, sizeof(buf), "%016d", i);
|
|
|
|
|
|
|
|
Slice key(buf, 16);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
|
|
|
|
iter->Seek(key);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().compare(key), 0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, ManagedTailingIteratorSeekToNext) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.managed = true;
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
std::string value(1024, 'a');
|
|
|
|
|
|
|
|
const int num_records = 1000;
|
|
|
|
for (int i = 1; i < num_records; ++i) {
|
|
|
|
char buf1[32];
|
|
|
|
char buf2[32];
|
|
|
|
snprintf(buf1, sizeof(buf1), "00a0%016d", i * 5);
|
|
|
|
|
|
|
|
Slice key(buf1, 20);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
|
|
|
|
if (i % 100 == 99) {
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
}
|
|
|
|
|
|
|
|
snprintf(buf2, sizeof(buf2), "00a0%016d", i * 5 - 2);
|
|
|
|
Slice target(buf2, 20);
|
|
|
|
iter->Seek(target);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().compare(key), 0);
|
|
|
|
}
|
|
|
|
for (int i = 2 * num_records; i > 0; --i) {
|
|
|
|
char buf1[32];
|
|
|
|
char buf2[32];
|
|
|
|
snprintf(buf1, sizeof(buf1), "00a0%016d", i * 5);
|
|
|
|
|
|
|
|
Slice key(buf1, 20);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
|
|
|
|
if (i % 100 == 99) {
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
}
|
|
|
|
|
|
|
|
snprintf(buf2, sizeof(buf2), "00a0%016d", i * 5 - 2);
|
|
|
|
Slice target(buf2, 20);
|
|
|
|
iter->Seek(target);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().compare(key), 0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, ManagedTailingIteratorDeletes) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.managed = true;
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
|
|
|
|
// write a single record, read it using the iterator, then delete it
|
|
|
|
ASSERT_OK(Put(1, "0test", "test"));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "0test");
|
|
|
|
ASSERT_OK(Delete(1, "0test"));
|
|
|
|
|
|
|
|
// write many more records
|
|
|
|
const int num_records = 10000;
|
|
|
|
std::string value(1024, 'A');
|
|
|
|
|
|
|
|
for (int i = 0; i < num_records; ++i) {
|
|
|
|
char buf[32];
|
|
|
|
snprintf(buf, sizeof(buf), "1%015d", i);
|
|
|
|
|
|
|
|
Slice key(buf, 16);
|
|
|
|
ASSERT_OK(Put(1, key, value));
|
|
|
|
}
|
|
|
|
|
|
|
|
// force a flush to make sure that no records are read from memtable
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
|
|
|
|
// skip "0test"
|
|
|
|
iter->Next();
|
|
|
|
|
|
|
|
// make sure we can read all new records using the existing iterator
|
|
|
|
int count = 0;
|
|
|
|
for (; iter->Valid(); iter->Next(), ++count) {
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT_EQ(count, num_records);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, ManagedTailingIteratorPrefixSeek) {
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.managed = true;
|
|
|
|
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.create_if_missing = true;
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
options.prefix_extractor.reset(NewFixedPrefixTransform(2));
|
|
|
|
options.memtable_factory.reset(NewHashSkipListRepFactory(16));
|
|
|
|
options.allow_concurrent_memtable_write = false;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options, handles_[1]));
|
|
|
|
ASSERT_OK(Put(1, "0101", "test"));
|
|
|
|
|
|
|
|
ASSERT_OK(Flush(1));
|
|
|
|
|
|
|
|
ASSERT_OK(Put(1, "0202", "test"));
|
|
|
|
|
|
|
|
// Seek(0102) shouldn't find any records since 0202 has a different prefix
|
|
|
|
iter->Seek("0102");
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
|
|
|
|
iter->Seek("0202");
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "0202");
|
|
|
|
|
|
|
|
iter->Next();
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, ManagedTailingIteratorIncomplete) {
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, CurrentOptions());
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.managed = true;
|
|
|
|
read_options.read_tier = kBlockCacheTier;
|
|
|
|
|
|
|
|
std::string key = "key";
|
|
|
|
std::string value = "value";
|
|
|
|
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), key, value));
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
// we either see the entry or it's not in cache
|
|
|
|
ASSERT_TRUE(iter->Valid() || iter->status().IsIncomplete());
|
|
|
|
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
// should still be true after compaction
|
|
|
|
ASSERT_TRUE(iter->Valid() || iter->status().IsIncomplete());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, ManagedTailingIteratorSeekToSame) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.compaction_style = kCompactionStyleUniversal;
|
|
|
|
options.write_buffer_size = 1000;
|
|
|
|
CreateAndReopenWithCF({"pikachu"}, options);
|
|
|
|
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
read_options.managed = true;
|
|
|
|
|
|
|
|
const int NROWS = 10000;
|
|
|
|
// Write rows with keys 00000, 00002, 00004 etc.
|
|
|
|
for (int i = 0; i < NROWS; ++i) {
|
|
|
|
char buf[100];
|
|
|
|
snprintf(buf, sizeof(buf), "%05d", 2 * i);
|
|
|
|
std::string key(buf);
|
|
|
|
std::string value("value");
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), key, value));
|
|
|
|
}
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
// Seek to 00001. We expect to find 00002.
|
|
|
|
std::string start_key = "00001";
|
|
|
|
iter->Seek(start_key);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
|
|
|
|
std::string found = iter->key().ToString();
|
|
|
|
ASSERT_EQ("00002", found);
|
|
|
|
|
|
|
|
// Now seek to the same key. The iterator should remain in the same
|
|
|
|
// position.
|
|
|
|
iter->Seek(found);
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(found, iter->key().ToString());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, ForwardIteratorVersionProperty) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.write_buffer_size = 1000;
|
|
|
|
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
|
|
|
|
Put("foo", "bar");
|
|
|
|
|
|
|
|
uint64_t v1, v2, v3, v4;
|
|
|
|
{
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
iter->Seek("foo");
|
|
|
|
std::string prop_value;
|
|
|
|
ASSERT_OK(iter->GetProperty("rocksdb.iterator.super-version-number",
|
|
|
|
&prop_value));
|
|
|
|
v1 = static_cast<uint64_t>(std::atoi(prop_value.c_str()));
|
|
|
|
|
|
|
|
Put("foo1", "bar1");
|
|
|
|
Flush();
|
|
|
|
|
|
|
|
ASSERT_OK(iter->GetProperty("rocksdb.iterator.super-version-number",
|
|
|
|
&prop_value));
|
|
|
|
v2 = static_cast<uint64_t>(std::atoi(prop_value.c_str()));
|
|
|
|
|
|
|
|
iter->Seek("f");
|
|
|
|
|
|
|
|
ASSERT_OK(iter->GetProperty("rocksdb.iterator.super-version-number",
|
|
|
|
&prop_value));
|
|
|
|
v3 = static_cast<uint64_t>(std::atoi(prop_value.c_str()));
|
|
|
|
|
|
|
|
ASSERT_EQ(v1, v2);
|
|
|
|
ASSERT_GT(v3, v2);
|
|
|
|
}
|
|
|
|
|
|
|
|
{
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
iter->Seek("foo");
|
|
|
|
std::string prop_value;
|
|
|
|
ASSERT_OK(iter->GetProperty("rocksdb.iterator.super-version-number",
|
|
|
|
&prop_value));
|
|
|
|
v4 = static_cast<uint64_t>(std::atoi(prop_value.c_str()));
|
|
|
|
}
|
|
|
|
ASSERT_EQ(v3, v4);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, SeekWithUpperBoundBug) {
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
const Slice upper_bound("cc", 3);
|
|
|
|
read_options.iterate_upper_bound = &upper_bound;
|
|
|
|
|
|
|
|
|
|
|
|
// 1st L0 file
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "aa", "SEEN"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
// 2nd L0 file
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "zz", "NOT-SEEN"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
|
|
|
|
iter->Seek("aa");
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "aa");
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBTestTailingIterator, SeekToFirstWithUpperBoundBug) {
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.tailing = true;
|
|
|
|
const Slice upper_bound("cc", 3);
|
|
|
|
read_options.iterate_upper_bound = &upper_bound;
|
|
|
|
|
|
|
|
|
|
|
|
// 1st L0 file
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "aa", "SEEN"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
// 2nd L0 file
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "zz", "NOT-SEEN"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(read_options));
|
|
|
|
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "aa");
|
|
|
|
|
|
|
|
iter->Next();
|
|
|
|
ASSERT_FALSE(iter->Valid());
|
|
|
|
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ(iter->key().ToString(), "aa");
|
|
|
|
}
|
|
|
|
|
|
|
|
} // namespace rocksdb
|
|
|
|
|
|
|
|
#endif // !defined(ROCKSDB_LITE)
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
|
|
#if !defined(ROCKSDB_LITE)
|
|
|
|
rocksdb::port::InstallStackTraceHandler();
|
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
|
|
return RUN_ALL_TESTS();
|
|
|
|
#else
|
|
|
|
return 0;
|
|
|
|
#endif
|
|
|
|
}
|