diff --git a/db/db_basic_test.cc b/db/db_basic_test.cc index 579f5a0fd..efb864c73 100644 --- a/db/db_basic_test.cc +++ b/db/db_basic_test.cc @@ -2407,6 +2407,36 @@ TEST_F(DBMultiGetAsyncIOTest, GetFromL2WithRangeDelInL1) { // Bloom filters in L0/L1 will avoid the coroutine calls in those levels ASSERT_EQ(statistics()->getTickerCount(MULTIGET_COROUTINE_COUNT), 2); } + +TEST_F(DBMultiGetAsyncIOTest, GetFromL1AndL2WithRangeDelInL1) { + std::vector key_strs; + std::vector keys; + std::vector values; + std::vector statuses; + + // 139 and 163 are in L2, but overlap with a range deletes in L1 + key_strs.push_back(Key(139)); + key_strs.push_back(Key(144)); + key_strs.push_back(Key(163)); + keys.push_back(key_strs[0]); + keys.push_back(key_strs[1]); + keys.push_back(key_strs[2]); + values.resize(keys.size()); + statuses.resize(keys.size()); + + ReadOptions ro; + ro.async_io = true; + dbfull()->MultiGet(ro, dbfull()->DefaultColumnFamily(), keys.size(), + keys.data(), values.data(), statuses.data()); + ASSERT_EQ(values.size(), keys.size()); + ASSERT_EQ(statuses[0], Status::NotFound()); + ASSERT_EQ(statuses[1], Status::OK()); + ASSERT_EQ(values[1], "val_l1_" + std::to_string(144)); + ASSERT_EQ(statuses[2], Status::NotFound()); + + // Bloom filters in L0/L1 will avoid the coroutine calls in those levels + ASSERT_EQ(statistics()->getTickerCount(MULTIGET_COROUTINE_COUNT), 3); +} #endif // USE_COROUTINES TEST_F(DBBasicTest, MultiGetStats) { diff --git a/db/table_cache.cc b/db/table_cache.cc index 1b5aa7e22..2dca38c01 100644 --- a/db/table_cache.cc +++ b/db/table_cache.cc @@ -557,17 +557,15 @@ Status TableCache::MultiGetFilter( if (s.ok()) { s = t->MultiGetFilter(options, prefix_extractor.get(), mget_range); } - if (mget_range->empty()) { - if (s.ok() && !options.ignore_range_deletions) { - // If all the keys have been filtered out by the bloom filter, then - // update the range tombstone sequence numbers for the keys as - // MultiGet() will not be called for this set of keys. - UpdateRangeTombstoneSeqnums(options, t, tombstone_range); - } - if (handle) { - ReleaseHandle(handle); - *table_handle = nullptr; - } + if (s.ok() && !options.ignore_range_deletions) { + // Update the range tombstone sequence numbers for the keys here + // as TableCache::MultiGet may or may not be called, and even if it + // is, it may be called with fewer keys in the rangedue to filtering. + UpdateRangeTombstoneSeqnums(options, t, tombstone_range); + } + if (mget_range->empty() && handle) { + ReleaseHandle(handle); + *table_handle = nullptr; } return s; diff --git a/db/table_cache.h b/db/table_cache.h index dae47ede0..2b0205b3e 100644 --- a/db/table_cache.h +++ b/db/table_cache.h @@ -135,7 +135,8 @@ class TableCache { const FileMetaData& file_meta, const MultiGetContext::Range* mget_range, const std::shared_ptr& prefix_extractor = nullptr, HistogramImpl* file_read_hist = nullptr, bool skip_filters = false, - int level = -1, Cache::Handle* table_handle = nullptr); + bool skip_range_deletions = false, int level = -1, + Cache::Handle* table_handle = nullptr); // Evict any entry for the specified file number static void Evict(Cache* cache, uint64_t file_number); diff --git a/db/table_cache_sync_and_async.h b/db/table_cache_sync_and_async.h index 4b3ec6c77..12f04453f 100644 --- a/db/table_cache_sync_and_async.h +++ b/db/table_cache_sync_and_async.h @@ -18,8 +18,8 @@ DEFINE_SYNC_AND_ASYNC(Status, TableCache::MultiGet) (const ReadOptions& options, const InternalKeyComparator& internal_comparator, const FileMetaData& file_meta, const MultiGetContext::Range* mget_range, const std::shared_ptr& prefix_extractor, - HistogramImpl* file_read_hist, bool skip_filters, int level, - Cache::Handle* table_handle) { + HistogramImpl* file_read_hist, bool skip_filters, bool skip_range_deletions, + int level, Cache::Handle* table_handle) { auto& fd = file_meta.fd; Status s; TableReader* t = fd.table_reader; @@ -79,7 +79,7 @@ DEFINE_SYNC_AND_ASYNC(Status, TableCache::MultiGet) assert(t); } } - if (s.ok() && !options.ignore_range_deletions) { + if (s.ok() && !options.ignore_range_deletions && !skip_range_deletions) { UpdateRangeTombstoneSeqnums(options, t, table_range); } if (s.ok()) { diff --git a/db/version_set.cc b/db/version_set.cc index 9d37db84f..5fba4a432 100644 --- a/db/version_set.cc +++ b/db/version_set.cc @@ -2219,7 +2219,8 @@ void Version::MultiGet(const ReadOptions& read_options, MultiGetRange* range, static_cast(fp.GetHitFileLevel()), fp.IsHitFileLastInLevel()); // Call MultiGetFromSST for looking up a single file s = MultiGetFromSST(read_options, fp.CurrentFileRange(), - fp.GetHitFileLevel(), skip_filters, f, blob_ctxs, + fp.GetHitFileLevel(), skip_filters, + /*skip_range_deletions=*/false, f, blob_ctxs, /*table_handle=*/nullptr, num_filter_read, num_index_read, num_sst_read); if (fp.GetHitFileLevel() == 0) { @@ -2237,12 +2238,14 @@ void Version::MultiGet(const ReadOptions& read_options, MultiGetRange* range, Cache::Handle* table_handle = nullptr; bool skip_filters = IsFilterSkipped( static_cast(fp.GetHitFileLevel()), fp.IsHitFileLastInLevel()); + bool skip_range_deletions = false; if (!skip_filters) { Status status = table_cache_->MultiGetFilter( read_options, *internal_comparator(), *f->file_metadata, mutable_cf_options_.prefix_extractor, cfd_->internal_stats()->GetFileReadHist(fp.GetHitFileLevel()), fp.GetHitFileLevel(), &file_range, &table_handle); + skip_range_deletions = true; if (status.ok()) { skip_filters = true; } else if (!status.IsNotSupported()) { @@ -2256,9 +2259,9 @@ void Version::MultiGet(const ReadOptions& read_options, MultiGetRange* range, if (!file_range.empty()) { mget_tasks.emplace_back(MultiGetFromSSTCoroutine( - read_options, file_range, fp.GetHitFileLevel(), skip_filters, f, - blob_ctxs, table_handle, num_filter_read, num_index_read, - num_sst_read)); + read_options, file_range, fp.GetHitFileLevel(), skip_filters, + skip_range_deletions, f, blob_ctxs, table_handle, num_filter_read, + num_index_read, num_sst_read)); } if (fp.KeyMaySpanNextFile()) { break; diff --git a/db/version_set.h b/db/version_set.h index 2b967c1c7..3ad992a78 100644 --- a/db/version_set.h +++ b/db/version_set.h @@ -990,7 +990,8 @@ class Version { DECLARE_SYNC_AND_ASYNC( /* ret_type */ Status, /* func_name */ MultiGetFromSST, const ReadOptions& read_options, MultiGetRange file_range, - int hit_file_level, bool skip_filters, FdWithKeyRange* f, + int hit_file_level, bool skip_filters, bool skip_range_deletions, + FdWithKeyRange* f, std::unordered_map& blob_ctxs, Cache::Handle* table_handle, uint64_t& num_filter_read, uint64_t& num_index_read, uint64_t& num_sst_read); diff --git a/db/version_set_sync_and_async.h b/db/version_set_sync_and_async.h index 954ae8d01..86f0c2077 100644 --- a/db/version_set_sync_and_async.h +++ b/db/version_set_sync_and_async.h @@ -14,7 +14,7 @@ namespace ROCKSDB_NAMESPACE { // Lookup a batch of keys in a single SST file DEFINE_SYNC_AND_ASYNC(Status, Version::MultiGetFromSST) (const ReadOptions& read_options, MultiGetRange file_range, int hit_file_level, - bool skip_filters, FdWithKeyRange* f, + bool skip_filters, bool skip_range_deletions, FdWithKeyRange* f, std::unordered_map& blob_ctxs, Cache::Handle* table_handle, uint64_t& num_filter_read, uint64_t& num_index_read, uint64_t& num_sst_read) { @@ -27,7 +27,7 @@ DEFINE_SYNC_AND_ASYNC(Status, Version::MultiGetFromSST) read_options, *internal_comparator(), *f->file_metadata, &file_range, mutable_cf_options_.prefix_extractor, cfd_->internal_stats()->GetFileReadHist(hit_file_level), skip_filters, - hit_file_level, table_handle); + skip_range_deletions, hit_file_level, table_handle); // TODO: examine the behavior for corrupted key if (timer_enabled) { PERF_COUNTER_BY_LEVEL_ADD(get_from_table_nanos, timer.ElapsedNanos(),