fork of https://github.com/rust-rocksdb/rust-rocksdb for nextgraph
				
			
			
		
			You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							1038 lines
						
					
					
						
							43 KiB
						
					
					
				
			
		
		
	
	
							1038 lines
						
					
					
						
							43 KiB
						
					
					
				| //  Copyright (c) 2011-present, Facebook, Inc.  All rights reserved.
 | |
| //  This source code is licensed under both the GPLv2 (found in the
 | |
| //  COPYING file in the root directory) and Apache 2.0 License
 | |
| //  (found in the LICENSE.Apache file in the root directory).
 | |
| 
 | |
| 
 | |
| #include "utilities/transactions/write_prepared_txn_db.h"
 | |
| 
 | |
| #include <algorithm>
 | |
| #include <cinttypes>
 | |
| #include <string>
 | |
| #include <unordered_set>
 | |
| #include <vector>
 | |
| 
 | |
| #include "db/arena_wrapped_db_iter.h"
 | |
| #include "db/db_impl/db_impl.h"
 | |
| #include "logging/logging.h"
 | |
| #include "rocksdb/db.h"
 | |
| #include "rocksdb/options.h"
 | |
| #include "rocksdb/utilities/transaction_db.h"
 | |
| #include "test_util/sync_point.h"
 | |
| #include "util/cast_util.h"
 | |
| #include "util/mutexlock.h"
 | |
| #include "util/string_util.h"
 | |
| #include "utilities/transactions/pessimistic_transaction.h"
 | |
| #include "utilities/transactions/transaction_db_mutex_impl.h"
 | |
| 
 | |
| // This function is for testing only. If it returns true, then all entries in
 | |
| // the commit cache will be evicted. Unit and/or stress tests (db_stress)
 | |
| // can implement this function and customize how frequently commit cache
 | |
| // eviction occurs.
 | |
| // TODO: remove this function once we can configure commit cache to be very
 | |
| // small so that eviction occurs very frequently. This requires the commit
 | |
| // cache entry to be able to encode prepare and commit sequence numbers so that
 | |
| // the commit sequence number does not have to be within a certain range of
 | |
| // prepare sequence number.
 | |
| extern "C" bool rocksdb_write_prepared_TEST_ShouldClearCommitCache(void)
 | |
|     __attribute__((__weak__));
 | |
| 
 | |
| namespace ROCKSDB_NAMESPACE {
 | |
| 
 | |
| Status WritePreparedTxnDB::Initialize(
 | |
|     const std::vector<size_t>& compaction_enabled_cf_indices,
 | |
|     const std::vector<ColumnFamilyHandle*>& handles) {
 | |
|   auto dbimpl = static_cast_with_check<DBImpl>(GetRootDB());
 | |
|   assert(dbimpl != nullptr);
 | |
|   auto rtxns = dbimpl->recovered_transactions();
 | |
|   std::map<SequenceNumber, SequenceNumber> ordered_seq_cnt;
 | |
|   for (auto rtxn : rtxns) {
 | |
|     // There should only one batch for WritePrepared policy.
 | |
|     assert(rtxn.second->batches_.size() == 1);
 | |
|     const auto& seq = rtxn.second->batches_.begin()->first;
 | |
|     const auto& batch_info = rtxn.second->batches_.begin()->second;
 | |
|     auto cnt = batch_info.batch_cnt_ ? batch_info.batch_cnt_ : 1;
 | |
|     ordered_seq_cnt[seq] = cnt;
 | |
|   }
 | |
|   // AddPrepared must be called in order
 | |
|   for (auto seq_cnt : ordered_seq_cnt) {
 | |
|     auto seq = seq_cnt.first;
 | |
|     auto cnt = seq_cnt.second;
 | |
|     for (size_t i = 0; i < cnt; i++) {
 | |
|       AddPrepared(seq + i);
 | |
|     }
 | |
|   }
 | |
|   SequenceNumber prev_max = max_evicted_seq_;
 | |
|   SequenceNumber last_seq = db_impl_->GetLatestSequenceNumber();
 | |
|   AdvanceMaxEvictedSeq(prev_max, last_seq);
 | |
|   // Create a gap between max and the next snapshot. This simplifies the logic
 | |
|   // in IsInSnapshot by not having to consider the special case of max ==
 | |
|   // snapshot after recovery. This is tested in IsInSnapshotEmptyMapTest.
 | |
|   if (last_seq) {
 | |
|     db_impl_->versions_->SetLastAllocatedSequence(last_seq + 1);
 | |
|     db_impl_->versions_->SetLastSequence(last_seq + 1);
 | |
|     db_impl_->versions_->SetLastPublishedSequence(last_seq + 1);
 | |
|   }
 | |
| 
 | |
|   db_impl_->SetSnapshotChecker(new WritePreparedSnapshotChecker(this));
 | |
|   // A callback to commit a single sub-batch
 | |
|   class CommitSubBatchPreReleaseCallback : public PreReleaseCallback {
 | |
|    public:
 | |
|     explicit CommitSubBatchPreReleaseCallback(WritePreparedTxnDB* db)
 | |
|         : db_(db) {}
 | |
|     Status Callback(SequenceNumber commit_seq,
 | |
|                     bool is_mem_disabled __attribute__((__unused__)), uint64_t,
 | |
|                     size_t /*index*/, size_t /*total*/) override {
 | |
|       assert(!is_mem_disabled);
 | |
|       db_->AddCommitted(commit_seq, commit_seq);
 | |
|       return Status::OK();
 | |
|     }
 | |
| 
 | |
|    private:
 | |
|     WritePreparedTxnDB* db_;
 | |
|   };
 | |
|   db_impl_->SetRecoverableStatePreReleaseCallback(
 | |
|       new CommitSubBatchPreReleaseCallback(this));
 | |
| 
 | |
|   auto s = PessimisticTransactionDB::Initialize(compaction_enabled_cf_indices,
 | |
|                                                 handles);
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| Status WritePreparedTxnDB::VerifyCFOptions(
 | |
|     const ColumnFamilyOptions& cf_options) {
 | |
|   Status s = PessimisticTransactionDB::VerifyCFOptions(cf_options);
 | |
|   if (!s.ok()) {
 | |
|     return s;
 | |
|   }
 | |
|   if (!cf_options.memtable_factory->CanHandleDuplicatedKey()) {
 | |
|     return Status::InvalidArgument(
 | |
|         "memtable_factory->CanHandleDuplicatedKey() cannot be false with "
 | |
|         "WritePrpeared transactions");
 | |
|   }
 | |
|   return Status::OK();
 | |
| }
 | |
| 
 | |
| Transaction* WritePreparedTxnDB::BeginTransaction(
 | |
|     const WriteOptions& write_options, const TransactionOptions& txn_options,
 | |
|     Transaction* old_txn) {
 | |
|   if (old_txn != nullptr) {
 | |
|     ReinitializeTransaction(old_txn, write_options, txn_options);
 | |
|     return old_txn;
 | |
|   } else {
 | |
|     return new WritePreparedTxn(this, write_options, txn_options);
 | |
|   }
 | |
| }
 | |
| 
 | |
| Status WritePreparedTxnDB::Write(const WriteOptions& opts,
 | |
|                                  WriteBatch* updates) {
 | |
|   if (txn_db_options_.skip_concurrency_control) {
 | |
|     // Skip locking the rows
 | |
|     const size_t UNKNOWN_BATCH_CNT = 0;
 | |
|     WritePreparedTxn* NO_TXN = nullptr;
 | |
|     return WriteInternal(opts, updates, UNKNOWN_BATCH_CNT, NO_TXN);
 | |
|   } else {
 | |
|     return PessimisticTransactionDB::WriteWithConcurrencyControl(opts, updates);
 | |
|   }
 | |
| }
 | |
| 
 | |
| Status WritePreparedTxnDB::Write(
 | |
|     const WriteOptions& opts,
 | |
|     const TransactionDBWriteOptimizations& optimizations, WriteBatch* updates) {
 | |
|   if (optimizations.skip_concurrency_control) {
 | |
|     // Skip locking the rows
 | |
|     const size_t UNKNOWN_BATCH_CNT = 0;
 | |
|     const size_t ONE_BATCH_CNT = 1;
 | |
|     const size_t batch_cnt = optimizations.skip_duplicate_key_check
 | |
|                                  ? ONE_BATCH_CNT
 | |
|                                  : UNKNOWN_BATCH_CNT;
 | |
|     WritePreparedTxn* NO_TXN = nullptr;
 | |
|     return WriteInternal(opts, updates, batch_cnt, NO_TXN);
 | |
|   } else {
 | |
|     // TODO(myabandeh): Make use of skip_duplicate_key_check hint
 | |
|     // Fall back to unoptimized version
 | |
|     return PessimisticTransactionDB::WriteWithConcurrencyControl(opts, updates);
 | |
|   }
 | |
| }
 | |
| 
 | |
| Status WritePreparedTxnDB::WriteInternal(const WriteOptions& write_options_orig,
 | |
|                                          WriteBatch* batch, size_t batch_cnt,
 | |
|                                          WritePreparedTxn* txn) {
 | |
|   ROCKS_LOG_DETAILS(db_impl_->immutable_db_options().info_log,
 | |
|                     "CommitBatchInternal");
 | |
|   if (batch->Count() == 0) {
 | |
|     // Otherwise our 1 seq per batch logic will break since there is no seq
 | |
|     // increased for this batch.
 | |
|     return Status::OK();
 | |
|   }
 | |
| 
 | |
|   if (write_options_orig.protection_bytes_per_key > 0) {
 | |
|     auto s = WriteBatchInternal::UpdateProtectionInfo(
 | |
|         batch, write_options_orig.protection_bytes_per_key);
 | |
|     if (!s.ok()) {
 | |
|       return s;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   if (batch_cnt == 0) {  // not provided, then compute it
 | |
|     // TODO(myabandeh): add an option to allow user skipping this cost
 | |
|     SubBatchCounter counter(*GetCFComparatorMap());
 | |
|     auto s = batch->Iterate(&counter);
 | |
|     if (!s.ok()) {
 | |
|       return s;
 | |
|     }
 | |
|     batch_cnt = counter.BatchCount();
 | |
|     WPRecordTick(TXN_DUPLICATE_KEY_OVERHEAD);
 | |
|     ROCKS_LOG_DETAILS(info_log_, "Duplicate key overhead: %" PRIu64 " batches",
 | |
|                       static_cast<uint64_t>(batch_cnt));
 | |
|   }
 | |
|   assert(batch_cnt);
 | |
| 
 | |
|   bool do_one_write = !db_impl_->immutable_db_options().two_write_queues;
 | |
|   WriteOptions write_options(write_options_orig);
 | |
|   // In the absence of Prepare markers, use Noop as a batch separator
 | |
|   auto s = WriteBatchInternal::InsertNoop(batch);
 | |
|   assert(s.ok());
 | |
|   const bool DISABLE_MEMTABLE = true;
 | |
|   const uint64_t no_log_ref = 0;
 | |
|   uint64_t seq_used = kMaxSequenceNumber;
 | |
|   const size_t ZERO_PREPARES = 0;
 | |
|   const bool kSeperatePrepareCommitBatches = true;
 | |
|   // Since this is not 2pc, there is no need for AddPrepared but having it in
 | |
|   // the PreReleaseCallback enables an optimization. Refer to
 | |
|   // SmallestUnCommittedSeq for more details.
 | |
|   AddPreparedCallback add_prepared_callback(
 | |
|       this, db_impl_, batch_cnt,
 | |
|       db_impl_->immutable_db_options().two_write_queues,
 | |
|       !kSeperatePrepareCommitBatches);
 | |
|   WritePreparedCommitEntryPreReleaseCallback update_commit_map(
 | |
|       this, db_impl_, kMaxSequenceNumber, ZERO_PREPARES, batch_cnt);
 | |
|   PreReleaseCallback* pre_release_callback;
 | |
|   if (do_one_write) {
 | |
|     pre_release_callback = &update_commit_map;
 | |
|   } else {
 | |
|     pre_release_callback = &add_prepared_callback;
 | |
|   }
 | |
|   s = db_impl_->WriteImpl(write_options, batch, nullptr, nullptr, no_log_ref,
 | |
|                           !DISABLE_MEMTABLE, &seq_used, batch_cnt,
 | |
|                           pre_release_callback);
 | |
|   assert(!s.ok() || seq_used != kMaxSequenceNumber);
 | |
|   uint64_t prepare_seq = seq_used;
 | |
|   if (txn != nullptr) {
 | |
|     txn->SetId(prepare_seq);
 | |
|   }
 | |
|   if (!s.ok()) {
 | |
|     return s;
 | |
|   }
 | |
|   if (do_one_write) {
 | |
|     return s;
 | |
|   }  // else do the 2nd write for commit
 | |
|   ROCKS_LOG_DETAILS(db_impl_->immutable_db_options().info_log,
 | |
|                     "CommitBatchInternal 2nd write prepare_seq: %" PRIu64,
 | |
|                     prepare_seq);
 | |
|   // Commit the batch by writing an empty batch to the 2nd queue that will
 | |
|   // release the commit sequence number to readers.
 | |
|   const size_t ZERO_COMMITS = 0;
 | |
|   WritePreparedCommitEntryPreReleaseCallback update_commit_map_with_prepare(
 | |
|       this, db_impl_, prepare_seq, batch_cnt, ZERO_COMMITS);
 | |
|   WriteBatch empty_batch;
 | |
|   write_options.disableWAL = true;
 | |
|   write_options.sync = false;
 | |
|   const size_t ONE_BATCH = 1;  // Just to inc the seq
 | |
|   s = db_impl_->WriteImpl(write_options, &empty_batch, nullptr, nullptr,
 | |
|                           no_log_ref, DISABLE_MEMTABLE, &seq_used, ONE_BATCH,
 | |
|                           &update_commit_map_with_prepare);
 | |
|   assert(!s.ok() || seq_used != kMaxSequenceNumber);
 | |
|   // Note: RemovePrepared is called from within PreReleaseCallback
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| Status WritePreparedTxnDB::Get(const ReadOptions& options,
 | |
|                                ColumnFamilyHandle* column_family,
 | |
|                                const Slice& key, PinnableSlice* value) {
 | |
|   if (options.io_activity != Env::IOActivity::kUnknown) {
 | |
|     return Status::InvalidArgument(
 | |
|         "Cannot call Get with `ReadOptions::io_activity` != "
 | |
|         "`Env::IOActivity::kUnknown`");
 | |
|   }
 | |
|   SequenceNumber min_uncommitted, snap_seq;
 | |
|   const SnapshotBackup backed_by_snapshot =
 | |
|       AssignMinMaxSeqs(options.snapshot, &min_uncommitted, &snap_seq);
 | |
|   WritePreparedTxnReadCallback callback(this, snap_seq, min_uncommitted,
 | |
|                                         backed_by_snapshot);
 | |
|   bool* dont_care = nullptr;
 | |
|   DBImpl::GetImplOptions get_impl_options;
 | |
|   get_impl_options.column_family = column_family;
 | |
|   get_impl_options.value = value;
 | |
|   get_impl_options.value_found = dont_care;
 | |
|   get_impl_options.callback = &callback;
 | |
|   auto res = db_impl_->GetImpl(options, key, get_impl_options);
 | |
|   if (LIKELY(callback.valid() && ValidateSnapshot(callback.max_visible_seq(),
 | |
|                                                   backed_by_snapshot))) {
 | |
|     return res;
 | |
|   } else {
 | |
|     res.PermitUncheckedError();
 | |
|     WPRecordTick(TXN_GET_TRY_AGAIN);
 | |
|     return Status::TryAgain();
 | |
|   }
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::UpdateCFComparatorMap(
 | |
|     const std::vector<ColumnFamilyHandle*>& handles) {
 | |
|   auto cf_map = new std::map<uint32_t, const Comparator*>();
 | |
|   auto handle_map = new std::map<uint32_t, ColumnFamilyHandle*>();
 | |
|   for (auto h : handles) {
 | |
|     auto id = h->GetID();
 | |
|     const Comparator* comparator = h->GetComparator();
 | |
|     (*cf_map)[id] = comparator;
 | |
|     if (id != 0) {
 | |
|       (*handle_map)[id] = h;
 | |
|     } else {
 | |
|       // The pointer to the default cf handle in the handles will be deleted.
 | |
|       // Use the pointer maintained by the db instead.
 | |
|       (*handle_map)[id] = DefaultColumnFamily();
 | |
|     }
 | |
|   }
 | |
|   cf_map_.reset(cf_map);
 | |
|   handle_map_.reset(handle_map);
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::UpdateCFComparatorMap(ColumnFamilyHandle* h) {
 | |
|   auto old_cf_map_ptr = cf_map_.get();
 | |
|   assert(old_cf_map_ptr);
 | |
|   auto cf_map = new std::map<uint32_t, const Comparator*>(*old_cf_map_ptr);
 | |
|   auto old_handle_map_ptr = handle_map_.get();
 | |
|   assert(old_handle_map_ptr);
 | |
|   auto handle_map =
 | |
|       new std::map<uint32_t, ColumnFamilyHandle*>(*old_handle_map_ptr);
 | |
|   auto id = h->GetID();
 | |
|   const Comparator* comparator = h->GetComparator();
 | |
|   (*cf_map)[id] = comparator;
 | |
|   (*handle_map)[id] = h;
 | |
|   cf_map_.reset(cf_map);
 | |
|   handle_map_.reset(handle_map);
 | |
| }
 | |
| 
 | |
| std::vector<Status> WritePreparedTxnDB::MultiGet(
 | |
|     const ReadOptions& options,
 | |
|     const std::vector<ColumnFamilyHandle*>& column_family,
 | |
|     const std::vector<Slice>& keys, std::vector<std::string>* values) {
 | |
|   assert(values);
 | |
|   size_t num_keys = keys.size();
 | |
|   values->resize(num_keys);
 | |
| 
 | |
|   std::vector<Status> stat_list(num_keys);
 | |
|   for (size_t i = 0; i < num_keys; ++i) {
 | |
|     stat_list[i] = this->Get(options, column_family[i], keys[i], &(*values)[i]);
 | |
|   }
 | |
|   return stat_list;
 | |
| }
 | |
| 
 | |
| // Struct to hold ownership of snapshot and read callback for iterator cleanup.
 | |
| struct WritePreparedTxnDB::IteratorState {
 | |
|   IteratorState(WritePreparedTxnDB* txn_db, SequenceNumber sequence,
 | |
|                 std::shared_ptr<ManagedSnapshot> s,
 | |
|                 SequenceNumber min_uncommitted)
 | |
|       : callback(txn_db, sequence, min_uncommitted, kBackedByDBSnapshot),
 | |
|         snapshot(s) {}
 | |
| 
 | |
|   WritePreparedTxnReadCallback callback;
 | |
|   std::shared_ptr<ManagedSnapshot> snapshot;
 | |
| };
 | |
| 
 | |
| namespace {
 | |
| static void CleanupWritePreparedTxnDBIterator(void* arg1, void* /*arg2*/) {
 | |
|   delete reinterpret_cast<WritePreparedTxnDB::IteratorState*>(arg1);
 | |
| }
 | |
| }  // anonymous namespace
 | |
| 
 | |
| Iterator* WritePreparedTxnDB::NewIterator(const ReadOptions& options,
 | |
|                                           ColumnFamilyHandle* column_family) {
 | |
|   if (options.io_activity != Env::IOActivity::kUnknown) {
 | |
|     return NewErrorIterator(Status::InvalidArgument(
 | |
|         "Cannot call NewIterator with `ReadOptions::io_activity` != "
 | |
|         "`Env::IOActivity::kUnknown`"));
 | |
|   }
 | |
|   constexpr bool expose_blob_index = false;
 | |
|   constexpr bool allow_refresh = false;
 | |
|   std::shared_ptr<ManagedSnapshot> own_snapshot = nullptr;
 | |
|   SequenceNumber snapshot_seq = kMaxSequenceNumber;
 | |
|   SequenceNumber min_uncommitted = 0;
 | |
|   if (options.snapshot != nullptr) {
 | |
|     snapshot_seq = options.snapshot->GetSequenceNumber();
 | |
|     min_uncommitted =
 | |
|         static_cast_with_check<const SnapshotImpl>(options.snapshot)
 | |
|             ->min_uncommitted_;
 | |
|   } else {
 | |
|     auto* snapshot = GetSnapshot();
 | |
|     // We take a snapshot to make sure that the related data in the commit map
 | |
|     // are not deleted.
 | |
|     snapshot_seq = snapshot->GetSequenceNumber();
 | |
|     min_uncommitted =
 | |
|         static_cast_with_check<const SnapshotImpl>(snapshot)->min_uncommitted_;
 | |
|     own_snapshot = std::make_shared<ManagedSnapshot>(db_impl_, snapshot);
 | |
|   }
 | |
|   assert(snapshot_seq != kMaxSequenceNumber);
 | |
|   auto* cfd =
 | |
|       static_cast_with_check<ColumnFamilyHandleImpl>(column_family)->cfd();
 | |
|   auto* state =
 | |
|       new IteratorState(this, snapshot_seq, own_snapshot, min_uncommitted);
 | |
|   auto* db_iter =
 | |
|       db_impl_->NewIteratorImpl(options, cfd, snapshot_seq, &state->callback,
 | |
|                                 expose_blob_index, allow_refresh);
 | |
|   db_iter->RegisterCleanup(CleanupWritePreparedTxnDBIterator, state, nullptr);
 | |
|   return db_iter;
 | |
| }
 | |
| 
 | |
| Status WritePreparedTxnDB::NewIterators(
 | |
|     const ReadOptions& options,
 | |
|     const std::vector<ColumnFamilyHandle*>& column_families,
 | |
|     std::vector<Iterator*>* iterators) {
 | |
|   constexpr bool expose_blob_index = false;
 | |
|   constexpr bool allow_refresh = false;
 | |
|   std::shared_ptr<ManagedSnapshot> own_snapshot = nullptr;
 | |
|   SequenceNumber snapshot_seq = kMaxSequenceNumber;
 | |
|   SequenceNumber min_uncommitted = 0;
 | |
|   if (options.snapshot != nullptr) {
 | |
|     snapshot_seq = options.snapshot->GetSequenceNumber();
 | |
|     min_uncommitted =
 | |
|         static_cast_with_check<const SnapshotImpl>(options.snapshot)
 | |
|             ->min_uncommitted_;
 | |
|   } else {
 | |
|     auto* snapshot = GetSnapshot();
 | |
|     // We take a snapshot to make sure that the related data in the commit map
 | |
|     // are not deleted.
 | |
|     snapshot_seq = snapshot->GetSequenceNumber();
 | |
|     own_snapshot = std::make_shared<ManagedSnapshot>(db_impl_, snapshot);
 | |
|     min_uncommitted =
 | |
|         static_cast_with_check<const SnapshotImpl>(snapshot)->min_uncommitted_;
 | |
|   }
 | |
|   iterators->clear();
 | |
|   iterators->reserve(column_families.size());
 | |
|   for (auto* column_family : column_families) {
 | |
|     auto* cfd =
 | |
|         static_cast_with_check<ColumnFamilyHandleImpl>(column_family)->cfd();
 | |
|     auto* state =
 | |
|         new IteratorState(this, snapshot_seq, own_snapshot, min_uncommitted);
 | |
|     auto* db_iter =
 | |
|         db_impl_->NewIteratorImpl(options, cfd, snapshot_seq, &state->callback,
 | |
|                                   expose_blob_index, allow_refresh);
 | |
|     db_iter->RegisterCleanup(CleanupWritePreparedTxnDBIterator, state, nullptr);
 | |
|     iterators->push_back(db_iter);
 | |
|   }
 | |
|   return Status::OK();
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::Init(const TransactionDBOptions& txn_db_opts) {
 | |
|   // Adcance max_evicted_seq_ no more than 100 times before the cache wraps
 | |
|   // around.
 | |
|   INC_STEP_FOR_MAX_EVICTED =
 | |
|       std::max(COMMIT_CACHE_SIZE / 100, static_cast<size_t>(1));
 | |
|   snapshot_cache_ = std::unique_ptr<std::atomic<SequenceNumber>[]>(
 | |
|       new std::atomic<SequenceNumber>[SNAPSHOT_CACHE_SIZE] {});
 | |
|   commit_cache_ = std::unique_ptr<std::atomic<CommitEntry64b>[]>(
 | |
|       new std::atomic<CommitEntry64b>[COMMIT_CACHE_SIZE] {});
 | |
|   dummy_max_snapshot_.number_ = kMaxSequenceNumber;
 | |
|   rollback_deletion_type_callback_ =
 | |
|       txn_db_opts.rollback_deletion_type_callback;
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::CheckPreparedAgainstMax(SequenceNumber new_max,
 | |
|                                                  bool locked) {
 | |
|   // When max_evicted_seq_ advances, move older entries from prepared_txns_
 | |
|   // to delayed_prepared_. This guarantees that if a seq is lower than max,
 | |
|   // then it is not in prepared_txns_ and save an expensive, synchronized
 | |
|   // lookup from a shared set. delayed_prepared_ is expected to be empty in
 | |
|   // normal cases.
 | |
|   ROCKS_LOG_DETAILS(
 | |
|       info_log_,
 | |
|       "CheckPreparedAgainstMax prepared_txns_.empty() %d top: %" PRIu64,
 | |
|       prepared_txns_.empty(),
 | |
|       prepared_txns_.empty() ? 0 : prepared_txns_.top());
 | |
|   const SequenceNumber prepared_top = prepared_txns_.top();
 | |
|   const bool empty = prepared_top == kMaxSequenceNumber;
 | |
|   // Preliminary check to avoid the synchronization cost
 | |
|   if (!empty && prepared_top <= new_max) {
 | |
|     if (locked) {
 | |
|       // Needed to avoid double locking in pop().
 | |
|       prepared_txns_.push_pop_mutex()->Unlock();
 | |
|     }
 | |
|     WriteLock wl(&prepared_mutex_);
 | |
|     // Need to fetch fresh values of ::top after mutex is acquired
 | |
|     while (!prepared_txns_.empty() && prepared_txns_.top() <= new_max) {
 | |
|       auto to_be_popped = prepared_txns_.top();
 | |
|       delayed_prepared_.insert(to_be_popped);
 | |
|       ROCKS_LOG_WARN(info_log_,
 | |
|                      "prepared_mutex_ overhead %" PRIu64 " (prep=%" PRIu64
 | |
|                      " new_max=%" PRIu64 ")",
 | |
|                      static_cast<uint64_t>(delayed_prepared_.size()),
 | |
|                      to_be_popped, new_max);
 | |
|       delayed_prepared_empty_.store(false, std::memory_order_release);
 | |
|       // Update prepared_txns_ after updating delayed_prepared_empty_ otherwise
 | |
|       // there will be a point in time that the entry is neither in
 | |
|       // prepared_txns_ nor in delayed_prepared_, which will not be checked if
 | |
|       // delayed_prepared_empty_ is false.
 | |
|       prepared_txns_.pop();
 | |
|     }
 | |
|     if (locked) {
 | |
|       prepared_txns_.push_pop_mutex()->Lock();
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::AddPrepared(uint64_t seq, bool locked) {
 | |
|   ROCKS_LOG_DETAILS(info_log_, "Txn %" PRIu64 " Preparing with max %" PRIu64,
 | |
|                     seq, max_evicted_seq_.load());
 | |
|   TEST_SYNC_POINT("AddPrepared::begin:pause");
 | |
|   TEST_SYNC_POINT("AddPrepared::begin:resume");
 | |
|   if (!locked) {
 | |
|     prepared_txns_.push_pop_mutex()->Lock();
 | |
|   }
 | |
|   prepared_txns_.push_pop_mutex()->AssertHeld();
 | |
|   prepared_txns_.push(seq);
 | |
|   auto new_max = future_max_evicted_seq_.load();
 | |
|   if (UNLIKELY(seq <= new_max)) {
 | |
|     // This should not happen in normal case
 | |
|     ROCKS_LOG_ERROR(
 | |
|         info_log_,
 | |
|         "Added prepare_seq is not larger than max_evicted_seq_: %" PRIu64
 | |
|         " <= %" PRIu64,
 | |
|         seq, new_max);
 | |
|     CheckPreparedAgainstMax(new_max, true /*locked*/);
 | |
|   }
 | |
|   if (!locked) {
 | |
|     prepared_txns_.push_pop_mutex()->Unlock();
 | |
|   }
 | |
|   TEST_SYNC_POINT("AddPrepared::end");
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::AddCommitted(uint64_t prepare_seq, uint64_t commit_seq,
 | |
|                                       uint8_t loop_cnt) {
 | |
|   ROCKS_LOG_DETAILS(info_log_, "Txn %" PRIu64 " Committing with %" PRIu64,
 | |
|                     prepare_seq, commit_seq);
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::AddCommitted:start");
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::AddCommitted:start:pause");
 | |
|   auto indexed_seq = prepare_seq % COMMIT_CACHE_SIZE;
 | |
|   CommitEntry64b evicted_64b;
 | |
|   CommitEntry evicted;
 | |
|   bool to_be_evicted = GetCommitEntry(indexed_seq, &evicted_64b, &evicted);
 | |
|   if (LIKELY(to_be_evicted)) {
 | |
|     assert(evicted.prep_seq != prepare_seq);
 | |
|     auto prev_max = max_evicted_seq_.load(std::memory_order_acquire);
 | |
|     ROCKS_LOG_DETAILS(info_log_,
 | |
|                       "Evicting %" PRIu64 ",%" PRIu64 " with max %" PRIu64,
 | |
|                       evicted.prep_seq, evicted.commit_seq, prev_max);
 | |
|     if (prev_max < evicted.commit_seq) {
 | |
|       auto last = db_impl_->GetLastPublishedSequence();  // could be 0
 | |
|       SequenceNumber max_evicted_seq;
 | |
|       if (LIKELY(evicted.commit_seq < last)) {
 | |
|         assert(last > 0);
 | |
|         // Inc max in larger steps to avoid frequent updates
 | |
|         max_evicted_seq =
 | |
|             std::min(evicted.commit_seq + INC_STEP_FOR_MAX_EVICTED, last - 1);
 | |
|       } else {
 | |
|         // legit when a commit entry in a write batch overwrite the previous one
 | |
|         max_evicted_seq = evicted.commit_seq;
 | |
|       }
 | |
| #ifdef OS_LINUX
 | |
|       if (rocksdb_write_prepared_TEST_ShouldClearCommitCache &&
 | |
|           rocksdb_write_prepared_TEST_ShouldClearCommitCache()) {
 | |
|         max_evicted_seq = last;
 | |
|       }
 | |
| #endif  // OS_LINUX
 | |
|       ROCKS_LOG_DETAILS(info_log_,
 | |
|                         "%lu Evicting %" PRIu64 ",%" PRIu64 " with max %" PRIu64
 | |
|                         " => %lu",
 | |
|                         prepare_seq, evicted.prep_seq, evicted.commit_seq,
 | |
|                         prev_max, max_evicted_seq);
 | |
|       AdvanceMaxEvictedSeq(prev_max, max_evicted_seq);
 | |
|     }
 | |
|     if (UNLIKELY(!delayed_prepared_empty_.load(std::memory_order_acquire))) {
 | |
|       WriteLock wl(&prepared_mutex_);
 | |
|       auto dp_iter = delayed_prepared_.find(evicted.prep_seq);
 | |
|       if (dp_iter != delayed_prepared_.end()) {
 | |
|         // This is a rare case that txn is committed but prepared_txns_ is not
 | |
|         // cleaned up yet. Refer to delayed_prepared_commits_ definition for
 | |
|         // why it should be kept updated.
 | |
|         delayed_prepared_commits_[evicted.prep_seq] = evicted.commit_seq;
 | |
|         ROCKS_LOG_DEBUG(info_log_,
 | |
|                         "delayed_prepared_commits_[%" PRIu64 "]=%" PRIu64,
 | |
|                         evicted.prep_seq, evicted.commit_seq);
 | |
|       }
 | |
|     }
 | |
|     // After each eviction from commit cache, check if the commit entry should
 | |
|     // be kept around because it overlaps with a live snapshot.
 | |
|     CheckAgainstSnapshots(evicted);
 | |
|   }
 | |
|   bool succ =
 | |
|       ExchangeCommitEntry(indexed_seq, evicted_64b, {prepare_seq, commit_seq});
 | |
|   if (UNLIKELY(!succ)) {
 | |
|     ROCKS_LOG_ERROR(info_log_,
 | |
|                     "ExchangeCommitEntry failed on [%" PRIu64 "] %" PRIu64
 | |
|                     ",%" PRIu64 " retrying...",
 | |
|                     indexed_seq, prepare_seq, commit_seq);
 | |
|     // A very rare event, in which the commit entry is updated before we do.
 | |
|     // Here we apply a very simple solution of retrying.
 | |
|     if (loop_cnt > 100) {
 | |
|       throw std::runtime_error("Infinite loop in AddCommitted!");
 | |
|     }
 | |
|     AddCommitted(prepare_seq, commit_seq, ++loop_cnt);
 | |
|     return;
 | |
|   }
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::AddCommitted:end");
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::AddCommitted:end:pause");
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::RemovePrepared(const uint64_t prepare_seq,
 | |
|                                         const size_t batch_cnt) {
 | |
|   TEST_SYNC_POINT_CALLBACK(
 | |
|       "RemovePrepared:Start",
 | |
|       const_cast<void*>(reinterpret_cast<const void*>(&prepare_seq)));
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::RemovePrepared:pause");
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::RemovePrepared:resume");
 | |
|   ROCKS_LOG_DETAILS(info_log_,
 | |
|                     "RemovePrepared %" PRIu64 " cnt: %" ROCKSDB_PRIszt,
 | |
|                     prepare_seq, batch_cnt);
 | |
|   WriteLock wl(&prepared_mutex_);
 | |
|   for (size_t i = 0; i < batch_cnt; i++) {
 | |
|     prepared_txns_.erase(prepare_seq + i);
 | |
|     bool was_empty = delayed_prepared_.empty();
 | |
|     if (!was_empty) {
 | |
|       delayed_prepared_.erase(prepare_seq + i);
 | |
|       auto it = delayed_prepared_commits_.find(prepare_seq + i);
 | |
|       if (it != delayed_prepared_commits_.end()) {
 | |
|         ROCKS_LOG_DETAILS(info_log_, "delayed_prepared_commits_.erase %" PRIu64,
 | |
|                           prepare_seq + i);
 | |
|         delayed_prepared_commits_.erase(it);
 | |
|       }
 | |
|       bool is_empty = delayed_prepared_.empty();
 | |
|       if (was_empty != is_empty) {
 | |
|         delayed_prepared_empty_.store(is_empty, std::memory_order_release);
 | |
|       }
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| bool WritePreparedTxnDB::GetCommitEntry(const uint64_t indexed_seq,
 | |
|                                         CommitEntry64b* entry_64b,
 | |
|                                         CommitEntry* entry) const {
 | |
|   *entry_64b = commit_cache_[static_cast<size_t>(indexed_seq)].load(
 | |
|       std::memory_order_acquire);
 | |
|   bool valid = entry_64b->Parse(indexed_seq, entry, FORMAT);
 | |
|   return valid;
 | |
| }
 | |
| 
 | |
| bool WritePreparedTxnDB::AddCommitEntry(const uint64_t indexed_seq,
 | |
|                                         const CommitEntry& new_entry,
 | |
|                                         CommitEntry* evicted_entry) {
 | |
|   CommitEntry64b new_entry_64b(new_entry, FORMAT);
 | |
|   CommitEntry64b evicted_entry_64b =
 | |
|       commit_cache_[static_cast<size_t>(indexed_seq)].exchange(
 | |
|           new_entry_64b, std::memory_order_acq_rel);
 | |
|   bool valid = evicted_entry_64b.Parse(indexed_seq, evicted_entry, FORMAT);
 | |
|   return valid;
 | |
| }
 | |
| 
 | |
| bool WritePreparedTxnDB::ExchangeCommitEntry(const uint64_t indexed_seq,
 | |
|                                              CommitEntry64b& expected_entry_64b,
 | |
|                                              const CommitEntry& new_entry) {
 | |
|   auto& atomic_entry = commit_cache_[static_cast<size_t>(indexed_seq)];
 | |
|   CommitEntry64b new_entry_64b(new_entry, FORMAT);
 | |
|   bool succ = atomic_entry.compare_exchange_strong(
 | |
|       expected_entry_64b, new_entry_64b, std::memory_order_acq_rel,
 | |
|       std::memory_order_acquire);
 | |
|   return succ;
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::AdvanceMaxEvictedSeq(const SequenceNumber& prev_max,
 | |
|                                               const SequenceNumber& new_max) {
 | |
|   ROCKS_LOG_DETAILS(info_log_,
 | |
|                     "AdvanceMaxEvictedSeq overhead %" PRIu64 " => %" PRIu64,
 | |
|                     prev_max, new_max);
 | |
|   // Declare the intention before getting snapshot from the DB. This helps a
 | |
|   // concurrent GetSnapshot to wait to catch up with future_max_evicted_seq_ if
 | |
|   // it has not already. Otherwise the new snapshot is when we ask DB for
 | |
|   // snapshots smaller than future max.
 | |
|   auto updated_future_max = prev_max;
 | |
|   while (updated_future_max < new_max &&
 | |
|          !future_max_evicted_seq_.compare_exchange_weak(
 | |
|              updated_future_max, new_max, std::memory_order_acq_rel,
 | |
|              std::memory_order_relaxed)) {
 | |
|   };
 | |
| 
 | |
|   CheckPreparedAgainstMax(new_max, false /*locked*/);
 | |
| 
 | |
|   // With each change to max_evicted_seq_ fetch the live snapshots behind it.
 | |
|   // We use max as the version of snapshots to identify how fresh are the
 | |
|   // snapshot list. This works because the snapshots are between 0 and
 | |
|   // max, so the larger the max, the more complete they are.
 | |
|   SequenceNumber new_snapshots_version = new_max;
 | |
|   std::vector<SequenceNumber> snapshots;
 | |
|   bool update_snapshots = false;
 | |
|   if (new_snapshots_version > snapshots_version_) {
 | |
|     // This is to avoid updating the snapshots_ if it already updated
 | |
|     // with a more recent vesion by a concrrent thread
 | |
|     update_snapshots = true;
 | |
|     // We only care about snapshots lower then max
 | |
|     snapshots = GetSnapshotListFromDB(new_max);
 | |
|   }
 | |
|   if (update_snapshots) {
 | |
|     UpdateSnapshots(snapshots, new_snapshots_version);
 | |
|     if (!snapshots.empty()) {
 | |
|       WriteLock wl(&old_commit_map_mutex_);
 | |
|       for (auto snap : snapshots) {
 | |
|         // This allows IsInSnapshot to tell apart the reads from in valid
 | |
|         // snapshots from the reads from committed values in valid snapshots.
 | |
|         old_commit_map_[snap];
 | |
|       }
 | |
|       old_commit_map_empty_.store(false, std::memory_order_release);
 | |
|     }
 | |
|   }
 | |
|   auto updated_prev_max = prev_max;
 | |
|   TEST_SYNC_POINT("AdvanceMaxEvictedSeq::update_max:pause");
 | |
|   TEST_SYNC_POINT("AdvanceMaxEvictedSeq::update_max:resume");
 | |
|   while (updated_prev_max < new_max &&
 | |
|          !max_evicted_seq_.compare_exchange_weak(updated_prev_max, new_max,
 | |
|                                                  std::memory_order_acq_rel,
 | |
|                                                  std::memory_order_relaxed)) {
 | |
|   };
 | |
| }
 | |
| 
 | |
| const Snapshot* WritePreparedTxnDB::GetSnapshot() {
 | |
|   const bool kForWWConflictCheck = true;
 | |
|   return GetSnapshotInternal(!kForWWConflictCheck);
 | |
| }
 | |
| 
 | |
| SnapshotImpl* WritePreparedTxnDB::GetSnapshotInternal(
 | |
|     bool for_ww_conflict_check) {
 | |
|   // Note: for this optimization setting the last sequence number and obtaining
 | |
|   // the smallest uncommitted seq should be done atomically. However to avoid
 | |
|   // the mutex overhead, we call SmallestUnCommittedSeq BEFORE taking the
 | |
|   // snapshot. Since we always updated the list of unprepared seq (via
 | |
|   // AddPrepared) AFTER the last sequence is updated, this guarantees that the
 | |
|   // smallest uncommitted seq that we pair with the snapshot is smaller or equal
 | |
|   // the value that would be obtained otherwise atomically. That is ok since
 | |
|   // this optimization works as long as min_uncommitted is less than or equal
 | |
|   // than the smallest uncommitted seq when the snapshot was taken.
 | |
|   auto min_uncommitted = WritePreparedTxnDB::SmallestUnCommittedSeq();
 | |
|   SnapshotImpl* snap_impl = db_impl_->GetSnapshotImpl(for_ww_conflict_check);
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::GetSnapshotInternal:first");
 | |
|   assert(snap_impl);
 | |
|   SequenceNumber snap_seq = snap_impl->GetSequenceNumber();
 | |
|   // Note: Check against future_max_evicted_seq_ (in contrast with
 | |
|   // max_evicted_seq_) in case there is a concurrent AdvanceMaxEvictedSeq.
 | |
|   if (UNLIKELY(snap_seq != 0 && snap_seq <= future_max_evicted_seq_)) {
 | |
|     // There is a very rare case in which the commit entry evicts another commit
 | |
|     // entry that is not published yet thus advancing max evicted seq beyond the
 | |
|     // last published seq. This case is not likely in real-world setup so we
 | |
|     // handle it with a few retries.
 | |
|     size_t retry = 0;
 | |
|     SequenceNumber max;
 | |
|     while ((max = future_max_evicted_seq_.load()) != 0 &&
 | |
|            snap_impl->GetSequenceNumber() <= max && retry < 100) {
 | |
|       ROCKS_LOG_WARN(info_log_,
 | |
|                      "GetSnapshot snap: %" PRIu64 " max: %" PRIu64
 | |
|                      " retry %" ROCKSDB_PRIszt,
 | |
|                      snap_impl->GetSequenceNumber(), max, retry);
 | |
|       ReleaseSnapshot(snap_impl);
 | |
|       // Wait for last visible seq to catch up with max, and also go beyond it
 | |
|       // by one.
 | |
|       AdvanceSeqByOne();
 | |
|       snap_impl = db_impl_->GetSnapshotImpl(for_ww_conflict_check);
 | |
|       assert(snap_impl);
 | |
|       retry++;
 | |
|     }
 | |
|     assert(snap_impl->GetSequenceNumber() > max);
 | |
|     if (snap_impl->GetSequenceNumber() <= max) {
 | |
|       throw std::runtime_error(
 | |
|           "Snapshot seq " + std::to_string(snap_impl->GetSequenceNumber()) +
 | |
|           " after " + std::to_string(retry) +
 | |
|           " retries is still less than futre_max_evicted_seq_" +
 | |
|           std::to_string(max));
 | |
|     }
 | |
|   }
 | |
|   EnhanceSnapshot(snap_impl, min_uncommitted);
 | |
|   ROCKS_LOG_DETAILS(
 | |
|       db_impl_->immutable_db_options().info_log,
 | |
|       "GetSnapshot %" PRIu64 " ww:%" PRIi32 " min_uncommitted: %" PRIu64,
 | |
|       snap_impl->GetSequenceNumber(), for_ww_conflict_check, min_uncommitted);
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::GetSnapshotInternal:end");
 | |
|   return snap_impl;
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::AdvanceSeqByOne() {
 | |
|   // Inserting an empty value will i) let the max evicted entry to be
 | |
|   // published, i.e., max == last_published, increase the last published to
 | |
|   // be one beyond max, i.e., max < last_published.
 | |
|   WriteOptions woptions;
 | |
|   TransactionOptions txn_options;
 | |
|   Transaction* txn0 = BeginTransaction(woptions, txn_options, nullptr);
 | |
|   std::hash<std::thread::id> hasher;
 | |
|   char name[64];
 | |
|   snprintf(name, 64, "txn%" ROCKSDB_PRIszt, hasher(std::this_thread::get_id()));
 | |
|   assert(strlen(name) < 64 - 1);
 | |
|   Status s = txn0->SetName(name);
 | |
|   assert(s.ok());
 | |
|   if (s.ok()) {
 | |
|     // Without prepare it would simply skip the commit
 | |
|     s = txn0->Prepare();
 | |
|   }
 | |
|   assert(s.ok());
 | |
|   if (s.ok()) {
 | |
|     s = txn0->Commit();
 | |
|   }
 | |
|   assert(s.ok());
 | |
|   delete txn0;
 | |
| }
 | |
| 
 | |
| const std::vector<SequenceNumber> WritePreparedTxnDB::GetSnapshotListFromDB(
 | |
|     SequenceNumber max) {
 | |
|   ROCKS_LOG_DETAILS(info_log_, "GetSnapshotListFromDB with max %" PRIu64, max);
 | |
|   InstrumentedMutexLock dblock(db_impl_->mutex());
 | |
|   db_impl_->mutex()->AssertHeld();
 | |
|   return db_impl_->snapshots().GetAll(nullptr, max);
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::ReleaseSnapshotInternal(
 | |
|     const SequenceNumber snap_seq) {
 | |
|   // TODO(myabandeh): relax should enough since the synchronizatin is already
 | |
|   // done by snapshots_mutex_ under which this function is called.
 | |
|   if (snap_seq <= max_evicted_seq_.load(std::memory_order_acquire)) {
 | |
|     // Then this is a rare case that transaction did not finish before max
 | |
|     // advances. It is expected for a few read-only backup snapshots. For such
 | |
|     // snapshots we might have kept around a couple of entries in the
 | |
|     // old_commit_map_. Check and do garbage collection if that is the case.
 | |
|     bool need_gc = false;
 | |
|     {
 | |
|       WPRecordTick(TXN_OLD_COMMIT_MAP_MUTEX_OVERHEAD);
 | |
|       ROCKS_LOG_WARN(info_log_, "old_commit_map_mutex_ overhead for %" PRIu64,
 | |
|                      snap_seq);
 | |
|       ReadLock rl(&old_commit_map_mutex_);
 | |
|       auto prep_set_entry = old_commit_map_.find(snap_seq);
 | |
|       need_gc = prep_set_entry != old_commit_map_.end();
 | |
|     }
 | |
|     if (need_gc) {
 | |
|       WPRecordTick(TXN_OLD_COMMIT_MAP_MUTEX_OVERHEAD);
 | |
|       ROCKS_LOG_WARN(info_log_, "old_commit_map_mutex_ overhead for %" PRIu64,
 | |
|                      snap_seq);
 | |
|       WriteLock wl(&old_commit_map_mutex_);
 | |
|       old_commit_map_.erase(snap_seq);
 | |
|       old_commit_map_empty_.store(old_commit_map_.empty(),
 | |
|                                   std::memory_order_release);
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::CleanupReleasedSnapshots(
 | |
|     const std::vector<SequenceNumber>& new_snapshots,
 | |
|     const std::vector<SequenceNumber>& old_snapshots) {
 | |
|   auto newi = new_snapshots.begin();
 | |
|   auto oldi = old_snapshots.begin();
 | |
|   for (; newi != new_snapshots.end() && oldi != old_snapshots.end();) {
 | |
|     assert(*newi >= *oldi);  // cannot have new snapshots with lower seq
 | |
|     if (*newi == *oldi) {    // still not released
 | |
|       auto value = *newi;
 | |
|       while (newi != new_snapshots.end() && *newi == value) {
 | |
|         newi++;
 | |
|       }
 | |
|       while (oldi != old_snapshots.end() && *oldi == value) {
 | |
|         oldi++;
 | |
|       }
 | |
|     } else {
 | |
|       assert(*newi > *oldi);  // *oldi is released
 | |
|       ReleaseSnapshotInternal(*oldi);
 | |
|       oldi++;
 | |
|     }
 | |
|   }
 | |
|   // Everything remained in old_snapshots is released and must be cleaned up
 | |
|   for (; oldi != old_snapshots.end(); oldi++) {
 | |
|     ReleaseSnapshotInternal(*oldi);
 | |
|   }
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::UpdateSnapshots(
 | |
|     const std::vector<SequenceNumber>& snapshots,
 | |
|     const SequenceNumber& version) {
 | |
|   ROCKS_LOG_DETAILS(info_log_, "UpdateSnapshots with version %" PRIu64,
 | |
|                     version);
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::UpdateSnapshots:p:start");
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::UpdateSnapshots:s:start");
 | |
| #ifndef NDEBUG
 | |
|   size_t sync_i = 0;
 | |
| #endif
 | |
|   ROCKS_LOG_DETAILS(info_log_, "snapshots_mutex_ overhead");
 | |
|   WriteLock wl(&snapshots_mutex_);
 | |
|   snapshots_version_ = version;
 | |
|   // We update the list concurrently with the readers.
 | |
|   // Both new and old lists are sorted and the new list is subset of the
 | |
|   // previous list plus some new items. Thus if a snapshot repeats in
 | |
|   // both new and old lists, it will appear upper in the new list. So if
 | |
|   // we simply insert the new snapshots in order, if an overwritten item
 | |
|   // is still valid in the new list is either written to the same place in
 | |
|   // the array or it is written in a higher palce before it gets
 | |
|   // overwritten by another item. This guarantess a reader that reads the
 | |
|   // list bottom-up will eventaully see a snapshot that repeats in the
 | |
|   // update, either before it gets overwritten by the writer or
 | |
|   // afterwards.
 | |
|   size_t i = 0;
 | |
|   auto it = snapshots.begin();
 | |
|   for (; it != snapshots.end() && i < SNAPSHOT_CACHE_SIZE; ++it, ++i) {
 | |
|     snapshot_cache_[i].store(*it, std::memory_order_release);
 | |
|     TEST_IDX_SYNC_POINT("WritePreparedTxnDB::UpdateSnapshots:p:", ++sync_i);
 | |
|     TEST_IDX_SYNC_POINT("WritePreparedTxnDB::UpdateSnapshots:s:", sync_i);
 | |
|   }
 | |
| #ifndef NDEBUG
 | |
|   // Release the remaining sync points since they are useless given that the
 | |
|   // reader would also use lock to access snapshots
 | |
|   for (++sync_i; sync_i <= 10; ++sync_i) {
 | |
|     TEST_IDX_SYNC_POINT("WritePreparedTxnDB::UpdateSnapshots:p:", sync_i);
 | |
|     TEST_IDX_SYNC_POINT("WritePreparedTxnDB::UpdateSnapshots:s:", sync_i);
 | |
|   }
 | |
| #endif
 | |
|   snapshots_.clear();
 | |
|   for (; it != snapshots.end(); ++it) {
 | |
|     // Insert them to a vector that is less efficient to access
 | |
|     // concurrently
 | |
|     snapshots_.push_back(*it);
 | |
|   }
 | |
|   // Update the size at the end. Otherwise a parallel reader might read
 | |
|   // items that are not set yet.
 | |
|   snapshots_total_.store(snapshots.size(), std::memory_order_release);
 | |
| 
 | |
|   // Note: this must be done after the snapshots data structures are updated
 | |
|   // with the new list of snapshots.
 | |
|   CleanupReleasedSnapshots(snapshots, snapshots_all_);
 | |
|   snapshots_all_ = snapshots;
 | |
| 
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::UpdateSnapshots:p:end");
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::UpdateSnapshots:s:end");
 | |
| }
 | |
| 
 | |
| void WritePreparedTxnDB::CheckAgainstSnapshots(const CommitEntry& evicted) {
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::CheckAgainstSnapshots:p:start");
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::CheckAgainstSnapshots:s:start");
 | |
| #ifndef NDEBUG
 | |
|   size_t sync_i = 0;
 | |
| #endif
 | |
|   // First check the snapshot cache that is efficient for concurrent access
 | |
|   auto cnt = snapshots_total_.load(std::memory_order_acquire);
 | |
|   // The list might get updated concurrently as we are reading from it. The
 | |
|   // reader should be able to read all the snapshots that are still valid
 | |
|   // after the update. Since the survived snapshots are written in a higher
 | |
|   // place before gets overwritten the reader that reads bottom-up will
 | |
|   // eventully see it.
 | |
|   const bool next_is_larger = true;
 | |
|   // We will set to true if the border line snapshot suggests that.
 | |
|   bool search_larger_list = false;
 | |
|   size_t ip1 = std::min(cnt, SNAPSHOT_CACHE_SIZE);
 | |
|   for (; 0 < ip1; ip1--) {
 | |
|     SequenceNumber snapshot_seq =
 | |
|         snapshot_cache_[ip1 - 1].load(std::memory_order_acquire);
 | |
|     TEST_IDX_SYNC_POINT("WritePreparedTxnDB::CheckAgainstSnapshots:p:",
 | |
|                         ++sync_i);
 | |
|     TEST_IDX_SYNC_POINT("WritePreparedTxnDB::CheckAgainstSnapshots:s:", sync_i);
 | |
|     if (ip1 == SNAPSHOT_CACHE_SIZE) {  // border line snapshot
 | |
|       // snapshot_seq < commit_seq => larger_snapshot_seq <= commit_seq
 | |
|       // then later also continue the search to larger snapshots
 | |
|       search_larger_list = snapshot_seq < evicted.commit_seq;
 | |
|     }
 | |
|     if (!MaybeUpdateOldCommitMap(evicted.prep_seq, evicted.commit_seq,
 | |
|                                  snapshot_seq, !next_is_larger)) {
 | |
|       break;
 | |
|     }
 | |
|   }
 | |
| #ifndef NDEBUG
 | |
|   // Release the remaining sync points before accquiring the lock
 | |
|   for (++sync_i; sync_i <= 10; ++sync_i) {
 | |
|     TEST_IDX_SYNC_POINT("WritePreparedTxnDB::CheckAgainstSnapshots:p:", sync_i);
 | |
|     TEST_IDX_SYNC_POINT("WritePreparedTxnDB::CheckAgainstSnapshots:s:", sync_i);
 | |
|   }
 | |
| #endif
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::CheckAgainstSnapshots:p:end");
 | |
|   TEST_SYNC_POINT("WritePreparedTxnDB::CheckAgainstSnapshots:s:end");
 | |
|   if (UNLIKELY(SNAPSHOT_CACHE_SIZE < cnt && search_larger_list)) {
 | |
|     // Then access the less efficient list of snapshots_
 | |
|     WPRecordTick(TXN_SNAPSHOT_MUTEX_OVERHEAD);
 | |
|     ROCKS_LOG_WARN(info_log_,
 | |
|                    "snapshots_mutex_ overhead for <%" PRIu64 ",%" PRIu64
 | |
|                    "> with %" ROCKSDB_PRIszt " snapshots",
 | |
|                    evicted.prep_seq, evicted.commit_seq, cnt);
 | |
|     ReadLock rl(&snapshots_mutex_);
 | |
|     // Items could have moved from the snapshots_ to snapshot_cache_ before
 | |
|     // accquiring the lock. To make sure that we do not miss a valid snapshot,
 | |
|     // read snapshot_cache_ again while holding the lock.
 | |
|     for (size_t i = 0; i < SNAPSHOT_CACHE_SIZE; i++) {
 | |
|       SequenceNumber snapshot_seq =
 | |
|           snapshot_cache_[i].load(std::memory_order_acquire);
 | |
|       if (!MaybeUpdateOldCommitMap(evicted.prep_seq, evicted.commit_seq,
 | |
|                                    snapshot_seq, next_is_larger)) {
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
|     for (auto snapshot_seq_2 : snapshots_) {
 | |
|       if (!MaybeUpdateOldCommitMap(evicted.prep_seq, evicted.commit_seq,
 | |
|                                    snapshot_seq_2, next_is_larger)) {
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
|   }
 | |
| }
 | |
| 
 | |
| bool WritePreparedTxnDB::MaybeUpdateOldCommitMap(
 | |
|     const uint64_t& prep_seq, const uint64_t& commit_seq,
 | |
|     const uint64_t& snapshot_seq, const bool next_is_larger = true) {
 | |
|   // If we do not store an entry in old_commit_map_ we assume it is committed in
 | |
|   // all snapshots. If commit_seq <= snapshot_seq, it is considered already in
 | |
|   // the snapshot so we need not to keep the entry around for this snapshot.
 | |
|   if (commit_seq <= snapshot_seq) {
 | |
|     // continue the search if the next snapshot could be smaller than commit_seq
 | |
|     return !next_is_larger;
 | |
|   }
 | |
|   // then snapshot_seq < commit_seq
 | |
|   if (prep_seq <= snapshot_seq) {  // overlapping range
 | |
|     WPRecordTick(TXN_OLD_COMMIT_MAP_MUTEX_OVERHEAD);
 | |
|     ROCKS_LOG_WARN(info_log_,
 | |
|                    "old_commit_map_mutex_ overhead for %" PRIu64
 | |
|                    " commit entry: <%" PRIu64 ",%" PRIu64 ">",
 | |
|                    snapshot_seq, prep_seq, commit_seq);
 | |
|     WriteLock wl(&old_commit_map_mutex_);
 | |
|     old_commit_map_empty_.store(false, std::memory_order_release);
 | |
|     auto& vec = old_commit_map_[snapshot_seq];
 | |
|     vec.insert(std::upper_bound(vec.begin(), vec.end(), prep_seq), prep_seq);
 | |
|     // We need to store it once for each overlapping snapshot. Returning true to
 | |
|     // continue the search if there is more overlapping snapshot.
 | |
|     return true;
 | |
|   }
 | |
|   // continue the search if the next snapshot could be larger than prep_seq
 | |
|   return next_is_larger;
 | |
| }
 | |
| 
 | |
| WritePreparedTxnDB::~WritePreparedTxnDB() {
 | |
|   // At this point there could be running compaction/flush holding a
 | |
|   // SnapshotChecker, which holds a pointer back to WritePreparedTxnDB.
 | |
|   // Make sure those jobs finished before destructing WritePreparedTxnDB.
 | |
|   if (!db_impl_->shutting_down_) {
 | |
|     db_impl_->CancelAllBackgroundWork(true /*wait*/);
 | |
|   }
 | |
| }
 | |
| 
 | |
| void SubBatchCounter::InitWithComp(const uint32_t cf) {
 | |
|   auto cmp = comparators_[cf];
 | |
|   keys_[cf] = CFKeys(SetComparator(cmp));
 | |
| }
 | |
| 
 | |
| void SubBatchCounter::AddKey(const uint32_t cf, const Slice& key) {
 | |
|   CFKeys& cf_keys = keys_[cf];
 | |
|   if (cf_keys.size() == 0) {  // just inserted
 | |
|     InitWithComp(cf);
 | |
|   }
 | |
|   auto it = cf_keys.insert(key);
 | |
|   if (it.second == false) {  // second is false if a element already existed.
 | |
|     batches_++;
 | |
|     keys_.clear();
 | |
|     InitWithComp(cf);
 | |
|     keys_[cf].insert(key);
 | |
|   }
 | |
| }
 | |
| 
 | |
| }  // namespace ROCKSDB_NAMESPACE
 | |
| 
 |