// Copyright (c) 2011-present, Facebook, Inc. All rights reserved. // This source code is licensed under both the GPLv2 (found in the // COPYING file in the root directory) and Apache 2.0 License // (found in the LICENSE.Apache file in the root directory). #ifndef ROCKSDB_LITE #ifdef GFLAGS #include "tools/block_cache_analyzer/block_cache_trace_analyzer.h" #include #include #include #include #include #include #include #include #include #include #include "monitoring/histogram.h" #include "util/gflags_compat.h" #include "util/string_util.h" using GFLAGS_NAMESPACE::ParseCommandLineFlags; DEFINE_string(block_cache_trace_path, "", "The trace file path."); DEFINE_bool(is_block_cache_human_readable_trace, false, "Is the trace file provided for analysis generated by running " "block_cache_trace_analyzer with " "FLAGS_human_readable_trace_file_path is specified."); DEFINE_string( block_cache_sim_config_path, "", "The config file path. One cache configuration per line. The format of a " "cache configuration is " "cache_name,num_shard_bits,ghost_capacity,cache_capacity_1,...,cache_" "capacity_N. Supported cache names are lru, lru_priority, lru_hybrid, and " "lru_hybrid_no_insert_on_row_miss. User may also add a prefix 'ghost_' to " "a cache_name to add a ghost cache in front of the real cache. " "ghost_capacity and cache_capacity can be xK, xM or xG where x is a " "positive number."); DEFINE_int32(block_cache_trace_downsample_ratio, 1, "The trace collected accesses on one in every " "block_cache_trace_downsample_ratio blocks. We scale " "down the simulated cache size by this ratio."); DEFINE_bool(print_block_size_stats, false, "Print block size distribution and the distribution break down by " "block type and column family."); DEFINE_bool(print_access_count_stats, false, "Print access count distribution and the distribution break down " "by block type and column family."); DEFINE_bool(print_data_block_access_count_stats, false, "Print data block accesses by user Get and Multi-Get."); DEFINE_int32(cache_sim_warmup_seconds, 0, "The number of seconds to warmup simulated caches. The hit/miss " "counters are reset after the warmup completes."); DEFINE_int32(analyze_bottom_k_access_count_blocks, 0, "Print out detailed access information for blocks with their " "number of accesses are the bottom k among all blocks."); DEFINE_int32(analyze_top_k_access_count_blocks, 0, "Print out detailed access information for blocks with their " "number of accesses are the top k among all blocks."); DEFINE_string(block_cache_analysis_result_dir, "", "The directory that saves block cache analysis results."); DEFINE_string( timeline_labels, "", "Group the number of accesses per block per second using these labels. " "Possible labels are a combination of the following: cf (column family), " "sst, level, bt (block type), caller, block. For example, label \"cf_bt\" " "means the number of acccess per second is grouped by unique pairs of " "\"cf_bt\". A label \"all\" contains the aggregated number of accesses per " "second across all possible labels."); DEFINE_string(reuse_distance_labels, "", "Group the reuse distance of a block using these labels. Reuse " "distance is defined as the cumulated size of unique blocks read " "between two consecutive accesses on the same block."); DEFINE_string( reuse_distance_buckets, "", "Group blocks by their reuse distances given these buckets. For " "example, if 'reuse_distance_buckets' is '1K,1M,1G', we will " "create four buckets. The first three buckets contain the number of " "blocks with reuse distance less than 1KB, between 1K and 1M, between 1M " "and 1G, respectively. The last bucket contains the number of blocks with " "reuse distance larger than 1G. "); DEFINE_string( reuse_interval_labels, "", "Group the reuse interval of a block using these labels. Reuse " "interval is defined as the time between two consecutive accesses " "on the same block."); DEFINE_string( reuse_interval_buckets, "", "Group blocks by their reuse interval given these buckets. For " "example, if 'reuse_distance_buckets' is '1,10,100', we will " "create four buckets. The first three buckets contain the number of " "blocks with reuse interval less than 1 second, between 1 second and 10 " "seconds, between 10 seconds and 100 seconds, respectively. The last " "bucket contains the number of blocks with reuse interval longer than 100 " "seconds."); DEFINE_string( reuse_lifetime_labels, "", "Group the reuse lifetime of a block using these labels. Reuse " "lifetime is defined as the time interval between the first access on a " "block and the last access on the same block. For blocks that are only " "accessed once, its lifetime is set to kMaxUint64."); DEFINE_string( reuse_lifetime_buckets, "", "Group blocks by their reuse lifetime given these buckets. For " "example, if 'reuse_lifetime_buckets' is '1,10,100', we will " "create four buckets. The first three buckets contain the number of " "blocks with reuse lifetime less than 1 second, between 1 second and 10 " "seconds, between 10 seconds and 100 seconds, respectively. The last " "bucket contains the number of blocks with reuse lifetime longer than 100 " "seconds."); DEFINE_string( analyze_callers, "", "The list of callers to perform a detailed analysis on. If speicfied, the " "analyzer will output a detailed percentage of accesses for each caller " "break down by column family, level, and block type. A list of available " "callers are: Get, MultiGet, Iterator, ApproximateSize, VerifyChecksum, " "SSTDumpTool, ExternalSSTIngestion, Repair, Prefetch, Compaction, " "CompactionRefill, Flush, SSTFileReader, Uncategorized."); DEFINE_string(access_count_buckets, "", "Group number of blocks by their access count given these " "buckets. If specified, the analyzer will output a detailed " "analysis on the number of blocks grouped by their access count " "break down by block type and column family."); DEFINE_int32(analyze_blocks_reuse_k_reuse_window, 0, "Analyze the percentage of blocks that are accessed in the " "[k, 2*k] seconds are accessed again in the next [2*k, 3*k], " "[3*k, 4*k],...,[k*(n-1), k*n] seconds. "); DEFINE_string(analyze_get_spatial_locality_labels, "", "Group data blocks using these labels."); DEFINE_string(analyze_get_spatial_locality_buckets, "", "Group data blocks by their statistics using these buckets."); DEFINE_string(skew_labels, "", "Group the access count of a block using these labels."); DEFINE_string(skew_buckets, "", "Group the skew labels using these buckets."); DEFINE_bool(mrc_only, false, "Evaluate alternative cache policies only. When this flag is true, " "the analyzer does NOT maintain states of each block in memory for " "analysis. It only feeds the accesses into the cache simulators."); DEFINE_string( analyze_correlation_coefficients_labels, "", "Analyze the correlation coefficients of features such as number of past " "accesses with regard to the number of accesses till the next access."); DEFINE_int32(analyze_correlation_coefficients_max_number_of_values, 1000000, "The maximum number of values for a feature. If the number of " "values for a feature is larger than this max, it randomly " "selects 'max' number of values."); DEFINE_string(human_readable_trace_file_path, "", "The filt path that saves human readable access records."); namespace ROCKSDB_NAMESPACE { namespace { const std::string kMissRatioCurveFileName = "mrc"; const std::string kGroupbyBlock = "block"; const std::string kGroupbyTable = "table"; const std::string kGroupbyColumnFamily = "cf"; const std::string kGroupbySSTFile = "sst"; const std::string kGroupbyBlockType = "bt"; const std::string kGroupbyCaller = "caller"; const std::string kGroupbyLevel = "level"; const std::string kGroupbyAll = "all"; const std::set kGroupbyLabels{ kGroupbyBlock, kGroupbyColumnFamily, kGroupbySSTFile, kGroupbyLevel, kGroupbyBlockType, kGroupbyCaller, kGroupbyAll}; const std::string kSupportedCacheNames = " lru ghost_lru lru_priority ghost_lru_priority lru_hybrid " "ghost_lru_hybrid lru_hybrid_no_insert_on_row_miss " "ghost_lru_hybrid_no_insert_on_row_miss "; // The suffix for the generated csv files. const std::string kFileNameSuffixMissRatioTimeline = "miss_ratio_timeline"; const std::string kFileNameSuffixMissTimeline = "miss_timeline"; const std::string kFileNameSuffixSkew = "skewness"; const std::string kFileNameSuffixAccessTimeline = "access_timeline"; const std::string kFileNameSuffixCorrelation = "correlation_input"; const std::string kFileNameSuffixAvgReuseIntervalNaccesses = "avg_reuse_interval_naccesses"; const std::string kFileNameSuffixAvgReuseInterval = "avg_reuse_interval"; const std::string kFileNameSuffixReuseInterval = "access_reuse_interval"; const std::string kFileNameSuffixReuseLifetime = "reuse_lifetime"; const std::string kFileNameSuffixAccessReuseBlocksTimeline = "reuse_blocks_timeline"; const std::string kFileNameSuffixPercentOfAccessSummary = "percentage_of_accesses_summary"; const std::string kFileNameSuffixPercentRefKeys = "percent_ref_keys"; const std::string kFileNameSuffixPercentDataSizeOnRefKeys = "percent_data_size_on_ref_keys"; const std::string kFileNameSuffixPercentAccessesOnRefKeys = "percent_accesses_on_ref_keys"; const std::string kFileNameSuffixAccessCountSummary = "access_count_summary"; std::string block_type_to_string(TraceType type) { switch (type) { case kBlockTraceFilterBlock: return "Filter"; case kBlockTraceDataBlock: return "Data"; case kBlockTraceIndexBlock: return "Index"; case kBlockTraceRangeDeletionBlock: return "RangeDeletion"; case kBlockTraceUncompressionDictBlock: return "UncompressionDict"; default: break; } // This cannot happen. return "InvalidType"; } std::string caller_to_string(TableReaderCaller caller) { switch (caller) { case kUserGet: return "Get"; case kUserMultiGet: return "MultiGet"; case kUserIterator: return "Iterator"; case kUserApproximateSize: return "ApproximateSize"; case kUserVerifyChecksum: return "VerifyChecksum"; case kSSTDumpTool: return "SSTDumpTool"; case kExternalSSTIngestion: return "ExternalSSTIngestion"; case kRepair: return "Repair"; case kPrefetch: return "Prefetch"; case kCompaction: return "Compaction"; case kCompactionRefill: return "CompactionRefill"; case kFlush: return "Flush"; case kSSTFileReader: return "SSTFileReader"; case kUncategorized: return "Uncategorized"; default: break; } // This cannot happen. return "InvalidCaller"; } TableReaderCaller string_to_caller(std::string caller_str) { if (caller_str == "Get") { return kUserGet; } else if (caller_str == "MultiGet") { return kUserMultiGet; } else if (caller_str == "Iterator") { return kUserIterator; } else if (caller_str == "ApproximateSize") { return kUserApproximateSize; } else if (caller_str == "VerifyChecksum") { return kUserVerifyChecksum; } else if (caller_str == "SSTDumpTool") { return kSSTDumpTool; } else if (caller_str == "ExternalSSTIngestion") { return kExternalSSTIngestion; } else if (caller_str == "Repair") { return kRepair; } else if (caller_str == "Prefetch") { return kPrefetch; } else if (caller_str == "Compaction") { return kCompaction; } else if (caller_str == "CompactionRefill") { return kCompactionRefill; } else if (caller_str == "Flush") { return kFlush; } else if (caller_str == "SSTFileReader") { return kSSTFileReader; } else if (caller_str == "Uncategorized") { return kUncategorized; } return TableReaderCaller::kMaxBlockCacheLookupCaller; } bool is_user_access(TableReaderCaller caller) { switch (caller) { case kUserGet: case kUserMultiGet: case kUserIterator: case kUserApproximateSize: case kUserVerifyChecksum: return true; default: break; } return false; } const char kBreakLine[] = "***************************************************************\n"; void print_break_lines(uint32_t num_break_lines) { for (uint32_t i = 0; i < num_break_lines; i++) { fprintf(stdout, kBreakLine); } } double percent(uint64_t numerator, uint64_t denomenator) { if (denomenator == 0) { return -1; } return static_cast(numerator * 100.0 / denomenator); } std::map adjust_time_unit( const std::map& time_stats, uint64_t time_unit) { if (time_unit == 1) { return time_stats; } std::map adjusted_time_stats; for (auto const& time : time_stats) { adjusted_time_stats[static_cast(time.first / time_unit)] += time.second; } return adjusted_time_stats; } } // namespace void BlockCacheTraceAnalyzer::WriteMissRatioCurves() const { if (!cache_simulator_) { return; } if (output_dir_.empty()) { return; } uint64_t trace_duration = trace_end_timestamp_in_seconds_ - trace_start_timestamp_in_seconds_; uint64_t total_accesses = access_sequence_number_; const std::string output_miss_ratio_curve_path = output_dir_ + "/" + std::to_string(trace_duration) + "_" + std::to_string(total_accesses) + "_" + kMissRatioCurveFileName; std::ofstream out(output_miss_ratio_curve_path); if (!out.is_open()) { return; } // Write header. const std::string header = "cache_name,num_shard_bits,ghost_capacity,capacity,miss_ratio,total_" "accesses"; out << header << std::endl; for (auto const& config_caches : cache_simulator_->sim_caches()) { const CacheConfiguration& config = config_caches.first; for (uint32_t i = 0; i < config.cache_capacities.size(); i++) { double miss_ratio = config_caches.second[i]->miss_ratio_stats().miss_ratio(); // Write the body. out << config.cache_name; out << ","; out << config.num_shard_bits; out << ","; out << config.ghost_cache_capacity; out << ","; out << config.cache_capacities[i]; out << ","; out << std::fixed << std::setprecision(4) << miss_ratio; out << ","; out << config_caches.second[i]->miss_ratio_stats().total_accesses(); out << std::endl; } } out.close(); } void BlockCacheTraceAnalyzer::UpdateFeatureVectors( const std::vector& access_sequence_number_timeline, const std::vector& access_timeline, const std::string& label, std::map* label_features, std::map* label_predictions) const { if (access_sequence_number_timeline.empty() || access_timeline.empty()) { return; } assert(access_timeline.size() == access_sequence_number_timeline.size()); uint64_t prev_access_sequence_number = access_sequence_number_timeline[0]; uint64_t prev_access_timestamp = access_timeline[0]; for (uint32_t i = 0; i < access_sequence_number_timeline.size(); i++) { uint64_t num_accesses_since_last_access = access_sequence_number_timeline[i] - prev_access_sequence_number; uint64_t elapsed_time_since_last_access = access_timeline[i] - prev_access_timestamp; prev_access_sequence_number = access_sequence_number_timeline[i]; prev_access_timestamp = access_timeline[i]; if (i < access_sequence_number_timeline.size() - 1) { (*label_features)[label].num_accesses_since_last_access.push_back( num_accesses_since_last_access); (*label_features)[label].num_past_accesses.push_back(i); (*label_features)[label].elapsed_time_since_last_access.push_back( elapsed_time_since_last_access); } if (i >= 1) { (*label_predictions)[label].num_accesses_till_next_access.push_back( num_accesses_since_last_access); (*label_predictions)[label].elapsed_time_till_next_access.push_back( elapsed_time_since_last_access); } } } void BlockCacheTraceAnalyzer::WriteMissRatioTimeline(uint64_t time_unit) const { if (!cache_simulator_ || output_dir_.empty()) { return; } std::map>> cs_name_timeline; uint64_t start_time = port::kMaxUint64; uint64_t end_time = 0; const std::map& trace_num_misses = adjust_time_unit(miss_ratio_stats_.num_misses_timeline(), time_unit); const std::map& trace_num_accesses = adjust_time_unit(miss_ratio_stats_.num_accesses_timeline(), time_unit); assert(trace_num_misses.size() == trace_num_accesses.size()); for (auto const& num_miss : trace_num_misses) { uint64_t time = num_miss.first; start_time = std::min(start_time, time); end_time = std::max(end_time, time); uint64_t miss = num_miss.second; auto it = trace_num_accesses.find(time); assert(it != trace_num_accesses.end()); uint64_t access = it->second; cs_name_timeline[port::kMaxUint64]["trace"][time] = percent(miss, access); } for (auto const& config_caches : cache_simulator_->sim_caches()) { const CacheConfiguration& config = config_caches.first; std::string cache_label = config.cache_name + "-" + std::to_string(config.num_shard_bits) + "-" + std::to_string(config.ghost_cache_capacity); for (uint32_t i = 0; i < config.cache_capacities.size(); i++) { const std::map& num_misses = adjust_time_unit( config_caches.second[i]->miss_ratio_stats().num_misses_timeline(), time_unit); const std::map& num_accesses = adjust_time_unit( config_caches.second[i]->miss_ratio_stats().num_accesses_timeline(), time_unit); assert(num_misses.size() == num_accesses.size()); for (auto const& num_miss : num_misses) { uint64_t time = num_miss.first; start_time = std::min(start_time, time); end_time = std::max(end_time, time); uint64_t miss = num_miss.second; auto it = num_accesses.find(time); assert(it != num_accesses.end()); uint64_t access = it->second; cs_name_timeline[config.cache_capacities[i]][cache_label][time] = percent(miss, access); } } } for (auto const& it : cs_name_timeline) { const std::string output_miss_ratio_timeline_path = output_dir_ + "/" + std::to_string(it.first) + "_" + std::to_string(time_unit) + "_" + kFileNameSuffixMissRatioTimeline; std::ofstream out(output_miss_ratio_timeline_path); if (!out.is_open()) { return; } std::string header("time"); for (uint64_t now = start_time; now <= end_time; now++) { header += ","; header += std::to_string(now); } out << header << std::endl; for (auto const& label : it.second) { std::string row(label.first); for (uint64_t now = start_time; now <= end_time; now++) { auto misses = label.second.find(now); row += ","; if (misses != label.second.end()) { row += std::to_string(misses->second); } else { row += "0"; } } out << row << std::endl; } out.close(); } } void BlockCacheTraceAnalyzer::WriteMissTimeline(uint64_t time_unit) const { if (!cache_simulator_ || output_dir_.empty()) { return; } std::map>> cs_name_timeline; uint64_t start_time = port::kMaxUint64; uint64_t end_time = 0; const std::map& trace_num_misses = adjust_time_unit(miss_ratio_stats_.num_misses_timeline(), time_unit); for (auto const& num_miss : trace_num_misses) { uint64_t time = num_miss.first; start_time = std::min(start_time, time); end_time = std::max(end_time, time); uint64_t miss = num_miss.second; cs_name_timeline[port::kMaxUint64]["trace"][time] = miss; } for (auto const& config_caches : cache_simulator_->sim_caches()) { const CacheConfiguration& config = config_caches.first; std::string cache_label = config.cache_name + "-" + std::to_string(config.num_shard_bits) + "-" + std::to_string(config.ghost_cache_capacity); for (uint32_t i = 0; i < config.cache_capacities.size(); i++) { const std::map& num_misses = adjust_time_unit( config_caches.second[i]->miss_ratio_stats().num_misses_timeline(), time_unit); for (auto const& num_miss : num_misses) { uint64_t time = num_miss.first; start_time = std::min(start_time, time); end_time = std::max(end_time, time); uint64_t miss = num_miss.second; cs_name_timeline[config.cache_capacities[i]][cache_label][time] = miss; } } } for (auto const& it : cs_name_timeline) { const std::string output_miss_ratio_timeline_path = output_dir_ + "/" + std::to_string(it.first) + "_" + std::to_string(time_unit) + "_" + kFileNameSuffixMissTimeline; std::ofstream out(output_miss_ratio_timeline_path); if (!out.is_open()) { return; } std::string header("time"); for (uint64_t now = start_time; now <= end_time; now++) { header += ","; header += std::to_string(now); } out << header << std::endl; for (auto const& label : it.second) { std::string row(label.first); for (uint64_t now = start_time; now <= end_time; now++) { auto misses = label.second.find(now); row += ","; if (misses != label.second.end()) { row += std::to_string(misses->second); } else { row += "0"; } } out << row << std::endl; } out.close(); } } void BlockCacheTraceAnalyzer::WriteSkewness( const std::string& label_str, const std::vector& percent_buckets, TraceType target_block_type) const { std::set labels = ParseLabelStr(label_str); std::map label_naccesses; uint64_t total_naccesses = 0; auto block_callback = [&](const std::string& cf_name, uint64_t fd, uint32_t level, TraceType type, const std::string& /*block_key*/, uint64_t block_id, const BlockAccessInfo& block) { if (target_block_type != TraceType::kTraceMax && target_block_type != type) { return; } const std::string label = BuildLabel( labels, cf_name, fd, level, type, TableReaderCaller::kMaxBlockCacheLookupCaller, block_id, block); label_naccesses[label] += block.num_accesses; total_naccesses += block.num_accesses; }; TraverseBlocks(block_callback, &labels); std::map> label_bucket_naccesses; std::vector> pairs; for (auto const& itr : label_naccesses) { pairs.push_back(itr); } // Sort in descending order. sort(pairs.begin(), pairs.end(), [=](const std::pair& a, const std::pair& b) { return b.second < a.second; }); size_t prev_start_index = 0; for (auto const& percent : percent_buckets) { label_bucket_naccesses[label_str][percent] = 0; size_t end_index = 0; if (percent == port::kMaxUint64) { end_index = label_naccesses.size(); } else { end_index = percent * label_naccesses.size() / 100; } for (size_t i = prev_start_index; i < end_index; i++) { label_bucket_naccesses[label_str][percent] += pairs[i].second; } prev_start_index = end_index; } std::string filename_suffix; if (target_block_type != TraceType::kTraceMax) { filename_suffix = block_type_to_string(target_block_type); filename_suffix += "_"; } filename_suffix += kFileNameSuffixSkew; WriteStatsToFile(label_str, percent_buckets, filename_suffix, label_bucket_naccesses, total_naccesses); } void BlockCacheTraceAnalyzer::WriteCorrelationFeatures( const std::string& label_str, uint32_t max_number_of_values) const { std::set labels = ParseLabelStr(label_str); std::map label_features; std::map label_predictions; auto block_callback = [&](const std::string& cf_name, uint64_t fd, uint32_t level, TraceType block_type, const std::string& /*block_key*/, uint64_t /*block_key_id*/, const BlockAccessInfo& block) { if (block.table_id == 0 && labels.find(kGroupbyTable) != labels.end()) { // We only know table id information for get requests. return; } if (labels.find(kGroupbyCaller) != labels.end()) { // Group by caller. for (auto const& caller_map : block.caller_access_timeline) { const std::string label = BuildLabel(labels, cf_name, fd, level, block_type, caller_map.first, /*block_id=*/0, block); auto it = block.caller_access_sequence__number_timeline.find( caller_map.first); assert(it != block.caller_access_sequence__number_timeline.end()); UpdateFeatureVectors(it->second, caller_map.second, label, &label_features, &label_predictions); } return; } const std::string label = BuildLabel(labels, cf_name, fd, level, block_type, TableReaderCaller::kMaxBlockCacheLookupCaller, /*block_id=*/0, block); UpdateFeatureVectors(block.access_sequence_number_timeline, block.access_timeline, label, &label_features, &label_predictions); }; TraverseBlocks(block_callback, &labels); WriteCorrelationFeaturesToFile(label_str, label_features, label_predictions, max_number_of_values); } void BlockCacheTraceAnalyzer::WriteCorrelationFeaturesToFile( const std::string& label, const std::map& label_features, const std::map& label_predictions, uint32_t max_number_of_values) const { std::default_random_engine rand_engine(static_cast(env_->NowMicros())); for (auto const& label_feature_vectors : label_features) { const Features& past = label_feature_vectors.second; auto it = label_predictions.find(label_feature_vectors.first); assert(it != label_predictions.end()); const Predictions& future = it->second; const std::string output_path = output_dir_ + "/" + label + "_" + label_feature_vectors.first + "_" + kFileNameSuffixCorrelation; std::ofstream out(output_path); if (!out.is_open()) { return; } std::string header( "num_accesses_since_last_access,elapsed_time_since_last_access,num_" "past_accesses,num_accesses_till_next_access,elapsed_time_till_next_" "access"); out << header << std::endl; std::vector indexes; for (uint32_t i = 0; i < past.num_accesses_since_last_access.size(); i++) { indexes.push_back(i); } std::shuffle(indexes.begin(), indexes.end(), rand_engine); for (uint32_t i = 0; i < max_number_of_values && i < indexes.size(); i++) { uint32_t rand_index = indexes[i]; out << std::to_string(past.num_accesses_since_last_access[rand_index]) << ","; out << std::to_string(past.elapsed_time_since_last_access[rand_index]) << ","; out << std::to_string(past.num_past_accesses[rand_index]) << ","; out << std::to_string(future.num_accesses_till_next_access[rand_index]) << ","; out << std::to_string(future.elapsed_time_till_next_access[rand_index]) << std::endl; } out.close(); } } void BlockCacheTraceAnalyzer::WriteCorrelationFeaturesForGet( uint32_t max_number_of_values) const { std::string label = "GetKeyInfo"; std::map label_features; std::map label_predictions; for (auto const& get_info : get_key_info_map_) { const GetKeyInfo& info = get_info.second; UpdateFeatureVectors(info.access_sequence_number_timeline, info.access_timeline, label, &label_features, &label_predictions); } WriteCorrelationFeaturesToFile(label, label_features, label_predictions, max_number_of_values); } std::set BlockCacheTraceAnalyzer::ParseLabelStr( const std::string& label_str) const { std::stringstream ss(label_str); std::set labels; // label_str is in the form of "label1_label2_label3", e.g., cf_bt. while (ss.good()) { std::string label_name; getline(ss, label_name, '_'); if (kGroupbyLabels.find(label_name) == kGroupbyLabels.end()) { // Unknown label name. fprintf(stderr, "Unknown label name %s, label string %s\n", label_name.c_str(), label_str.c_str()); return {}; } labels.insert(label_name); } return labels; } std::string BlockCacheTraceAnalyzer::BuildLabel( const std::set& labels, const std::string& cf_name, uint64_t fd, uint32_t level, TraceType type, TableReaderCaller caller, uint64_t block_key, const BlockAccessInfo& block) const { std::map label_value_map; label_value_map[kGroupbyAll] = kGroupbyAll; label_value_map[kGroupbyLevel] = std::to_string(level); label_value_map[kGroupbyCaller] = caller_to_string(caller); label_value_map[kGroupbySSTFile] = std::to_string(fd); label_value_map[kGroupbyBlockType] = block_type_to_string(type); label_value_map[kGroupbyColumnFamily] = cf_name; label_value_map[kGroupbyBlock] = std::to_string(block_key); label_value_map[kGroupbyTable] = std::to_string(block.table_id); // Concatenate the label values. std::string label; for (auto const& l : labels) { label += label_value_map[l]; label += "-"; } if (!label.empty()) { label.pop_back(); } return label; } void BlockCacheTraceAnalyzer::TraverseBlocks( std::function block_callback, std::set* labels) const { for (auto const& cf_aggregates : cf_aggregates_map_) { // Stats per column family. const std::string& cf_name = cf_aggregates.first; for (auto const& file_aggregates : cf_aggregates.second.fd_aggregates_map) { // Stats per SST file. const uint64_t fd = file_aggregates.first; const uint32_t level = file_aggregates.second.level; for (auto const& block_type_aggregates : file_aggregates.second.block_type_aggregates_map) { // Stats per block type. const TraceType type = block_type_aggregates.first; for (auto const& block_access_info : block_type_aggregates.second.block_access_info_map) { // Stats per block. if (labels && block_access_info.second.table_id == 0 && labels->find(kGroupbyTable) != labels->end()) { // We only know table id information for get requests. return; } block_callback(cf_name, fd, level, type, block_access_info.first, block_access_info.second.block_id, block_access_info.second); } } } } } void BlockCacheTraceAnalyzer::WriteGetSpatialLocality( const std::string& label_str, const std::vector& percent_buckets) const { std::set labels = ParseLabelStr(label_str); std::map> label_pnrefkeys_nblocks; std::map> label_pnrefs_nblocks; std::map> label_pndatasize_nblocks; uint64_t nblocks = 0; auto block_callback = [&](const std::string& cf_name, uint64_t fd, uint32_t level, TraceType /*block_type*/, const std::string& /*block_key*/, uint64_t /*block_key_id*/, const BlockAccessInfo& block) { if (block.num_keys == 0) { return; } uint64_t naccesses = 0; for (auto const& key_access : block.key_num_access_map) { for (auto const& caller_access : key_access.second) { if (caller_access.first == TableReaderCaller::kUserGet) { naccesses += caller_access.second; } } } const std::string label = BuildLabel(labels, cf_name, fd, level, TraceType::kBlockTraceDataBlock, TableReaderCaller::kUserGet, /*block_id=*/0, block); const uint64_t percent_referenced_for_existing_keys = static_cast(std::max( percent(block.key_num_access_map.size(), block.num_keys), 0.0)); const uint64_t percent_accesses_for_existing_keys = static_cast(std::max( percent(block.num_referenced_key_exist_in_block, naccesses), 0.0)); const uint64_t percent_referenced_data_size = static_cast( std::max(percent(block.referenced_data_size, block.block_size), 0.0)); if (label_pnrefkeys_nblocks.find(label) == label_pnrefkeys_nblocks.end()) { for (auto const& percent_bucket : percent_buckets) { label_pnrefkeys_nblocks[label][percent_bucket] = 0; label_pnrefs_nblocks[label][percent_bucket] = 0; label_pndatasize_nblocks[label][percent_bucket] = 0; } } label_pnrefkeys_nblocks[label] .upper_bound(percent_referenced_for_existing_keys) ->second += 1; label_pnrefs_nblocks[label] .upper_bound(percent_accesses_for_existing_keys) ->second += 1; label_pndatasize_nblocks[label] .upper_bound(percent_referenced_data_size) ->second += 1; nblocks += 1; }; TraverseBlocks(block_callback, &labels); WriteStatsToFile(label_str, percent_buckets, kFileNameSuffixPercentRefKeys, label_pnrefkeys_nblocks, nblocks); WriteStatsToFile(label_str, percent_buckets, kFileNameSuffixPercentAccessesOnRefKeys, label_pnrefs_nblocks, nblocks); WriteStatsToFile(label_str, percent_buckets, kFileNameSuffixPercentDataSizeOnRefKeys, label_pndatasize_nblocks, nblocks); } void BlockCacheTraceAnalyzer::WriteAccessTimeline(const std::string& label_str, uint64_t time_unit, bool user_access_only) const { std::set labels = ParseLabelStr(label_str); uint64_t start_time = port::kMaxUint64; uint64_t end_time = 0; std::map> label_access_timeline; std::map> access_count_block_id_map; auto block_callback = [&](const std::string& cf_name, uint64_t fd, uint32_t level, TraceType type, const std::string& /*block_key*/, uint64_t block_id, const BlockAccessInfo& block) { uint64_t naccesses = 0; for (auto const& timeline : block.caller_num_accesses_timeline) { const TableReaderCaller caller = timeline.first; if (user_access_only && !is_user_access(caller)) { continue; } const std::string label = BuildLabel(labels, cf_name, fd, level, type, caller, block_id, block); for (auto const& naccess : timeline.second) { const uint64_t timestamp = naccess.first / time_unit; const uint64_t num = naccess.second; label_access_timeline[label][timestamp] += num; start_time = std::min(start_time, timestamp); end_time = std::max(end_time, timestamp); naccesses += num; } } if (naccesses > 0) { access_count_block_id_map[naccesses].push_back(std::to_string(block_id)); } }; TraverseBlocks(block_callback, &labels); // We have label_access_timeline now. Write them into a file. const std::string user_access_prefix = user_access_only ? "user_access_only_" : "all_access_"; const std::string output_path = output_dir_ + "/" + user_access_prefix + label_str + "_" + std::to_string(time_unit) + "_" + kFileNameSuffixAccessTimeline; std::ofstream out(output_path); if (!out.is_open()) { return; } std::string header("time"); if (labels.find("block") != labels.end()) { for (uint64_t now = start_time; now <= end_time; now++) { header += ","; header += std::to_string(now); } out << header << std::endl; // Write the most frequently accessed blocks first. for (auto naccess_it = access_count_block_id_map.rbegin(); naccess_it != access_count_block_id_map.rend(); naccess_it++) { for (auto& block_id_it : naccess_it->second) { std::string row(block_id_it); for (uint64_t now = start_time; now <= end_time; now++) { auto it = label_access_timeline[block_id_it].find(now); row += ","; if (it != label_access_timeline[block_id_it].end()) { row += std::to_string(it->second); } else { row += "0"; } } out << row << std::endl; } } out.close(); return; } for (uint64_t now = start_time; now <= end_time; now++) { header += ","; header += std::to_string(now); } out << header << std::endl; for (auto const& label : label_access_timeline) { std::string row(label.first); for (uint64_t now = start_time; now <= end_time; now++) { auto it = label.second.find(now); row += ","; if (it != label.second.end()) { row += std::to_string(it->second); } else { row += "0"; } } out << row << std::endl; } out.close(); } void BlockCacheTraceAnalyzer::WriteReuseDistance( const std::string& label_str, const std::vector& distance_buckets) const { std::set labels = ParseLabelStr(label_str); std::map> label_distance_num_reuses; uint64_t total_num_reuses = 0; auto block_callback = [&](const std::string& cf_name, uint64_t fd, uint32_t level, TraceType type, const std::string& /*block_key*/, uint64_t block_id, const BlockAccessInfo& block) { const std::string label = BuildLabel( labels, cf_name, fd, level, type, TableReaderCaller::kMaxBlockCacheLookupCaller, block_id, block); if (label_distance_num_reuses.find(label) == label_distance_num_reuses.end()) { // The first time we encounter this label. for (auto const& distance_bucket : distance_buckets) { label_distance_num_reuses[label][distance_bucket] = 0; } } for (auto const& reuse_distance : block.reuse_distance_count) { label_distance_num_reuses[label] .upper_bound(reuse_distance.first) ->second += reuse_distance.second; total_num_reuses += reuse_distance.second; } }; TraverseBlocks(block_callback, &labels); // We have label_naccesses and label_distance_num_reuses now. Write them into // a file. const std::string output_path = output_dir_ + "/" + label_str + "_reuse_distance"; std::ofstream out(output_path); if (!out.is_open()) { return; } std::string header("bucket"); for (auto const& label_it : label_distance_num_reuses) { header += ","; header += label_it.first; } out << header << std::endl; for (auto const& bucket : distance_buckets) { std::string row(std::to_string(bucket)); for (auto const& label_it : label_distance_num_reuses) { auto const& it = label_it.second.find(bucket); assert(it != label_it.second.end()); row += ","; row += std::to_string(percent(it->second, total_num_reuses)); } out << row << std::endl; } out.close(); } void BlockCacheTraceAnalyzer::UpdateReuseIntervalStats( const std::string& label, const std::vector& time_buckets, const std::map timeline, std::map>* label_time_num_reuses, uint64_t* total_num_reuses) const { assert(label_time_num_reuses); assert(total_num_reuses); if (label_time_num_reuses->find(label) == label_time_num_reuses->end()) { // The first time we encounter this label. for (auto const& time_bucket : time_buckets) { (*label_time_num_reuses)[label][time_bucket] = 0; } } auto it = timeline.begin(); uint64_t prev_timestamp = it->first; const uint64_t prev_num = it->second; it++; // Reused within one second. if (prev_num > 1) { (*label_time_num_reuses)[label].upper_bound(0)->second += prev_num - 1; *total_num_reuses += prev_num - 1; } while (it != timeline.end()) { const uint64_t timestamp = it->first; const uint64_t num = it->second; const uint64_t reuse_interval = timestamp - prev_timestamp; (*label_time_num_reuses)[label].upper_bound(reuse_interval)->second += 1; if (num > 1) { (*label_time_num_reuses)[label].upper_bound(0)->second += num - 1; } prev_timestamp = timestamp; *total_num_reuses += num; it++; } } void BlockCacheTraceAnalyzer::WriteStatsToFile( const std::string& label_str, const std::vector& time_buckets, const std::string& filename_suffix, const std::map>& label_data, uint64_t ntotal) const { const std::string output_path = output_dir_ + "/" + label_str + "_" + filename_suffix; std::ofstream out(output_path); if (!out.is_open()) { return; } std::string header("bucket"); for (auto const& label_it : label_data) { header += ","; header += label_it.first; } out << header << std::endl; for (auto const& bucket : time_buckets) { std::string row(std::to_string(bucket)); for (auto const& label_it : label_data) { auto const& it = label_it.second.find(bucket); assert(it != label_it.second.end()); row += ","; row += std::to_string(percent(it->second, ntotal)); } out << row << std::endl; } out.close(); } void BlockCacheTraceAnalyzer::WriteReuseInterval( const std::string& label_str, const std::vector& time_buckets) const { std::set labels = ParseLabelStr(label_str); std::map> label_time_num_reuses; std::map> label_avg_reuse_nblocks; std::map> label_avg_reuse_naccesses; uint64_t total_num_reuses = 0; uint64_t total_nblocks = 0; uint64_t total_accesses = 0; auto block_callback = [&](const std::string& cf_name, uint64_t fd, uint32_t level, TraceType type, const std::string& /*block_key*/, uint64_t block_id, const BlockAccessInfo& block) { total_nblocks++; total_accesses += block.num_accesses; uint64_t avg_reuse_interval = 0; if (block.num_accesses > 1) { avg_reuse_interval = ((block.last_access_time - block.first_access_time) / kMicrosInSecond) / block.num_accesses; } else { avg_reuse_interval = port::kMaxUint64 - 1; } if (labels.find(kGroupbyCaller) != labels.end()) { for (auto const& timeline : block.caller_num_accesses_timeline) { const TableReaderCaller caller = timeline.first; const std::string label = BuildLabel(labels, cf_name, fd, level, type, caller, block_id, block); UpdateReuseIntervalStats(label, time_buckets, timeline.second, &label_time_num_reuses, &total_num_reuses); } return; } // Does not group by caller so we need to flatten the access timeline. const std::string label = BuildLabel( labels, cf_name, fd, level, type, TableReaderCaller::kMaxBlockCacheLookupCaller, block_id, block); std::map timeline; for (auto const& caller_timeline : block.caller_num_accesses_timeline) { for (auto const& time_naccess : caller_timeline.second) { timeline[time_naccess.first] += time_naccess.second; } } UpdateReuseIntervalStats(label, time_buckets, timeline, &label_time_num_reuses, &total_num_reuses); if (label_avg_reuse_nblocks.find(label) == label_avg_reuse_nblocks.end()) { for (auto const& time_bucket : time_buckets) { label_avg_reuse_nblocks[label][time_bucket] = 0; label_avg_reuse_naccesses[label][time_bucket] = 0; } } label_avg_reuse_nblocks[label].upper_bound(avg_reuse_interval)->second += 1; label_avg_reuse_naccesses[label].upper_bound(avg_reuse_interval)->second += block.num_accesses; }; TraverseBlocks(block_callback, &labels); // Write the stats into files. WriteStatsToFile(label_str, time_buckets, kFileNameSuffixReuseInterval, label_time_num_reuses, total_num_reuses); WriteStatsToFile(label_str, time_buckets, kFileNameSuffixAvgReuseInterval, label_avg_reuse_nblocks, total_nblocks); WriteStatsToFile(label_str, time_buckets, kFileNameSuffixAvgReuseIntervalNaccesses, label_avg_reuse_naccesses, total_accesses); } void BlockCacheTraceAnalyzer::WriteReuseLifetime( const std::string& label_str, const std::vector& time_buckets) const { std::set labels = ParseLabelStr(label_str); std::map> label_lifetime_nblocks; uint64_t total_nblocks = 0; auto block_callback = [&](const std::string& cf_name, uint64_t fd, uint32_t level, TraceType type, const std::string& /*block_key*/, uint64_t block_id, const BlockAccessInfo& block) { uint64_t lifetime = 0; if (block.num_accesses > 1) { lifetime = (block.last_access_time - block.first_access_time) / kMicrosInSecond; } else { lifetime = port::kMaxUint64 - 1; } const std::string label = BuildLabel( labels, cf_name, fd, level, type, TableReaderCaller::kMaxBlockCacheLookupCaller, block_id, block); if (label_lifetime_nblocks.find(label) == label_lifetime_nblocks.end()) { // The first time we encounter this label. for (auto const& time_bucket : time_buckets) { label_lifetime_nblocks[label][time_bucket] = 0; } } label_lifetime_nblocks[label].upper_bound(lifetime)->second += 1; total_nblocks += 1; }; TraverseBlocks(block_callback, &labels); WriteStatsToFile(label_str, time_buckets, kFileNameSuffixReuseLifetime, label_lifetime_nblocks, total_nblocks); } void BlockCacheTraceAnalyzer::WriteBlockReuseTimeline( const uint64_t reuse_window, bool user_access_only, TraceType block_type) const { // A map from block key to an array of bools that states whether a block is // accessed in a time window. std::map> block_accessed; const uint64_t trace_duration = trace_end_timestamp_in_seconds_ - trace_start_timestamp_in_seconds_; const uint64_t reuse_vector_size = (trace_duration / reuse_window); if (reuse_vector_size < 2) { // The reuse window is less than 2. We cannot calculate the reused // percentage of blocks. return; } auto block_callback = [&](const std::string& /*cf_name*/, uint64_t /*fd*/, uint32_t /*level*/, TraceType /*type*/, const std::string& /*block_key*/, uint64_t block_id, const BlockAccessInfo& block) { if (block_accessed.find(block_id) == block_accessed.end()) { block_accessed[block_id].resize(reuse_vector_size); for (uint64_t i = 0; i < reuse_vector_size; i++) { block_accessed[block_id][i] = false; } } for (auto const& caller_num : block.caller_num_accesses_timeline) { const TableReaderCaller caller = caller_num.first; for (auto const& timeline : caller_num.second) { const uint64_t timestamp = timeline.first; const uint64_t elapsed_time = timestamp - trace_start_timestamp_in_seconds_; if (!user_access_only || is_user_access(caller)) { uint64_t index = std::min(elapsed_time / reuse_window, reuse_vector_size - 1); block_accessed[block_id][index] = true; } } } }; TraverseBlocks(block_callback); // A cell is the number of blocks accessed in a reuse window. std::unique_ptr reuse_table(new uint64_t[reuse_vector_size * reuse_vector_size]); for (uint64_t start_time = 0; start_time < reuse_vector_size; start_time++) { // Initialize the reuse_table. for (uint64_t i = 0; i < reuse_vector_size; i++) { reuse_table[start_time * reuse_vector_size + i] = 0; } // Examine all blocks. for (auto const& block : block_accessed) { for (uint64_t i = start_time; i < reuse_vector_size; i++) { if (block.second[start_time] && block.second[i]) { // This block is accessed at start time and at the current time. We // increment reuse_table[start_time][i] since it is reused at the ith // window. reuse_table[start_time * reuse_vector_size + i]++; } } } } const std::string user_access_prefix = user_access_only ? "_user_access_only_" : "_all_access_"; const std::string output_path = output_dir_ + "/" + block_type_to_string(block_type) + user_access_prefix + std::to_string(reuse_window) + "_" + kFileNameSuffixAccessReuseBlocksTimeline; std::ofstream out(output_path); if (!out.is_open()) { return; } std::string header("start_time"); for (uint64_t start_time = 0; start_time < reuse_vector_size; start_time++) { header += ","; header += std::to_string(start_time); } out << header << std::endl; for (uint64_t start_time = 0; start_time < reuse_vector_size; start_time++) { std::string row(std::to_string(start_time * reuse_window)); for (uint64_t j = 0; j < reuse_vector_size; j++) { row += ","; if (j < start_time) { row += "100.0"; } else { row += std::to_string(percent(reuse_table[start_time * reuse_vector_size + j], reuse_table[start_time * reuse_vector_size + start_time])); } } out << row << std::endl; } out.close(); } std::string BlockCacheTraceAnalyzer::OutputPercentAccessStats( uint64_t total_accesses, const std::map& cf_access_count) const { std::string row; for (auto const& cf_aggregates : cf_aggregates_map_) { const std::string& cf_name = cf_aggregates.first; const auto& naccess = cf_access_count.find(cf_name); row += ","; if (naccess != cf_access_count.end()) { row += std::to_string(percent(naccess->second, total_accesses)); } else { row += "0"; } } return row; } void BlockCacheTraceAnalyzer::WritePercentAccessSummaryStats() const { std::map> caller_cf_accesses; uint64_t total_accesses = 0; auto block_callback = [&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/, TraceType /*type*/, const std::string& /*block_key*/, uint64_t /*block_id*/, const BlockAccessInfo& block) { for (auto const& caller_num : block.caller_num_access_map) { const TableReaderCaller caller = caller_num.first; const uint64_t naccess = caller_num.second; caller_cf_accesses[caller][cf_name] += naccess; total_accesses += naccess; } }; TraverseBlocks(block_callback); const std::string output_path = output_dir_ + "/" + kFileNameSuffixPercentOfAccessSummary; std::ofstream out(output_path); if (!out.is_open()) { return; } std::string header("caller"); for (auto const& cf_name : cf_aggregates_map_) { header += ","; header += cf_name.first; } out << header << std::endl; for (auto const& cf_naccess_it : caller_cf_accesses) { const TableReaderCaller caller = cf_naccess_it.first; std::string row; row += caller_to_string(caller); row += OutputPercentAccessStats(total_accesses, cf_naccess_it.second); out << row << std::endl; } out.close(); } void BlockCacheTraceAnalyzer::WriteDetailedPercentAccessSummaryStats( TableReaderCaller analyzing_caller) const { std::map> level_cf_accesses; std::map> bt_cf_accesses; uint64_t total_accesses = 0; auto block_callback = [&](const std::string& cf_name, uint64_t /*fd*/, uint32_t level, TraceType type, const std::string& /*block_key*/, uint64_t /*block_id*/, const BlockAccessInfo& block) { for (auto const& caller_num : block.caller_num_access_map) { const TableReaderCaller caller = caller_num.first; if (caller == analyzing_caller) { const uint64_t naccess = caller_num.second; level_cf_accesses[level][cf_name] += naccess; bt_cf_accesses[type][cf_name] += naccess; total_accesses += naccess; } } }; TraverseBlocks(block_callback); { const std::string output_path = output_dir_ + "/" + caller_to_string(analyzing_caller) + "_level_" + kFileNameSuffixPercentOfAccessSummary; std::ofstream out(output_path); if (!out.is_open()) { return; } std::string header("level"); for (auto const& cf_name : cf_aggregates_map_) { header += ","; header += cf_name.first; } out << header << std::endl; for (auto const& level_naccess_it : level_cf_accesses) { const uint32_t level = level_naccess_it.first; std::string row; row += std::to_string(level); row += OutputPercentAccessStats(total_accesses, level_naccess_it.second); out << row << std::endl; } out.close(); } { const std::string output_path = output_dir_ + "/" + caller_to_string(analyzing_caller) + "_bt_" + kFileNameSuffixPercentOfAccessSummary; std::ofstream out(output_path); if (!out.is_open()) { return; } std::string header("bt"); for (auto const& cf_name : cf_aggregates_map_) { header += ","; header += cf_name.first; } out << header << std::endl; for (auto const& bt_naccess_it : bt_cf_accesses) { const TraceType bt = bt_naccess_it.first; std::string row; row += block_type_to_string(bt); row += OutputPercentAccessStats(total_accesses, bt_naccess_it.second); out << row << std::endl; } out.close(); } } void BlockCacheTraceAnalyzer::WriteAccessCountSummaryStats( const std::vector& access_count_buckets, bool user_access_only) const { // x: buckets. // y: # of accesses. std::map> bt_access_nblocks; std::map> cf_access_nblocks; uint64_t total_nblocks = 0; auto block_callback = [&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/, TraceType type, const std::string& /*block_key*/, uint64_t /*block_id*/, const BlockAccessInfo& block) { const std::string type_str = block_type_to_string(type); if (cf_access_nblocks.find(cf_name) == cf_access_nblocks.end()) { // initialize. for (auto& access : access_count_buckets) { cf_access_nblocks[cf_name][access] = 0; } } if (bt_access_nblocks.find(type_str) == bt_access_nblocks.end()) { // initialize. for (auto& access : access_count_buckets) { bt_access_nblocks[type_str][access] = 0; } } uint64_t naccesses = 0; for (auto const& caller_access : block.caller_num_access_map) { if (!user_access_only || is_user_access(caller_access.first)) { naccesses += caller_access.second; } } if (naccesses == 0) { return; } total_nblocks += 1; bt_access_nblocks[type_str].upper_bound(naccesses)->second += 1; cf_access_nblocks[cf_name].upper_bound(naccesses)->second += 1; }; TraverseBlocks(block_callback); const std::string user_access_prefix = user_access_only ? "user_access_only_" : "all_access_"; WriteStatsToFile("cf", access_count_buckets, user_access_prefix + kFileNameSuffixAccessCountSummary, cf_access_nblocks, total_nblocks); WriteStatsToFile("bt", access_count_buckets, user_access_prefix + kFileNameSuffixAccessCountSummary, bt_access_nblocks, total_nblocks); } BlockCacheTraceAnalyzer::BlockCacheTraceAnalyzer( const std::string& trace_file_path, const std::string& output_dir, const std::string& human_readable_trace_file_path, bool compute_reuse_distance, bool mrc_only, bool is_human_readable_trace_file, std::unique_ptr&& cache_simulator) : env_(ROCKSDB_NAMESPACE::Env::Default()), trace_file_path_(trace_file_path), output_dir_(output_dir), human_readable_trace_file_path_(human_readable_trace_file_path), compute_reuse_distance_(compute_reuse_distance), mrc_only_(mrc_only), is_human_readable_trace_file_(is_human_readable_trace_file), cache_simulator_(std::move(cache_simulator)) {} void BlockCacheTraceAnalyzer::ComputeReuseDistance( BlockAccessInfo* info) const { assert(info); if (info->num_accesses == 0) { return; } uint64_t reuse_distance = 0; for (auto const& block_key : info->unique_blocks_since_last_access) { auto const& it = block_info_map_.find(block_key); // This block must exist. assert(it != block_info_map_.end()); reuse_distance += it->second->block_size; } info->reuse_distance_count[reuse_distance] += 1; // We clear this hash set since this is the second access on this block. info->unique_blocks_since_last_access.clear(); } Status BlockCacheTraceAnalyzer::RecordAccess( const BlockCacheTraceRecord& access) { ColumnFamilyAccessInfoAggregate& cf_aggr = cf_aggregates_map_[access.cf_name]; SSTFileAccessInfoAggregate& file_aggr = cf_aggr.fd_aggregates_map[access.sst_fd_number]; file_aggr.level = access.level; BlockTypeAccessInfoAggregate& block_type_aggr = file_aggr.block_type_aggregates_map[access.block_type]; if (block_type_aggr.block_access_info_map.find(access.block_key) == block_type_aggr.block_access_info_map.end()) { block_type_aggr.block_access_info_map[access.block_key].block_id = unique_block_id_; unique_block_id_++; } BlockAccessInfo& block_access_info = block_type_aggr.block_access_info_map[access.block_key]; if (compute_reuse_distance_) { ComputeReuseDistance(&block_access_info); } block_access_info.AddAccess(access, access_sequence_number_); block_info_map_[access.block_key] = &block_access_info; uint64_t get_key_id = 0; if (access.caller == TableReaderCaller::kUserGet && access.get_id != BlockCacheTraceHelper::kReservedGetId) { std::string user_key = ExtractUserKey(access.referenced_key).ToString(); if (get_key_info_map_.find(user_key) == get_key_info_map_.end()) { get_key_info_map_[user_key].key_id = unique_get_key_id_; unique_get_key_id_++; } get_key_id = get_key_info_map_[user_key].key_id; get_key_info_map_[user_key].AddAccess(access, access_sequence_number_); } if (compute_reuse_distance_) { // Add this block to all existing blocks. for (auto& cf_aggregates : cf_aggregates_map_) { for (auto& file_aggregates : cf_aggregates.second.fd_aggregates_map) { for (auto& block_type_aggregates : file_aggregates.second.block_type_aggregates_map) { for (auto& existing_block : block_type_aggregates.second.block_access_info_map) { existing_block.second.unique_blocks_since_last_access.insert( access.block_key); } } } } } return human_readable_trace_writer_.WriteHumanReadableTraceRecord( access, block_access_info.block_id, get_key_id); } Status BlockCacheTraceAnalyzer::Analyze() { std::unique_ptr reader; Status s = Status::OK(); if (is_human_readable_trace_file_) { reader.reset(new BlockCacheHumanReadableTraceReader(trace_file_path_)); } else { std::unique_ptr trace_reader; s = NewFileTraceReader(env_, EnvOptions(), trace_file_path_, &trace_reader); if (!s.ok()) { return s; } reader.reset(new BlockCacheTraceReader(std::move(trace_reader))); s = reader->ReadHeader(&header_); if (!s.ok()) { return s; } } if (!human_readable_trace_file_path_.empty()) { s = human_readable_trace_writer_.NewWritableFile( human_readable_trace_file_path_, env_); if (!s.ok()) { return s; } } uint64_t start = env_->NowMicros(); uint64_t time_interval = 0; while (s.ok()) { BlockCacheTraceRecord access; s = reader->ReadAccess(&access); if (!s.ok()) { break; } if (!mrc_only_) { s = RecordAccess(access); if (!s.ok()) { break; } } if (trace_start_timestamp_in_seconds_ == 0) { trace_start_timestamp_in_seconds_ = access.access_timestamp / kMicrosInSecond; } trace_end_timestamp_in_seconds_ = access.access_timestamp / kMicrosInSecond; miss_ratio_stats_.UpdateMetrics(access.access_timestamp, is_user_access(access.caller), access.is_cache_hit == Boolean::kFalse); if (cache_simulator_) { cache_simulator_->Access(access); } access_sequence_number_++; uint64_t now = env_->NowMicros(); uint64_t duration = (now - start) / kMicrosInSecond; if (duration > 10 * time_interval) { uint64_t trace_duration = trace_end_timestamp_in_seconds_ - trace_start_timestamp_in_seconds_; fprintf(stdout, "Running for %" PRIu64 " seconds: Processed %" PRIu64 " records/second. Trace duration %" PRIu64 " seconds. Observed miss ratio %.2f\n", duration, duration > 0 ? access_sequence_number_ / duration : 0, trace_duration, miss_ratio_stats_.miss_ratio()); time_interval++; } } uint64_t now = env_->NowMicros(); uint64_t duration = (now - start) / kMicrosInSecond; uint64_t trace_duration = trace_end_timestamp_in_seconds_ - trace_start_timestamp_in_seconds_; fprintf(stdout, "Running for %" PRIu64 " seconds: Processed %" PRIu64 " records/second. Trace duration %" PRIu64 " seconds. Observed miss ratio %.2f\n", duration, duration > 0 ? access_sequence_number_ / duration : 0, trace_duration, miss_ratio_stats_.miss_ratio()); return s; } void BlockCacheTraceAnalyzer::PrintBlockSizeStats() const { HistogramStat bs_stats; std::map bt_stats_map; std::map> cf_bt_stats_map; auto block_callback = [&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/, TraceType type, const std::string& /*block_key*/, uint64_t /*block_id*/, const BlockAccessInfo& block) { if (block.block_size == 0) { // Block size may be 0 when 1) compaction observes a cache miss and // does not insert the missing block into the cache again. 2) // fetching filter blocks in SST files at the last level. return; } bs_stats.Add(block.block_size); bt_stats_map[type].Add(block.block_size); cf_bt_stats_map[cf_name][type].Add(block.block_size); }; TraverseBlocks(block_callback); fprintf(stdout, "Block size stats: \n%s", bs_stats.ToString().c_str()); for (auto const& bt_stats : bt_stats_map) { print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Block size stats for block type %s: \n%s", block_type_to_string(bt_stats.first).c_str(), bt_stats.second.ToString().c_str()); } for (auto const& cf_bt_stats : cf_bt_stats_map) { const std::string& cf_name = cf_bt_stats.first; for (auto const& bt_stats : cf_bt_stats.second) { print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Block size stats for column family %s and block type %s: \n%s", cf_name.c_str(), block_type_to_string(bt_stats.first).c_str(), bt_stats.second.ToString().c_str()); } } } void BlockCacheTraceAnalyzer::PrintAccessCountStats(bool user_access_only, uint32_t bottom_k, uint32_t top_k) const { HistogramStat access_stats; std::map bt_stats_map; std::map> cf_bt_stats_map; std::map> access_count_blocks; auto block_callback = [&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/, TraceType type, const std::string& block_key, uint64_t /*block_id*/, const BlockAccessInfo& block) { uint64_t naccesses = 0; for (auto const& caller_access : block.caller_num_access_map) { if (!user_access_only || is_user_access(caller_access.first)) { naccesses += caller_access.second; } } if (naccesses == 0) { return; } if (type == TraceType::kBlockTraceDataBlock) { access_count_blocks[naccesses].push_back(block_key); } access_stats.Add(naccesses); bt_stats_map[type].Add(naccesses); cf_bt_stats_map[cf_name][type].Add(naccesses); }; TraverseBlocks(block_callback); fprintf(stdout, "Block access count stats: The number of accesses per block. %s\n%s", user_access_only ? "User accesses only" : "All accesses", access_stats.ToString().c_str()); uint32_t bottom_k_index = 0; for (auto naccess_it = access_count_blocks.begin(); naccess_it != access_count_blocks.end(); naccess_it++) { bottom_k_index++; if (bottom_k_index >= bottom_k) { break; } std::map caller_naccesses; uint64_t naccesses = 0; for (auto const& block_id : naccess_it->second) { BlockAccessInfo* block = block_info_map_.find(block_id)->second; for (auto const& caller_access : block->caller_num_access_map) { if (!user_access_only || is_user_access(caller_access.first)) { caller_naccesses[caller_access.first] += caller_access.second; naccesses += caller_access.second; } } } std::string statistics("Caller:"); for (auto const& caller_naccessess_it : caller_naccesses) { statistics += caller_to_string(caller_naccessess_it.first); statistics += ":"; statistics += std::to_string(percent(caller_naccessess_it.second, naccesses)); statistics += ","; } fprintf(stdout, "Bottom %" PRIu32 " access count. Access count=%" PRIu64 " nblocks=%" ROCKSDB_PRIszt " %s\n", bottom_k, naccess_it->first, naccess_it->second.size(), statistics.c_str()); } uint32_t top_k_index = 0; for (auto naccess_it = access_count_blocks.rbegin(); naccess_it != access_count_blocks.rend(); naccess_it++) { top_k_index++; if (top_k_index >= top_k) { break; } for (auto const& block_id : naccess_it->second) { BlockAccessInfo* block = block_info_map_.find(block_id)->second; std::string statistics("Caller:"); uint64_t naccesses = 0; for (auto const& caller_access : block->caller_num_access_map) { if (!user_access_only || is_user_access(caller_access.first)) { naccesses += caller_access.second; } } assert(naccesses > 0); for (auto const& caller_access : block->caller_num_access_map) { if (!user_access_only || is_user_access(caller_access.first)) { statistics += ","; statistics += caller_to_string(caller_access.first); statistics += ":"; statistics += std::to_string(percent(caller_access.second, naccesses)); } } uint64_t ref_keys_accesses = 0; uint64_t ref_keys_does_not_exist_accesses = 0; for (auto const& ref_key_caller_access : block->key_num_access_map) { for (auto const& caller_access : ref_key_caller_access.second) { if (!user_access_only || is_user_access(caller_access.first)) { ref_keys_accesses += caller_access.second; } } } for (auto const& ref_key_caller_access : block->non_exist_key_num_access_map) { for (auto const& caller_access : ref_key_caller_access.second) { if (!user_access_only || is_user_access(caller_access.first)) { ref_keys_does_not_exist_accesses += caller_access.second; } } } statistics += ",nkeys="; statistics += std::to_string(block->num_keys); statistics += ",block_size="; statistics += std::to_string(block->block_size); statistics += ",num_ref_keys="; statistics += std::to_string(block->key_num_access_map.size()); statistics += ",percent_access_ref_keys="; statistics += std::to_string(percent(ref_keys_accesses, naccesses)); statistics += ",num_ref_keys_does_not_exist="; statistics += std::to_string(block->non_exist_key_num_access_map.size()); statistics += ",percent_access_ref_keys_does_not_exist="; statistics += std::to_string(percent(ref_keys_does_not_exist_accesses, naccesses)); statistics += ",ref_data_size="; statistics += std::to_string(block->referenced_data_size); fprintf(stdout, "Top %" PRIu32 " access count blocks access_count=%" PRIu64 " %s\n", top_k, naccess_it->first, statistics.c_str()); } } for (auto const& bt_stats : bt_stats_map) { print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Break down by block type %s: \n%s", block_type_to_string(bt_stats.first).c_str(), bt_stats.second.ToString().c_str()); } for (auto const& cf_bt_stats : cf_bt_stats_map) { const std::string& cf_name = cf_bt_stats.first; for (auto const& bt_stats : cf_bt_stats.second) { print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Break down by column family %s and block type " "%s: \n%s", cf_name.c_str(), block_type_to_string(bt_stats.first).c_str(), bt_stats.second.ToString().c_str()); } } } void BlockCacheTraceAnalyzer::PrintDataBlockAccessStats() const { HistogramStat existing_keys_stats; std::map cf_existing_keys_stats_map; HistogramStat non_existing_keys_stats; std::map cf_non_existing_keys_stats_map; HistogramStat block_access_stats; std::map cf_block_access_info; HistogramStat percent_referenced_bytes; std::map cf_percent_referenced_bytes; // Total number of accesses in a data block / number of keys in a data block. HistogramStat avg_naccesses_per_key_in_a_data_block; std::map cf_avg_naccesses_per_key_in_a_data_block; // The standard deviation on the number of accesses of a key in a data block. HistogramStat stdev_naccesses_per_key_in_a_data_block; std::map cf_stdev_naccesses_per_key_in_a_data_block; auto block_callback = [&](const std::string& cf_name, uint64_t /*fd*/, uint32_t /*level*/, TraceType /*type*/, const std::string& /*block_key*/, uint64_t /*block_id*/, const BlockAccessInfo& block) { if (block.num_keys == 0) { return; } // Use four decimal points. uint64_t percent_referenced_for_existing_keys = (uint64_t)( ((double)block.key_num_access_map.size() / (double)block.num_keys) * 10000.0); uint64_t percent_referenced_for_non_existing_keys = (uint64_t)(((double)block.non_exist_key_num_access_map.size() / (double)block.num_keys) * 10000.0); uint64_t percent_accesses_for_existing_keys = (uint64_t)(((double)block.num_referenced_key_exist_in_block / (double)block.num_accesses) * 10000.0); HistogramStat hist_naccess_per_key; for (auto const& key_access : block.key_num_access_map) { for (auto const& caller_access : key_access.second) { hist_naccess_per_key.Add(caller_access.second); } } uint64_t avg_accesses = static_cast(hist_naccess_per_key.Average()); uint64_t stdev_accesses = static_cast(hist_naccess_per_key.StandardDeviation()); avg_naccesses_per_key_in_a_data_block.Add(avg_accesses); cf_avg_naccesses_per_key_in_a_data_block[cf_name].Add(avg_accesses); stdev_naccesses_per_key_in_a_data_block.Add(stdev_accesses); cf_stdev_naccesses_per_key_in_a_data_block[cf_name].Add(stdev_accesses); existing_keys_stats.Add(percent_referenced_for_existing_keys); cf_existing_keys_stats_map[cf_name].Add( percent_referenced_for_existing_keys); non_existing_keys_stats.Add(percent_referenced_for_non_existing_keys); cf_non_existing_keys_stats_map[cf_name].Add( percent_referenced_for_non_existing_keys); block_access_stats.Add(percent_accesses_for_existing_keys); cf_block_access_info[cf_name].Add(percent_accesses_for_existing_keys); }; TraverseBlocks(block_callback); fprintf(stdout, "Histogram on the number of referenced keys existing in a block over " "the total number of keys in a block: \n%s", existing_keys_stats.ToString().c_str()); for (auto const& cf_stats : cf_existing_keys_stats_map) { print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Break down by column family %s: \n%s", cf_stats.first.c_str(), cf_stats.second.ToString().c_str()); } print_break_lines(/*num_break_lines=*/1); fprintf( stdout, "Histogram on the number of referenced keys DO NOT exist in a block over " "the total number of keys in a block: \n%s", non_existing_keys_stats.ToString().c_str()); for (auto const& cf_stats : cf_non_existing_keys_stats_map) { print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Break down by column family %s: \n%s", cf_stats.first.c_str(), cf_stats.second.ToString().c_str()); } print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Histogram on the number of accesses on keys exist in a block over " "the total number of accesses in a block: \n%s", block_access_stats.ToString().c_str()); for (auto const& cf_stats : cf_block_access_info) { print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Break down by column family %s: \n%s", cf_stats.first.c_str(), cf_stats.second.ToString().c_str()); } print_break_lines(/*num_break_lines=*/1); fprintf( stdout, "Histogram on the average number of accesses per key in a block: \n%s", avg_naccesses_per_key_in_a_data_block.ToString().c_str()); for (auto const& cf_stats : cf_avg_naccesses_per_key_in_a_data_block) { fprintf(stdout, "Break down by column family %s: \n%s", cf_stats.first.c_str(), cf_stats.second.ToString().c_str()); } print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Histogram on the standard deviation of the number of accesses per " "key in a block: \n%s", stdev_naccesses_per_key_in_a_data_block.ToString().c_str()); for (auto const& cf_stats : cf_stdev_naccesses_per_key_in_a_data_block) { fprintf(stdout, "Break down by column family %s: \n%s", cf_stats.first.c_str(), cf_stats.second.ToString().c_str()); } } void BlockCacheTraceAnalyzer::PrintStatsSummary() const { uint64_t total_num_files = 0; uint64_t total_num_blocks = 0; uint64_t total_num_accesses = 0; std::map bt_num_blocks_map; std::map caller_num_access_map; std::map> caller_bt_num_access_map; std::map> caller_level_num_access_map; for (auto const& cf_aggregates : cf_aggregates_map_) { // Stats per column family. const std::string& cf_name = cf_aggregates.first; uint64_t cf_num_files = 0; uint64_t cf_num_blocks = 0; std::map cf_bt_blocks; uint64_t cf_num_accesses = 0; std::map cf_caller_num_accesses_map; std::map> cf_caller_level_num_accesses_map; std::map> cf_caller_file_num_accesses_map; std::map> cf_caller_bt_num_accesses_map; total_num_files += cf_aggregates.second.fd_aggregates_map.size(); for (auto const& file_aggregates : cf_aggregates.second.fd_aggregates_map) { // Stats per SST file. const uint64_t fd = file_aggregates.first; const uint32_t level = file_aggregates.second.level; cf_num_files++; for (auto const& block_type_aggregates : file_aggregates.second.block_type_aggregates_map) { // Stats per block type. const TraceType type = block_type_aggregates.first; cf_bt_blocks[type] += block_type_aggregates.second.block_access_info_map.size(); total_num_blocks += block_type_aggregates.second.block_access_info_map.size(); bt_num_blocks_map[type] += block_type_aggregates.second.block_access_info_map.size(); for (auto const& block_access_info : block_type_aggregates.second.block_access_info_map) { // Stats per block. cf_num_blocks++; for (auto const& stats : block_access_info.second.caller_num_access_map) { // Stats per caller. const TableReaderCaller caller = stats.first; const uint64_t num_accesses = stats.second; // Overall stats. total_num_accesses += num_accesses; caller_num_access_map[caller] += num_accesses; caller_bt_num_access_map[caller][type] += num_accesses; caller_level_num_access_map[caller][level] += num_accesses; // Column Family stats. cf_num_accesses += num_accesses; cf_caller_num_accesses_map[caller] += num_accesses; cf_caller_level_num_accesses_map[caller][level] += num_accesses; cf_caller_file_num_accesses_map[caller][fd] += num_accesses; cf_caller_bt_num_accesses_map[caller][type] += num_accesses; } } } } // Print stats. print_break_lines(/*num_break_lines=*/3); fprintf(stdout, "Statistics for column family %s:\n", cf_name.c_str()); fprintf(stdout, " Number of files:%" PRIu64 " Number of blocks: %" PRIu64 " Number of accesses: %" PRIu64 "\n", cf_num_files, cf_num_blocks, cf_num_accesses); for (auto block_type : cf_bt_blocks) { fprintf(stdout, "Number of %s blocks: %" PRIu64 " Percent: %.2f\n", block_type_to_string(block_type.first).c_str(), block_type.second, percent(block_type.second, cf_num_blocks)); } for (auto caller : cf_caller_num_accesses_map) { const uint64_t naccesses = caller.second; print_break_lines(/*num_break_lines=*/1); fprintf(stdout, "Caller %s: Number of accesses %" PRIu64 " Percent: %.2f\n", caller_to_string(caller.first).c_str(), naccesses, percent(naccesses, cf_num_accesses)); fprintf(stdout, "Caller %s: Number of accesses per level break down\n", caller_to_string(caller.first).c_str()); for (auto naccess_level : cf_caller_level_num_accesses_map[caller.first]) { fprintf(stdout, "\t Level %" PRIu64 ": Number of accesses: %" PRIu64 " Percent: %.2f\n", naccess_level.first, naccess_level.second, percent(naccess_level.second, naccesses)); } fprintf(stdout, "Caller %s: Number of accesses per file break down\n", caller_to_string(caller.first).c_str()); for (auto naccess_file : cf_caller_file_num_accesses_map[caller.first]) { fprintf(stdout, "\t File %" PRIu64 ": Number of accesses: %" PRIu64 " Percent: %.2f\n", naccess_file.first, naccess_file.second, percent(naccess_file.second, naccesses)); } fprintf(stdout, "Caller %s: Number of accesses per block type break down\n", caller_to_string(caller.first).c_str()); for (auto naccess_type : cf_caller_bt_num_accesses_map[caller.first]) { fprintf(stdout, "\t Block Type %s: Number of accesses: %" PRIu64 " Percent: %.2f\n", block_type_to_string(naccess_type.first).c_str(), naccess_type.second, percent(naccess_type.second, naccesses)); } } } print_break_lines(/*num_break_lines=*/3); fprintf(stdout, "Overall statistics:\n"); fprintf(stdout, "Number of files: %" PRIu64 " Number of blocks: %" PRIu64 " Number of accesses: %" PRIu64 "\n", total_num_files, total_num_blocks, total_num_accesses); for (auto block_type : bt_num_blocks_map) { fprintf(stdout, "Number of %s blocks: %" PRIu64 " Percent: %.2f\n", block_type_to_string(block_type.first).c_str(), block_type.second, percent(block_type.second, total_num_blocks)); } for (auto caller : caller_num_access_map) { print_break_lines(/*num_break_lines=*/1); uint64_t naccesses = caller.second; fprintf(stdout, "Caller %s: Number of accesses %" PRIu64 " Percent: %.2f\n", caller_to_string(caller.first).c_str(), naccesses, percent(naccesses, total_num_accesses)); fprintf(stdout, "Caller %s: Number of accesses per level break down\n", caller_to_string(caller.first).c_str()); for (auto naccess_level : caller_level_num_access_map[caller.first]) { fprintf(stdout, "\t Level %d: Number of accesses: %" PRIu64 " Percent: %.2f\n", naccess_level.first, naccess_level.second, percent(naccess_level.second, naccesses)); } fprintf(stdout, "Caller %s: Number of accesses per block type break down\n", caller_to_string(caller.first).c_str()); for (auto naccess_type : caller_bt_num_access_map[caller.first]) { fprintf(stdout, "\t Block Type %s: Number of accesses: %" PRIu64 " Percent: %.2f\n", block_type_to_string(naccess_type.first).c_str(), naccess_type.second, percent(naccess_type.second, naccesses)); } } } std::vector parse_cache_config_file( const std::string& config_path) { std::ifstream file(config_path); if (!file.is_open()) { return {}; } std::vector configs; std::string line; while (getline(file, line)) { CacheConfiguration cache_config; std::stringstream ss(line); std::vector config_strs; while (ss.good()) { std::string substr; getline(ss, substr, ','); config_strs.push_back(substr); } // Sanity checks. if (config_strs.size() < 4) { fprintf(stderr, "Invalid cache simulator configuration %s\n", line.c_str()); exit(1); } if (kSupportedCacheNames.find(" " + config_strs[0] + " ") == std::string::npos) { fprintf(stderr, "Invalid cache name %s. Supported cache names are %s\n", line.c_str(), kSupportedCacheNames.c_str()); exit(1); } cache_config.cache_name = config_strs[0]; cache_config.num_shard_bits = ParseUint32(config_strs[1]); cache_config.ghost_cache_capacity = ParseUint64(config_strs[2]); for (uint32_t i = 3; i < config_strs.size(); i++) { uint64_t capacity = ParseUint64(config_strs[i]); if (capacity == 0) { fprintf(stderr, "Invalid cache capacity %s, %s\n", config_strs[i].c_str(), line.c_str()); exit(1); } cache_config.cache_capacities.push_back(capacity); } configs.push_back(cache_config); } file.close(); return configs; } std::vector parse_buckets(const std::string& bucket_str) { std::vector buckets; std::stringstream ss(bucket_str); while (ss.good()) { std::string bucket; getline(ss, bucket, ','); buckets.push_back(ParseUint64(bucket)); } buckets.push_back(port::kMaxUint64); return buckets; } int block_cache_trace_analyzer_tool(int argc, char** argv) { ParseCommandLineFlags(&argc, &argv, true); if (FLAGS_block_cache_trace_path.empty()) { fprintf(stderr, "block cache trace path is empty\n"); exit(1); } uint64_t warmup_seconds = FLAGS_cache_sim_warmup_seconds > 0 ? FLAGS_cache_sim_warmup_seconds : 0; uint32_t downsample_ratio = FLAGS_block_cache_trace_downsample_ratio > 0 ? FLAGS_block_cache_trace_downsample_ratio : 0; std::vector cache_configs = parse_cache_config_file(FLAGS_block_cache_sim_config_path); std::unique_ptr cache_simulator; if (!cache_configs.empty()) { cache_simulator.reset(new BlockCacheTraceSimulator( warmup_seconds, downsample_ratio, cache_configs)); Status s = cache_simulator->InitializeCaches(); if (!s.ok()) { fprintf(stderr, "Cannot initialize cache simulators %s\n", s.ToString().c_str()); exit(1); } } BlockCacheTraceAnalyzer analyzer( FLAGS_block_cache_trace_path, FLAGS_block_cache_analysis_result_dir, FLAGS_human_readable_trace_file_path, !FLAGS_reuse_distance_labels.empty(), FLAGS_mrc_only, FLAGS_is_block_cache_human_readable_trace, std::move(cache_simulator)); Status s = analyzer.Analyze(); if (!s.IsIncomplete() && !s.ok()) { // Read all traces. fprintf(stderr, "Cannot process the trace %s\n", s.ToString().c_str()); exit(1); } fprintf(stdout, "Status: %s\n", s.ToString().c_str()); analyzer.WriteMissRatioCurves(); analyzer.WriteMissRatioTimeline(1); analyzer.WriteMissRatioTimeline(kSecondInMinute); analyzer.WriteMissRatioTimeline(kSecondInHour); analyzer.WriteMissTimeline(1); analyzer.WriteMissTimeline(kSecondInMinute); analyzer.WriteMissTimeline(kSecondInHour); if (FLAGS_mrc_only) { fprintf(stdout, "Skipping the analysis statistics since the user wants to compute " "MRC only"); return 0; } analyzer.PrintStatsSummary(); if (FLAGS_print_access_count_stats) { print_break_lines(/*num_break_lines=*/3); analyzer.PrintAccessCountStats( /*user_access_only=*/false, FLAGS_analyze_bottom_k_access_count_blocks, FLAGS_analyze_top_k_access_count_blocks); print_break_lines(/*num_break_lines=*/3); analyzer.PrintAccessCountStats( /*user_access_only=*/true, FLAGS_analyze_bottom_k_access_count_blocks, FLAGS_analyze_top_k_access_count_blocks); } if (FLAGS_print_block_size_stats) { print_break_lines(/*num_break_lines=*/3); analyzer.PrintBlockSizeStats(); } if (FLAGS_print_data_block_access_count_stats) { print_break_lines(/*num_break_lines=*/3); analyzer.PrintDataBlockAccessStats(); } print_break_lines(/*num_break_lines=*/3); if (!FLAGS_timeline_labels.empty()) { std::stringstream ss(FLAGS_timeline_labels); while (ss.good()) { std::string label; getline(ss, label, ','); if (label.find("block") != std::string::npos) { analyzer.WriteAccessTimeline(label, kSecondInMinute, true); analyzer.WriteAccessTimeline(label, kSecondInMinute, false); analyzer.WriteAccessTimeline(label, kSecondInHour, true); analyzer.WriteAccessTimeline(label, kSecondInHour, false); } else { analyzer.WriteAccessTimeline(label, kSecondInMinute, false); analyzer.WriteAccessTimeline(label, kSecondInHour, false); } } } if (!FLAGS_analyze_callers.empty()) { analyzer.WritePercentAccessSummaryStats(); std::stringstream ss(FLAGS_analyze_callers); while (ss.good()) { std::string caller; getline(ss, caller, ','); analyzer.WriteDetailedPercentAccessSummaryStats(string_to_caller(caller)); } } if (!FLAGS_access_count_buckets.empty()) { std::vector buckets = parse_buckets(FLAGS_access_count_buckets); analyzer.WriteAccessCountSummaryStats(buckets, /*user_access_only=*/true); analyzer.WriteAccessCountSummaryStats(buckets, /*user_access_only=*/false); } if (!FLAGS_reuse_distance_labels.empty() && !FLAGS_reuse_distance_buckets.empty()) { std::vector buckets = parse_buckets(FLAGS_reuse_distance_buckets); std::stringstream ss(FLAGS_reuse_distance_labels); while (ss.good()) { std::string label; getline(ss, label, ','); analyzer.WriteReuseDistance(label, buckets); } } if (!FLAGS_reuse_interval_labels.empty() && !FLAGS_reuse_interval_buckets.empty()) { std::vector buckets = parse_buckets(FLAGS_reuse_interval_buckets); std::stringstream ss(FLAGS_reuse_interval_labels); while (ss.good()) { std::string label; getline(ss, label, ','); analyzer.WriteReuseInterval(label, buckets); } } if (!FLAGS_reuse_lifetime_labels.empty() && !FLAGS_reuse_lifetime_buckets.empty()) { std::vector buckets = parse_buckets(FLAGS_reuse_lifetime_buckets); std::stringstream ss(FLAGS_reuse_lifetime_labels); while (ss.good()) { std::string label; getline(ss, label, ','); analyzer.WriteReuseLifetime(label, buckets); } } if (FLAGS_analyze_blocks_reuse_k_reuse_window != 0) { std::vector block_types{TraceType::kBlockTraceIndexBlock, TraceType::kBlockTraceDataBlock, TraceType::kBlockTraceFilterBlock}; for (auto block_type : block_types) { analyzer.WriteBlockReuseTimeline( FLAGS_analyze_blocks_reuse_k_reuse_window, /*user_access_only=*/true, block_type); analyzer.WriteBlockReuseTimeline( FLAGS_analyze_blocks_reuse_k_reuse_window, /*user_access_only=*/false, block_type); } } if (!FLAGS_analyze_get_spatial_locality_labels.empty() && !FLAGS_analyze_get_spatial_locality_buckets.empty()) { std::vector buckets = parse_buckets(FLAGS_analyze_get_spatial_locality_buckets); std::stringstream ss(FLAGS_analyze_get_spatial_locality_labels); while (ss.good()) { std::string label; getline(ss, label, ','); analyzer.WriteGetSpatialLocality(label, buckets); } } if (!FLAGS_analyze_correlation_coefficients_labels.empty()) { std::stringstream ss(FLAGS_analyze_correlation_coefficients_labels); while (ss.good()) { std::string label; getline(ss, label, ','); analyzer.WriteCorrelationFeatures( label, FLAGS_analyze_correlation_coefficients_max_number_of_values); } analyzer.WriteCorrelationFeaturesForGet( FLAGS_analyze_correlation_coefficients_max_number_of_values); } if (!FLAGS_skew_labels.empty() && !FLAGS_skew_buckets.empty()) { std::vector buckets = parse_buckets(FLAGS_skew_buckets); std::stringstream ss(FLAGS_skew_labels); while (ss.good()) { std::string label; getline(ss, label, ','); if (label.find("block") != std::string::npos) { analyzer.WriteSkewness(label, buckets, TraceType::kBlockTraceIndexBlock); analyzer.WriteSkewness(label, buckets, TraceType::kBlockTraceFilterBlock); analyzer.WriteSkewness(label, buckets, TraceType::kBlockTraceDataBlock); analyzer.WriteSkewness(label, buckets, TraceType::kTraceMax); } else { analyzer.WriteSkewness(label, buckets, TraceType::kTraceMax); } } } return 0; } } // namespace ROCKSDB_NAMESPACE #endif // GFLAGS #endif // ROCKSDB_LITE