From abb405227848581d3e6d2ba40d94dbc0a5513902 Mon Sep 17 00:00:00 2001 From: Igor Canadi Date: Mon, 13 Apr 2015 10:17:42 -0700 Subject: [PATCH] Kill benchharness Summary: 1. it doesn't work 2. we're not using it In the future, if we need general benchmark framework, we should probably use https://github.com/google/benchmark Test Plan: make all Reviewers: yhchiang, rven, anthony, sdong Reviewed By: sdong Subscribers: dhruba, leveldb Differential Revision: https://reviews.facebook.net/D36777 --- Makefile | 12 +- db/log_and_apply_bench.cc | 96 --------- util/benchharness.cc | 407 -------------------------------------- util/benchharness.h | 356 --------------------------------- util/benchharness_test.cc | 67 ------- 5 files changed, 2 insertions(+), 936 deletions(-) delete mode 100644 db/log_and_apply_bench.cc delete mode 100644 util/benchharness.cc delete mode 100644 util/benchharness.h delete mode 100644 util/benchharness_test.cc diff --git a/Makefile b/Makefile index 01245b01e..124cd228f 100644 --- a/Makefile +++ b/Makefile @@ -188,7 +188,6 @@ MOCKOBJECTS = $(MOCK_SOURCES:.cc=.o) GTEST = $(GTEST_DIR)/gtest/gtest-all.o TESTUTIL = ./util/testutil.o TESTHARNESS = ./util/testharness.o $(TESTUTIL) $(MOCKOBJECTS) $(GTEST) -BENCHHARNESS = ./util/benchharness.o VALGRIND_ERROR = 2 VALGRIND_DIR = build_tools/VALGRIND_LOGS VALGRIND_VER := $(join $(VALGRIND_VER),valgrind) @@ -204,7 +203,6 @@ TESTS = \ table_properties_collector_test \ arena_test \ auto_roll_logger_test \ - benchharness_test \ block_test \ bloom_test \ dynamic_bloom_test \ @@ -277,7 +275,7 @@ TOOLS = \ ldb \ db_repl_stress -BENCHMARKS = db_bench table_reader_bench log_and_apply_bench cache_bench memtablerep_bench +BENCHMARKS = db_bench table_reader_bench cache_bench memtablerep_bench # The library name is configurable since we are maintaining libraries of both # debug/release mode. @@ -620,9 +618,6 @@ stringappend_test: utilities/merge_operators/string_append/stringappend_test.o $ redis_test: utilities/redis/redis_lists_test.o $(LIBOBJECTS) $(TESTHARNESS) $(AM_LINK) -benchharness_test: util/benchharness_test.o $(LIBOBJECTS) $(TESTHARNESS) $(BENCHHARNESS) - $(AM_LINK) - histogram_test: util/histogram_test.o $(LIBOBJECTS) $(TESTHARNESS) $(AM_LINK) @@ -657,9 +652,6 @@ comparator_db_test: db/comparator_db_test.o $(LIBOBJECTS) $(TESTHARNESS) table_reader_bench: table/table_reader_bench.o $(LIBOBJECTS) $(TESTHARNESS) $(AM_LINK) $(pg) -log_and_apply_bench: db/log_and_apply_bench.o $(LIBOBJECTS) $(TESTHARNESS) $(BENCHHARNESS) - $(AM_LINK) $(pg) - perf_context_test: db/perf_context_test.o $(LIBOBJECTS) $(TESTHARNESS) $(AM_V_CCLD)$(CXX) $^ $(EXEC_LDFLAGS) -o $@ $(LDFLAGS) @@ -765,7 +757,7 @@ geodb_test: utilities/geodb/geodb_test.o $(LIBOBJECTS) $(TESTHARNESS) cuckoo_table_builder_test: table/cuckoo_table_builder_test.o $(LIBOBJECTS) $(TESTHARNESS) $(AM_LINK) -cuckoo_table_reader_test: table/cuckoo_table_reader_test.o $(LIBOBJECTS) $(TESTHARNESS) $(BENCHHARNESS) +cuckoo_table_reader_test: table/cuckoo_table_reader_test.o $(LIBOBJECTS) $(TESTHARNESS) $(AM_LINK) cuckoo_table_db_test: db/cuckoo_table_db_test.o $(LIBOBJECTS) $(TESTHARNESS) diff --git a/db/log_and_apply_bench.cc b/db/log_and_apply_bench.cc deleted file mode 100644 index e5e271a1f..000000000 --- a/db/log_and_apply_bench.cc +++ /dev/null @@ -1,96 +0,0 @@ -// Copyright (c) 2013, Facebook, Inc. All rights reserved. -// This source code is licensed under the BSD-style license found in the -// LICENSE file in the root directory of this source tree. An additional grant -// of patent rights can be found in the PATENTS file in the same directory. - - -#include - -#ifndef __STDC_FORMAT_MACROS -#define __STDC_FORMAT_MACROS -#endif - -#include -#include "util/testharness.h" -#include "util/benchharness.h" -#include "db/version_set.h" -#include "db/write_controller.h" -#include "db/writebuffer.h" -#include "util/mutexlock.h" - -namespace rocksdb { - -std::string MakeKey(uint64_t num) { - char buf[30]; - snprintf(buf, sizeof(buf), "%016" PRIu64, num); - return std::string(buf); -} - -void BM_LogAndApply(int iters, int num_base_files) { - VersionSet* vset; - WriteController wc; - ColumnFamilyData* default_cfd; - uint64_t fnum = 1; - InstrumentedMutex mu; - InstrumentedMutexLock l(&mu); - - BENCHMARK_SUSPEND { - std::string dbname = test::TmpDir() + "/rocksdb_test_benchmark"; - ASSERT_OK(DestroyDB(dbname, Options())); - - DB* db = nullptr; - Options opts; - opts.create_if_missing = true; - Status s = DB::Open(opts, dbname, &db); - ASSERT_OK(s); - ASSERT_TRUE(db != nullptr); - - delete db; - db = nullptr; - - Options options; - EnvOptions sopt; - // Notice we are using the default options not through SanitizeOptions(). - // We might want to initialize some options manually if needed. - options.db_paths.emplace_back(dbname, 0); - WriteBuffer wb(options.db_write_buffer_size); - // The parameter of table cache is passed in as null, so any file I/O - // operation is likely to fail. - vset = new VersionSet(dbname, &options, sopt, nullptr, &wb, &wc); - std::vector dummy; - dummy.push_back(ColumnFamilyDescriptor()); - ASSERT_OK(vset->Recover(dummy)); - default_cfd = vset->GetColumnFamilySet()->GetDefault(); - VersionEdit vbase; - for (int i = 0; i < num_base_files; i++) { - InternalKey start(MakeKey(2 * fnum), 1, kTypeValue); - InternalKey limit(MakeKey(2 * fnum + 1), 1, kTypeDeletion); - vbase.AddFile(2, ++fnum, 0, 1 /* file size */, start, limit, 1, 1); - } - ASSERT_OK(vset->LogAndApply(default_cfd, - *default_cfd->GetLatestMutableCFOptions(), &vbase, &mu)); - } - - for (int i = 0; i < iters; i++) { - VersionEdit vedit; - vedit.DeleteFile(2, fnum); - InternalKey start(MakeKey(2 * fnum), 1, kTypeValue); - InternalKey limit(MakeKey(2 * fnum + 1), 1, kTypeDeletion); - vedit.AddFile(2, ++fnum, 0, 1 /* file size */, start, limit, 1, 1); - vset->LogAndApply(default_cfd, *default_cfd->GetLatestMutableCFOptions(), - &vedit, &mu); - } - delete vset; -} - -BENCHMARK_NAMED_PARAM(BM_LogAndApply, 1000_iters_1_file, 1000, 1) -BENCHMARK_NAMED_PARAM(BM_LogAndApply, 1000_iters_100_files, 1000, 100) -BENCHMARK_NAMED_PARAM(BM_LogAndApply, 1000_iters_10000_files, 1000, 10000) -BENCHMARK_NAMED_PARAM(BM_LogAndApply, 100_iters_100000_files, 100, 100000) - -} // namespace rocksdb - -int main(int argc, char** argv) { - rocksdb::benchmark::RunBenchmarks(); - return 0; -} diff --git a/util/benchharness.cc b/util/benchharness.cc deleted file mode 100644 index e533ed454..000000000 --- a/util/benchharness.cc +++ /dev/null @@ -1,407 +0,0 @@ -// Copyright (c) 2013, Facebook, Inc. All rights reserved. -// This source code is licensed under the BSD-style license found in the -// LICENSE file in the root directory of this source tree. An additional grant -// of patent rights can be found in the PATENTS file in the same directory. -// -// This code is derived from Benchmark.cpp implemented in Folly, the opensourced -// Facebook C++ library available at https://github.com/facebook/folly -// The code has removed any dependence on other folly and boost libraries - -#include "util/benchharness.h" - -#include -#include -#include -#include -#include -#include -#include -#include "util/string_util.h" - -#ifndef GFLAGS -bool FLAGS_benchmark = false; -uint64_t FLAGS_bm_min_usec = 100; -int64_t FLAGS_bm_min_iters = 1; -int32_t FLAGS_bm_max_secs = 1; -#else -#include -DEFINE_bool(benchmark, false, "Run benchmarks."); - -DEFINE_uint64(bm_min_usec, 100, - "Minimum # of microseconds we'll accept for each benchmark."); - -DEFINE_int64(bm_min_iters, 1, - "Minimum # of iterations we'll try for each benchmark."); - -DEFINE_int32(bm_max_secs, 1, - "Maximum # of seconds we'll spend on each benchmark."); -#endif // GFLAGS - -using std::function; -using std::get; -using std::make_pair; -using std::max; -using std::min; -using std::pair; -using std::sort; -using std::string; -using std::tuple; -using std::vector; - -namespace rocksdb { -namespace benchmark { - -BenchmarkSuspender::NanosecondsSpent BenchmarkSuspender::nsSpent; - -typedef function BenchmarkFun; -static vector> benchmarks; - -// Add the global baseline -BENCHMARK(globalBenchmarkBaseline) { - asm volatile(""); -} - -void detail::AddBenchmarkImpl(const char* file, const char* name, - BenchmarkFun fun) { - benchmarks.emplace_back(file, name, std::move(fun)); -} - -/** - * Given a point, gives density at that point as a number 0.0 < x <= - * 1.0. The result is 1.0 if all samples are equal to where, and - * decreases near 0 if all points are far away from it. The density is - * computed with the help of a radial basis function. - */ -static double Density(const double * begin, const double *const end, - const double where, const double bandwidth) { - assert(begin < end); - assert(bandwidth > 0.0); - double sum = 0.0; - for (auto i = begin; i < end; i++) { - auto d = (*i - where) / bandwidth; - sum += exp(- d * d); - } - return sum / (end - begin); -} - -/** - * Computes mean and variance for a bunch of data points. Note that - * mean is currently not being used. - */ -static pair -MeanVariance(const double * begin, const double *const end) { - assert(begin < end); - double sum = 0.0, sum2 = 0.0; - for (auto i = begin; i < end; i++) { - sum += *i; - sum2 += *i * *i; - } - auto const n = end - begin; - return make_pair(sum / n, sqrt((sum2 - sum * sum / n) / n)); -} - -/** - * Computes the mode of a sample set through brute force. Assumes - * input is sorted. - */ -static double Mode(const double * begin, const double *const end) { - assert(begin < end); - // Lower bound and upper bound for result and their respective - // densities. - auto - result = 0.0, - bestDensity = 0.0; - - // Get the variance so we pass it down to Density() - auto const sigma = MeanVariance(begin, end).second; - if (!sigma) { - // No variance means constant signal - return *begin; - } - - for (auto i = begin; i < end; i++) { - assert(i == begin || *i >= i[-1]); - auto candidate = Density(begin, end, *i, sigma * sqrt(2.0)); - if (candidate > bestDensity) { - // Found a new best - bestDensity = candidate; - result = *i; - } else { - // Density is decreasing... we could break here if we definitely - // knew this is unimodal. - } - } - - return result; -} - -/** - * Given a bunch of benchmark samples, estimate the actual run time. - */ -static double EstimateTime(double * begin, double * end) { - assert(begin < end); - - // Current state of the art: get the minimum. After some - // experimentation, it seems taking the minimum is the best. - - return *std::min_element(begin, end); - - // What follows after estimates the time as the mode of the - // distribution. - - // Select the awesomest (i.e. most frequent) result. We do this by - // sorting and then computing the longest run length. - sort(begin, end); - - // Eliminate outliers. A time much larger than the minimum time is - // considered an outlier. - while (end[-1] > 2.0 * *begin) { - --end; - if (begin == end) { -// LOG(INFO) << *begin; - } - assert(begin < end); - } - - double result = 0; - - /* Code used just for comparison purposes */ { - unsigned bestFrequency = 0; - unsigned candidateFrequency = 1; - double candidateValue = *begin; - for (auto current = begin + 1; ; ++current) { - if (current == end || *current != candidateValue) { - // Done with the current run, see if it was best - if (candidateFrequency > bestFrequency) { - bestFrequency = candidateFrequency; - result = candidateValue; - } - if (current == end) { - break; - } - // Start a new run - candidateValue = *current; - candidateFrequency = 1; - } else { - // Cool, inside a run, increase the frequency - ++candidateFrequency; - } - } - } - - result = Mode(begin, end); - - return result; -} - -static double RunBenchmarkGetNSPerIteration(const BenchmarkFun& fun, - const double globalBaseline) { - // They key here is accuracy; too low numbers means the accuracy was - // coarse. We up the ante until we get to at least minNanoseconds - // timings. - static const auto minNanoseconds = FLAGS_bm_min_usec * 1000UL; - - // We do measurements in several epochs and take the minimum, to - // account for jitter. - static const unsigned int epochs = 1000; - // We establish a total time budget as we don't want a measurement - // to take too long. This will curtail the number of actual epochs. - const uint64_t timeBudgetInNs = FLAGS_bm_max_secs * 1000000000; - auto env = Env::Default(); - uint64_t global = env->NowNanos(); - - double epochResults[epochs] = { 0 }; - size_t actualEpochs = 0; - - for (; actualEpochs < epochs; ++actualEpochs) { - for (unsigned int n = static_cast(FLAGS_bm_min_iters); - n < (1UL << 30); n *= 2) { - auto const nsecs = fun(n); - if (nsecs < minNanoseconds) { - continue; - } - // We got an accurate enough timing, done. But only save if - // smaller than the current result. - epochResults[actualEpochs] = max(0.0, - static_cast(nsecs) / n - globalBaseline); - // Done with the current epoch, we got a meaningful timing. - break; - } - uint64_t now = env->NowNanos(); - if ((now - global) >= timeBudgetInNs) { - // No more time budget available. - ++actualEpochs; - break; - } - } - - // If the benchmark was basically drowned in baseline noise, it's - // possible it became negative. - return max(0.0, EstimateTime(epochResults, epochResults + actualEpochs)); -} - -struct ScaleInfo { - double boundary; - const char* suffix; -}; - -static const ScaleInfo kTimeSuffixes[] { - { 365.25 * 24 * 3600, "years" }, - { 24 * 3600, "days" }, - { 3600, "hr" }, - { 60, "min" }, - { 1, "s" }, - { 1E-3, "ms" }, - { 1E-6, "us" }, - { 1E-9, "ns" }, - { 1E-12, "ps" }, - { 1E-15, "fs" }, - { 0, nullptr }, -}; - -static const ScaleInfo kMetricSuffixes[] { - { 1E24, "Y" }, // yotta - { 1E21, "Z" }, // zetta - { 1E18, "X" }, // "exa" written with suffix 'X' so as to not create - // confusion with scientific notation - { 1E15, "P" }, // peta - { 1E12, "T" }, // terra - { 1E9, "G" }, // giga - { 1E6, "M" }, // mega - { 1E3, "K" }, // kilo - { 1, "" }, - { 1E-3, "m" }, // milli - { 1E-6, "u" }, // micro - { 1E-9, "n" }, // nano - { 1E-12, "p" }, // pico - { 1E-15, "f" }, // femto - { 1E-18, "a" }, // atto - { 1E-21, "z" }, // zepto - { 1E-24, "y" }, // yocto - { 0, nullptr }, -}; - -static string HumanReadable(double n, unsigned int decimals, - const ScaleInfo* scales) { - if (std::isinf(n) || std::isnan(n)) { - return ToString(n); - } - - const double absValue = fabs(n); - const ScaleInfo* scale = scales; - while (absValue < scale[0].boundary && scale[1].suffix != nullptr) { - ++scale; - } - - const double scaledValue = n / scale->boundary; - char a[80]; - snprintf(a, sizeof(a), "%.*f%s", decimals, scaledValue, scale->suffix); - return a; -} - -static string ReadableTime(double n, unsigned int decimals) { - return HumanReadable(n, decimals, kTimeSuffixes); -} - -static string MetricReadable(double n, unsigned int decimals) { - return HumanReadable(n, decimals, kMetricSuffixes); -} - -static void PrintBenchmarkResultsAsTable( - const vector >& data) { - // Width available - static const uint columns = 76; - - // Compute the longest benchmark name - size_t longestName = 0; - for (size_t i = 1; i < benchmarks.size(); i++) { - longestName = max(longestName, strlen(get<1>(benchmarks[i]))); - } - - // Print a horizontal rule - auto separator = [&](char pad) { - puts(string(columns, pad).c_str()); - }; - - // Print header for a file - auto header = [&](const char* file) { - separator('='); - printf("%-*srelative time/iter iters/s\n", - columns - 28, file); - separator('='); - }; - - double baselineNsPerIter = std::numeric_limits::max(); - const char* lastFile = ""; - - for (auto& datum : data) { - auto file = get<0>(datum); - if (strcmp(file, lastFile)) { - // New file starting - header(file); - lastFile = file; - } - - string s = get<1>(datum); - if (s == "-") { - separator('-'); - continue; - } - bool useBaseline /* = void */; - if (s[0] == '%') { - s.erase(0, 1); - useBaseline = true; - } else { - baselineNsPerIter = get<2>(datum); - useBaseline = false; - } - s.resize(columns - 29, ' '); - auto nsPerIter = get<2>(datum); - auto secPerIter = nsPerIter / 1E9; - auto itersPerSec = 1 / secPerIter; - if (!useBaseline) { - // Print without baseline - printf("%*s %9s %7s\n", - static_cast(s.size()), s.c_str(), - ReadableTime(secPerIter, 2).c_str(), - MetricReadable(itersPerSec, 2).c_str()); - } else { - // Print with baseline - auto rel = baselineNsPerIter / nsPerIter * 100.0; - printf("%*s %7.2f%% %9s %7s\n", - static_cast(s.size()), s.c_str(), - rel, - ReadableTime(secPerIter, 2).c_str(), - MetricReadable(itersPerSec, 2).c_str()); - } - } - separator('='); -} - -void RunBenchmarks() { - ASSERT_TRUE(!benchmarks.empty()); - - vector> results; - results.reserve(benchmarks.size() - 1); - - // PLEASE KEEP QUIET. MEASUREMENTS IN PROGRESS. - - auto const globalBaseline = RunBenchmarkGetNSPerIteration( - get<2>(benchmarks.front()), 0); - for (size_t i = 1; i < benchmarks.size(); i++) { - double elapsed = 0.0; - if (strcmp(get<1>(benchmarks[i]), "-") != 0) { // skip separators - elapsed = RunBenchmarkGetNSPerIteration(get<2>(benchmarks[i]), - globalBaseline); - } - results.emplace_back(get<0>(benchmarks[i]), - get<1>(benchmarks[i]), elapsed); - } - - // PLEASE MAKE NOISE. MEASUREMENTS DONE. - - PrintBenchmarkResultsAsTable(results); -} - -} // namespace benchmark -} // namespace rocksdb diff --git a/util/benchharness.h b/util/benchharness.h deleted file mode 100644 index b416f31db..000000000 --- a/util/benchharness.h +++ /dev/null @@ -1,356 +0,0 @@ -// Copyright (c) 2013, Facebook, Inc. All rights reserved. -// This source code is licensed under the BSD-style license found in the -// LICENSE file in the root directory of this source tree. An additional grant -// of patent rights can be found in the PATENTS file in the same directory. -// -// This code is derived from Benchmark.h implemented in Folly, the opensourced -// Facebook C++ library available at https://github.com/facebook/folly -// The code has removed any dependence on other folly and boost libraries - -#pragma once - -#include -#include -#include - -#include "util/testharness.h" -#include "rocksdb/env.h" - -namespace rocksdb { -namespace benchmark { - -/** - * Runs all benchmarks defined. Usually put in main(). - */ -void RunBenchmarks(); - -namespace detail { - -/** - * Adds a benchmark wrapped in a std::function. Only used - * internally. Pass by value is intentional. - */ -void AddBenchmarkImpl(const char* file, - const char* name, - std::function); - -} // namespace detail - - -/** - * Supporting type for BENCHMARK_SUSPEND defined below. - */ -struct BenchmarkSuspender { - BenchmarkSuspender() { start_ = Env::Default()->NowNanos(); } - - BenchmarkSuspender(const BenchmarkSuspender&) = delete; - BenchmarkSuspender(BenchmarkSuspender && rhs) { - start_ = rhs.start_; - rhs.start_ = 0; - } - - BenchmarkSuspender& operator=(const BenchmarkSuspender &) = delete; - BenchmarkSuspender& operator=(BenchmarkSuspender && rhs) { - if (start_ > 0) { - tally(); - } - start_ = rhs.start_; - rhs.start_ = 0; - return *this; - } - - ~BenchmarkSuspender() { - if (start_ > 0) { - tally(); - } - } - - void Dismiss() { - assert(start_ > 0); - tally(); - start_ = 0; - } - - void Rehire() { start_ = Env::Default()->NowNanos(); } - - /** - * This helps the macro definition. To get around the dangers of - * operator bool, returns a pointer to member (which allows no - * arithmetic). - */ - /* implicit */ - operator int BenchmarkSuspender::*() const { return nullptr; } - - /** - * Accumulates nanoseconds spent outside benchmark. - */ - typedef uint64_t NanosecondsSpent; - static NanosecondsSpent nsSpent; - - private: - void tally() { - uint64_t end = Env::Default()->NowNanos(); - nsSpent += start_ - end; - start_ = end; - } - - uint64_t start_; -}; - -/** - * Adds a benchmark. Usually not called directly but instead through - * the macro BENCHMARK defined below. The lambda function involved - * must take exactly one parameter of type unsigned, and the benchmark - * uses it with counter semantics (iteration occurs inside the - * function). - */ -template -void -AddBenchmark_n(const char* file, const char* name, Lambda&& lambda) { - auto execute = [=](unsigned int times) -> uint64_t { - BenchmarkSuspender::nsSpent = 0; - uint64_t start, end; - auto env = Env::Default(); - - // CORE MEASUREMENT STARTS - start = env->NowNanos(); - lambda(times); - end = env->NowNanos(); - // CORE MEASUREMENT ENDS - return (end - start) - BenchmarkSuspender::nsSpent; - }; - - detail::AddBenchmarkImpl(file, name, - std::function(execute)); -} - -/** - * Adds a benchmark. Usually not called directly but instead through - * the macro BENCHMARK defined below. The lambda function involved - * must take zero parameters, and the benchmark calls it repeatedly - * (iteration occurs outside the function). - */ -template -void -AddBenchmark(const char* file, const char* name, Lambda&& lambda) { - AddBenchmark_n(file, name, [=](unsigned int times) { - while (times-- > 0) { - lambda(); - } - }); -} - -} // namespace benchmark -} // namespace rocksdb - -/** - * FB_ONE_OR_NONE(hello, world) expands to hello and - * FB_ONE_OR_NONE(hello) expands to nothing. This macro is used to - * insert or eliminate text based on the presence of another argument. - */ -#define FB_ONE_OR_NONE(a, ...) FB_THIRD(a, ## __VA_ARGS__, a) -#define FB_THIRD(a, b, ...) __VA_ARGS__ - -#define FB_CONCATENATE_IMPL(s1, s2) s1##s2 -#define FB_CONCATENATE(s1, s2) FB_CONCATENATE_IMPL(s1, s2) - -#define FB_ANONYMOUS_VARIABLE(str) \ - FB_CONCATENATE(str, __LINE__ __attribute__((__unused__))) - -#define FB_STRINGIZE(x) #x - -/** - * Introduces a benchmark function. Used internally, see BENCHMARK and - * friends below. - */ -#define BENCHMARK_IMPL_N(funName, stringName, paramType, paramName) \ - static void funName(paramType); \ - static bool FB_ANONYMOUS_VARIABLE(rocksdbBenchmarkUnused) = ( \ - ::rocksdb::benchmark::AddBenchmark_n(__FILE__, stringName, \ - [](paramType paramName) { funName(paramName); }), \ - true); \ - static void funName(paramType paramName) - -#define BENCHMARK_IMPL(funName, stringName) \ - static void funName(); \ - static bool FB_ANONYMOUS_VARIABLE(rocksdbBenchmarkUnused) = ( \ - ::rocksdb::benchmark::AddBenchmark(__FILE__, stringName, \ - []() { funName(); }), \ - true); \ - static void funName() - -/** - * Introduces a benchmark function. Use with either one one or two - * arguments. The first is the name of the benchmark. Use something - * descriptive, such as insertVectorBegin. The second argument may be - * missing, or could be a symbolic counter. The counter dictates how - * many internal iteration the benchmark does. Example: - * - * BENCHMARK(vectorPushBack) { - * vector v; - * v.push_back(42); - * } - * - * BENCHMARK_N(insertVectorBegin, n) { - * vector v; - * FOR_EACH_RANGE (i, 0, n) { - * v.insert(v.begin(), 42); - * } - * } - */ -#define BENCHMARK_N(name, ...) \ - BENCHMARK_IMPL_N( \ - name, \ - FB_STRINGIZE(name), \ - FB_ONE_OR_NONE(unsigned, ## __VA_ARGS__), \ - __VA_ARGS__) - -#define BENCHMARK(name) \ - BENCHMARK_IMPL( \ - name, \ - FB_STRINGIZE(name)) - -/** - * Defines a benchmark that passes a parameter to another one. This is - * common for benchmarks that need a "problem size" in addition to - * "number of iterations". Consider: - * - * void pushBack(uint n, size_t initialSize) { - * vector v; - * BENCHMARK_SUSPEND { - * v.resize(initialSize); - * } - * FOR_EACH_RANGE (i, 0, n) { - * v.push_back(i); - * } - * } - * BENCHMARK_PARAM(pushBack, 0) - * BENCHMARK_PARAM(pushBack, 1000) - * BENCHMARK_PARAM(pushBack, 1000000) - * - * The benchmark above estimates the speed of push_back at different - * initial sizes of the vector. The framework will pass 0, 1000, and - * 1000000 for initialSize, and the iteration count for n. - */ -#define BENCHMARK_PARAM(name, param) \ - BENCHMARK_NAMED_PARAM(name, param, param) - -/* - * Like BENCHMARK_PARAM(), but allows a custom name to be specified for each - * parameter, rather than using the parameter value. - * - * Useful when the parameter value is not a valid token for string pasting, - * of when you want to specify multiple parameter arguments. - * - * For example: - * - * void addValue(uint n, int64_t bucketSize, int64_t min, int64_t max) { - * Histogram hist(bucketSize, min, max); - * int64_t num = min; - * FOR_EACH_RANGE (i, 0, n) { - * hist.addValue(num); - * ++num; - * if (num > max) { num = min; } - * } - * } - * - * BENCHMARK_NAMED_PARAM(addValue, 0_to_100, 1, 0, 100) - * BENCHMARK_NAMED_PARAM(addValue, 0_to_1000, 10, 0, 1000) - * BENCHMARK_NAMED_PARAM(addValue, 5k_to_20k, 250, 5000, 20000) - */ -#define BENCHMARK_NAMED_PARAM(name, param_name, ...) \ - BENCHMARK_IMPL( \ - FB_CONCATENATE(name, FB_CONCATENATE(_, param_name)), \ - FB_STRINGIZE(name) "(" FB_STRINGIZE(param_name) ")") { \ - name(__VA_ARGS__); \ - } - -#define BENCHMARK_NAMED_PARAM_N(name, param_name, ...) \ - BENCHMARK_IMPL_N( \ - FB_CONCATENATE(name, FB_CONCATENATE(_, param_name)), \ - FB_STRINGIZE(name) "(" FB_STRINGIZE(param_name) ")", \ - unsigned, \ - iters) { \ - name(iters, ## __VA_ARGS__); \ - } - -/** - * Just like BENCHMARK, but prints the time relative to a - * baseline. The baseline is the most recent BENCHMARK() seen in - * lexical order. Example: - * - * // This is the baseline - * BENCHMARK_N(insertVectorBegin, n) { - * vector v; - * FOR_EACH_RANGE (i, 0, n) { - * v.insert(v.begin(), 42); - * } - * } - * - * BENCHMARK_RELATIVE_N(insertListBegin, n) { - * list s; - * FOR_EACH_RANGE (i, 0, n) { - * s.insert(s.begin(), 42); - * } - * } - * - * Any number of relative benchmark can be associated with a - * baseline. Another BENCHMARK() occurrence effectively establishes a - * new baseline. - */ -#define BENCHMARK_RELATIVE_N(name, ...) \ - BENCHMARK_IMPL_N( \ - name, \ - "%" FB_STRINGIZE(name), \ - FB_ONE_OR_NONE(unsigned, ## __VA_ARGS__), \ - __VA_ARGS__) - -#define BENCHMARK_RELATIVE(name) \ - BENCHMARK_IMPL( \ - name, \ - "%" FB_STRINGIZE(name)) - -/** - * A combination of BENCHMARK_RELATIVE and BENCHMARK_PARAM. - */ -#define BENCHMARK_RELATIVE_PARAM(name, param) \ - BENCHMARK_RELATIVE_NAMED_PARAM(name, param, param) - -/** - * A combination of BENCHMARK_RELATIVE and BENCHMARK_NAMED_PARAM. - */ -#define BENCHMARK_RELATIVE_NAMED_PARAM(name, param_name, ...) \ - BENCHMARK_IMPL_N( \ - FB_CONCATENATE(name, FB_CONCATENATE(_, param_name)), \ - "%" FB_STRINGIZE(name) "(" FB_STRINGIZE(param_name) ")", \ - unsigned, \ - iters) { \ - name(iters, ## __VA_ARGS__); \ - } - -/** - * Draws a line of dashes. - */ -#define BENCHMARK_DRAW_LINE() \ - static bool FB_ANONYMOUS_VARIABLE(rocksdbBenchmarkUnused) = ( \ - ::rocksdb::benchmark::AddBenchmark(__FILE__, "-", []() { }), \ - true); - -/** - * Allows execution of code that doesn't count torward the benchmark's - * time budget. Example: - * - * BENCHMARK_START_GROUP(insertVectorBegin, n) { - * vector v; - * BENCHMARK_SUSPEND { - * v.reserve(n); - * } - * FOR_EACH_RANGE (i, 0, n) { - * v.insert(v.begin(), 42); - * } - * } - */ -#define BENCHMARK_SUSPEND \ - if (auto FB_ANONYMOUS_VARIABLE(BENCHMARK_SUSPEND) = \ - ::rocksdb::benchmark::BenchmarkSuspender()) {} \ - else diff --git a/util/benchharness_test.cc b/util/benchharness_test.cc deleted file mode 100644 index f2c910edb..000000000 --- a/util/benchharness_test.cc +++ /dev/null @@ -1,67 +0,0 @@ -// Copyright (c) 2013, Facebook, Inc. All rights reserved. -// This source code is licensed under the BSD-style license found in the -// LICENSE file in the root directory of this source tree. An additional grant -// of patent rights can be found in the PATENTS file in the same directory. -// - -#include "util/benchharness.h" -#include - -namespace rocksdb { - -BENCHMARK(insertFrontVector) { - std::vector v; - for (size_t i = 0; i < 100; i++) { - v.insert(v.begin(), i); - } -} - -BENCHMARK_RELATIVE(insertBackVector) { - std::vector v; - for (size_t i = 0; i < 100; i++) { - v.insert(v.end(), i); - } -} - -BENCHMARK_N(insertFrontVector_n, n) { - std::vector v; - for (size_t i = 0; i < n; i++) { - v.insert(v.begin(), i); - } -} - -BENCHMARK_RELATIVE_N(insertBackVector_n, n) { - std::vector v; - for (size_t i = 0; i < n; i++) { - v.insert(v.end(), i); - } -} - -BENCHMARK_N(insertFrontEnd_n, n) { - std::vector v; - for (size_t i = 0; i < n; i++) { - v.insert(v.begin(), i); - } - for (size_t i = 0; i < n; i++) { - v.insert(v.end(), i); - } -} - -BENCHMARK_RELATIVE_N(insertFrontEndSuspend_n, n) { - std::vector v; - for (size_t i = 0; i < n; i++) { - v.insert(v.begin(), i); - } - BENCHMARK_SUSPEND { - for (size_t i = 0; i < n; i++) { - v.insert(v.end(), i); - } - } -} - -} // namespace rocksdb - -int main(int argc, char** argv) { - rocksdb::benchmark::RunBenchmarks(); - return 0; -}