Summary:
fixed typo
Closes https://github.com/facebook/rocksdb/pull/2376

Differential Revision: D5183630

Pulled By: ajkr

fbshipit-source-id: 133cfd0445959e70aa2cd1a12151bf3c0c5c3ac5
main
hyunwoo 8 years ago committed by Facebook Github Bot
parent 7e8d95cc0e
commit c7662a44a4
  1. 2
      build_tools/run_ci_db_test.ps1
  2. 2
      db/column_family.cc
  3. 2
      db/compaction_picker_universal.cc
  4. 2
      db/db_impl.cc
  5. 2
      db/external_sst_file_ingestion_job.h
  6. 2
      include/rocksdb/options.h
  7. 10
      java/rocksjni/portal.h
  8. 4
      java/rocksjni/rocksjni.cc
  9. 2
      java/rocksjni/ttl.cc
  10. 4
      java/src/main/java/org/rocksdb/CompactionOptionsUniversal.java
  11. 4
      options/options_test.cc
  12. 2
      tools/rdb/API.md
  13. 2
      tools/rdb/README.md
  14. 2
      util/build_version.h
  15. 2
      util/xxhash.cc
  16. 2
      utilities/document/document_db.cc

@ -333,7 +333,7 @@ function RunJobs($Suites, $TestCmds, [int]$ConcurrencyVal)
$Cases = @{} $Cases = @{}
if($Suites.Count -gt 0) { if($Suites.Count -gt 0) {
# Will teh first one # Will the first one
ForEach($e in $Suites.Keys) { ForEach($e in $Suites.Keys) {
$exe_name = $e $exe_name = $e
$Cases = $Suites[$e] $Cases = $Suites[$e]

@ -551,7 +551,7 @@ std::unique_ptr<WriteControllerToken> SetupDelay(
// If DB just falled into the stop condition, we need to further reduce // If DB just falled into the stop condition, we need to further reduce
// the write rate to avoid the stop condition. // the write rate to avoid the stop condition.
if (penalize_stop) { if (penalize_stop) {
// Penalize the near stop or stop condition by more agressive slowdown. // Penalize the near stop or stop condition by more aggressive slowdown.
// This is to provide the long term slowdown increase signal. // This is to provide the long term slowdown increase signal.
// The penalty is more than the reward of recovering to the normal // The penalty is more than the reward of recovering to the normal
// condition. // condition.

@ -621,7 +621,7 @@ Compaction* UniversalCompactionPicker::PickCompactionToReduceSizeAmp(
const std::string& cf_name, const MutableCFOptions& mutable_cf_options, const std::string& cf_name, const MutableCFOptions& mutable_cf_options,
VersionStorageInfo* vstorage, double score, VersionStorageInfo* vstorage, double score,
const std::vector<SortedRun>& sorted_runs, LogBuffer* log_buffer) { const std::vector<SortedRun>& sorted_runs, LogBuffer* log_buffer) {
// percentage flexibilty while reducing size amplification // percentage flexibility while reducing size amplification
uint64_t ratio = uint64_t ratio =
ioptions_.compaction_options_universal.max_size_amplification_percent; ioptions_.compaction_options_universal.max_size_amplification_percent;

@ -2448,7 +2448,7 @@ void DBImpl::EraseThreadStatusDbInfo() const {
// A global method that can dump out the build version // A global method that can dump out the build version
void DumpRocksDBBuildVersion(Logger * log) { void DumpRocksDBBuildVersion(Logger * log) {
#if !defined(IOS_CROSS_COMPILE) #if !defined(IOS_CROSS_COMPILE)
// if we compile with Xcode, we don't run build_detect_vesion, so we don't // if we compile with Xcode, we don't run build_detect_version, so we don't
// generate util/build_version.cc // generate util/build_version.cc
ROCKS_LOG_HEADER(log, "RocksDB version: %d.%d.%d\n", ROCKSDB_MAJOR, ROCKS_LOG_HEADER(log, "RocksDB version: %d.%d.%d\n", ROCKSDB_MAJOR,
ROCKSDB_MINOR, ROCKSDB_PATCH); ROCKSDB_MINOR, ROCKSDB_PATCH);

@ -97,7 +97,7 @@ class ExternalSstFileIngestionJob {
// REQUIRES: Mutex held // REQUIRES: Mutex held
void UpdateStats(); void UpdateStats();
// Cleanup after successfull/failed job // Cleanup after successful/failed job
void Cleanup(const Status& status); void Cleanup(const Status& status);
VersionEdit* edit() { return &edit_; } VersionEdit* edit() { return &edit_; }

@ -771,7 +771,7 @@ struct DBOptions {
// If enable_pipelined_write is true, separate write thread queue is // If enable_pipelined_write is true, separate write thread queue is
// maintained for WAL write and memtable write. A write thread first enter WAL // maintained for WAL write and memtable write. A write thread first enter WAL
// writer queue and then memtable writer queue. Pending thread on the WAL // writer queue and then memtable writer queue. Pending thread on the WAL
// writer queue thus only have to wait for previous writers to finish thier // writer queue thus only have to wait for previous writers to finish their
// WAL writing but not the memtable writing. Enabling the feature may improve // WAL writing but not the memtable writing. Enabling the feature may improve
// write throughput and reduce latency of the prepare phase of two-phase // write throughput and reduce latency of the prepare phase of two-phase
// commit. // commit.

@ -1635,7 +1635,7 @@ class BackupInfoListJni {
jobject jbackup_info_handle_list = jobject jbackup_info_handle_list =
env->NewObject(jarray_list_clazz, cstr_mid, backup_infos.size()); env->NewObject(jarray_list_clazz, cstr_mid, backup_infos.size());
if(env->ExceptionCheck()) { if(env->ExceptionCheck()) {
// exception occured constructing object // exception occurred constructing object
return nullptr; return nullptr;
} }
@ -1650,7 +1650,7 @@ class BackupInfoListJni {
backup_info.size, backup_info.size,
backup_info.number_files); backup_info.number_files);
if(env->ExceptionCheck()) { if(env->ExceptionCheck()) {
// exception occured constructing object // exception occurred constructing object
if(obj != nullptr) { if(obj != nullptr) {
env->DeleteLocalRef(obj); env->DeleteLocalRef(obj);
} }
@ -1663,7 +1663,7 @@ class BackupInfoListJni {
jboolean rs = jboolean rs =
env->CallBooleanMethod(jbackup_info_handle_list, add_mid, obj); env->CallBooleanMethod(jbackup_info_handle_list, add_mid, obj);
if(env->ExceptionCheck() || rs == JNI_FALSE) { if(env->ExceptionCheck() || rs == JNI_FALSE) {
// exception occured calling method, or could not add // exception occurred calling method, or could not add
if(obj != nullptr) { if(obj != nullptr) {
env->DeleteLocalRef(obj); env->DeleteLocalRef(obj);
} }
@ -1825,7 +1825,7 @@ class WriteTypeJni : public JavaClass {
env->GetStaticFieldID(jclazz, name, env->GetStaticFieldID(jclazz, name,
"Lorg/rocksdb/WBWIRocksIterator$WriteType;"); "Lorg/rocksdb/WBWIRocksIterator$WriteType;");
if(env->ExceptionCheck()) { if(env->ExceptionCheck()) {
// exception occured while getting field // exception occurred while getting field
return nullptr; return nullptr;
} else if(jfid == nullptr) { } else if(jfid == nullptr) {
return nullptr; return nullptr;
@ -1962,7 +1962,7 @@ class InfoLogLevelJni : public JavaClass {
jfieldID jfid = jfieldID jfid =
env->GetStaticFieldID(jclazz, name, "Lorg/rocksdb/InfoLogLevel;"); env->GetStaticFieldID(jclazz, name, "Lorg/rocksdb/InfoLogLevel;");
if(env->ExceptionCheck()) { if(env->ExceptionCheck()) {
// exception occured while getting field // exception occurred while getting field
return nullptr; return nullptr;
} else if(jfid == nullptr) { } else if(jfid == nullptr) {
return nullptr; return nullptr;

@ -122,7 +122,7 @@ jlongArray rocksdb_open_helper(JNIEnv* env, jlong jopt_handle,
env->ReleaseLongArrayElements(jcolumn_options, jco, JNI_ABORT); env->ReleaseLongArrayElements(jcolumn_options, jco, JNI_ABORT);
if(has_exception == JNI_TRUE) { if(has_exception == JNI_TRUE) {
// exception occured // exception occurred
env->ReleaseStringUTFChars(jdb_path, db_path); env->ReleaseStringUTFChars(jdb_path, db_path);
return nullptr; return nullptr;
} }
@ -2180,7 +2180,7 @@ void Java_org_rocksdb_RocksDB_ingestExternalFile(
rocksdb::JniUtil::copyStrings(env, jfile_path_list, jfile_path_list_len, rocksdb::JniUtil::copyStrings(env, jfile_path_list, jfile_path_list_len,
&has_exception); &has_exception);
if(has_exception == JNI_TRUE) { if(has_exception == JNI_TRUE) {
// exception occured // exception occurred
return; return;
} }

@ -91,7 +91,7 @@ jlongArray
env->ReleaseLongArrayElements(jcolumn_options, jco, JNI_ABORT); env->ReleaseLongArrayElements(jcolumn_options, jco, JNI_ABORT);
if(has_exception == JNI_TRUE) { if(has_exception == JNI_TRUE) {
// exception occured // exception occurred
env->ReleaseStringUTFChars(jdb_path, db_path); env->ReleaseStringUTFChars(jdb_path, db_path);
return nullptr; return nullptr;
} }

@ -15,7 +15,7 @@ public class CompactionOptionsUniversal extends RocksObject {
} }
/** /**
* Percentage flexibilty while comparing file size. If the candidate file(s) * Percentage flexibility while comparing file size. If the candidate file(s)
* size is 1% smaller than the next file's size, then include next file into * size is 1% smaller than the next file's size, then include next file into
* this candidate set. * this candidate set.
* *
@ -31,7 +31,7 @@ public class CompactionOptionsUniversal extends RocksObject {
} }
/** /**
* Percentage flexibilty while comparing file size. If the candidate file(s) * Percentage flexibility while comparing file size. If the candidate file(s)
* size is 1% smaller than the next file's size, then include next file into * size is 1% smaller than the next file's size, then include next file into
* this candidate set. * this candidate set.
* *

@ -336,7 +336,7 @@ TEST_F(OptionsTest, GetColumnFamilyOptionsFromStringTest) {
ASSERT_EQ(new_cf_opt.arena_block_size, 21 * tera); ASSERT_EQ(new_cf_opt.arena_block_size, 21 * tera);
// Nested block based table options // Nested block based table options
// Emtpy // Empty
ASSERT_OK(GetColumnFamilyOptionsFromString(base_cf_opt, ASSERT_OK(GetColumnFamilyOptionsFromString(base_cf_opt,
"write_buffer_size=10;max_write_buffer_number=16;" "write_buffer_size=10;max_write_buffer_number=16;"
"block_based_table_factory={};arena_block_size=1024", "block_based_table_factory={};arena_block_size=1024",
@ -404,7 +404,7 @@ TEST_F(OptionsTest, GetColumnFamilyOptionsFromStringTest) {
ASSERT_OK(RocksDBOptionsParser::VerifyCFOptions(base_cf_opt, new_cf_opt)); ASSERT_OK(RocksDBOptionsParser::VerifyCFOptions(base_cf_opt, new_cf_opt));
// Nested plain table options // Nested plain table options
// Emtpy // Empty
ASSERT_OK(GetColumnFamilyOptionsFromString(base_cf_opt, ASSERT_OK(GetColumnFamilyOptionsFromString(base_cf_opt,
"write_buffer_size=10;max_write_buffer_number=16;" "write_buffer_size=10;max_write_buffer_number=16;"
"plain_table_factory={};arena_block_size=1024", "plain_table_factory={};arena_block_size=1024",

@ -92,7 +92,7 @@
### CreateColumnFamily ### CreateColumnFamily
# Create a new column familiy for the database. # Create a new column family for the database.
# #
# column_family_name (string) - Name of the new column family. # column_family_name (string) - Name of the new column family.
# #

@ -16,7 +16,7 @@ JavaScript binding for RocksDB within a Node application.
### Installation ### Installation
NOTE: If your default `python` binary is not a version of python2, add NOTE: If your default `python` binary is not a version of python2, add
the arguments `--python /path/to/python2` to the the `node-gyp` commands. the arguments `--python /path/to/python2` to the `node-gyp` commands.
1. Make sure you have the static library (i.e. "librocksdb.a") in the root 1. Make sure you have the static library (i.e. "librocksdb.a") in the root
directory of your rocksdb installation. If not, `cd` there and run directory of your rocksdb installation. If not, `cd` there and run

@ -7,7 +7,7 @@
// //
#pragma once #pragma once
#if !defined(IOS_CROSS_COMPILE) #if !defined(IOS_CROSS_COMPILE)
// if we compile with Xcode, we don't run build_detect_vesion, so we don't // if we compile with Xcode, we don't run build_detect_version, so we don't
// generate these variables // generate these variables
// this variable tells us about the git revision // this variable tells us about the git revision
extern const char* rocksdb_build_git_sha; extern const char* rocksdb_build_git_sha;

@ -53,7 +53,7 @@ You can contact the author at :
// By default, xxHash library provides endian-independent Hash values, based on little-endian convention. // By default, xxHash library provides endian-independent Hash values, based on little-endian convention.
// Results are therefore identical for little-endian and big-endian CPU. // Results are therefore identical for little-endian and big-endian CPU.
// This comes at a performance cost for big-endian CPU, since some swapping is required to emulate little-endian format. // This comes at a performance cost for big-endian CPU, since some swapping is required to emulate little-endian format.
// Should endian-independance be of no importance for your application, you may set the #define below to 1. // Should endian-independence be of no importance for your application, you may set the #define below to 1.
// It will improve speed for Big-endian CPU. // It will improve speed for Big-endian CPU.
// This option has no impact on Little_Endian CPU. // This option has no impact on Little_Endian CPU.
#define XXH_FORCE_NATIVE_FORMAT 0 #define XXH_FORCE_NATIVE_FORMAT 0

@ -25,7 +25,7 @@ namespace rocksdb {
// IMPORTANT NOTE: Secondary index column families should be very small and // IMPORTANT NOTE: Secondary index column families should be very small and
// generally fit in memory. Assume that accessing secondary index column // generally fit in memory. Assume that accessing secondary index column
// families is much faster than accessing primary index (data heap) column // families is much faster than accessing primary index (data heap) column
// family. Accessing a key (i.e. checking for existance) from a column family in // family. Accessing a key (i.e. checking for existence) from a column family in
// RocksDB is not much faster than accessing both key and value since they are // RocksDB is not much faster than accessing both key and value since they are
// kept together and loaded from storage together. // kept together and loaded from storage together.

Loading…
Cancel
Save