Summary:
fixed typo
Closes https://github.com/facebook/rocksdb/pull/2376

Differential Revision: D5183630

Pulled By: ajkr

fbshipit-source-id: 133cfd0445959e70aa2cd1a12151bf3c0c5c3ac5
main
hyunwoo 8 years ago committed by Facebook Github Bot
parent 7e8d95cc0e
commit c7662a44a4
  1. 2
      build_tools/run_ci_db_test.ps1
  2. 2
      db/column_family.cc
  3. 2
      db/compaction_picker_universal.cc
  4. 2
      db/db_impl.cc
  5. 2
      db/external_sst_file_ingestion_job.h
  6. 2
      include/rocksdb/options.h
  7. 10
      java/rocksjni/portal.h
  8. 4
      java/rocksjni/rocksjni.cc
  9. 2
      java/rocksjni/ttl.cc
  10. 4
      java/src/main/java/org/rocksdb/CompactionOptionsUniversal.java
  11. 4
      options/options_test.cc
  12. 2
      tools/rdb/API.md
  13. 2
      tools/rdb/README.md
  14. 2
      util/build_version.h
  15. 2
      util/xxhash.cc
  16. 2
      utilities/document/document_db.cc

@ -333,7 +333,7 @@ function RunJobs($Suites, $TestCmds, [int]$ConcurrencyVal)
$Cases = @{}
if($Suites.Count -gt 0) {
# Will teh first one
# Will the first one
ForEach($e in $Suites.Keys) {
$exe_name = $e
$Cases = $Suites[$e]

@ -551,7 +551,7 @@ std::unique_ptr<WriteControllerToken> SetupDelay(
// If DB just falled into the stop condition, we need to further reduce
// the write rate to avoid the stop condition.
if (penalize_stop) {
// Penalize the near stop or stop condition by more agressive slowdown.
// Penalize the near stop or stop condition by more aggressive slowdown.
// This is to provide the long term slowdown increase signal.
// The penalty is more than the reward of recovering to the normal
// condition.

@ -621,7 +621,7 @@ Compaction* UniversalCompactionPicker::PickCompactionToReduceSizeAmp(
const std::string& cf_name, const MutableCFOptions& mutable_cf_options,
VersionStorageInfo* vstorage, double score,
const std::vector<SortedRun>& sorted_runs, LogBuffer* log_buffer) {
// percentage flexibilty while reducing size amplification
// percentage flexibility while reducing size amplification
uint64_t ratio =
ioptions_.compaction_options_universal.max_size_amplification_percent;

@ -2448,7 +2448,7 @@ void DBImpl::EraseThreadStatusDbInfo() const {
// A global method that can dump out the build version
void DumpRocksDBBuildVersion(Logger * log) {
#if !defined(IOS_CROSS_COMPILE)
// if we compile with Xcode, we don't run build_detect_vesion, so we don't
// if we compile with Xcode, we don't run build_detect_version, so we don't
// generate util/build_version.cc
ROCKS_LOG_HEADER(log, "RocksDB version: %d.%d.%d\n", ROCKSDB_MAJOR,
ROCKSDB_MINOR, ROCKSDB_PATCH);

@ -97,7 +97,7 @@ class ExternalSstFileIngestionJob {
// REQUIRES: Mutex held
void UpdateStats();
// Cleanup after successfull/failed job
// Cleanup after successful/failed job
void Cleanup(const Status& status);
VersionEdit* edit() { return &edit_; }

@ -771,7 +771,7 @@ struct DBOptions {
// If enable_pipelined_write is true, separate write thread queue is
// maintained for WAL write and memtable write. A write thread first enter WAL
// writer queue and then memtable writer queue. Pending thread on the WAL
// writer queue thus only have to wait for previous writers to finish thier
// writer queue thus only have to wait for previous writers to finish their
// WAL writing but not the memtable writing. Enabling the feature may improve
// write throughput and reduce latency of the prepare phase of two-phase
// commit.

@ -1635,7 +1635,7 @@ class BackupInfoListJni {
jobject jbackup_info_handle_list =
env->NewObject(jarray_list_clazz, cstr_mid, backup_infos.size());
if(env->ExceptionCheck()) {
// exception occured constructing object
// exception occurred constructing object
return nullptr;
}
@ -1650,7 +1650,7 @@ class BackupInfoListJni {
backup_info.size,
backup_info.number_files);
if(env->ExceptionCheck()) {
// exception occured constructing object
// exception occurred constructing object
if(obj != nullptr) {
env->DeleteLocalRef(obj);
}
@ -1663,7 +1663,7 @@ class BackupInfoListJni {
jboolean rs =
env->CallBooleanMethod(jbackup_info_handle_list, add_mid, obj);
if(env->ExceptionCheck() || rs == JNI_FALSE) {
// exception occured calling method, or could not add
// exception occurred calling method, or could not add
if(obj != nullptr) {
env->DeleteLocalRef(obj);
}
@ -1825,7 +1825,7 @@ class WriteTypeJni : public JavaClass {
env->GetStaticFieldID(jclazz, name,
"Lorg/rocksdb/WBWIRocksIterator$WriteType;");
if(env->ExceptionCheck()) {
// exception occured while getting field
// exception occurred while getting field
return nullptr;
} else if(jfid == nullptr) {
return nullptr;
@ -1962,7 +1962,7 @@ class InfoLogLevelJni : public JavaClass {
jfieldID jfid =
env->GetStaticFieldID(jclazz, name, "Lorg/rocksdb/InfoLogLevel;");
if(env->ExceptionCheck()) {
// exception occured while getting field
// exception occurred while getting field
return nullptr;
} else if(jfid == nullptr) {
return nullptr;

@ -122,7 +122,7 @@ jlongArray rocksdb_open_helper(JNIEnv* env, jlong jopt_handle,
env->ReleaseLongArrayElements(jcolumn_options, jco, JNI_ABORT);
if(has_exception == JNI_TRUE) {
// exception occured
// exception occurred
env->ReleaseStringUTFChars(jdb_path, db_path);
return nullptr;
}
@ -2180,7 +2180,7 @@ void Java_org_rocksdb_RocksDB_ingestExternalFile(
rocksdb::JniUtil::copyStrings(env, jfile_path_list, jfile_path_list_len,
&has_exception);
if(has_exception == JNI_TRUE) {
// exception occured
// exception occurred
return;
}

@ -91,7 +91,7 @@ jlongArray
env->ReleaseLongArrayElements(jcolumn_options, jco, JNI_ABORT);
if(has_exception == JNI_TRUE) {
// exception occured
// exception occurred
env->ReleaseStringUTFChars(jdb_path, db_path);
return nullptr;
}

@ -15,7 +15,7 @@ public class CompactionOptionsUniversal extends RocksObject {
}
/**
* Percentage flexibilty while comparing file size. If the candidate file(s)
* Percentage flexibility while comparing file size. If the candidate file(s)
* size is 1% smaller than the next file's size, then include next file into
* this candidate set.
*
@ -31,7 +31,7 @@ public class CompactionOptionsUniversal extends RocksObject {
}
/**
* Percentage flexibilty while comparing file size. If the candidate file(s)
* Percentage flexibility while comparing file size. If the candidate file(s)
* size is 1% smaller than the next file's size, then include next file into
* this candidate set.
*

@ -336,7 +336,7 @@ TEST_F(OptionsTest, GetColumnFamilyOptionsFromStringTest) {
ASSERT_EQ(new_cf_opt.arena_block_size, 21 * tera);
// Nested block based table options
// Emtpy
// Empty
ASSERT_OK(GetColumnFamilyOptionsFromString(base_cf_opt,
"write_buffer_size=10;max_write_buffer_number=16;"
"block_based_table_factory={};arena_block_size=1024",
@ -404,7 +404,7 @@ TEST_F(OptionsTest, GetColumnFamilyOptionsFromStringTest) {
ASSERT_OK(RocksDBOptionsParser::VerifyCFOptions(base_cf_opt, new_cf_opt));
// Nested plain table options
// Emtpy
// Empty
ASSERT_OK(GetColumnFamilyOptionsFromString(base_cf_opt,
"write_buffer_size=10;max_write_buffer_number=16;"
"plain_table_factory={};arena_block_size=1024",

@ -92,7 +92,7 @@
### CreateColumnFamily
# Create a new column familiy for the database.
# Create a new column family for the database.
#
# column_family_name (string) - Name of the new column family.
#

@ -16,7 +16,7 @@ JavaScript binding for RocksDB within a Node application.
### Installation
NOTE: If your default `python` binary is not a version of python2, add
the arguments `--python /path/to/python2` to the the `node-gyp` commands.
the arguments `--python /path/to/python2` to the `node-gyp` commands.
1. Make sure you have the static library (i.e. "librocksdb.a") in the root
directory of your rocksdb installation. If not, `cd` there and run

@ -7,7 +7,7 @@
//
#pragma once
#if !defined(IOS_CROSS_COMPILE)
// if we compile with Xcode, we don't run build_detect_vesion, so we don't
// if we compile with Xcode, we don't run build_detect_version, so we don't
// generate these variables
// this variable tells us about the git revision
extern const char* rocksdb_build_git_sha;

@ -53,7 +53,7 @@ You can contact the author at :
// By default, xxHash library provides endian-independent Hash values, based on little-endian convention.
// Results are therefore identical for little-endian and big-endian CPU.
// This comes at a performance cost for big-endian CPU, since some swapping is required to emulate little-endian format.
// Should endian-independance be of no importance for your application, you may set the #define below to 1.
// Should endian-independence be of no importance for your application, you may set the #define below to 1.
// It will improve speed for Big-endian CPU.
// This option has no impact on Little_Endian CPU.
#define XXH_FORCE_NATIVE_FORMAT 0

@ -25,7 +25,7 @@ namespace rocksdb {
// IMPORTANT NOTE: Secondary index column families should be very small and
// generally fit in memory. Assume that accessing secondary index column
// families is much faster than accessing primary index (data heap) column
// family. Accessing a key (i.e. checking for existance) from a column family in
// family. Accessing a key (i.e. checking for existence) from a column family in
// RocksDB is not much faster than accessing both key and value since they are
// kept together and loaded from storage together.

Loading…
Cancel
Save