Summary:
fixed typo
Closes https://github.com/facebook/rocksdb/pull/2430

Differential Revision: D5242471

Pulled By: IslamAbdelRahman

fbshipit-source-id: 832eb3a4c70221444ccd2ae63217823fec56c748
main
hyunwoo 8 years ago committed by Facebook Github Bot
parent 0f228be3bb
commit 6b5a5dc5d8
  1. 2
      USERS.md
  2. 2
      db/db_compaction_test.cc
  3. 2
      db/db_impl_compaction_flush.cc
  4. 4
      include/rocksdb/advanced_options.h
  5. 2
      include/rocksdb/sst_file_writer.h
  6. 4
      java/src/main/java/org/rocksdb/AdvancedMutableColumnFamilyOptionsInterface.java
  7. 4
      third-party/gtest-1.7.0/fused-src/gtest/gtest-all.cc
  8. 2
      util/threadpool_imp.h

@ -76,7 +76,7 @@ quasardb uses a heavily tuned RocksDB as its persistence layer.
[Dgraph](https://github.com/dgraph-io/dgraph) is an open-source, scalable, distributed, low latency, high throughput Graph database .They use RocksDB to store state locally on a machine. [Dgraph](https://github.com/dgraph-io/dgraph) is an open-source, scalable, distributed, low latency, high throughput Graph database .They use RocksDB to store state locally on a machine.
## Uber ## Uber
[Uber](http://eng.uber.com/cherami/) uses RocksDB as a durable and and scalable task queue. [Uber](http://eng.uber.com/cherami/) uses RocksDB as a durable and scalable task queue.
## 360 Pika ## 360 Pika
[360](http://www.360.cn/) [Pika](https://github.com/Qihoo360/pika) is a nosql compatible with redis. With the huge amount of data stored, redis may suffer for a capacity bottleneck, and pika was born for solving it. It has widely been widely used in many company [360](http://www.360.cn/) [Pika](https://github.com/Qihoo360/pika) is a nosql compatible with redis. With the huge amount of data stored, redis may suffer for a capacity bottleneck, and pika was born for solving it. It has widely been widely used in many company

@ -1919,7 +1919,7 @@ TEST_F(DBCompactionTest, ManualAutoRace) {
// The auto compaction is scheduled but waited until here // The auto compaction is scheduled but waited until here
TEST_SYNC_POINT("DBCompactionTest::ManualAutoRace:1"); TEST_SYNC_POINT("DBCompactionTest::ManualAutoRace:1");
// The auto compaction will wait until the the manual compaction is registerd // The auto compaction will wait until the manual compaction is registerd
// before processing so that it will be cancelled. // before processing so that it will be cancelled.
dbfull()->CompactRange(CompactRangeOptions(), handles_[1], nullptr, nullptr); dbfull()->CompactRange(CompactRangeOptions(), handles_[1], nullptr, nullptr);
ASSERT_EQ("0,1", FilesPerLevel(1)); ASSERT_EQ("0,1", FilesPerLevel(1));

@ -1730,7 +1730,7 @@ bool DBImpl::HaveManualCompaction(ColumnFamilyData* cfd) {
} }
if ((cfd == (*it)->cfd) && (!((*it)->in_progress || (*it)->done))) { if ((cfd == (*it)->cfd) && (!((*it)->in_progress || (*it)->done))) {
// Allow automatic compaction if manual compaction is // Allow automatic compaction if manual compaction is
// is in progress // in progress
return true; return true;
} }
it++; it++;

@ -252,7 +252,7 @@ struct AdvancedColumnFamilyOptions {
// ignore the option. // ignore the option.
// //
// The option is best suited for workloads where keys will likely to insert // The option is best suited for workloads where keys will likely to insert
// to a location close the the last inserted key with the same prefix. // to a location close the last inserted key with the same prefix.
// One example could be inserting keys of the form (prefix + timestamp), // One example could be inserting keys of the form (prefix + timestamp),
// and keys of the same prefix always comes in with time order. Another // and keys of the same prefix always comes in with time order. Another
// example would be updating the same key over and over again, in which case // example would be updating the same key over and over again, in which case
@ -274,7 +274,7 @@ struct AdvancedColumnFamilyOptions {
// If <= 0, a proper value is automatically calculated (usually 1/8 of // If <= 0, a proper value is automatically calculated (usually 1/8 of
// writer_buffer_size, rounded up to a multiple of 4KB). // writer_buffer_size, rounded up to a multiple of 4KB).
// //
// There are two additional restriction of the The specified size: // There are two additional restriction of the specified size:
// (1) size should be in the range of [4096, 2 << 30] and // (1) size should be in the range of [4096, 2 << 30] and
// (2) be the multiple of the CPU word (which helps with the memory // (2) be the multiple of the CPU word (which helps with the memory
// alignment). // alignment).

@ -57,7 +57,7 @@ struct ExternalSstFileInfo {
// All keys in files generated by SstFileWriter will have sequence number = 0. // All keys in files generated by SstFileWriter will have sequence number = 0.
class SstFileWriter { class SstFileWriter {
public: public:
// User can pass `column_family` to specify that the the generated file will // User can pass `column_family` to specify that the generated file will
// be ingested into this column_family, note that passing nullptr means that // be ingested into this column_family, note that passing nullptr means that
// the column_family is unknown. // the column_family is unknown.
// If invalidate_page_cache is set to true, SstFileWriter will give the OS a // If invalidate_page_cache is set to true, SstFileWriter will give the OS a

@ -111,7 +111,7 @@ public interface AdvancedMutableColumnFamilyOptionsInterface
* If &le; 0, a proper value is automatically calculated (usually 1/10 of * If &le; 0, a proper value is automatically calculated (usually 1/10 of
* writer_buffer_size). * writer_buffer_size).
* *
* There are two additional restriction of the The specified size: * There are two additional restriction of the specified size:
* (1) size should be in the range of [4096, 2 &lt;&lt; 30] and * (1) size should be in the range of [4096, 2 &lt;&lt; 30] and
* (2) be the multiple of the CPU word (which helps with the memory * (2) be the multiple of the CPU word (which helps with the memory
* alignment). * alignment).
@ -132,7 +132,7 @@ public interface AdvancedMutableColumnFamilyOptionsInterface
* If &le; 0, a proper value is automatically calculated (usually 1/10 of * If &le; 0, a proper value is automatically calculated (usually 1/10 of
* writer_buffer_size). * writer_buffer_size).
* *
* There are two additional restriction of the The specified size: * There are two additional restriction of the specified size:
* (1) size should be in the range of [4096, 2 &lt;&lt; 30] and * (1) size should be in the range of [4096, 2 &lt;&lt; 30] and
* (2) be the multiple of the CPU word (which helps with the memory * (2) be the multiple of the CPU word (which helps with the memory
* alignment). * alignment).

@ -673,7 +673,7 @@ GTEST_API_ bool ShouldShard(const char* total_shards_str,
// Parses the environment variable var as an Int32. If it is unset, // Parses the environment variable var as an Int32. If it is unset,
// returns default_val. If it is not an Int32, prints an error and // returns default_val. If it is not an Int32, prints an error and
// and aborts. // aborts.
GTEST_API_ Int32 Int32FromEnvOrDie(const char* env_var, Int32 default_val); GTEST_API_ Int32 Int32FromEnvOrDie(const char* env_var, Int32 default_val);
// Given the total number of shards, the shard index, and the test id, // Given the total number of shards, the shard index, and the test id,
@ -3207,7 +3207,7 @@ std::string CodePointToUtf8(UInt32 code_point) {
return str; return str;
} }
// The following two functions only make sense if the the system // The following two functions only make sense if the system
// uses UTF-16 for wide string encoding. All supported systems // uses UTF-16 for wide string encoding. All supported systems
// with 16 bit wchar_t (Windows, Cygwin, Symbian OS) do use UTF-16. // with 16 bit wchar_t (Windows, Cygwin, Symbian OS) do use UTF-16.

@ -30,7 +30,7 @@ class ThreadPoolImpl : public ThreadPool {
// Implement ThreadPool interfaces // Implement ThreadPool interfaces
// Wait for all threads to finish. // Wait for all threads to finish.
// Discards all the the jobs that did not // Discards all the jobs that did not
// start executing and waits for those running // start executing and waits for those running
// to complete // to complete
void JoinAllThreads() override; void JoinAllThreads() override;

Loading…
Cancel
Save