|
|
|
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
|
|
|
// This source code is licensed under the BSD-style license found in the
|
|
|
|
// LICENSE file in the root directory of this source tree. An additional grant
|
|
|
|
// of patent rights can be found in the PATENTS file in the same directory.
|
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
//
|
|
|
|
// An Env is an interface used by the rocksdb implementation to access
|
|
|
|
// operating system functionality like the filesystem etc. Callers
|
|
|
|
// may wish to provide a custom Env object when opening a database to
|
|
|
|
// get fine gain control; e.g., to rate limit file system operations.
|
|
|
|
//
|
|
|
|
// All Env implementations are safe for concurrent access from
|
|
|
|
// multiple threads without any external synchronization.
|
|
|
|
|
|
|
|
#ifndef STORAGE_ROCKSDB_INCLUDE_ENV_H_
|
|
|
|
#define STORAGE_ROCKSDB_INCLUDE_ENV_H_
|
|
|
|
|
|
|
|
#include <cstdarg>
|
|
|
|
#include <string>
|
|
|
|
#include <memory>
|
|
|
|
#include <limits>
|
|
|
|
#include <vector>
|
|
|
|
#include <stdint.h>
|
|
|
|
#include "rocksdb/status.h"
|
|
|
|
|
|
|
|
namespace rocksdb {
|
|
|
|
|
|
|
|
class FileLock;
|
|
|
|
class Logger;
|
|
|
|
class RandomAccessFile;
|
|
|
|
class SequentialFile;
|
|
|
|
class Slice;
|
|
|
|
class WritableFile;
|
|
|
|
class RandomRWFile;
|
|
|
|
class Directory;
|
|
|
|
struct DBOptions;
|
|
|
|
class RateLimiter;
|
|
|
|
|
|
|
|
using std::unique_ptr;
|
|
|
|
using std::shared_ptr;
|
|
|
|
|
|
|
|
|
|
|
|
// Options while opening a file to read/write
|
|
|
|
struct EnvOptions {
|
|
|
|
|
|
|
|
// construct with default Options
|
|
|
|
EnvOptions();
|
|
|
|
|
|
|
|
// construct from Options
|
|
|
|
explicit EnvOptions(const DBOptions& options);
|
|
|
|
|
|
|
|
// If true, then allow caching of data in environment buffers
|
|
|
|
bool use_os_buffer = true;
|
|
|
|
|
|
|
|
// If true, then use mmap to read data
|
|
|
|
bool use_mmap_reads = false;
|
|
|
|
|
|
|
|
// If true, then use mmap to write data
|
|
|
|
bool use_mmap_writes = true;
|
|
|
|
|
|
|
|
// If true, set the FD_CLOEXEC on open fd.
|
|
|
|
bool set_fd_cloexec = true;
|
|
|
|
|
|
|
|
// Allows OS to incrementally sync files to disk while they are being
|
|
|
|
// written, in the background. Issue one request for every bytes_per_sync
|
|
|
|
// written. 0 turns it off.
|
|
|
|
// Default: 0
|
|
|
|
uint64_t bytes_per_sync = 0;
|
|
|
|
|
|
|
|
// If true, we will preallocate the file with FALLOC_FL_KEEP_SIZE flag, which
|
|
|
|
// means that file size won't change as part of preallocation.
|
|
|
|
// If false, preallocation will also change the file size. This option will
|
|
|
|
// improve the performance in workloads where you sync the data on every
|
|
|
|
// write. By default, we set it to true for MANIFEST writes and false for
|
|
|
|
// WAL writes
|
|
|
|
bool fallocate_with_keep_size = true;
|
|
|
|
|
|
|
|
// If not nullptr, write rate limiting is enabled for flush and compaction
|
|
|
|
RateLimiter* rate_limiter = nullptr;
|
|
|
|
};
|
|
|
|
|
|
|
|
class Env {
|
|
|
|
public:
|
|
|
|
Env() { }
|
|
|
|
virtual ~Env();
|
|
|
|
|
|
|
|
// Return a default environment suitable for the current operating
|
|
|
|
// system. Sophisticated users may wish to provide their own Env
|
|
|
|
// implementation instead of relying on this default environment.
|
|
|
|
//
|
|
|
|
// The result of Default() belongs to rocksdb and must never be deleted.
|
|
|
|
static Env* Default();
|
|
|
|
|
|
|
|
// Create a brand new sequentially-readable file with the specified name.
|
|
|
|
// On success, stores a pointer to the new file in *result and returns OK.
|
|
|
|
// On failure stores nullptr in *result and returns non-OK. If the file does
|
|
|
|
// not exist, returns a non-OK status.
|
|
|
|
//
|
|
|
|
// The returned file will only be accessed by one thread at a time.
|
|
|
|
virtual Status NewSequentialFile(const std::string& fname,
|
|
|
|
unique_ptr<SequentialFile>* result,
|
|
|
|
const EnvOptions& options)
|
|
|
|
= 0;
|
|
|
|
|
|
|
|
// Create a brand new random access read-only file with the
|
|
|
|
// specified name. On success, stores a pointer to the new file in
|
|
|
|
// *result and returns OK. On failure stores nullptr in *result and
|
|
|
|
// returns non-OK. If the file does not exist, returns a non-OK
|
|
|
|
// status.
|
|
|
|
//
|
|
|
|
// The returned file may be concurrently accessed by multiple threads.
|
|
|
|
virtual Status NewRandomAccessFile(const std::string& fname,
|
|
|
|
unique_ptr<RandomAccessFile>* result,
|
|
|
|
const EnvOptions& options)
|
|
|
|
= 0;
|
|
|
|
|
|
|
|
// Create an object that writes to a new file with the specified
|
|
|
|
// name. Deletes any existing file with the same name and creates a
|
|
|
|
// new file. On success, stores a pointer to the new file in
|
|
|
|
// *result and returns OK. On failure stores nullptr in *result and
|
|
|
|
// returns non-OK.
|
|
|
|
//
|
|
|
|
// The returned file will only be accessed by one thread at a time.
|
|
|
|
virtual Status NewWritableFile(const std::string& fname,
|
|
|
|
unique_ptr<WritableFile>* result,
|
|
|
|
const EnvOptions& options) = 0;
|
|
|
|
|
|
|
|
// Create an object that both reads and writes to a file on
|
|
|
|
// specified offsets (random access). If file already exists,
|
|
|
|
// does not overwrite it. On success, stores a pointer to the
|
|
|
|
// new file in *result and returns OK. On failure stores nullptr
|
|
|
|
// in *result and returns non-OK.
|
|
|
|
virtual Status NewRandomRWFile(const std::string& fname,
|
|
|
|
unique_ptr<RandomRWFile>* result,
|
|
|
|
const EnvOptions& options) = 0;
|
|
|
|
|
|
|
|
// Create an object that represents a directory. Will fail if directory
|
|
|
|
// doesn't exist. If the directory exists, it will open the directory
|
|
|
|
// and create a new Directory object.
|
|
|
|
//
|
|
|
|
// On success, stores a pointer to the new Directory in
|
|
|
|
// *result and returns OK. On failure stores nullptr in *result and
|
|
|
|
// returns non-OK.
|
|
|
|
virtual Status NewDirectory(const std::string& name,
|
|
|
|
unique_ptr<Directory>* result) = 0;
|
|
|
|
|
|
|
|
// Returns true iff the named file exists.
|
|
|
|
virtual bool FileExists(const std::string& fname) = 0;
|
|
|
|
|
|
|
|
// Store in *result the names of the children of the specified directory.
|
|
|
|
// The names are relative to "dir".
|
|
|
|
// Original contents of *results are dropped.
|
|
|
|
virtual Status GetChildren(const std::string& dir,
|
|
|
|
std::vector<std::string>* result) = 0;
|
|
|
|
|
|
|
|
// Delete the named file.
|
|
|
|
virtual Status DeleteFile(const std::string& fname) = 0;
|
|
|
|
|
|
|
|
// Create the specified directory. Returns error if directory exists.
|
|
|
|
virtual Status CreateDir(const std::string& dirname) = 0;
|
|
|
|
|
|
|
|
// Creates directory if missing. Return Ok if it exists, or successful in
|
|
|
|
// Creating.
|
|
|
|
virtual Status CreateDirIfMissing(const std::string& dirname) = 0;
|
|
|
|
|
|
|
|
// Delete the specified directory.
|
|
|
|
virtual Status DeleteDir(const std::string& dirname) = 0;
|
|
|
|
|
|
|
|
// Store the size of fname in *file_size.
|
|
|
|
virtual Status GetFileSize(const std::string& fname, uint64_t* file_size) = 0;
|
|
|
|
|
|
|
|
// Store the last modification time of fname in *file_mtime.
|
|
|
|
virtual Status GetFileModificationTime(const std::string& fname,
|
|
|
|
uint64_t* file_mtime) = 0;
|
|
|
|
// Rename file src to target.
|
|
|
|
virtual Status RenameFile(const std::string& src,
|
|
|
|
const std::string& target) = 0;
|
|
|
|
|
|
|
|
// Lock the specified file. Used to prevent concurrent access to
|
|
|
|
// the same db by multiple processes. On failure, stores nullptr in
|
|
|
|
// *lock and returns non-OK.
|
|
|
|
//
|
|
|
|
// On success, stores a pointer to the object that represents the
|
|
|
|
// acquired lock in *lock and returns OK. The caller should call
|
|
|
|
// UnlockFile(*lock) to release the lock. If the process exits,
|
|
|
|
// the lock will be automatically released.
|
|
|
|
//
|
|
|
|
// If somebody else already holds the lock, finishes immediately
|
|
|
|
// with a failure. I.e., this call does not wait for existing locks
|
|
|
|
// to go away.
|
|
|
|
//
|
|
|
|
// May create the named file if it does not already exist.
|
|
|
|
virtual Status LockFile(const std::string& fname, FileLock** lock) = 0;
|
|
|
|
|
|
|
|
// Release the lock acquired by a previous successful call to LockFile.
|
|
|
|
// REQUIRES: lock was returned by a successful LockFile() call
|
|
|
|
// REQUIRES: lock has not already been unlocked.
|
|
|
|
virtual Status UnlockFile(FileLock* lock) = 0;
|
|
|
|
|
generic rate limiter
Summary:
A generic rate limiter that can be shared by threads and rocksdb
instances. Will use this to smooth out write traffic generated by
compaction and flush. This will help us get better p99 behavior on flash
storage.
Test Plan:
unit test output
==== Test RateLimiterTest.Rate
request size [1 - 1023], limit 10 KB/sec, actual rate: 10.374969 KB/sec, elapsed 2002265
request size [1 - 2047], limit 20 KB/sec, actual rate: 20.771242 KB/sec, elapsed 2002139
request size [1 - 4095], limit 40 KB/sec, actual rate: 41.285299 KB/sec, elapsed 2202424
request size [1 - 8191], limit 80 KB/sec, actual rate: 81.371605 KB/sec, elapsed 2402558
request size [1 - 16383], limit 160 KB/sec, actual rate: 162.541268 KB/sec, elapsed 3303500
Reviewers: yhchiang, igor, sdong
Reviewed By: sdong
Subscribers: leveldb
Differential Revision: https://reviews.facebook.net/D19359
11 years ago
|
|
|
// Priority for scheduling job in thread pool
|
|
|
|
enum Priority { LOW, HIGH, TOTAL };
|
|
|
|
|
|
|
|
// Priority for requesting bytes in rate limiter scheduler
|
generic rate limiter
Summary:
A generic rate limiter that can be shared by threads and rocksdb
instances. Will use this to smooth out write traffic generated by
compaction and flush. This will help us get better p99 behavior on flash
storage.
Test Plan:
unit test output
==== Test RateLimiterTest.Rate
request size [1 - 1023], limit 10 KB/sec, actual rate: 10.374969 KB/sec, elapsed 2002265
request size [1 - 2047], limit 20 KB/sec, actual rate: 20.771242 KB/sec, elapsed 2002139
request size [1 - 4095], limit 40 KB/sec, actual rate: 41.285299 KB/sec, elapsed 2202424
request size [1 - 8191], limit 80 KB/sec, actual rate: 81.371605 KB/sec, elapsed 2402558
request size [1 - 16383], limit 160 KB/sec, actual rate: 162.541268 KB/sec, elapsed 3303500
Reviewers: yhchiang, igor, sdong
Reviewed By: sdong
Subscribers: leveldb
Differential Revision: https://reviews.facebook.net/D19359
11 years ago
|
|
|
enum IOPriority {
|
|
|
|
IO_LOW = 0,
|
|
|
|
IO_HIGH = 1,
|
|
|
|
IO_TOTAL = 2
|
|
|
|
};
|
|
|
|
|
|
|
|
// Arrange to run "(*function)(arg)" once in a background thread, in
|
|
|
|
// the thread pool specified by pri. By default, jobs go to the 'LOW'
|
|
|
|
// priority thread pool.
|
|
|
|
|
|
|
|
// "function" may run in an unspecified thread. Multiple functions
|
|
|
|
// added to the same Env may run concurrently in different threads.
|
|
|
|
// I.e., the caller may not assume that background work items are
|
|
|
|
// serialized.
|
|
|
|
virtual void Schedule(
|
|
|
|
void (*function)(void* arg),
|
|
|
|
void* arg,
|
|
|
|
Priority pri = LOW) = 0;
|
|
|
|
|
|
|
|
// Start a new thread, invoking "function(arg)" within the new thread.
|
|
|
|
// When "function(arg)" returns, the thread will be destroyed.
|
|
|
|
virtual void StartThread(void (*function)(void* arg), void* arg) = 0;
|
|
|
|
|
|
|
|
// Wait for all threads started by StartThread to terminate.
|
|
|
|
virtual void WaitForJoin() {}
|
|
|
|
|
|
|
|
// Get thread pool queue length for specific thrad pool.
|
|
|
|
virtual unsigned int GetThreadPoolQueueLen(Priority pri = LOW) const {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
// *path is set to a temporary directory that can be used for testing. It may
|
|
|
|
// or many not have just been created. The directory may or may not differ
|
|
|
|
// between runs of the same process, but subsequent calls will return the
|
|
|
|
// same directory.
|
|
|
|
virtual Status GetTestDirectory(std::string* path) = 0;
|
|
|
|
|
|
|
|
// Create and return a log file for storing informational messages.
|
|
|
|
virtual Status NewLogger(const std::string& fname,
|
|
|
|
shared_ptr<Logger>* result) = 0;
|
|
|
|
|
|
|
|
// Returns the number of micro-seconds since some fixed point in time. Only
|
|
|
|
// useful for computing deltas of time.
|
|
|
|
virtual uint64_t NowMicros() = 0;
|
|
|
|
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
12 years ago
|
|
|
// Returns the number of nano-seconds since some fixed point in time. Only
|
|
|
|
// useful for computing deltas of time in one run.
|
|
|
|
// Default implementation simply relies on NowMicros
|
|
|
|
virtual uint64_t NowNanos() {
|
|
|
|
return NowMicros() * 1000;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Sleep/delay the thread for the perscribed number of micro-seconds.
|
|
|
|
virtual void SleepForMicroseconds(int micros) = 0;
|
|
|
|
|
|
|
|
// Get the current host name.
|
|
|
|
virtual Status GetHostName(char* name, uint64_t len) = 0;
|
|
|
|
|
|
|
|
// Get the number of seconds since the Epoch, 1970-01-01 00:00:00 (UTC).
|
|
|
|
virtual Status GetCurrentTime(int64_t* unix_time) = 0;
|
|
|
|
|
|
|
|
// Get full directory name for this db.
|
|
|
|
virtual Status GetAbsolutePath(const std::string& db_path,
|
|
|
|
std::string* output_path) = 0;
|
|
|
|
|
|
|
|
// The number of background worker threads of a specific thread pool
|
|
|
|
// for this environment. 'LOW' is the default pool.
|
|
|
|
// default number: 1
|
|
|
|
virtual void SetBackgroundThreads(int number, Priority pri = LOW) = 0;
|
|
|
|
|
|
|
|
// Enlarge number of background worker threads of a specific thread pool
|
|
|
|
// for this environment if it is smaller than specified. 'LOW' is the default
|
|
|
|
// pool.
|
|
|
|
virtual void IncBackgroundThreadsIfNeeded(int number, Priority pri) = 0;
|
|
|
|
|
|
|
|
// Lower IO priority for threads from the specified pool.
|
|
|
|
virtual void LowerThreadPoolIOPriority(Priority pool = LOW) {}
|
|
|
|
|
|
|
|
// Converts seconds-since-Jan-01-1970 to a printable string
|
|
|
|
virtual std::string TimeToString(uint64_t time) = 0;
|
|
|
|
|
|
|
|
// Generates a unique id that can be used to identify a db
|
|
|
|
virtual std::string GenerateUniqueId();
|
|
|
|
|
|
|
|
// OptimizeForLogWrite will create a new EnvOptions object that is a copy of
|
|
|
|
// the EnvOptions in the parameters, but is optimized for writing log files.
|
|
|
|
// Default implementation returns the copy of the same object.
|
|
|
|
virtual EnvOptions OptimizeForLogWrite(const EnvOptions& env_options) const;
|
|
|
|
// OptimizeForManifestWrite will create a new EnvOptions object that is a copy
|
|
|
|
// of the EnvOptions in the parameters, but is optimized for writing manifest
|
|
|
|
// files. Default implementation returns the copy of the same object.
|
|
|
|
virtual EnvOptions OptimizeForManifestWrite(const EnvOptions& env_options)
|
|
|
|
const;
|
|
|
|
|
|
|
|
private:
|
|
|
|
// No copying allowed
|
|
|
|
Env(const Env&);
|
|
|
|
void operator=(const Env&);
|
|
|
|
};
|
|
|
|
|
|
|
|
// A file abstraction for reading sequentially through a file
|
|
|
|
class SequentialFile {
|
|
|
|
public:
|
|
|
|
SequentialFile() { }
|
|
|
|
virtual ~SequentialFile();
|
|
|
|
|
|
|
|
// Read up to "n" bytes from the file. "scratch[0..n-1]" may be
|
|
|
|
// written by this routine. Sets "*result" to the data that was
|
|
|
|
// read (including if fewer than "n" bytes were successfully read).
|
|
|
|
// May set "*result" to point at data in "scratch[0..n-1]", so
|
|
|
|
// "scratch[0..n-1]" must be live when "*result" is used.
|
|
|
|
// If an error was encountered, returns a non-OK status.
|
|
|
|
//
|
|
|
|
// REQUIRES: External synchronization
|
|
|
|
virtual Status Read(size_t n, Slice* result, char* scratch) = 0;
|
|
|
|
|
|
|
|
// Skip "n" bytes from the file. This is guaranteed to be no
|
|
|
|
// slower that reading the same data, but may be faster.
|
|
|
|
//
|
|
|
|
// If end of file is reached, skipping will stop at the end of the
|
|
|
|
// file, and Skip will return OK.
|
|
|
|
//
|
|
|
|
// REQUIRES: External synchronization
|
|
|
|
virtual Status Skip(uint64_t n) = 0;
|
|
|
|
|
|
|
|
// Remove any kind of caching of data from the offset to offset+length
|
|
|
|
// of this file. If the length is 0, then it refers to the end of file.
|
|
|
|
// If the system is not caching the file contents, then this is a noop.
|
|
|
|
virtual Status InvalidateCache(size_t offset, size_t length) {
|
|
|
|
return Status::NotSupported("InvalidateCache not supported.");
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// A file abstraction for randomly reading the contents of a file.
|
|
|
|
class RandomAccessFile {
|
|
|
|
public:
|
|
|
|
RandomAccessFile() { }
|
|
|
|
virtual ~RandomAccessFile();
|
|
|
|
|
|
|
|
// Read up to "n" bytes from the file starting at "offset".
|
|
|
|
// "scratch[0..n-1]" may be written by this routine. Sets "*result"
|
|
|
|
// to the data that was read (including if fewer than "n" bytes were
|
|
|
|
// successfully read). May set "*result" to point at data in
|
|
|
|
// "scratch[0..n-1]", so "scratch[0..n-1]" must be live when
|
|
|
|
// "*result" is used. If an error was encountered, returns a non-OK
|
|
|
|
// status.
|
|
|
|
//
|
|
|
|
// Safe for concurrent use by multiple threads.
|
|
|
|
virtual Status Read(uint64_t offset, size_t n, Slice* result,
|
|
|
|
char* scratch) const = 0;
|
|
|
|
|
|
|
|
// Tries to get an unique ID for this file that will be the same each time
|
|
|
|
// the file is opened (and will stay the same while the file is open).
|
|
|
|
// Furthermore, it tries to make this ID at most "max_size" bytes. If such an
|
|
|
|
// ID can be created this function returns the length of the ID and places it
|
|
|
|
// in "id"; otherwise, this function returns 0, in which case "id"
|
|
|
|
// may not have been modified.
|
|
|
|
//
|
|
|
|
// This function guarantees, for IDs from a given environment, two unique ids
|
|
|
|
// cannot be made equal to eachother by adding arbitrary bytes to one of
|
|
|
|
// them. That is, no unique ID is the prefix of another.
|
|
|
|
//
|
|
|
|
// This function guarantees that the returned ID will not be interpretable as
|
|
|
|
// a single varint.
|
|
|
|
//
|
|
|
|
// Note: these IDs are only valid for the duration of the process.
|
|
|
|
virtual size_t GetUniqueId(char* id, size_t max_size) const {
|
|
|
|
return 0; // Default implementation to prevent issues with backwards
|
|
|
|
// compatibility.
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
enum AccessPattern { NORMAL, RANDOM, SEQUENTIAL, WILLNEED, DONTNEED };
|
|
|
|
|
|
|
|
virtual void Hint(AccessPattern pattern) {}
|
|
|
|
|
|
|
|
// Remove any kind of caching of data from the offset to offset+length
|
|
|
|
// of this file. If the length is 0, then it refers to the end of file.
|
|
|
|
// If the system is not caching the file contents, then this is a noop.
|
|
|
|
virtual Status InvalidateCache(size_t offset, size_t length) {
|
|
|
|
return Status::NotSupported("InvalidateCache not supported.");
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// A file abstraction for sequential writing. The implementation
|
|
|
|
// must provide buffering since callers may append small fragments
|
|
|
|
// at a time to the file.
|
|
|
|
class WritableFile {
|
|
|
|
public:
|
|
|
|
WritableFile()
|
|
|
|
: last_preallocated_block_(0),
|
|
|
|
preallocation_block_size_(0),
|
|
|
|
io_priority_(Env::IO_TOTAL) {
|
|
|
|
}
|
|
|
|
virtual ~WritableFile();
|
|
|
|
|
|
|
|
virtual Status Append(const Slice& data) = 0;
|
|
|
|
virtual Status Close() = 0;
|
|
|
|
virtual Status Flush() = 0;
|
|
|
|
virtual Status Sync() = 0; // sync data
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Sync data and/or metadata as well.
|
|
|
|
* By default, sync only data.
|
|
|
|
* Override this method for environments where we need to sync
|
|
|
|
* metadata as well.
|
|
|
|
*/
|
|
|
|
virtual Status Fsync() {
|
|
|
|
return Sync();
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Change the priority in rate limiter if rate limiting is enabled.
|
|
|
|
* If rate limiting is not enabled, this call has no effect.
|
|
|
|
*/
|
|
|
|
virtual void SetIOPriority(Env::IOPriority pri) {
|
|
|
|
io_priority_ = pri;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Get the size of valid data in the file.
|
|
|
|
*/
|
|
|
|
virtual uint64_t GetFileSize() {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Get and set the default pre-allocation block size for writes to
|
|
|
|
* this file. If non-zero, then Allocate will be used to extend the
|
|
|
|
* underlying storage of a file (generally via fallocate) if the Env
|
|
|
|
* instance supports it.
|
|
|
|
*/
|
|
|
|
void SetPreallocationBlockSize(size_t size) {
|
|
|
|
preallocation_block_size_ = size;
|
|
|
|
}
|
|
|
|
|
|
|
|
virtual void GetPreallocationStatus(size_t* block_size,
|
|
|
|
size_t* last_allocated_block) {
|
|
|
|
*last_allocated_block = last_preallocated_block_;
|
|
|
|
*block_size = preallocation_block_size_;
|
|
|
|
}
|
|
|
|
|
|
|
|
// For documentation, refer to RandomAccessFile::GetUniqueId()
|
|
|
|
virtual size_t GetUniqueId(char* id, size_t max_size) const {
|
|
|
|
return 0; // Default implementation to prevent issues with backwards
|
|
|
|
}
|
|
|
|
|
|
|
|
// Remove any kind of caching of data from the offset to offset+length
|
|
|
|
// of this file. If the length is 0, then it refers to the end of file.
|
|
|
|
// If the system is not caching the file contents, then this is a noop.
|
|
|
|
// This call has no effect on dirty pages in the cache.
|
|
|
|
virtual Status InvalidateCache(size_t offset, size_t length) {
|
|
|
|
return Status::NotSupported("InvalidateCache not supported.");
|
|
|
|
}
|
|
|
|
|
|
|
|
protected:
|
|
|
|
// PrepareWrite performs any necessary preparation for a write
|
|
|
|
// before the write actually occurs. This allows for pre-allocation
|
|
|
|
// of space on devices where it can result in less file
|
|
|
|
// fragmentation and/or less waste from over-zealous filesystem
|
|
|
|
// pre-allocation.
|
|
|
|
void PrepareWrite(size_t offset, size_t len) {
|
|
|
|
if (preallocation_block_size_ == 0) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
// If this write would cross one or more preallocation blocks,
|
|
|
|
// determine what the last preallocation block necesessary to
|
|
|
|
// cover this write would be and Allocate to that point.
|
|
|
|
const auto block_size = preallocation_block_size_;
|
|
|
|
size_t new_last_preallocated_block =
|
|
|
|
(offset + len + block_size - 1) / block_size;
|
|
|
|
if (new_last_preallocated_block > last_preallocated_block_) {
|
|
|
|
size_t num_spanned_blocks =
|
|
|
|
new_last_preallocated_block - last_preallocated_block_;
|
|
|
|
Allocate(static_cast<off_t>(block_size * last_preallocated_block_),
|
|
|
|
static_cast<off_t>(block_size * num_spanned_blocks));
|
|
|
|
last_preallocated_block_ = new_last_preallocated_block;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Pre-allocate space for a file.
|
|
|
|
*/
|
|
|
|
virtual Status Allocate(off_t offset, off_t len) {
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
[RocksDB] Sync file to disk incrementally
Summary:
During compaction, we sync the output files after they are fully written out. This causes unnecessary blocking of the compaction thread and burstiness of the write traffic.
This diff simply asks the OS to sync data incrementally as they are written, on the background. The hope is that, at the final sync, most of the data are already on disk and we would block less on the sync call. Thus, each compaction runs faster and we could use fewer number of compaction threads to saturate IO.
In addition, the write traffic will be smoothed out, hopefully reducing the IO P99 latency too.
Some quick tests show 10~20% improvement in per thread compaction throughput. Combined with posix advice on compaction read, just 5 threads are enough to almost saturate the udb flash bandwidth for 800 bytes write only benchmark.
What's more promising is that, with saturated IO, iostat shows average wait time is actually smoother and much smaller.
For the write only test 800bytes test:
Before the change: await occillate between 10ms and 3ms
After the change: await ranges 1-3ms
Will test against read-modify-write workload too, see if high read latency P99 could be resolved.
Will introduce a parameter to control the sync interval in a follow up diff after cleaning up EnvOptions.
Test Plan: make check; db_bench; db_stress
Reviewers: dhruba
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11115
12 years ago
|
|
|
// Sync a file range with disk.
|
|
|
|
// offset is the starting byte of the file range to be synchronized.
|
|
|
|
// nbytes specifies the length of the range to be synchronized.
|
|
|
|
// This asks the OS to initiate flushing the cached data to disk,
|
|
|
|
// without waiting for completion.
|
|
|
|
// Default implementation does nothing.
|
|
|
|
virtual Status RangeSync(off_t offset, off_t nbytes) {
|
[RocksDB] Sync file to disk incrementally
Summary:
During compaction, we sync the output files after they are fully written out. This causes unnecessary blocking of the compaction thread and burstiness of the write traffic.
This diff simply asks the OS to sync data incrementally as they are written, on the background. The hope is that, at the final sync, most of the data are already on disk and we would block less on the sync call. Thus, each compaction runs faster and we could use fewer number of compaction threads to saturate IO.
In addition, the write traffic will be smoothed out, hopefully reducing the IO P99 latency too.
Some quick tests show 10~20% improvement in per thread compaction throughput. Combined with posix advice on compaction read, just 5 threads are enough to almost saturate the udb flash bandwidth for 800 bytes write only benchmark.
What's more promising is that, with saturated IO, iostat shows average wait time is actually smoother and much smaller.
For the write only test 800bytes test:
Before the change: await occillate between 10ms and 3ms
After the change: await ranges 1-3ms
Will test against read-modify-write workload too, see if high read latency P99 could be resolved.
Will introduce a parameter to control the sync interval in a follow up diff after cleaning up EnvOptions.
Test Plan: make check; db_bench; db_stress
Reviewers: dhruba
CC: leveldb
Differential Revision: https://reviews.facebook.net/D11115
12 years ago
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
size_t last_preallocated_block_;
|
|
|
|
size_t preallocation_block_size_;
|
|
|
|
// No copying allowed
|
|
|
|
WritableFile(const WritableFile&);
|
|
|
|
void operator=(const WritableFile&);
|
|
|
|
|
|
|
|
protected:
|
|
|
|
Env::IOPriority io_priority_;
|
|
|
|
};
|
|
|
|
|
|
|
|
// A file abstraction for random reading and writing.
|
|
|
|
class RandomRWFile {
|
|
|
|
public:
|
|
|
|
RandomRWFile() {}
|
|
|
|
virtual ~RandomRWFile() {}
|
|
|
|
|
|
|
|
// Write data from Slice data to file starting from offset
|
|
|
|
// Returns IOError on failure, but does not guarantee
|
|
|
|
// atomicity of a write. Returns OK status on success.
|
|
|
|
//
|
|
|
|
// Safe for concurrent use.
|
|
|
|
virtual Status Write(uint64_t offset, const Slice& data) = 0;
|
|
|
|
// Read up to "n" bytes from the file starting at "offset".
|
|
|
|
// "scratch[0..n-1]" may be written by this routine. Sets "*result"
|
|
|
|
// to the data that was read (including if fewer than "n" bytes were
|
|
|
|
// successfully read). May set "*result" to point at data in
|
|
|
|
// "scratch[0..n-1]", so "scratch[0..n-1]" must be live when
|
|
|
|
// "*result" is used. If an error was encountered, returns a non-OK
|
|
|
|
// status.
|
|
|
|
//
|
|
|
|
// Safe for concurrent use by multiple threads.
|
|
|
|
virtual Status Read(uint64_t offset, size_t n, Slice* result,
|
|
|
|
char* scratch) const = 0;
|
|
|
|
virtual Status Close() = 0; // closes the file
|
|
|
|
virtual Status Sync() = 0; // sync data
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Sync data and/or metadata as well.
|
|
|
|
* By default, sync only data.
|
|
|
|
* Override this method for environments where we need to sync
|
|
|
|
* metadata as well.
|
|
|
|
*/
|
|
|
|
virtual Status Fsync() {
|
|
|
|
return Sync();
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Pre-allocate space for a file.
|
|
|
|
*/
|
|
|
|
virtual Status Allocate(off_t offset, off_t len) {
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
// No copying allowed
|
|
|
|
RandomRWFile(const RandomRWFile&);
|
|
|
|
void operator=(const RandomRWFile&);
|
|
|
|
};
|
|
|
|
|
|
|
|
// Directory object represents collection of files and implements
|
|
|
|
// filesystem operations that can be executed on directories.
|
|
|
|
class Directory {
|
|
|
|
public:
|
|
|
|
virtual ~Directory() {}
|
|
|
|
// Fsync directory
|
|
|
|
virtual Status Fsync() = 0;
|
|
|
|
};
|
|
|
|
|
|
|
|
enum InfoLogLevel : unsigned char {
|
|
|
|
DEBUG_LEVEL = 0,
|
|
|
|
INFO_LEVEL,
|
|
|
|
WARN_LEVEL,
|
|
|
|
ERROR_LEVEL,
|
|
|
|
FATAL_LEVEL,
|
|
|
|
NUM_INFO_LOG_LEVELS,
|
|
|
|
};
|
|
|
|
|
|
|
|
// An interface for writing log messages.
|
|
|
|
class Logger {
|
|
|
|
public:
|
|
|
|
size_t kDoNotSupportGetLogFileSize = std::numeric_limits<size_t>::max();
|
|
|
|
|
|
|
|
explicit Logger(const InfoLogLevel log_level = InfoLogLevel::INFO_LEVEL)
|
|
|
|
: log_level_(log_level) {}
|
|
|
|
virtual ~Logger();
|
|
|
|
|
|
|
|
// Write an entry to the log file with the specified format.
|
|
|
|
virtual void Logv(const char* format, va_list ap) = 0;
|
|
|
|
|
|
|
|
// Write an entry to the log file with the specified log level
|
|
|
|
// and format. Any log with level under the internal log level
|
|
|
|
// of *this (see @SetInfoLogLevel and @GetInfoLogLevel) will not be
|
|
|
|
// printed.
|
|
|
|
void Logv(const InfoLogLevel log_level, const char* format, va_list ap) {
|
|
|
|
static const char* kInfoLogLevelNames[5] = {"DEBUG", "INFO", "WARN",
|
|
|
|
"ERROR", "FATAL"};
|
|
|
|
if (log_level < log_level_) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (log_level == InfoLogLevel::INFO_LEVEL) {
|
|
|
|
// Doesn't print log level if it is INFO level.
|
|
|
|
// This is to avoid unexpected performance regression after we add
|
|
|
|
// the feature of log level. All the logs before we add the feature
|
|
|
|
// are INFO level. We don't want to add extra costs to those existing
|
|
|
|
// logging.
|
|
|
|
Logv(format, ap);
|
|
|
|
} else {
|
|
|
|
char new_format[500];
|
|
|
|
snprintf(new_format, sizeof(new_format) - 1, "[%s] %s",
|
|
|
|
kInfoLogLevelNames[log_level], format);
|
|
|
|
Logv(new_format, ap);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
virtual size_t GetLogFileSize() const { return kDoNotSupportGetLogFileSize; }
|
|
|
|
// Flush to the OS buffers
|
|
|
|
virtual void Flush() {}
|
|
|
|
virtual InfoLogLevel GetInfoLogLevel() const { return log_level_; }
|
|
|
|
virtual void SetInfoLogLevel(const InfoLogLevel log_level) {
|
|
|
|
log_level_ = log_level;
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
// No copying allowed
|
|
|
|
Logger(const Logger&);
|
|
|
|
void operator=(const Logger&);
|
|
|
|
InfoLogLevel log_level_;
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
// Identifies a locked file.
|
|
|
|
class FileLock {
|
|
|
|
public:
|
|
|
|
FileLock() { }
|
|
|
|
virtual ~FileLock();
|
|
|
|
private:
|
|
|
|
// No copying allowed
|
|
|
|
FileLock(const FileLock&);
|
|
|
|
void operator=(const FileLock&);
|
|
|
|
};
|
|
|
|
|
|
|
|
extern void LogFlush(const shared_ptr<Logger>& info_log);
|
|
|
|
|
|
|
|
extern void Log(const InfoLogLevel log_level,
|
|
|
|
const shared_ptr<Logger>& info_log, const char* format, ...);
|
|
|
|
|
|
|
|
// a set of log functions with different log levels.
|
|
|
|
extern void Debug(const shared_ptr<Logger>& info_log, const char* format, ...);
|
|
|
|
extern void Info(const shared_ptr<Logger>& info_log, const char* format, ...);
|
|
|
|
extern void Warn(const shared_ptr<Logger>& info_log, const char* format, ...);
|
|
|
|
extern void Error(const shared_ptr<Logger>& info_log, const char* format, ...);
|
|
|
|
extern void Fatal(const shared_ptr<Logger>& info_log, const char* format, ...);
|
|
|
|
|
|
|
|
// Log the specified data to *info_log if info_log is non-nullptr.
|
|
|
|
// The default info log level is InfoLogLevel::ERROR.
|
|
|
|
extern void Log(const shared_ptr<Logger>& info_log, const char* format, ...)
|
|
|
|
# if defined(__GNUC__) || defined(__clang__)
|
|
|
|
__attribute__((__format__ (__printf__, 2, 3)))
|
|
|
|
# endif
|
|
|
|
;
|
|
|
|
|
|
|
|
extern void LogFlush(Logger *info_log);
|
|
|
|
|
|
|
|
extern void Log(const InfoLogLevel log_level, Logger* info_log,
|
|
|
|
const char* format, ...);
|
|
|
|
|
|
|
|
// The default info log level is InfoLogLevel::ERROR.
|
|
|
|
extern void Log(Logger* info_log, const char* format, ...)
|
|
|
|
# if defined(__GNUC__) || defined(__clang__)
|
|
|
|
__attribute__((__format__ (__printf__, 2, 3)))
|
|
|
|
# endif
|
|
|
|
;
|
|
|
|
|
|
|
|
// a set of log functions with different log levels.
|
|
|
|
extern void Debug(Logger* info_log, const char* format, ...);
|
|
|
|
extern void Info(Logger* info_log, const char* format, ...);
|
|
|
|
extern void Warn(Logger* info_log, const char* format, ...);
|
|
|
|
extern void Error(Logger* info_log, const char* format, ...);
|
|
|
|
extern void Fatal(Logger* info_log, const char* format, ...);
|
|
|
|
|
|
|
|
// A utility routine: write "data" to the named file.
|
|
|
|
extern Status WriteStringToFile(Env* env, const Slice& data,
|
|
|
|
const std::string& fname,
|
|
|
|
bool should_sync = false);
|
|
|
|
|
|
|
|
// A utility routine: read contents of named file into *data
|
|
|
|
extern Status ReadFileToString(Env* env, const std::string& fname,
|
|
|
|
std::string* data);
|
|
|
|
|
|
|
|
// An implementation of Env that forwards all calls to another Env.
|
|
|
|
// May be useful to clients who wish to override just part of the
|
|
|
|
// functionality of another Env.
|
|
|
|
class EnvWrapper : public Env {
|
|
|
|
public:
|
|
|
|
// Initialize an EnvWrapper that delegates all calls to *t
|
|
|
|
explicit EnvWrapper(Env* t) : target_(t) { }
|
|
|
|
virtual ~EnvWrapper();
|
|
|
|
|
|
|
|
// Return the target to which this Env forwards all calls
|
|
|
|
Env* target() const { return target_; }
|
|
|
|
|
|
|
|
// The following text is boilerplate that forwards all methods to target()
|
|
|
|
Status NewSequentialFile(const std::string& f,
|
|
|
|
unique_ptr<SequentialFile>* r,
|
|
|
|
const EnvOptions& options) {
|
|
|
|
return target_->NewSequentialFile(f, r, options);
|
|
|
|
}
|
|
|
|
Status NewRandomAccessFile(const std::string& f,
|
|
|
|
unique_ptr<RandomAccessFile>* r,
|
|
|
|
const EnvOptions& options) {
|
|
|
|
return target_->NewRandomAccessFile(f, r, options);
|
|
|
|
}
|
|
|
|
Status NewWritableFile(const std::string& f, unique_ptr<WritableFile>* r,
|
|
|
|
const EnvOptions& options) {
|
|
|
|
return target_->NewWritableFile(f, r, options);
|
|
|
|
}
|
|
|
|
Status NewRandomRWFile(const std::string& f, unique_ptr<RandomRWFile>* r,
|
|
|
|
const EnvOptions& options) {
|
|
|
|
return target_->NewRandomRWFile(f, r, options);
|
|
|
|
}
|
|
|
|
virtual Status NewDirectory(const std::string& name,
|
|
|
|
unique_ptr<Directory>* result) {
|
|
|
|
return target_->NewDirectory(name, result);
|
|
|
|
}
|
|
|
|
bool FileExists(const std::string& f) { return target_->FileExists(f); }
|
|
|
|
Status GetChildren(const std::string& dir, std::vector<std::string>* r) {
|
|
|
|
return target_->GetChildren(dir, r);
|
|
|
|
}
|
|
|
|
Status DeleteFile(const std::string& f) { return target_->DeleteFile(f); }
|
|
|
|
Status CreateDir(const std::string& d) { return target_->CreateDir(d); }
|
|
|
|
Status CreateDirIfMissing(const std::string& d) {
|
|
|
|
return target_->CreateDirIfMissing(d);
|
|
|
|
}
|
|
|
|
Status DeleteDir(const std::string& d) { return target_->DeleteDir(d); }
|
|
|
|
Status GetFileSize(const std::string& f, uint64_t* s) {
|
|
|
|
return target_->GetFileSize(f, s);
|
|
|
|
}
|
|
|
|
|
|
|
|
Status GetFileModificationTime(const std::string& fname,
|
|
|
|
uint64_t* file_mtime) {
|
|
|
|
return target_->GetFileModificationTime(fname, file_mtime);
|
|
|
|
}
|
|
|
|
|
|
|
|
Status RenameFile(const std::string& s, const std::string& t) {
|
|
|
|
return target_->RenameFile(s, t);
|
|
|
|
}
|
|
|
|
Status LockFile(const std::string& f, FileLock** l) {
|
|
|
|
return target_->LockFile(f, l);
|
|
|
|
}
|
|
|
|
Status UnlockFile(FileLock* l) { return target_->UnlockFile(l); }
|
|
|
|
void Schedule(void (*f)(void*), void* a, Priority pri) {
|
|
|
|
return target_->Schedule(f, a, pri);
|
|
|
|
}
|
|
|
|
void StartThread(void (*f)(void*), void* a) {
|
|
|
|
return target_->StartThread(f, a);
|
|
|
|
}
|
|
|
|
void WaitForJoin() { return target_->WaitForJoin(); }
|
|
|
|
virtual unsigned int GetThreadPoolQueueLen(Priority pri = LOW) const {
|
|
|
|
return target_->GetThreadPoolQueueLen(pri);
|
|
|
|
}
|
|
|
|
virtual Status GetTestDirectory(std::string* path) {
|
|
|
|
return target_->GetTestDirectory(path);
|
|
|
|
}
|
|
|
|
virtual Status NewLogger(const std::string& fname,
|
|
|
|
shared_ptr<Logger>* result) {
|
|
|
|
return target_->NewLogger(fname, result);
|
|
|
|
}
|
|
|
|
uint64_t NowMicros() {
|
|
|
|
return target_->NowMicros();
|
|
|
|
}
|
|
|
|
void SleepForMicroseconds(int micros) {
|
|
|
|
target_->SleepForMicroseconds(micros);
|
|
|
|
}
|
|
|
|
Status GetHostName(char* name, uint64_t len) {
|
|
|
|
return target_->GetHostName(name, len);
|
|
|
|
}
|
|
|
|
Status GetCurrentTime(int64_t* unix_time) {
|
|
|
|
return target_->GetCurrentTime(unix_time);
|
|
|
|
}
|
|
|
|
Status GetAbsolutePath(const std::string& db_path,
|
|
|
|
std::string* output_path) {
|
|
|
|
return target_->GetAbsolutePath(db_path, output_path);
|
|
|
|
}
|
|
|
|
void SetBackgroundThreads(int num, Priority pri) {
|
|
|
|
return target_->SetBackgroundThreads(num, pri);
|
|
|
|
}
|
|
|
|
|
|
|
|
void IncBackgroundThreadsIfNeeded(int num, Priority pri) {
|
|
|
|
return target_->IncBackgroundThreadsIfNeeded(num, pri);
|
|
|
|
}
|
|
|
|
|
|
|
|
void LowerThreadPoolIOPriority(Priority pool = LOW) override {
|
|
|
|
target_->LowerThreadPoolIOPriority(pool);
|
|
|
|
}
|
|
|
|
std::string TimeToString(uint64_t time) {
|
|
|
|
return target_->TimeToString(time);
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
Env* target_;
|
|
|
|
};
|
|
|
|
|
|
|
|
// Returns a new environment that stores its data in memory and delegates
|
|
|
|
// all non-file-storage tasks to base_env. The caller must delete the result
|
|
|
|
// when it is no longer needed.
|
|
|
|
// *base_env must remain live while the result is in use.
|
|
|
|
Env* NewMemEnv(Env* base_env);
|
|
|
|
|
|
|
|
} // namespace rocksdb
|
|
|
|
|
|
|
|
#endif // STORAGE_ROCKSDB_INCLUDE_ENV_H_
|