@ -98,6 +98,7 @@ struct DBImpl::CompactionState {
// Files produced by compaction
// Files produced by compaction
struct Output {
struct Output {
uint64_t number ;
uint64_t number ;
uint32_t path_id ;
uint64_t file_size ;
uint64_t file_size ;
InternalKey smallest , largest ;
InternalKey smallest , largest ;
SequenceNumber smallest_seqno , largest_seqno ;
SequenceNumber smallest_seqno , largest_seqno ;
@ -294,6 +295,10 @@ DBOptions SanitizeOptions(const std::string& dbname, const DBOptions& src) {
result . wal_dir = result . wal_dir . substr ( 0 , result . wal_dir . size ( ) - 1 ) ;
result . wal_dir = result . wal_dir . substr ( 0 , result . wal_dir . size ( ) - 1 ) ;
}
}
if ( result . db_paths . size ( ) = = 0 ) {
result . db_paths . push_back ( dbname ) ;
}
return result ;
return result ;
}
}
@ -573,30 +578,48 @@ void DBImpl::FindObsoleteFiles(DeletionState& deletion_state,
}
}
// don't delete live files
// don't delete live files
deletion_state . sst_live . assign ( pending_outputs_ . begin ( ) ,
for ( auto pair : pending_outputs_ ) {
pending_outputs_ . end ( ) ) ;
deletion_state . sst_live . emplace_back ( pair . first , pair . second , 0 ) ;
}
/* deletion_state.sst_live.insert(pending_outputs_.begin(),
pending_outputs_ . end ( ) ) ; */
versions_ - > AddLiveFiles ( & deletion_state . sst_live ) ;
versions_ - > AddLiveFiles ( & deletion_state . sst_live ) ;
if ( doing_the_full_scan ) {
if ( doing_the_full_scan ) {
// set of all files in the directory. We'll exclude files that are still
for ( uint32_t path_id = 0 ; path_id < options_ . db_paths . size ( ) ; path_id + + ) {
// alive in the subsequent processings.
// set of all files in the directory. We'll exclude files that are still
env_ - > GetChildren (
// alive in the subsequent processings.
dbname_ , & deletion_state . candidate_files
std : : vector < std : : string > files ;
) ; // Ignore errors
env_ - > GetChildren ( dbname_ , & files ) ; // Ignore errors
for ( std : : string file : files ) {
deletion_state . candidate_files . emplace_back ( file , path_id ) ;
}
}
//Add log files in wal_dir
//Add log files in wal_dir
if ( options_ . wal_dir ! = dbname_ ) {
if ( options_ . wal_dir ! = dbname_ ) {
std : : vector < std : : string > log_files ;
std : : vector < std : : string > log_files ;
env_ - > GetChildren ( options_ . wal_dir , & log_files ) ; // Ignore errors
env_ - > GetChildren ( options_ . wal_dir , & log_files ) ; // Ignore errors
deletion_state . candidate_files . insert (
for ( std : : string log_file : log_files ) {
deletion_state . candidate_files . end ( ) ,
deletion_state . candidate_files . emplace_back ( log_file , 0 ) ;
log_files . begin ( ) ,
}
log_files . end ( )
) ;
}
}
}
}
}
}
namespace {
bool CompareCandidateFile ( const rocksdb : : DBImpl : : CandidateFileInfo & first ,
const rocksdb : : DBImpl : : CandidateFileInfo & second ) {
if ( first . file_name > second . file_name ) {
return true ;
} else if ( first . file_name < second . file_name ) {
return false ;
} else {
return ( first . path_id > first . path_id ) ;
}
}
} ; // namespace
// Diffs the files listed in filenames and those that do not
// Diffs the files listed in filenames and those that do not
// belong to live files are posibly removed. Also, removes all the
// belong to live files are posibly removed. Also, removes all the
// files in sst_delete_files and log_delete_files.
// files in sst_delete_files and log_delete_files.
@ -612,10 +635,12 @@ void DBImpl::PurgeObsoleteFiles(DeletionState& state) {
return ;
return ;
}
}
// Now, convert live list to an unordered set , WITHOUT mutex held;
// Now, convert live list to an unordered map , WITHOUT mutex held;
// set is slow.
// set is slow.
std : : unordered_set < uint64_t > sst_live ( state . sst_live . begin ( ) ,
std : : unordered_map < uint64_t , const FileDescriptor * > sst_live_map ;
state . sst_live . end ( ) ) ;
for ( FileDescriptor & fd : state . sst_live ) {
sst_live_map [ fd . GetNumber ( ) ] = & fd ;
}
auto & candidate_files = state . candidate_files ;
auto & candidate_files = state . candidate_files ;
candidate_files . reserve (
candidate_files . reserve (
@ -625,26 +650,30 @@ void DBImpl::PurgeObsoleteFiles(DeletionState& state) {
// We may ignore the dbname when generating the file names.
// We may ignore the dbname when generating the file names.
const char * kDumbDbName = " " ;
const char * kDumbDbName = " " ;
for ( auto file : state . sst_delete_files ) {
for ( auto file : state . sst_delete_files ) {
candidate_files . push_back (
candidate_files . emplace_back (
TableFileName ( kDumbDbName , file - > fd . GetNumber ( ) ) . substr ( 1 ) ) ;
MakeTableFileName ( kDumbDbName , file - > fd . GetNumber ( ) ) ,
file - > fd . GetPathId ( ) ) ;
delete file ;
delete file ;
}
}
for ( auto file_num : state . log_delete_files ) {
for ( auto file_num : state . log_delete_files ) {
if ( file_num > 0 ) {
if ( file_num > 0 ) {
candidate_files . push_back ( LogFileName ( kDumbDbName , file_num ) . substr ( 1 ) ) ;
candidate_files . emplace_back ( LogFileName ( kDumbDbName , file_num ) . substr ( 1 ) ,
0 ) ;
}
}
}
}
// dedup state.candidate_files so we don't try to delete the same
// dedup state.candidate_files so we don't try to delete the same
// file twice
// file twice
sort ( candidate_files . begin ( ) , candidate_files . end ( ) ) ;
sort ( candidate_files . begin ( ) , candidate_files . end ( ) , CompareCandidateFile ) ;
candidate_files . erase ( unique ( candidate_files . begin ( ) , candidate_files . end ( ) ) ,
candidate_files . erase ( unique ( candidate_files . begin ( ) , candidate_files . end ( ) ) ,
candidate_files . end ( ) ) ;
candidate_files . end ( ) ) ;
std : : vector < std : : string > old_info_log_files ;
std : : vector < std : : string > old_info_log_files ;
for ( const auto & to_delete : candidate_files ) {
for ( const auto & candidate_file : candidate_files ) {
std : : string to_delete = candidate_file . file_name ;
uint32_t path_id = candidate_file . path_id ;
uint64_t number ;
uint64_t number ;
FileType type ;
FileType type ;
// Ignore file if we cannot recognize it.
// Ignore file if we cannot recognize it.
@ -664,7 +693,7 @@ void DBImpl::PurgeObsoleteFiles(DeletionState& state) {
keep = ( number > = state . manifest_file_number ) ;
keep = ( number > = state . manifest_file_number ) ;
break ;
break ;
case kTableFile :
case kTableFile :
keep = ( sst_live . find ( number ) ! = sst_live . end ( ) ) ;
keep = ( sst_live_map . find ( number ) ! = sst_live_map . end ( ) ) ;
break ;
break ;
case kTempFile :
case kTempFile :
// Any temp files that are currently being written to must
// Any temp files that are currently being written to must
@ -672,7 +701,7 @@ void DBImpl::PurgeObsoleteFiles(DeletionState& state) {
// Also, SetCurrentFile creates a temp file when writing out new
// Also, SetCurrentFile creates a temp file when writing out new
// manifest, which is equal to state.pending_manifest_file_number. We
// manifest, which is equal to state.pending_manifest_file_number. We
// should not delete that file
// should not delete that file
keep = ( sst_live . find ( number ) ! = sst_live . end ( ) ) | |
keep = ( sst_live_map . find ( number ) ! = sst_live_map . end ( ) ) | |
( number = = state . pending_manifest_file_number ) ;
( number = = state . pending_manifest_file_number ) ;
break ;
break ;
case kInfoLogFile :
case kInfoLogFile :
@ -693,13 +722,16 @@ void DBImpl::PurgeObsoleteFiles(DeletionState& state) {
continue ;
continue ;
}
}
std : : string fname ;
if ( type = = kTableFile ) {
if ( type = = kTableFile ) {
// evict from cache
// evict from cache
TableCache : : Evict ( table_cache_ . get ( ) , number ) ;
TableCache : : Evict ( table_cache_ . get ( ) , number ) ;
fname = TableFileName ( options_ . db_paths , number , path_id ) ;
} else {
fname =
( ( type = = kLogFile ) ? options_ . wal_dir : dbname_ ) + " / " + to_delete ;
}
}
std : : string fname = ( ( type = = kLogFile ) ? options_ . wal_dir : dbname_ ) +
" / " + to_delete ;
if ( type = = kLogFile & &
if ( type = = kLogFile & &
( options_ . WAL_ttl_seconds > 0 | | options_ . WAL_size_limit_MB > 0 ) ) {
( options_ . WAL_ttl_seconds > 0 | | options_ . WAL_size_limit_MB > 0 ) ) {
auto archived_log_name = ArchivedLogFileName ( options_ . wal_dir , number ) ;
auto archived_log_name = ArchivedLogFileName ( options_ . wal_dir , number ) ;
@ -1084,6 +1116,13 @@ Status DBImpl::Recover(
return s ;
return s ;
}
}
for ( auto db_path : options_ . db_paths ) {
s = env_ - > CreateDirIfMissing ( db_path ) ;
if ( ! s . ok ( ) ) {
return s ;
}
}
s = env_ - > NewDirectory ( dbname_ , & db_directory_ ) ;
s = env_ - > NewDirectory ( dbname_ , & db_directory_ ) ;
if ( ! s . ok ( ) ) {
if ( ! s . ok ( ) ) {
return s ;
return s ;
@ -1349,8 +1388,8 @@ Status DBImpl::WriteLevel0TableForRecovery(ColumnFamilyData* cfd, MemTable* mem,
mutex_ . AssertHeld ( ) ;
mutex_ . AssertHeld ( ) ;
const uint64_t start_micros = env_ - > NowMicros ( ) ;
const uint64_t start_micros = env_ - > NowMicros ( ) ;
FileMetaData meta ;
FileMetaData meta ;
meta . fd . number = versions_ - > NewFileNumber ( ) ;
meta . fd = FileDescriptor ( versions_ - > NewFileNumber ( ) , 0 , 0 ) ;
pending_outputs_ . insert ( meta . fd . GetNumber ( ) ) ;
pending_outputs_ [ meta . fd . GetNumber ( ) ] = 0 ; // path 0 for level 0 file.
Iterator * iter = mem - > NewIterator ( ReadOptions ( ) , true ) ;
Iterator * iter = mem - > NewIterator ( ReadOptions ( ) , true ) ;
const SequenceNumber newest_snapshot = snapshots_ . GetNewest ( ) ;
const SequenceNumber newest_snapshot = snapshots_ . GetNewest ( ) ;
const SequenceNumber earliest_seqno_in_memtable =
const SequenceNumber earliest_seqno_in_memtable =
@ -1381,9 +1420,9 @@ Status DBImpl::WriteLevel0TableForRecovery(ColumnFamilyData* cfd, MemTable* mem,
// should not be added to the manifest.
// should not be added to the manifest.
int level = 0 ;
int level = 0 ;
if ( s . ok ( ) & & meta . fd . GetFileSize ( ) > 0 ) {
if ( s . ok ( ) & & meta . fd . GetFileSize ( ) > 0 ) {
edit - > AddFile ( level , meta . fd . GetNumber ( ) , meta . fd . GetFileSize ( ) ,
edit - > AddFile ( level , meta . fd . GetNumber ( ) , meta . fd . GetPathId ( ) ,
meta . smallest , meta . largest , meta . smallest_seqno ,
meta . fd . GetFileSize ( ) , meta . smallest , meta . largest ,
meta . largest_seqno ) ;
meta . smallest_seqno , meta . largest_seqno ) ;
}
}
InternalStats : : CompactionStats stats ;
InternalStats : : CompactionStats stats ;
@ -1402,9 +1441,10 @@ Status DBImpl::WriteLevel0Table(ColumnFamilyData* cfd,
mutex_ . AssertHeld ( ) ;
mutex_ . AssertHeld ( ) ;
const uint64_t start_micros = env_ - > NowMicros ( ) ;
const uint64_t start_micros = env_ - > NowMicros ( ) ;
FileMetaData meta ;
FileMetaData meta ;
meta . fd . number = versions_ - > NewFileNumber ( ) ;
meta . fd = FileDescriptor ( versions_ - > NewFileNumber ( ) , 0 , 0 ) ;
* filenumber = meta . fd . GetNumber ( ) ;
* filenumber = meta . fd . GetNumber ( ) ;
pending_outputs_ . insert ( meta . fd . GetNumber ( ) ) ;
pending_outputs_ [ meta . fd . GetNumber ( ) ] = 0 ; // path 0 for level 0 file.
const SequenceNumber newest_snapshot = snapshots_ . GetNewest ( ) ;
const SequenceNumber newest_snapshot = snapshots_ . GetNewest ( ) ;
const SequenceNumber earliest_seqno_in_memtable =
const SequenceNumber earliest_seqno_in_memtable =
@ -1471,9 +1511,9 @@ Status DBImpl::WriteLevel0Table(ColumnFamilyData* cfd,
cfd - > options ( ) - > compaction_style = = kCompactionStyleLevel ) {
cfd - > options ( ) - > compaction_style = = kCompactionStyleLevel ) {
level = base - > PickLevelForMemTableOutput ( min_user_key , max_user_key ) ;
level = base - > PickLevelForMemTableOutput ( min_user_key , max_user_key ) ;
}
}
edit - > AddFile ( level , meta . fd . GetNumber ( ) , meta . fd . GetFileSize ( ) ,
edit - > AddFile ( level , meta . fd . GetNumber ( ) , meta . fd . GetPathId ( ) ,
meta . smallest , meta . largest , meta . smallest_seqno ,
meta . fd . GetFileSize ( ) , meta . smallest , meta . largest ,
meta . largest_seqno ) ;
meta . smallest_seqno , meta . largest_seqno ) ;
}
}
InternalStats : : CompactionStats stats ;
InternalStats : : CompactionStats stats ;
@ -1529,7 +1569,7 @@ Status DBImpl::FlushMemTableToOutputFile(ColumnFamilyData* cfd,
// Replace immutable memtable with the generated Table
// Replace immutable memtable with the generated Table
s = cfd - > imm ( ) - > InstallMemtableFlushResults (
s = cfd - > imm ( ) - > InstallMemtableFlushResults (
cfd , mems , versions_ . get ( ) , & mutex_ , options_ . info_log . get ( ) ,
cfd , mems , versions_ . get ( ) , & mutex_ , options_ . info_log . get ( ) ,
file_number , pending_outputs_ , & deletion_state . memtables_to_free ,
file_number , & pending_outputs_ , & deletion_state . memtables_to_free ,
db_directory_ . get ( ) , log_buffer ) ;
db_directory_ . get ( ) , log_buffer ) ;
}
}
@ -1673,9 +1713,9 @@ Status DBImpl::ReFitLevel(ColumnFamilyData* cfd, int level, int target_level) {
edit . SetColumnFamily ( cfd - > GetID ( ) ) ;
edit . SetColumnFamily ( cfd - > GetID ( ) ) ;
for ( const auto & f : cfd - > current ( ) - > files_ [ level ] ) {
for ( const auto & f : cfd - > current ( ) - > files_ [ level ] ) {
edit . DeleteFile ( level , f - > fd . GetNumber ( ) ) ;
edit . DeleteFile ( level , f - > fd . GetNumber ( ) ) ;
edit . AddFile ( to_level , f - > fd . GetNumber ( ) , f - > fd . GetFileSize ( ) ,
edit . AddFile ( to_level , f - > fd . GetNumber ( ) , f - > fd . GetPathId ( ) ,
f - > smallest , f - > largest , f - > smallest_seqno ,
f - > fd . GetFileSize ( ) , f - > smallest , f - > largest ,
f - > largest_seqno ) ;
f - > smallest_seqno , f - > largest_seqno ) ;
}
}
Log ( options_ . info_log , " [%s] Apply version edit: \n %s " ,
Log ( options_ . info_log , " [%s] Apply version edit: \n %s " ,
cfd - > GetName ( ) . c_str ( ) , edit . DebugString ( ) . data ( ) ) ;
cfd - > GetName ( ) . c_str ( ) , edit . DebugString ( ) . data ( ) ) ;
@ -2172,9 +2212,9 @@ Status DBImpl::BackgroundCompaction(bool* madeProgress,
assert ( c - > num_input_files ( 0 ) = = 1 ) ;
assert ( c - > num_input_files ( 0 ) = = 1 ) ;
FileMetaData * f = c - > input ( 0 , 0 ) ;
FileMetaData * f = c - > input ( 0 , 0 ) ;
c - > edit ( ) - > DeleteFile ( c - > level ( ) , f - > fd . GetNumber ( ) ) ;
c - > edit ( ) - > DeleteFile ( c - > level ( ) , f - > fd . GetNumber ( ) ) ;
c - > edit ( ) - > AddFile ( c - > level ( ) + 1 , f - > fd . GetNumber ( ) , f - > fd . GetFileSize ( ) ,
c - > edit ( ) - > AddFile ( c - > level ( ) + 1 , f - > fd . GetNumber ( ) , f - > fd . GetPathId ( ) ,
f - > smallest , f - > largest , f - > smallest_seqno ,
f - > fd . GetFileSize ( ) , f - > smallest , f - > largest ,
f - > largest_seqno ) ;
f - > smallest_seqno , f - > largest_seqno ) ;
status = versions_ - > LogAndApply ( c - > column_family_data ( ) , c - > edit ( ) , & mutex_ ,
status = versions_ - > LogAndApply ( c - > column_family_data ( ) , c - > edit ( ) , & mutex_ ,
db_directory_ . get ( ) ) ;
db_directory_ . get ( ) ) ;
InstallSuperVersion ( c - > column_family_data ( ) , deletion_state ) ;
InstallSuperVersion ( c - > column_family_data ( ) , deletion_state ) ;
@ -2280,7 +2320,7 @@ void DBImpl::AllocateCompactionOutputFileNumbers(CompactionState* compact) {
int filesNeeded = compact - > compaction - > num_input_files ( 1 ) ;
int filesNeeded = compact - > compaction - > num_input_files ( 1 ) ;
for ( int i = 0 ; i < std : : max ( filesNeeded , 1 ) ; i + + ) {
for ( int i = 0 ; i < std : : max ( filesNeeded , 1 ) ; i + + ) {
uint64_t file_number = versions_ - > NewFileNumber ( ) ;
uint64_t file_number = versions_ - > NewFileNumber ( ) ;
pending_outputs_ . insert ( file_number ) ;
pending_outputs_ [ file_number ] = compact - > compaction - > GetOutputPathId ( ) ;
compact - > allocated_file_numbers . push_back ( file_number ) ;
compact - > allocated_file_numbers . push_back ( file_number ) ;
}
}
}
}
@ -2306,18 +2346,20 @@ Status DBImpl::OpenCompactionOutputFile(CompactionState* compact) {
} else {
} else {
mutex_ . Lock ( ) ;
mutex_ . Lock ( ) ;
file_number = versions_ - > NewFileNumber ( ) ;
file_number = versions_ - > NewFileNumber ( ) ;
pending_outputs_ . insert ( file_number ) ;
pending_outputs_ [ file_number ] = compact - > compaction - > GetOutputPathId ( ) ;
mutex_ . Unlock ( ) ;
mutex_ . Unlock ( ) ;
}
}
CompactionState : : Output out ;
CompactionState : : Output out ;
out . number = file_number ;
out . number = file_number ;
out . path_id = compact - > compaction - > GetOutputPathId ( ) ;
out . smallest . Clear ( ) ;
out . smallest . Clear ( ) ;
out . largest . Clear ( ) ;
out . largest . Clear ( ) ;
out . smallest_seqno = out . largest_seqno = 0 ;
out . smallest_seqno = out . largest_seqno = 0 ;
compact - > outputs . push_back ( out ) ;
compact - > outputs . push_back ( out ) ;
// Make the output file
// Make the output file
std : : string fname = TableFileName ( dbname_ , file_number ) ;
std : : string fname = TableFileName ( options_ . db_paths , file_number ,
compact - > compaction - > GetOutputPathId ( ) ) ;
Status s = env_ - > NewWritableFile ( fname , & compact - > outfile , storage_options_ ) ;
Status s = env_ - > NewWritableFile ( fname , & compact - > outfile , storage_options_ ) ;
if ( s . ok ( ) ) {
if ( s . ok ( ) ) {
@ -2340,6 +2382,7 @@ Status DBImpl::FinishCompactionOutputFile(CompactionState* compact,
assert ( compact - > builder ! = nullptr ) ;
assert ( compact - > builder ! = nullptr ) ;
const uint64_t output_number = compact - > current_output ( ) - > number ;
const uint64_t output_number = compact - > current_output ( ) - > number ;
const uint32_t output_path_id = compact - > current_output ( ) - > path_id ;
assert ( output_number ! = 0 ) ;
assert ( output_number ! = 0 ) ;
// Check for iterator errors
// Check for iterator errors
@ -2375,9 +2418,9 @@ Status DBImpl::FinishCompactionOutputFile(CompactionState* compact,
if ( s . ok ( ) & & current_entries > 0 ) {
if ( s . ok ( ) & & current_entries > 0 ) {
// Verify that the table is usable
// Verify that the table is usable
ColumnFamilyData * cfd = compact - > compaction - > column_family_data ( ) ;
ColumnFamilyData * cfd = compact - > compaction - > column_family_data ( ) ;
FileDescriptor meta ( output_number , current_bytes ) ;
FileDescriptor fd ( output_number , output_path_id , current_bytes ) ;
Iterator * iter = cfd - > table_cache ( ) - > NewIterator (
Iterator * iter = cfd - > table_cache ( ) - > NewIterator (
ReadOptions ( ) , storage_options_ , cfd - > internal_comparator ( ) , meta ) ;
ReadOptions ( ) , storage_options_ , cfd - > internal_comparator ( ) , fd ) ;
s = iter - > status ( ) ;
s = iter - > status ( ) ;
delete iter ;
delete iter ;
if ( s . ok ( ) ) {
if ( s . ok ( ) ) {
@ -2420,9 +2463,10 @@ Status DBImpl::InstallCompactionResults(CompactionState* compact,
compact - > compaction - > AddInputDeletions ( compact - > compaction - > edit ( ) ) ;
compact - > compaction - > AddInputDeletions ( compact - > compaction - > edit ( ) ) ;
for ( size_t i = 0 ; i < compact - > outputs . size ( ) ; i + + ) {
for ( size_t i = 0 ; i < compact - > outputs . size ( ) ; i + + ) {
const CompactionState : : Output & out = compact - > outputs [ i ] ;
const CompactionState : : Output & out = compact - > outputs [ i ] ;
compact - > compaction - > edit ( ) - > AddFile (
compact - > compaction - > edit ( ) - > AddFile ( compact - > compaction - > output_level ( ) ,
compact - > compaction - > output_level ( ) , out . number , out . file_size ,
out . number , out . path_id , out . file_size ,
out . smallest , out . largest , out . smallest_seqno , out . largest_seqno ) ;
out . smallest , out . largest ,
out . smallest_seqno , out . largest_seqno ) ;
}
}
return versions_ - > LogAndApply ( compact - > compaction - > column_family_data ( ) ,
return versions_ - > LogAndApply ( compact - > compaction - > column_family_data ( ) ,
compact - > compaction - > edit ( ) , & mutex_ ,
compact - > compaction - > edit ( ) , & mutex_ ,
@ -4118,7 +4162,7 @@ Status DBImpl::MakeRoomForWrite(
// how do we fail if we're not creating new log?
// how do we fail if we're not creating new log?
assert ( creating_new_log ) ;
assert ( creating_new_log ) ;
// Avoid chewing through file number space in a tight loop.
// Avoid chewing through file number space in a tight loop.
versions_ - > ReuseFileNumber ( new_log_number ) ;
versions_ - > ReuseLog FileNumber ( new_log_number ) ;
assert ( ! new_mem ) ;
assert ( ! new_mem ) ;
assert ( ! new_log ) ;
assert ( ! new_log ) ;
break ;
break ;
@ -4361,14 +4405,15 @@ Status DBImpl::CheckConsistency() {
std : : string corruption_messages ;
std : : string corruption_messages ;
for ( const auto & md : metadata ) {
for ( const auto & md : metadata ) {
std : : string file_path = dbname_ + md . name ;
std : : string file_path = md . db_path + " / " + md . name ;
uint64_t fsize = 0 ;
uint64_t fsize = 0 ;
Status s = env_ - > GetFileSize ( file_path , & fsize ) ;
Status s = env_ - > GetFileSize ( file_path , & fsize ) ;
if ( ! s . ok ( ) ) {
if ( ! s . ok ( ) ) {
corruption_messages + =
corruption_messages + =
" Can't access " + md . name + " : " + s . ToString ( ) + " \n " ;
" Can't access " + md . name + " : " + s . ToString ( ) + " \n " ;
} else if ( fsize ! = md . size ) {
} else if ( fsize ! = md . size ) {
corruption_messages + = " Sst file size mismatch: " + md . name +
corruption_messages + = " Sst file size mismatch: " + file_path +
" . Size recorded in manifest " +
" . Size recorded in manifest " +
std : : to_string ( md . size ) + " , actual size " +
std : : to_string ( md . size ) + " , actual size " +
std : : to_string ( fsize ) + " \n " ;
std : : to_string ( fsize ) + " \n " ;
@ -4466,6 +4511,11 @@ Status DB::Open(const Options& options, const std::string& dbname, DB** dbptr) {
Status DB : : Open ( const DBOptions & db_options , const std : : string & dbname ,
Status DB : : Open ( const DBOptions & db_options , const std : : string & dbname ,
const std : : vector < ColumnFamilyDescriptor > & column_families ,
const std : : vector < ColumnFamilyDescriptor > & column_families ,
std : : vector < ColumnFamilyHandle * > * handles , DB * * dbptr ) {
std : : vector < ColumnFamilyHandle * > * handles , DB * * dbptr ) {
if ( db_options . db_paths . size ( ) > 1 ) {
return Status : : NotSupported (
" More than one DB paths are not supported yet. " ) ;
}
* dbptr = nullptr ;
* dbptr = nullptr ;
handles - > clear ( ) ;
handles - > clear ( ) ;
@ -4481,6 +4531,15 @@ Status DB::Open(const DBOptions& db_options, const std::string& dbname,
DBImpl * impl = new DBImpl ( db_options , dbname ) ;
DBImpl * impl = new DBImpl ( db_options , dbname ) ;
Status s = impl - > env_ - > CreateDirIfMissing ( impl - > options_ . wal_dir ) ;
Status s = impl - > env_ - > CreateDirIfMissing ( impl - > options_ . wal_dir ) ;
if ( s . ok ( ) ) {
for ( auto path : impl - > options_ . db_paths ) {
s = impl - > env_ - > CreateDirIfMissing ( path ) ;
if ( ! s . ok ( ) ) {
break ;
}
}
}
if ( ! s . ok ( ) ) {
if ( ! s . ok ( ) ) {
delete impl ;
delete impl ;
return s ;
return s ;
@ -4643,6 +4702,21 @@ Status DestroyDB(const std::string& dbname, const Options& options) {
}
}
}
}
for ( auto db_path : options . db_paths ) {
env - > GetChildren ( db_path , & filenames ) ;
uint64_t number ;
FileType type ;
for ( size_t i = 0 ; i < filenames . size ( ) ; i + + ) {
if ( ParseFileName ( filenames [ i ] , & number , & type ) & &
type = = kTableFile ) { // Lock file will be deleted at end
Status del = env - > DeleteFile ( db_path + " / " + filenames [ i ] ) ;
if ( result . ok ( ) & & ! del . ok ( ) ) {
result = del ;
}
}
}
}
env - > GetChildren ( archivedir , & archiveFiles ) ;
env - > GetChildren ( archivedir , & archiveFiles ) ;
// Delete archival files.
// Delete archival files.
for ( size_t i = 0 ; i < archiveFiles . size ( ) ; + + i ) {
for ( size_t i = 0 ; i < archiveFiles . size ( ) ; + + i ) {