|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
//
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
|
|
|
|
#include "rocksdb/sst_dump_tool.h"
|
|
|
|
|
|
|
|
#include <cinttypes>
|
|
|
|
#include <iostream>
|
|
|
|
|
|
|
|
#include "port/port.h"
|
|
|
|
#include "rocksdb/utilities/ldb_cmd.h"
|
|
|
|
#include "table/sst_file_dumper.h"
|
|
|
|
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
|
|
|
|
static const std::vector<std::pair<CompressionType, const char*>>
|
|
|
|
kCompressions = {
|
|
|
|
{CompressionType::kNoCompression, "kNoCompression"},
|
|
|
|
{CompressionType::kSnappyCompression, "kSnappyCompression"},
|
|
|
|
{CompressionType::kZlibCompression, "kZlibCompression"},
|
|
|
|
{CompressionType::kBZip2Compression, "kBZip2Compression"},
|
|
|
|
{CompressionType::kLZ4Compression, "kLZ4Compression"},
|
|
|
|
{CompressionType::kLZ4HCCompression, "kLZ4HCCompression"},
|
|
|
|
{CompressionType::kXpressCompression, "kXpressCompression"},
|
|
|
|
{CompressionType::kZSTD, "kZSTD"}};
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
|
|
|
|
void print_help(bool to_stderr) {
|
|
|
|
fprintf(
|
|
|
|
to_stderr ? stderr : stdout,
|
|
|
|
R"(sst_dump --file=<data_dir_OR_sst_file> [--command=check|scan|raw|recompress|identify]
|
|
|
|
--file=<data_dir_OR_sst_file>
|
|
|
|
Path to SST file or directory containing SST files
|
|
|
|
|
|
|
|
--env_uri=<uri of underlying Env>
|
|
|
|
URI of underlying Env
|
|
|
|
|
|
|
|
--command=check|scan|raw|verify|identify
|
|
|
|
check: Iterate over entries in files but don't print anything except if an error is encountered (default command)
|
|
|
|
scan: Iterate over entries in files and print them to screen
|
|
|
|
raw: Dump all the table contents to <file_name>_dump.txt
|
|
|
|
verify: Iterate all the blocks in files verifying checksum to detect possible corruption but don't print anything except if a corruption is encountered
|
|
|
|
recompress: reports the SST file size if recompressed with different
|
|
|
|
compression types
|
|
|
|
identify: Reports a file is a valid SST file or lists all valid SST files under a directory
|
|
|
|
|
|
|
|
--output_hex
|
|
|
|
Can be combined with scan command to print the keys and values in Hex
|
|
|
|
|
|
|
|
--decode_blob_index
|
|
|
|
Decode blob indexes and print them in a human-readable format during scans.
|
|
|
|
|
|
|
|
--from=<user_key>
|
|
|
|
Key to start reading from when executing check|scan
|
|
|
|
|
|
|
|
--to=<user_key>
|
|
|
|
Key to stop reading at when executing check|scan
|
|
|
|
|
|
|
|
--prefix=<user_key>
|
|
|
|
Returns all keys with this prefix when executing check|scan
|
|
|
|
Cannot be used in conjunction with --from
|
|
|
|
|
|
|
|
--read_num=<num>
|
|
|
|
Maximum number of entries to read when executing check|scan
|
|
|
|
|
|
|
|
--verify_checksum
|
|
|
|
Verify file checksum when executing check|scan
|
|
|
|
|
|
|
|
--input_key_hex
|
|
|
|
Can be combined with --from and --to to indicate that these values are encoded in Hex
|
|
|
|
|
|
|
|
--show_properties
|
|
|
|
Print table properties after iterating over the file when executing
|
|
|
|
check|scan|raw|identify
|
|
|
|
|
|
|
|
--set_block_size=<block_size>
|
|
|
|
Can be combined with --command=recompress to set the block size that will
|
|
|
|
be used when trying different compression algorithms
|
|
|
|
|
|
|
|
--compression_types=<comma-separated list of CompressionType members, e.g.,
|
|
|
|
kSnappyCompression>
|
|
|
|
Can be combined with --command=recompress to run recompression for this
|
|
|
|
list of compression types
|
|
|
|
|
|
|
|
--parse_internal_key=<0xKEY>
|
|
|
|
Convenience option to parse an internal key on the command line. Dumps the
|
|
|
|
internal key in hex format {'key' @ SN: type}
|
|
|
|
|
|
|
|
--compression_level_from=<compression_level>
|
|
|
|
Compression level to start compressing when executing recompress. One compression type
|
|
|
|
and compression_level_to must also be specified
|
|
|
|
|
|
|
|
--compression_level_to=<compression_level>
|
|
|
|
Compression level to stop compressing when executing recompress. One compression type
|
|
|
|
and compression_level_from must also be specified
|
|
|
|
|
|
|
|
--compression_max_dict_bytes=<uint32_t>
|
|
|
|
Maximum size of dictionary used to prime the compression library
|
|
|
|
|
|
|
|
--compression_zstd_max_train_bytes=<uint32_t>
|
|
|
|
Maximum size of training data passed to zstd's dictionary trainer
|
|
|
|
)");
|
|
|
|
}
|
|
|
|
|
|
|
|
// arg_name would include all prefix, e.g. "--my_arg="
|
|
|
|
// arg_val is the parses value.
|
|
|
|
// True if there is a match. False otherwise.
|
|
|
|
// Woud exit after printing errmsg if cannot be parsed.
|
|
|
|
bool ParseIntArg(const char* arg, const std::string arg_name,
|
|
|
|
const std::string err_msg, int64_t* arg_val) {
|
|
|
|
if (strncmp(arg, arg_name.c_str(), arg_name.size()) == 0) {
|
|
|
|
std::string input_str = arg + arg_name.size();
|
|
|
|
std::istringstream iss(input_str);
|
|
|
|
iss >> *arg_val;
|
|
|
|
if (iss.fail()) {
|
|
|
|
fprintf(stderr, "%s\n", err_msg.c_str());
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
} // namespace
|
|
|
|
|
|
|
|
int SSTDumpTool::Run(int argc, char const* const* argv, Options options) {
|
|
|
|
const char* env_uri = nullptr;
|
|
|
|
const char* dir_or_file = nullptr;
|
|
|
|
uint64_t read_num = std::numeric_limits<uint64_t>::max();
|
|
|
|
std::string command;
|
|
|
|
|
|
|
|
char junk;
|
|
|
|
uint64_t n;
|
|
|
|
bool verify_checksum = false;
|
|
|
|
bool output_hex = false;
|
|
|
|
bool decode_blob_index = false;
|
|
|
|
bool input_key_hex = false;
|
|
|
|
bool has_from = false;
|
|
|
|
bool has_to = false;
|
|
|
|
bool use_from_as_prefix = false;
|
|
|
|
bool show_properties = false;
|
|
|
|
bool show_summary = false;
|
|
|
|
bool set_block_size = false;
|
|
|
|
bool has_compression_level_from = false;
|
|
|
|
bool has_compression_level_to = false;
|
|
|
|
bool has_specified_compression_types = false;
|
|
|
|
std::string from_key;
|
|
|
|
std::string to_key;
|
|
|
|
std::string block_size_str;
|
|
|
|
std::string compression_level_from_str;
|
|
|
|
std::string compression_level_to_str;
|
|
|
|
size_t block_size = 0;
|
|
|
|
size_t readahead_size = 2 * 1024 * 1024;
|
|
|
|
std::vector<std::pair<CompressionType, const char*>> compression_types;
|
|
|
|
uint64_t total_num_files = 0;
|
|
|
|
uint64_t total_num_data_blocks = 0;
|
|
|
|
uint64_t total_data_block_size = 0;
|
|
|
|
uint64_t total_index_block_size = 0;
|
|
|
|
uint64_t total_filter_block_size = 0;
|
|
|
|
int32_t compress_level_from = CompressionOptions::kDefaultCompressionLevel;
|
|
|
|
int32_t compress_level_to = CompressionOptions::kDefaultCompressionLevel;
|
|
|
|
uint32_t compression_max_dict_bytes =
|
|
|
|
ROCKSDB_NAMESPACE::CompressionOptions().max_dict_bytes;
|
|
|
|
uint32_t compression_zstd_max_train_bytes =
|
|
|
|
ROCKSDB_NAMESPACE::CompressionOptions().zstd_max_train_bytes;
|
|
|
|
|
|
|
|
int64_t tmp_val;
|
|
|
|
|
|
|
|
for (int i = 1; i < argc; i++) {
|
|
|
|
if (strncmp(argv[i], "--env_uri=", 10) == 0) {
|
|
|
|
env_uri = argv[i] + 10;
|
|
|
|
} else if (strncmp(argv[i], "--file=", 7) == 0) {
|
|
|
|
dir_or_file = argv[i] + 7;
|
|
|
|
} else if (strcmp(argv[i], "--output_hex") == 0) {
|
|
|
|
output_hex = true;
|
|
|
|
} else if (strcmp(argv[i], "--decode_blob_index") == 0) {
|
|
|
|
decode_blob_index = true;
|
|
|
|
} else if (strcmp(argv[i], "--input_key_hex") == 0) {
|
|
|
|
input_key_hex = true;
|
|
|
|
} else if (sscanf(argv[i], "--read_num=%lu%c", (unsigned long*)&n, &junk) ==
|
|
|
|
1) {
|
|
|
|
read_num = n;
|
|
|
|
} else if (strcmp(argv[i], "--verify_checksum") == 0) {
|
|
|
|
verify_checksum = true;
|
|
|
|
} else if (strncmp(argv[i], "--command=", 10) == 0) {
|
|
|
|
command = argv[i] + 10;
|
|
|
|
} else if (strncmp(argv[i], "--from=", 7) == 0) {
|
|
|
|
from_key = argv[i] + 7;
|
|
|
|
has_from = true;
|
|
|
|
} else if (strncmp(argv[i], "--to=", 5) == 0) {
|
|
|
|
to_key = argv[i] + 5;
|
|
|
|
has_to = true;
|
|
|
|
} else if (strncmp(argv[i], "--prefix=", 9) == 0) {
|
|
|
|
from_key = argv[i] + 9;
|
|
|
|
use_from_as_prefix = true;
|
|
|
|
} else if (strcmp(argv[i], "--show_properties") == 0) {
|
|
|
|
show_properties = true;
|
|
|
|
} else if (strcmp(argv[i], "--show_summary") == 0) {
|
|
|
|
show_summary = true;
|
|
|
|
} else if (ParseIntArg(argv[i], "--set_block_size=",
|
|
|
|
"block size must be numeric", &tmp_val)) {
|
|
|
|
set_block_size = true;
|
|
|
|
block_size = static_cast<size_t>(tmp_val);
|
|
|
|
} else if (ParseIntArg(argv[i], "--readahead_size=",
|
|
|
|
"readahead_size must be numeric", &tmp_val)) {
|
|
|
|
readahead_size = static_cast<size_t>(tmp_val);
|
|
|
|
} else if (strncmp(argv[i], "--compression_types=", 20) == 0) {
|
|
|
|
std::string compression_types_csv = argv[i] + 20;
|
|
|
|
std::istringstream iss(compression_types_csv);
|
|
|
|
std::string compression_type;
|
|
|
|
has_specified_compression_types = true;
|
|
|
|
while (std::getline(iss, compression_type, ',')) {
|
|
|
|
auto iter = std::find_if(
|
|
|
|
kCompressions.begin(), kCompressions.end(),
|
|
|
|
[&compression_type](std::pair<CompressionType, const char*> curr) {
|
|
|
|
return curr.second == compression_type;
|
|
|
|
});
|
|
|
|
if (iter == kCompressions.end()) {
|
|
|
|
fprintf(stderr, "%s is not a valid CompressionType\n",
|
|
|
|
compression_type.c_str());
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
compression_types.emplace_back(*iter);
|
|
|
|
}
|
|
|
|
} else if (strncmp(argv[i], "--parse_internal_key=", 21) == 0) {
|
|
|
|
std::string in_key(argv[i] + 21);
|
|
|
|
try {
|
|
|
|
in_key = ROCKSDB_NAMESPACE::LDBCommand::HexToString(in_key);
|
|
|
|
} catch (...) {
|
|
|
|
std::cerr << "ERROR: Invalid key input '"
|
|
|
|
<< in_key
|
|
|
|
<< "' Use 0x{hex representation of internal rocksdb key}" << std::endl;
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
Slice sl_key = ROCKSDB_NAMESPACE::Slice(in_key);
|
|
|
|
ParsedInternalKey ikey;
|
|
|
|
int retc = 0;
|
|
|
|
Status pik_status =
|
|
|
|
ParseInternalKey(sl_key, &ikey, true /* log_err_key */);
|
|
|
|
if (!pik_status.ok()) {
|
|
|
|
std::cerr << pik_status.getState() << "\n";
|
|
|
|
retc = -1;
|
|
|
|
}
|
|
|
|
fprintf(stdout, "key=%s\n", ikey.DebugString(true, true).c_str());
|
|
|
|
return retc;
|
|
|
|
} else if (ParseIntArg(argv[i], "--compression_level_from=",
|
|
|
|
"compression_level_from must be numeric",
|
|
|
|
&tmp_val)) {
|
|
|
|
has_compression_level_from = true;
|
|
|
|
compress_level_from = static_cast<int>(tmp_val);
|
|
|
|
} else if (ParseIntArg(argv[i], "--compression_level_to=",
|
|
|
|
"compression_level_to must be numeric", &tmp_val)) {
|
|
|
|
has_compression_level_to = true;
|
|
|
|
compress_level_to = static_cast<int>(tmp_val);
|
|
|
|
} else if (ParseIntArg(argv[i], "--compression_max_dict_bytes=",
|
|
|
|
"compression_max_dict_bytes must be numeric",
|
|
|
|
&tmp_val)) {
|
|
|
|
if (tmp_val < 0 || tmp_val > port::kMaxUint32) {
|
|
|
|
fprintf(stderr, "compression_max_dict_bytes must be a uint32_t: '%s'\n",
|
|
|
|
argv[i]);
|
|
|
|
print_help(/*to_stderr*/ true);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
compression_max_dict_bytes = static_cast<uint32_t>(tmp_val);
|
|
|
|
} else if (ParseIntArg(argv[i], "--compression_zstd_max_train_bytes=",
|
|
|
|
"compression_zstd_max_train_bytes must be numeric",
|
|
|
|
&tmp_val)) {
|
|
|
|
if (tmp_val < 0 || tmp_val > port::kMaxUint32) {
|
|
|
|
fprintf(stderr,
|
|
|
|
"compression_zstd_max_train_bytes must be a uint32_t: '%s'\n",
|
|
|
|
argv[i]);
|
|
|
|
print_help(/*to_stderr*/ true);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
compression_zstd_max_train_bytes = static_cast<uint32_t>(tmp_val);
|
|
|
|
} else if (strcmp(argv[i], "--help") == 0) {
|
|
|
|
print_help(/*to_stderr*/ false);
|
|
|
|
return 0;
|
|
|
|
} else if (strcmp(argv[i], "--version") == 0) {
|
|
|
|
printf("%s\n", GetRocksBuildInfoAsString("sst_dump").c_str());
|
|
|
|
return 0;
|
|
|
|
} else {
|
|
|
|
fprintf(stderr, "Unrecognized argument '%s'\n\n", argv[i]);
|
|
|
|
print_help(/*to_stderr*/ true);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if(has_compression_level_from && has_compression_level_to) {
|
|
|
|
if(!has_specified_compression_types || compression_types.size() != 1) {
|
|
|
|
fprintf(stderr, "Specify one compression type.\n\n");
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
} else if(has_compression_level_from || has_compression_level_to) {
|
|
|
|
fprintf(stderr, "Specify both --compression_level_from and "
|
|
|
|
"--compression_level_to.\n\n");
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (use_from_as_prefix && has_from) {
|
|
|
|
fprintf(stderr, "Cannot specify --prefix and --from\n\n");
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (input_key_hex) {
|
|
|
|
if (has_from || use_from_as_prefix) {
|
|
|
|
from_key = ROCKSDB_NAMESPACE::LDBCommand::HexToString(from_key);
|
|
|
|
}
|
|
|
|
if (has_to) {
|
|
|
|
to_key = ROCKSDB_NAMESPACE::LDBCommand::HexToString(to_key);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (dir_or_file == nullptr) {
|
|
|
|
fprintf(stderr, "file or directory must be specified.\n\n");
|
|
|
|
print_help(/*to_stderr*/ true);
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
std::shared_ptr<ROCKSDB_NAMESPACE::Env> env_guard;
|
|
|
|
|
|
|
|
// If caller of SSTDumpTool::Run(...) does not specify a different env other
|
|
|
|
// than Env::Default(), then try to load custom env based on dir_or_file.
|
|
|
|
// Otherwise, the caller is responsible for creating custom env.
|
|
|
|
if (!options.env || options.env == ROCKSDB_NAMESPACE::Env::Default()) {
|
|
|
|
Env* env = Env::Default();
|
|
|
|
Status s = Env::LoadEnv(env_uri ? env_uri : "", &env, &env_guard);
|
|
|
|
if (!s.ok() && !s.IsNotFound()) {
|
|
|
|
fprintf(stderr, "LoadEnv: %s\n", s.ToString().c_str());
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
options.env = env;
|
|
|
|
} else {
|
|
|
|
fprintf(stdout, "options.env is %p\n", options.env);
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<std::string> filenames;
|
|
|
|
ROCKSDB_NAMESPACE::Env* env = options.env;
|
|
|
|
ROCKSDB_NAMESPACE::Status st = env->GetChildren(dir_or_file, &filenames);
|
|
|
|
bool dir = true;
|
|
|
|
if (!st.ok() || filenames.empty()) {
|
|
|
|
// dir_or_file does not exist or does not contain children
|
|
|
|
// Check its existence first
|
|
|
|
Status s = env->FileExists(dir_or_file);
|
|
|
|
// dir_or_file does not exist
|
|
|
|
if (!s.ok()) {
|
|
|
|
fprintf(stderr, "%s%s: No such file or directory\n", s.ToString().c_str(),
|
|
|
|
dir_or_file);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
// dir_or_file exists and is treated as a "file"
|
|
|
|
// since it has no children
|
|
|
|
// This is ok since later it will be checked
|
|
|
|
// that whether it is a valid sst or not
|
|
|
|
// (A directory "file" is not a valid sst)
|
|
|
|
filenames.clear();
|
|
|
|
filenames.push_back(dir_or_file);
|
|
|
|
dir = false;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t total_read = 0;
|
|
|
|
// List of RocksDB SST file without corruption
|
|
|
|
std::vector<std::string> valid_sst_files;
|
|
|
|
for (size_t i = 0; i < filenames.size(); i++) {
|
|
|
|
std::string filename = filenames.at(i);
|
|
|
|
if (filename.length() <= 4 ||
|
|
|
|
filename.rfind(".sst") != filename.length() - 4) {
|
|
|
|
// ignore
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (dir) {
|
|
|
|
filename = std::string(dir_or_file) + "/" + filename;
|
|
|
|
}
|
|
|
|
|
|
|
|
ROCKSDB_NAMESPACE::SstFileDumper dumper(options, filename, readahead_size,
|
|
|
|
verify_checksum, output_hex,
|
|
|
|
decode_blob_index);
|
|
|
|
// Not a valid SST
|
|
|
|
if (!dumper.getStatus().ok()) {
|
|
|
|
fprintf(stderr, "%s: %s\n", filename.c_str(),
|
|
|
|
dumper.getStatus().ToString().c_str());
|
|
|
|
continue;
|
|
|
|
} else {
|
|
|
|
valid_sst_files.push_back(filename);
|
|
|
|
// Print out from and to key information once
|
|
|
|
// where there is at least one valid SST
|
|
|
|
if (valid_sst_files.size() == 1) {
|
|
|
|
// from_key and to_key are only used for "check", "scan", or ""
|
|
|
|
if (command == "check" || command == "scan" || command == "") {
|
|
|
|
fprintf(stdout, "from [%s] to [%s]\n",
|
|
|
|
ROCKSDB_NAMESPACE::Slice(from_key).ToString(true).c_str(),
|
|
|
|
ROCKSDB_NAMESPACE::Slice(to_key).ToString(true).c_str());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (command == "recompress") {
|
|
|
|
st = dumper.ShowAllCompressionSizes(
|
|
|
|
set_block_size ? block_size : 16384,
|
|
|
|
compression_types.empty() ? kCompressions : compression_types,
|
|
|
|
compress_level_from, compress_level_to, compression_max_dict_bytes,
|
|
|
|
compression_zstd_max_train_bytes);
|
|
|
|
if (!st.ok()) {
|
|
|
|
fprintf(stderr, "Failed to recompress: %s\n", st.ToString().c_str());
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (command == "raw") {
|
|
|
|
std::string out_filename = filename.substr(0, filename.length() - 4);
|
|
|
|
out_filename.append("_dump.txt");
|
|
|
|
|
|
|
|
st = dumper.DumpTable(out_filename);
|
|
|
|
if (!st.ok()) {
|
|
|
|
fprintf(stderr, "%s: %s\n", filename.c_str(), st.ToString().c_str());
|
|
|
|
exit(1);
|
|
|
|
} else {
|
|
|
|
fprintf(stdout, "raw dump written to file %s\n", &out_filename[0]);
|
|
|
|
}
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// scan all files in give file path.
|
|
|
|
if (command == "" || command == "scan" || command == "check") {
|
|
|
|
st = dumper.ReadSequential(
|
|
|
|
command == "scan", read_num > 0 ? (read_num - total_read) : read_num,
|
|
|
|
has_from || use_from_as_prefix, from_key, has_to, to_key,
|
|
|
|
use_from_as_prefix);
|
|
|
|
if (!st.ok()) {
|
|
|
|
fprintf(stderr, "%s: %s\n", filename.c_str(),
|
|
|
|
st.ToString().c_str());
|
|
|
|
}
|
|
|
|
total_read += dumper.GetReadNumber();
|
|
|
|
if (read_num > 0 && total_read > read_num) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (command == "verify") {
|
|
|
|
st = dumper.VerifyChecksum();
|
|
|
|
if (!st.ok()) {
|
|
|
|
fprintf(stderr, "%s is corrupted: %s\n", filename.c_str(),
|
|
|
|
st.ToString().c_str());
|
|
|
|
} else {
|
|
|
|
fprintf(stdout, "The file is ok\n");
|
|
|
|
}
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (show_properties || show_summary) {
|
|
|
|
const ROCKSDB_NAMESPACE::TableProperties* table_properties;
|
|
|
|
|
|
|
|
std::shared_ptr<const ROCKSDB_NAMESPACE::TableProperties>
|
|
|
|
table_properties_from_reader;
|
|
|
|
st = dumper.ReadTableProperties(&table_properties_from_reader);
|
|
|
|
if (!st.ok()) {
|
|
|
|
fprintf(stderr, "%s: %s\n", filename.c_str(), st.ToString().c_str());
|
|
|
|
fprintf(stderr, "Try to use initial table properties\n");
|
|
|
|
table_properties = dumper.GetInitTableProperties();
|
|
|
|
} else {
|
|
|
|
table_properties = table_properties_from_reader.get();
|
|
|
|
}
|
|
|
|
if (table_properties != nullptr) {
|
|
|
|
if (show_properties) {
|
|
|
|
fprintf(stdout,
|
|
|
|
"Table Properties:\n"
|
|
|
|
"------------------------------\n"
|
|
|
|
" %s",
|
|
|
|
table_properties->ToString("\n ", ": ").c_str());
|
|
|
|
}
|
|
|
|
total_num_files += 1;
|
|
|
|
total_num_data_blocks += table_properties->num_data_blocks;
|
|
|
|
total_data_block_size += table_properties->data_size;
|
|
|
|
total_index_block_size += table_properties->index_size;
|
|
|
|
total_filter_block_size += table_properties->filter_size;
|
|
|
|
if (show_properties) {
|
|
|
|
fprintf(stdout,
|
|
|
|
"Raw user collected properties\n"
|
|
|
|
"------------------------------\n");
|
|
|
|
for (const auto& kv : table_properties->user_collected_properties) {
|
|
|
|
std::string prop_name = kv.first;
|
|
|
|
std::string prop_val = Slice(kv.second).ToString(true);
|
|
|
|
fprintf(stdout, " # %s: 0x%s\n", prop_name.c_str(),
|
|
|
|
prop_val.c_str());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
fprintf(stderr, "Reader unexpectedly returned null properties\n");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (show_summary) {
|
|
|
|
fprintf(stdout, "total number of files: %" PRIu64 "\n", total_num_files);
|
|
|
|
fprintf(stdout, "total number of data blocks: %" PRIu64 "\n",
|
|
|
|
total_num_data_blocks);
|
|
|
|
fprintf(stdout, "total data block size: %" PRIu64 "\n",
|
|
|
|
total_data_block_size);
|
|
|
|
fprintf(stdout, "total index block size: %" PRIu64 "\n",
|
|
|
|
total_index_block_size);
|
|
|
|
fprintf(stdout, "total filter block size: %" PRIu64 "\n",
|
|
|
|
total_filter_block_size);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (valid_sst_files.empty()) {
|
|
|
|
// No valid SST files are found
|
|
|
|
// Exit with an error state
|
|
|
|
if (dir) {
|
|
|
|
fprintf(stdout, "------------------------------\n");
|
|
|
|
fprintf(stderr, "No valid SST files found in %s\n", dir_or_file);
|
|
|
|
} else {
|
|
|
|
fprintf(stderr, "%s is not a valid SST file\n", dir_or_file);
|
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
} else {
|
|
|
|
if (command == "identify") {
|
|
|
|
if (dir) {
|
|
|
|
fprintf(stdout, "------------------------------\n");
|
|
|
|
fprintf(stdout, "List of valid SST files found in %s:\n", dir_or_file);
|
|
|
|
for (const auto& f : valid_sst_files) {
|
|
|
|
fprintf(stdout, "%s\n", f.c_str());
|
|
|
|
}
|
|
|
|
fprintf(stdout, "Number of valid SST files: %zu\n",
|
|
|
|
valid_sst_files.size());
|
|
|
|
} else {
|
|
|
|
fprintf(stdout, "%s is a valid SST file\n", dir_or_file);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// At least one valid SST
|
|
|
|
// exit with a success state
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
|
|
|
|
|
|
#endif // ROCKSDB_LITE
|