You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
Changyu Bi
cc23b46da1
Support using ZDICT_finalizeDictionary to generate zstd dictionary (#9857)
Summary:
An untrained dictionary is currently simply the concatenation of several samples. The ZSTD API, ZDICT_finalizeDictionary(), can improve such a dictionary's effectiveness at low cost. This PR changes how dictionary is created by calling the ZSTD ZDICT_finalizeDictionary() API instead of creating raw content dictionary (when max_dict_buffer_bytes > 0), and pass in all buffered uncompressed data blocks as samples.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/9857
Test Plan:
#### db_bench test for cpu/memory of compression+decompression and space saving on synthetic data:
Set up: change the parameter [here](fb9a167a55/tools/db_bench_tool.cc (L1766) ) to 16384 to make synthetic data more compressible.
```
# linked local ZSTD with version 1.5.2
# DEBUG_LEVEL=0 ROCKSDB_NO_FBCODE=1 ROCKSDB_DISABLE_ZSTD=1 EXTRA_CXXFLAGS="-DZSTD_STATIC_LINKING_ONLY -DZSTD -I/data/users/changyubi/install/include/" EXTRA_LDFLAGS="-L/data/users/changyubi/install/lib/ -l:libzstd.a" make -j32 db_bench
dict_bytes=16384
train_bytes=1048576
echo "========== No Dictionary =========="
TEST_TMPDIR=/dev/shm ./db_bench -benchmarks=filluniquerandom,compact -num=10000000 -compression_type=zstd -compression_max_dict_bytes=0 -block_size=4096 -max_background_jobs=24 -memtablerep=vector -allow_concurrent_memtable_write=false -disable_wal=true -max_write_buffer_number=8 >/dev/null 2>&1
TEST_TMPDIR=/dev/shm /usr/bin/time ./db_bench -use_existing_db=true -benchmarks=compact -compression_type=zstd -compression_max_dict_bytes=0 -block_size=4096 2>&1 | grep elapsed
du -hc /dev/shm/dbbench/*sst | grep total
echo "========== Raw Content Dictionary =========="
TEST_TMPDIR=/dev/shm ./db_bench_main -benchmarks=filluniquerandom,compact -num=10000000 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -block_size=4096 -max_background_jobs=24 -memtablerep=vector -allow_concurrent_memtable_write=false -disable_wal=true -max_write_buffer_number=8 >/dev/null 2>&1
TEST_TMPDIR=/dev/shm /usr/bin/time ./db_bench_main -use_existing_db=true -benchmarks=compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -block_size=4096 2>&1 | grep elapsed
du -hc /dev/shm/dbbench/*sst | grep total
echo "========== FinalizeDictionary =========="
TEST_TMPDIR=/dev/shm ./db_bench -benchmarks=filluniquerandom,compact -num=10000000 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -compression_use_zstd_dict_trainer=false -block_size=4096 -max_background_jobs=24 -memtablerep=vector -allow_concurrent_memtable_write=false -disable_wal=true -max_write_buffer_number=8 >/dev/null 2>&1
TEST_TMPDIR=/dev/shm /usr/bin/time ./db_bench -use_existing_db=true -benchmarks=compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -compression_use_zstd_dict_trainer=false -block_size=4096 2>&1 | grep elapsed
du -hc /dev/shm/dbbench/*sst | grep total
echo "========== TrainDictionary =========="
TEST_TMPDIR=/dev/shm ./db_bench -benchmarks=filluniquerandom,compact -num=10000000 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -block_size=4096 -max_background_jobs=24 -memtablerep=vector -allow_concurrent_memtable_write=false -disable_wal=true -max_write_buffer_number=8 >/dev/null 2>&1
TEST_TMPDIR=/dev/shm /usr/bin/time ./db_bench -use_existing_db=true -benchmarks=compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -block_size=4096 2>&1 | grep elapsed
du -hc /dev/shm/dbbench/*sst | grep total
# Result: TrainDictionary is much better on space saving, but FinalizeDictionary seems to use less memory.
# before compression data size: 1.2GB
dict_bytes=16384
max_dict_buffer_bytes = 1048576
space cpu/memory
No Dictionary 468M 14.93user 1.00system 0:15.92elapsed 100%CPU (0avgtext+0avgdata 23904maxresident)k
Raw Dictionary 251M 15.81user 0.80system 0:16.56elapsed 100%CPU (0avgtext+0avgdata 156808maxresident)k
FinalizeDictionary 236M 11.93user 0.64system 0:12.56elapsed 100%CPU (0avgtext+0avgdata 89548maxresident)k
TrainDictionary 84M 7.29user 0.45system 0:07.75elapsed 100%CPU (0avgtext+0avgdata 97288maxresident)k
```
#### Benchmark on 10 sample SST files for spacing saving and CPU time on compression:
FinalizeDictionary is comparable to TrainDictionary in terms of space saving, and takes less time in compression.
```
dict_bytes=16384
train_bytes=1048576
for sst_file in `ls ../temp/myrock-sst/`
do
echo "********** $sst_file **********"
echo "========== No Dictionary =========="
./sst_dump --file="../temp/myrock-sst/$sst_file" --command=recompress --compression_level_from=6 --compression_level_to=6 --compression_types=kZSTD
echo "========== Raw Content Dictionary =========="
./sst_dump --file="../temp/myrock-sst/$sst_file" --command=recompress --compression_level_from=6 --compression_level_to=6 --compression_types=kZSTD --compression_max_dict_bytes=$dict_bytes
echo "========== FinalizeDictionary =========="
./sst_dump --file="../temp/myrock-sst/$sst_file" --command=recompress --compression_level_from=6 --compression_level_to=6 --compression_types=kZSTD --compression_max_dict_bytes=$dict_bytes --compression_zstd_max_train_bytes=$train_bytes --compression_use_zstd_finalize_dict
echo "========== TrainDictionary =========="
./sst_dump --file="../temp/myrock-sst/$sst_file" --command=recompress --compression_level_from=6 --compression_level_to=6 --compression_types=kZSTD --compression_max_dict_bytes=$dict_bytes --compression_zstd_max_train_bytes=$train_bytes
done
010240.sst (Size/Time) 011029.sst 013184.sst 021552.sst 185054.sst 185137.sst 191666.sst 7560381.sst 7604174.sst 7635312.sst
No Dictionary 28165569 / 2614419 32899411 / 2976832 32977848 / 3055542 31966329 / 2004590 33614351 / 1755877 33429029 / 1717042 33611933 / 1776936 33634045 / 2771417 33789721 / 2205414 33592194 / 388254
Raw Content Dictionary 28019950 / 2697961 33748665 / 3572422 33896373 / 3534701 26418431 / 2259658 28560825 / 1839168 28455030 / 1846039 28494319 / 1861349 32391599 / 3095649 33772142 / 2407843 33592230 / 474523
FinalizeDictionary 27896012 / 2650029 33763886 / 3719427 33904283 / 3552793 26008225 / 2198033 28111872 / 1869530 28014374 / 1789771 28047706 / 1848300 32296254 / 3204027 33698698 / 2381468 33592344 / 517433
TrainDictionary 28046089 / 2740037 33706480 / 3679019 33885741 / 3629351 25087123 / 2204558 27194353 / 1970207 27234229 / 1896811 27166710 / 1903119 32011041 / 3322315 32730692 / 2406146 33608631 / 570593
```
#### Decompression/Read test:
With FinalizeDictionary/TrainDictionary, some data structure used for decompression are in stored in dictionary, so they are expected to be faster in terms of decompression/reads.
```
dict_bytes=16384
train_bytes=1048576
echo "No Dictionary"
TEST_TMPDIR=/dev/shm/ ./db_bench -benchmarks=filluniquerandom,compact -compression_type=zstd -compression_max_dict_bytes=0 > /dev/null 2>&1
TEST_TMPDIR=/dev/shm/ ./db_bench -use_existing_db=true -benchmarks=readrandom -cache_size=0 -compression_type=zstd -compression_max_dict_bytes=0 2>&1 | grep MB/s
echo "Raw Dictionary"
TEST_TMPDIR=/dev/shm/ ./db_bench -benchmarks=filluniquerandom,compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes > /dev/null 2>&1
TEST_TMPDIR=/dev/shm/ ./db_bench -use_existing_db=true -benchmarks=readrandom -cache_size=0 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes 2>&1 | grep MB/s
echo "FinalizeDict"
TEST_TMPDIR=/dev/shm/ ./db_bench -benchmarks=filluniquerandom,compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -compression_use_zstd_dict_trainer=false > /dev/null 2>&1
TEST_TMPDIR=/dev/shm/ ./db_bench -use_existing_db=true -benchmarks=readrandom -cache_size=0 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -compression_use_zstd_dict_trainer=false 2>&1 | grep MB/s
echo "Train Dictionary"
TEST_TMPDIR=/dev/shm/ ./db_bench -benchmarks=filluniquerandom,compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes > /dev/null 2>&1
TEST_TMPDIR=/dev/shm/ ./db_bench -use_existing_db=true -benchmarks=readrandom -cache_size=0 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes 2>&1 | grep MB/s
No Dictionary
readrandom : 12.183 micros/op 82082 ops/sec 12.183 seconds 1000000 operations; 9.1 MB/s (1000000 of 1000000 found)
Raw Dictionary
readrandom : 12.314 micros/op 81205 ops/sec 12.314 seconds 1000000 operations; 9.0 MB/s (1000000 of 1000000 found)
FinalizeDict
readrandom : 9.787 micros/op 102180 ops/sec 9.787 seconds 1000000 operations; 11.3 MB/s (1000000 of 1000000 found)
Train Dictionary
readrandom : 9.698 micros/op 103108 ops/sec 9.699 seconds 1000000 operations; 11.4 MB/s (1000000 of 1000000 found)
```
Reviewed By: ajkr
Differential Revision: D35720026
Pulled By: cbi42
fbshipit-source-id: 24d230fdff0fd28a1bb650658798f00dfcfb2a1f
|
3 years ago |
.. |
advisor
|
Update branch as "main" in tools/advisor/README.md (#8744)
|
3 years ago |
block_cache_analyzer
|
Use std::numeric_limits<> (#9954)
|
3 years ago |
dump
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
CMakeLists.txt
|
Mark dependencies as PRIVATE and fix missing dependencies in tools. (#6790)
|
5 years ago |
Dockerfile
|
adding docker build script and dockerfile
|
10 years ago |
analyze_txn_stress_test.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
auto_sanity_test.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
backup_db.sh
|
Revamp check_format_compatible.sh (#8012)
|
4 years ago |
benchmark.sh
|
Remove deprecated API AdvancedColumnFamilyOptions::rate_limit_delay_max_milliseconds (#9455)
|
3 years ago |
benchmark_leveldb.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
blob_dump.cc
|
Remove using namespace (#9369)
|
3 years ago |
check_all_python.py
|
Allow missing "unversioned" python, as in CentOS 8 (#6883)
|
5 years ago |
check_format_compatible.sh
|
Fix format_compatible blowing away its TEST_TMPDIR (#9970)
|
3 years ago |
db_bench.cc
|
Add (& fix) some simple source code checks (#8821)
|
3 years ago |
db_bench_tool.cc
|
Support using ZDICT_finalizeDictionary to generate zstd dictionary (#9857)
|
3 years ago |
db_bench_tool_test.cc
|
Remove BlockBasedTableOptions.hash_index_allow_collision (#9454)
|
3 years ago |
db_crashtest.py
|
Support using ZDICT_finalizeDictionary to generate zstd dictionary (#9857)
|
3 years ago |
db_repl_stress.cc
|
Remove using namespace (#9369)
|
3 years ago |
db_sanity_test.cc
|
Remove own ToString() (#9955)
|
3 years ago |
dbench_monitor
|
Fix /bin/bash shebangs
|
7 years ago |
generate_random_db.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
ingest_external_sst.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
io_tracer_parser.cc
|
Add IO Tracer Parser (#7333)
|
4 years ago |
io_tracer_parser_test.cc
|
Cleanup includes in dbformat.h (#8930)
|
3 years ago |
io_tracer_parser_tool.cc
|
Add request_id in IODebugContext. (#8045)
|
4 years ago |
io_tracer_parser_tool.h
|
Add IO Tracer Parser (#7333)
|
4 years ago |
ldb.cc
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
ldb_cmd.cc
|
Support single delete in ldb (#9469)
|
3 years ago |
ldb_cmd_impl.h
|
Support single delete in ldb (#9469)
|
3 years ago |
ldb_cmd_test.cc
|
Remove own ToString() (#9955)
|
3 years ago |
ldb_test.py
|
Add blob dump support to the dump_live_files command (#9896)
|
3 years ago |
ldb_tool.cc
|
Default `try_load_options` to true when DB is specified (#9937)
|
3 years ago |
pflag
|
Fix /bin/bash shebangs
|
7 years ago |
reduce_levels_test.cc
|
Remove own ToString() (#9955)
|
3 years ago |
regression_test.sh
|
Fix COMMIT_ID in regression_test.sh (#9047)
|
3 years ago |
restore_db.sh
|
Revamp check_format_compatible.sh (#8012)
|
4 years ago |
rocksdb_dump_test.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
run_blob_bench.sh
|
Add blob compaction readahead size to the BlobDB benchmark script (#9566)
|
3 years ago |
run_flash_bench.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
run_leveldb.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
sample-dump.dmp
|
First version of rocksdb_dump and rocksdb_undump.
|
10 years ago |
simulated_hybrid_file_system.cc
|
Improve SimulatedHybridFileSystem (#9301)
|
3 years ago |
simulated_hybrid_file_system.h
|
Improve SimulatedHybridFileSystem (#9301)
|
3 years ago |
sst_dump.cc
|
Implement a new subcommand "identify" for sst_dump (#6943)
|
4 years ago |
sst_dump_test.cc
|
Use the comparator from the sst file table properties in sst_dump_tool (#9491)
|
3 years ago |
sst_dump_tool.cc
|
Support using ZDICT_finalizeDictionary to generate zstd dictionary (#9857)
|
3 years ago |
trace_analyzer.cc
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
trace_analyzer_test.cc
|
Disable the QPS verification in test temporally (#9190)
|
3 years ago |
trace_analyzer_tool.cc
|
Use std::numeric_limits<> (#9954)
|
3 years ago |
trace_analyzer_tool.h
|
Add commit marker with timestamp (#9266)
|
3 years ago |
verify_random_db.sh
|
Add copyright headers per FB open-source checkup tool. (#5199)
|
6 years ago |
write_external_sst.sh
|
Revamp check_format_compatible.sh (#8012)
|
4 years ago |
write_stress.cc
|
Add a SystemClock class to capture the time functions of an Env (#7858)
|
4 years ago |
write_stress_runner.py
|
Allow missing "unversioned" python, as in CentOS 8 (#6883)
|
5 years ago |