You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
Changyu Bi
cc23b46da1
Support using ZDICT_finalizeDictionary to generate zstd dictionary (#9857)
Summary:
An untrained dictionary is currently simply the concatenation of several samples. The ZSTD API, ZDICT_finalizeDictionary(), can improve such a dictionary's effectiveness at low cost. This PR changes how dictionary is created by calling the ZSTD ZDICT_finalizeDictionary() API instead of creating raw content dictionary (when max_dict_buffer_bytes > 0), and pass in all buffered uncompressed data blocks as samples.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/9857
Test Plan:
#### db_bench test for cpu/memory of compression+decompression and space saving on synthetic data:
Set up: change the parameter [here](fb9a167a55/tools/db_bench_tool.cc (L1766) ) to 16384 to make synthetic data more compressible.
```
# linked local ZSTD with version 1.5.2
# DEBUG_LEVEL=0 ROCKSDB_NO_FBCODE=1 ROCKSDB_DISABLE_ZSTD=1 EXTRA_CXXFLAGS="-DZSTD_STATIC_LINKING_ONLY -DZSTD -I/data/users/changyubi/install/include/" EXTRA_LDFLAGS="-L/data/users/changyubi/install/lib/ -l:libzstd.a" make -j32 db_bench
dict_bytes=16384
train_bytes=1048576
echo "========== No Dictionary =========="
TEST_TMPDIR=/dev/shm ./db_bench -benchmarks=filluniquerandom,compact -num=10000000 -compression_type=zstd -compression_max_dict_bytes=0 -block_size=4096 -max_background_jobs=24 -memtablerep=vector -allow_concurrent_memtable_write=false -disable_wal=true -max_write_buffer_number=8 >/dev/null 2>&1
TEST_TMPDIR=/dev/shm /usr/bin/time ./db_bench -use_existing_db=true -benchmarks=compact -compression_type=zstd -compression_max_dict_bytes=0 -block_size=4096 2>&1 | grep elapsed
du -hc /dev/shm/dbbench/*sst | grep total
echo "========== Raw Content Dictionary =========="
TEST_TMPDIR=/dev/shm ./db_bench_main -benchmarks=filluniquerandom,compact -num=10000000 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -block_size=4096 -max_background_jobs=24 -memtablerep=vector -allow_concurrent_memtable_write=false -disable_wal=true -max_write_buffer_number=8 >/dev/null 2>&1
TEST_TMPDIR=/dev/shm /usr/bin/time ./db_bench_main -use_existing_db=true -benchmarks=compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -block_size=4096 2>&1 | grep elapsed
du -hc /dev/shm/dbbench/*sst | grep total
echo "========== FinalizeDictionary =========="
TEST_TMPDIR=/dev/shm ./db_bench -benchmarks=filluniquerandom,compact -num=10000000 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -compression_use_zstd_dict_trainer=false -block_size=4096 -max_background_jobs=24 -memtablerep=vector -allow_concurrent_memtable_write=false -disable_wal=true -max_write_buffer_number=8 >/dev/null 2>&1
TEST_TMPDIR=/dev/shm /usr/bin/time ./db_bench -use_existing_db=true -benchmarks=compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -compression_use_zstd_dict_trainer=false -block_size=4096 2>&1 | grep elapsed
du -hc /dev/shm/dbbench/*sst | grep total
echo "========== TrainDictionary =========="
TEST_TMPDIR=/dev/shm ./db_bench -benchmarks=filluniquerandom,compact -num=10000000 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -block_size=4096 -max_background_jobs=24 -memtablerep=vector -allow_concurrent_memtable_write=false -disable_wal=true -max_write_buffer_number=8 >/dev/null 2>&1
TEST_TMPDIR=/dev/shm /usr/bin/time ./db_bench -use_existing_db=true -benchmarks=compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -block_size=4096 2>&1 | grep elapsed
du -hc /dev/shm/dbbench/*sst | grep total
# Result: TrainDictionary is much better on space saving, but FinalizeDictionary seems to use less memory.
# before compression data size: 1.2GB
dict_bytes=16384
max_dict_buffer_bytes = 1048576
space cpu/memory
No Dictionary 468M 14.93user 1.00system 0:15.92elapsed 100%CPU (0avgtext+0avgdata 23904maxresident)k
Raw Dictionary 251M 15.81user 0.80system 0:16.56elapsed 100%CPU (0avgtext+0avgdata 156808maxresident)k
FinalizeDictionary 236M 11.93user 0.64system 0:12.56elapsed 100%CPU (0avgtext+0avgdata 89548maxresident)k
TrainDictionary 84M 7.29user 0.45system 0:07.75elapsed 100%CPU (0avgtext+0avgdata 97288maxresident)k
```
#### Benchmark on 10 sample SST files for spacing saving and CPU time on compression:
FinalizeDictionary is comparable to TrainDictionary in terms of space saving, and takes less time in compression.
```
dict_bytes=16384
train_bytes=1048576
for sst_file in `ls ../temp/myrock-sst/`
do
echo "********** $sst_file **********"
echo "========== No Dictionary =========="
./sst_dump --file="../temp/myrock-sst/$sst_file" --command=recompress --compression_level_from=6 --compression_level_to=6 --compression_types=kZSTD
echo "========== Raw Content Dictionary =========="
./sst_dump --file="../temp/myrock-sst/$sst_file" --command=recompress --compression_level_from=6 --compression_level_to=6 --compression_types=kZSTD --compression_max_dict_bytes=$dict_bytes
echo "========== FinalizeDictionary =========="
./sst_dump --file="../temp/myrock-sst/$sst_file" --command=recompress --compression_level_from=6 --compression_level_to=6 --compression_types=kZSTD --compression_max_dict_bytes=$dict_bytes --compression_zstd_max_train_bytes=$train_bytes --compression_use_zstd_finalize_dict
echo "========== TrainDictionary =========="
./sst_dump --file="../temp/myrock-sst/$sst_file" --command=recompress --compression_level_from=6 --compression_level_to=6 --compression_types=kZSTD --compression_max_dict_bytes=$dict_bytes --compression_zstd_max_train_bytes=$train_bytes
done
010240.sst (Size/Time) 011029.sst 013184.sst 021552.sst 185054.sst 185137.sst 191666.sst 7560381.sst 7604174.sst 7635312.sst
No Dictionary 28165569 / 2614419 32899411 / 2976832 32977848 / 3055542 31966329 / 2004590 33614351 / 1755877 33429029 / 1717042 33611933 / 1776936 33634045 / 2771417 33789721 / 2205414 33592194 / 388254
Raw Content Dictionary 28019950 / 2697961 33748665 / 3572422 33896373 / 3534701 26418431 / 2259658 28560825 / 1839168 28455030 / 1846039 28494319 / 1861349 32391599 / 3095649 33772142 / 2407843 33592230 / 474523
FinalizeDictionary 27896012 / 2650029 33763886 / 3719427 33904283 / 3552793 26008225 / 2198033 28111872 / 1869530 28014374 / 1789771 28047706 / 1848300 32296254 / 3204027 33698698 / 2381468 33592344 / 517433
TrainDictionary 28046089 / 2740037 33706480 / 3679019 33885741 / 3629351 25087123 / 2204558 27194353 / 1970207 27234229 / 1896811 27166710 / 1903119 32011041 / 3322315 32730692 / 2406146 33608631 / 570593
```
#### Decompression/Read test:
With FinalizeDictionary/TrainDictionary, some data structure used for decompression are in stored in dictionary, so they are expected to be faster in terms of decompression/reads.
```
dict_bytes=16384
train_bytes=1048576
echo "No Dictionary"
TEST_TMPDIR=/dev/shm/ ./db_bench -benchmarks=filluniquerandom,compact -compression_type=zstd -compression_max_dict_bytes=0 > /dev/null 2>&1
TEST_TMPDIR=/dev/shm/ ./db_bench -use_existing_db=true -benchmarks=readrandom -cache_size=0 -compression_type=zstd -compression_max_dict_bytes=0 2>&1 | grep MB/s
echo "Raw Dictionary"
TEST_TMPDIR=/dev/shm/ ./db_bench -benchmarks=filluniquerandom,compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes > /dev/null 2>&1
TEST_TMPDIR=/dev/shm/ ./db_bench -use_existing_db=true -benchmarks=readrandom -cache_size=0 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes 2>&1 | grep MB/s
echo "FinalizeDict"
TEST_TMPDIR=/dev/shm/ ./db_bench -benchmarks=filluniquerandom,compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -compression_use_zstd_dict_trainer=false > /dev/null 2>&1
TEST_TMPDIR=/dev/shm/ ./db_bench -use_existing_db=true -benchmarks=readrandom -cache_size=0 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes -compression_use_zstd_dict_trainer=false 2>&1 | grep MB/s
echo "Train Dictionary"
TEST_TMPDIR=/dev/shm/ ./db_bench -benchmarks=filluniquerandom,compact -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes > /dev/null 2>&1
TEST_TMPDIR=/dev/shm/ ./db_bench -use_existing_db=true -benchmarks=readrandom -cache_size=0 -compression_type=zstd -compression_max_dict_bytes=$dict_bytes -compression_zstd_max_train_bytes=$train_bytes 2>&1 | grep MB/s
No Dictionary
readrandom : 12.183 micros/op 82082 ops/sec 12.183 seconds 1000000 operations; 9.1 MB/s (1000000 of 1000000 found)
Raw Dictionary
readrandom : 12.314 micros/op 81205 ops/sec 12.314 seconds 1000000 operations; 9.0 MB/s (1000000 of 1000000 found)
FinalizeDict
readrandom : 9.787 micros/op 102180 ops/sec 9.787 seconds 1000000 operations; 11.3 MB/s (1000000 of 1000000 found)
Train Dictionary
readrandom : 9.698 micros/op 103108 ops/sec 9.699 seconds 1000000 operations; 11.4 MB/s (1000000 of 1000000 found)
```
Reviewed By: ajkr
Differential Revision: D35720026
Pulled By: cbi42
fbshipit-source-id: 24d230fdff0fd28a1bb650658798f00dfcfb2a1f
|
3 years ago |
.. |
binary_search_index_reader.cc
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
binary_search_index_reader.h
|
Extend Get/MultiGet deadline support to table open (#6982)
|
5 years ago |
block.cc
|
Use std::numeric_limits<> (#9954)
|
3 years ago |
block.h
|
Reduce comparator objects init cost in BlockIter (#9611)
|
3 years ago |
block_based_filter_block.cc
|
Remove own ToString() (#9955)
|
3 years ago |
block_based_filter_block.h
|
FilterPolicy API changes for 7.0 (#9501)
|
3 years ago |
block_based_filter_block_test.cc
|
FilterPolicy API changes for 7.0 (#9501)
|
3 years ago |
block_based_table_builder.cc
|
Support using ZDICT_finalizeDictionary to generate zstd dictionary (#9857)
|
3 years ago |
block_based_table_builder.h
|
Fix segmentation fault in table_options.prepopulate_block_cache when used with partition_filters (#9263)
|
3 years ago |
block_based_table_factory.cc
|
Mark old reserve* option deprecated (#10016)
|
3 years ago |
block_based_table_factory.h
|
Account memory of big memory users in BlockBasedTable in global memory limit (#9748)
|
3 years ago |
block_based_table_iterator.cc
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
block_based_table_iterator.h
|
Make initial auto readahead_size configurable (#9836)
|
3 years ago |
block_based_table_reader.cc
|
Multi file concurrency in MultiGet using coroutines and async IO (#9968)
|
3 years ago |
block_based_table_reader.h
|
Multi file concurrency in MultiGet using coroutines and async IO (#9968)
|
3 years ago |
block_based_table_reader_impl.h
|
Fb 9718 verify checksums is ignored (#9767)
|
3 years ago |
block_based_table_reader_sync_and_async.h
|
Multi file concurrency in MultiGet using coroutines and async IO (#9968)
|
3 years ago |
block_based_table_reader_test.cc
|
Multi file concurrency in MultiGet using coroutines and async IO (#9968)
|
3 years ago |
block_builder.cc
|
Improve data block construction performance (#9040)
|
4 years ago |
block_builder.h
|
Improve data block construction performance (#9040)
|
4 years ago |
block_like_traits.h
|
Change type of cache buffer passed to `Cache::CreateCallback()` to `const void*` (#9595)
|
3 years ago |
block_prefetcher.cc
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
block_prefetcher.h
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
block_prefix_index.cc
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
block_prefix_index.h
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
block_test.cc
|
Improve / clean up meta block code & integrity (#9163)
|
3 years ago |
block_type.h
|
Fix and detect headers with missing dependencies (#8893)
|
4 years ago |
cachable_entry.h
|
Meta-internal folly integration with F14FastMap (#9546)
|
3 years ago |
data_block_footer.cc
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
data_block_footer.h
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
data_block_hash_index.cc
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
data_block_hash_index.h
|
Replace namespace name "rocksdb" with ROCKSDB_NAMESPACE (#6433)
|
5 years ago |
data_block_hash_index_test.cc
|
Fast path for detecting unchanged prefix_extractor (#9407)
|
3 years ago |
filter_block.h
|
Fix auto_prefix_mode performance with partitioned filters (#10012)
|
3 years ago |
filter_block_reader_common.cc
|
Fix auto_prefix_mode performance with partitioned filters (#10012)
|
3 years ago |
filter_block_reader_common.h
|
Fix auto_prefix_mode performance with partitioned filters (#10012)
|
3 years ago |
filter_policy.cc
|
Rewrite memory-charging feature's option API (#9926)
|
3 years ago |
filter_policy_internal.h
|
Fix a major performance bug in 7.0 re: filter compatibility (#9736)
|
3 years ago |
flush_block_policy.cc
|
Restore Regex support for ObjectLibrary::Register, rename new APIs to allow old one to be deprecated in the future (#9362)
|
3 years ago |
flush_block_policy.h
|
Make FlushBlockPolicyFactory into a Customizable class (#8432)
|
4 years ago |
full_filter_block.cc
|
Fix auto_prefix_mode performance with partitioned filters (#10012)
|
3 years ago |
full_filter_block.h
|
Fix auto_prefix_mode performance with partitioned filters (#10012)
|
3 years ago |
full_filter_block_test.cc
|
Fix a major performance bug in 7.0 re: filter compatibility (#9736)
|
3 years ago |
hash_index_reader.cc
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
hash_index_reader.h
|
Extend Get/MultiGet deadline support to table open (#6982)
|
5 years ago |
index_builder.cc
|
Add (& fix) some simple source code checks (#8821)
|
4 years ago |
index_builder.h
|
Remove own ToString() (#9955)
|
3 years ago |
index_reader_common.cc
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
index_reader_common.h
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
mock_block_based_table.h
|
Refactor FilterPolicies toward Customizable (#9567)
|
3 years ago |
parsed_full_filter_block.cc
|
Hide FilterBits{Builder,Reader} from public API (#9592)
|
3 years ago |
parsed_full_filter_block.h
|
Use new Insert and Lookup APIs in table reader to support secondary cache (#8315)
|
4 years ago |
partitioned_filter_block.cc
|
Enable READ_BLOCK_COMPACTION_MICROS to track stats (#9722)
|
3 years ago |
partitioned_filter_block.h
|
Meta-internal folly integration with F14FastMap (#9546)
|
3 years ago |
partitioned_filter_block_test.cc
|
More refactoring ahead of footer & meta changes (#9240)
|
3 years ago |
partitioned_index_iterator.cc
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
partitioned_index_iterator.h
|
Make initial auto readahead_size configurable (#9836)
|
3 years ago |
partitioned_index_reader.cc
|
Set Read rate limiter priority dynamically and pass it to FS (#9996)
|
3 years ago |
partitioned_index_reader.h
|
Meta-internal folly integration with F14FastMap (#9546)
|
3 years ago |
reader_common.cc
|
Remove own ToString() (#9955)
|
3 years ago |
reader_common.h
|
Bring the Configurable options together (#5753)
|
5 years ago |
uncompression_dict_reader.cc
|
Fb 9718 verify checksums is ignored (#9767)
|
3 years ago |
uncompression_dict_reader.h
|
Fb 9718 verify checksums is ignored (#9767)
|
3 years ago |