
Originally Posted by
Christian
Thank you. Does rz perform better with a bigger dictionary (like 256M or 512M)? rz's deduplication ability scales with the size of the selected lz-dictionary. Btw. you have a very nice machine.
About the small dictionary: Yes, it's funny, I understand. But 1M dictionary + x86 is not really rz's use case, is it?
It may not be the intended use case, but I find advances in data compression very interesting, more so when the same hardware limitations are used for testing. For example, cmix achieves amazing levels of compression, but only because it uses amazing amounts of RAM.
Here are some more benchmarks - I hope these are helpful:
WINDOWS 3.0 MULTIMEDIA EDITION
Code:
PROGRAM SIZE C.TIME D.TIME
(CPU) (CPU)
======================================
org 330,078,282
rz 1m 92,542,512 497.082 6.131
rz 64m 90,398,131 626.874 5.819
rz 128m 90,351,664 653.597 5.600
rz 256m 90,344,006 648.309 5.897
FIREFOX PROGRAM FILES DIRECTORY
Code:
PROGRAM SIZE C.TIME D.TIME
(CPU) (CPU)
======================================
org 245,683,256
rz 1m 75,949,354 305.216 3.775
rz 64m 65,149,620 425.898 3.853
rz 128m 65,104,141 443.402 3.557
rz 256m 65,070,761 432.310 3.526
ENCODE'S COMPRESSION CORPUS (ENCCC)
Code:
PROGRAM SIZE C.TIME D.TIME
(CPU) (CPU)
======================================
org 124,458,915
rz 1m 33,254,389 166.765 1.903
rz 64m 31,925,561 174.035 1.903
rz 128m 31,926,575 169.448 1.825
rz 256m 31,926,575 172.490 1.872