One key thing about blocking is that it permits random access within a compressed file if you couple it with some virtual offset mapping. Eg see the bgzf deflate variant. Attempts to use the previous header as a context for shrinking the current header prevent such things, and if you're preventing it then I agree that there aren't so many benefits to using a block based system. I agree with nburns - where huffman is being used sensibly (mainly simple systems without complex modelling), ANS can also be used sensibly and is often a better alternative. Sometimes using a model with context will give higher compression ratios of our data, but that doesn't always make it *better*. If it did the world would use paq8 everywhere, but it doesn't. It depends what we want to achieve.
I also still stand by my view that practical implementations of block-based entropy encoders vs adaptive streaming entropy encoders show the block-based systems to be faster. Agreed it needs multiple passes so in theory a streaming system has an advantage, so maybe there are commercial tools that prove otherwise, or hardware based systems too. However I've yet to see any free command-line tools that do nothing other than entropy encoding to demonstrate lightning quick arithmetic coding.