> you can partition your dataset and process each partition on separate core, which will produce some massive XX or even XXX GB/s?
Yes, but as I mentioned:
> multicore decompression is not really possible unless you manually tweak the compression
That is, there is no stable implementation out there that does it. You will have to do that manually and painfully. In which case, you're opening the doors for exotic/niche compression/decompression, and there are better alternatives than LZ4 if you're in the niche market.
> this is obviously depends on your data pattern. If it is some low cardinality IDs, they can be compressed by ratio 100 easily.
Everything is possible in theory. Yet we have to agree on what is a reasonable expectation. A compression factor of around 3:1 is, from my experience, what you would get from a reasonable compression speed on reasonably distributed data.