←back to thread

373 points ibobev | 1 comments | | HN request time: 0.231s | source
Show context
mingtianzhang ◴[] No.45067101[source]
It would be interesting to add more lossless compression stuff, which has a close connection to generative AI.

This PhD thesis gives a very good introduction: https://arxiv.org/abs/2104.10544

replies(1): >>45067799 #
roadside_picnic ◴[] No.45067799[source]
You don't need to restrict it to lossless compression, in fact nearly all machine learning can be understood as a type of compression (typically lossy). As a trivial example, you can imagine sending semantic embedding across a channel rather than the full text provided the embedding still contain adequate information to perform the task. Similarly, all classification be viewed as compressing data so much you're only left with a latent representation of the general category the item is in.

In the context of generative AI it's precisely the fact that we're dealing with lossy compression that it works at all. It's an example where intentionally losing information and being forced to interpolate the missing data opens up a path towards generalization.

Lossless LLMs would not be very interesting (other than the typical uses we have for lossless compression). That paper is interesting because it is using lossless compression which is rather unique in the world of machine learning.

replies(3): >>45068195 #>>45071246 #>>45073444 #
mingtianzhang ◴[] No.45073444[source]
I mean, all likelihood-based generative models can be used as lossless compressors (by using arithmetic coding). The likelihood of a generated text corresponds exactly to its minimal code length under the model in practice. Thus, all current likelihood-based generative models are exact lossless compressors.
replies(1): >>45073450 #
1. mingtianzhang ◴[] No.45073450[source]
For other AI systems like recognition/classification models, they are lossy.