←back to thread

373 points ibobev | 1 comments | | HN request time: 0.22s | source
Show context
mingtianzhang ◴[] No.45067101[source]
It would be interesting to add more lossless compression stuff, which has a close connection to generative AI.

This PhD thesis gives a very good introduction: https://arxiv.org/abs/2104.10544

replies(1): >>45067799 #
roadside_picnic ◴[] No.45067799[source]
You don't need to restrict it to lossless compression, in fact nearly all machine learning can be understood as a type of compression (typically lossy). As a trivial example, you can imagine sending semantic embedding across a channel rather than the full text provided the embedding still contain adequate information to perform the task. Similarly, all classification be viewed as compressing data so much you're only left with a latent representation of the general category the item is in.

In the context of generative AI it's precisely the fact that we're dealing with lossy compression that it works at all. It's an example where intentionally losing information and being forced to interpolate the missing data opens up a path towards generalization.

Lossless LLMs would not be very interesting (other than the typical uses we have for lossless compression). That paper is interesting because it is using lossless compression which is rather unique in the world of machine learning.

replies(3): >>45068195 #>>45071246 #>>45073444 #
1. andoando ◴[] No.45071246[source]
All learning, human or AI is a lossy compression.

It is by generalizing data that we form mental conceptions. A square is a square despite its size or color or material. A house is a house so long as something lives there.