←back to thread

237 points JnBrymn | 2 comments | | HN request time: 0.001s | source
Show context
sabareesh ◴[] No.45675879[source]
It might be that our current tokenization is inefficient compared to how well image pipeline does. Language already does lot of compression but there might be even better way to represent it in latent space
replies(3): >>45675953 #>>45676049 #>>45677115 #
ACCount37 ◴[] No.45675953[source]
People in the industry know that tokenizers suck and there's room to do better. But actually doing it better? At scale? Now that's hard.
replies(1): >>45676189 #
typpilol ◴[] No.45676189[source]
It will require like 20x the compute
replies(3): >>45676906 #>>45676935 #>>45676964 #
1. ACCount37 ◴[] No.45676935[source]
A lot of cool things are shot down by "it requires more compute, and by a lot, and we're already compute starved on any day of the week that ends in y, so, not worth it".

If we had a million times the compute? We might have brute forced our way to AGI by now.

replies(1): >>45676998 #
2. Jensson ◴[] No.45676998[source]
But we don't have a million times the compute, we have the compute we have so its fair to argue that we want to prioritize other things.