Most active commenters
  • taeric(3)
  • TZubiri(3)

←back to thread

688 points crescit_eundo | 11 comments | | HN request time: 1.1s | source | bottom
Show context
azeirah ◴[] No.42141993[source]
Maybe I'm really stupid... but perhaps if we want really intelligent models we need to stop tokenizing at all? We're literally limiting what a model can see and how it percieves the world by limiting the structure of the information streams that come into the model from the very beginning.

I know working with raw bits or bytes is slower, but it should be relatively cheap and easy to at least falsify this hypothesis that many huge issues might be due to tokenization problems but... yeah.

Surprised I don't see more research into radicaly different tokenization.

replies(14): >>42142033 #>>42142384 #>>42143197 #>>42143338 #>>42143381 #>>42144059 #>>42144207 #>>42144582 #>>42144600 #>>42145725 #>>42146419 #>>42146444 #>>42149355 #>>42151016 #
cschep ◴[] No.42142033[source]
How would we train it? Don't we need it to understand the heaps and heaps of data we already have "tokenized" e.g. the internet? Written words for humans? Genuinely curious how we could approach it differently?
replies(2): >>42142126 #>>42142146 #
1. skylerwiernik ◴[] No.42142146[source]
Couldn't we just make every human readable character a token?

OpenAI's tokenizer makes "chess" "ch" and "ess". We could just make it into "c" "h" "e" "s" "s"

replies(3): >>42142199 #>>42142203 #>>42142835 #
2. taeric ◴[] No.42142199[source]
This is just more tokens? And probably requires the model to learn about common groups. Consider, "ess" makes sense to see as a group. "Wss" does not.

That is, the groups are encoding something the model doesn't have to learn.

This is not much astray from "sight words" we teach kids.

replies(2): >>42143246 #>>42145899 #
3. tchalla ◴[] No.42142203[source]
aka Character Language Models which have existed for a while now.
4. cco ◴[] No.42142835[source]
We can, tokenization is literally just to maximize resources and provide as much "space" as possible in the context window.

There is no advantage to tokenization, it just helps solve limitations in context windows and training.

replies(1): >>42143249 #
5. TZubiri ◴[] No.42143246[source]
This is just more tokens?

Yup. Just let the actual ML git gud

replies(1): >>42143334 #
6. TZubiri ◴[] No.42143249[source]
I like this explanation
7. taeric ◴[] No.42143334{3}[source]
So, put differently, this is just more expensive?
replies(1): >>42152806 #
8. Hendrikto ◴[] No.42145899[source]
No, actually much fewer tokens. 256 tokens cover all bytes. See the ByT5 paper: https://arxiv.org/abs/2105.13626
replies(1): >>42145925 #
9. taeric ◴[] No.42145925{3}[source]
More tokens to a sequence, though. And since it is learning sequences...
replies(1): >>42147484 #
10. loa_in_ ◴[] No.42147484{4}[source]
Yeah, suddenly 16k tokens is just 16kb of ASCII instead of ~6kwords
11. TZubiri ◴[] No.42152806{4}[source]
Expensive in terms of computationally expensive, time expensive, and yes cost expensive.

Worth noting that the relationship between characters to token ratio is probably quadratic or cubic or some other polynomial. So the difference in terms of computational difficulty is probably huge when compared to a character per token.