←back to thread

114 points cmcconomy | 3 comments | | HN request time: 0.539s | source
1. lr1970 ◴[] No.42178968[source]
> We have extended the model’s context length from 128k to 1M, which is approximately 1 million English words

Actually English language tokenizers map on average 3 words into 4 tokens. Hence 1M tokens is about 750K English words not a million as claimed.

replies(2): >>42179102 #>>42179262 #
2. ◴[] No.42179102[source]
3. swyx ◴[] No.42179262[source]
good, its been hours since i saw a "well actually" comment on HN