←back to thread

114 points cmcconomy | 1 comments | | HN request time: 0s | source
Show context
lr1970 ◴[] No.42178968[source]
> We have extended the model’s context length from 128k to 1M, which is approximately 1 million English words

Actually English language tokenizers map on average 3 words into 4 tokens. Hence 1M tokens is about 750K English words not a million as claimed.

replies(2): >>42179102 #>>42179262 #
1. ◴[] No.42179102[source]