←back to thread

296 points todsacerdoti | 1 comments | | HN request time: 0.206s | source
Show context
kgeist ◴[] No.44372056[source]
>From a domain point of view, some are skeptical that bytes are adequate for modelling natural language

If I remember correctly, GPT3.5's tokenizer treated Cyrillic as individual characters, and GPT3.5 was pretty good at Russian.

replies(1): >>44373669 #
1. dgfitz ◴[] No.44373669[source]
I wonder if they treat each letter as a Unicode code point, and each of those is a token? I could see the same being true of other languages.