←back to thread

350 points kashifr | 3 comments | | HN request time: 0.579s | source
Show context
bitwize ◴[] No.44502203[source]
There's a British comedy skit lurking in here.

"So it's a small large language model?"

"Oh yes, very small."

"How can it be small and large at the same time?"

"Well, it's small by the standards of a large language model."

"So it's large."

"Oh yes, very large."

"Large compared to what?"

"Small language models."

"And so something like ChatGPT, what would that be exactly? A large large language model?"

"Yes, precisely. An LLLM."

replies(6): >>44502878 #>>44502893 #>>44502939 #>>44503785 #>>44504888 #>>44506490 #
janalsncm ◴[] No.44502893[source]
Standards have shifted as well. Gpt2 used to be considered “large” but it is half the size of this. Oh and also Sam Altman said it was too dangerous to release. At this point I consider anything too big to run on consumer grade hardware to be large, but an exact definition is a little silly to argue about.
replies(2): >>44504325 #>>44507408 #
a_wild_dandan ◴[] No.44504325[source]
Altman released GPT-2 despite expressing that doing so was a bad idea? That's wild.
replies(1): >>44504590 #
Alifatisk ◴[] No.44504590[source]
I think Altman meant it's too dangerous to open-source GPT-2, therefore locked it in behind a service.
replies(1): >>44505067 #
1. janalsncm ◴[] No.44505067[source]
It’s not locked behind a service though.

https://huggingface.co/openai-community/gpt2/blob/main/model...

replies(1): >>44507162 #
2. Alifatisk ◴[] No.44507162[source]
That’s only 124M param
replies(1): >>44508906 #
3. thatjoeoverthr ◴[] No.44508906[source]
Behold https://huggingface.co/openai-community/gpt2-xl