Most active commenters
  • verdverm(3)

←back to thread

GPT-5.2

(openai.com)
1019 points atgctg | 27 comments | | HN request time: 0.002s | source | bottom
1. sfmike ◴[] No.46234974[source]
Everything is still based on 4 4o still right? is a new model training just too expensive? They can consult deepseek team maybe for cost constrained new models.
replies(4): >>46235000 #>>46235052 #>>46235127 #>>46235143 #
2. verdverm ◴[] No.46235000[source]
Apparently they have not had a successful pre training run in 1.5 years
replies(2): >>46235068 #>>46235299 #
3. Wowfunhappy ◴[] No.46235052[source]
I thought whenever the knowledge cutoff increased that meant they’d trained a new model, I guess that’s completely wrong?
replies(2): >>46235181 #>>46236200 #
4. fouronnes3 ◴[] No.46235068[source]
I want to read a short scify story set in 2150 about how, mysteriously, no one has been able to train a better LLM for 125 years. The binary weights are studied with unbelievably advanced quantum computers but no one can really train a new AI from scratch. This starts cults, wars and legends and ultimately (by the third book) leads to the main protagonist learning to code by hand, something that no human left alive still knows how to do. Could this be the secret to making a new AI from scratch, more than a century later?
replies(6): >>46235128 #>>46235237 #>>46235306 #>>46235386 #>>46235429 #>>46235455 #
5. elgatolopez ◴[] No.46235127[source]
Where did you get that from? Cutoff date says august 2025. Looks like a newly pretrained model
replies(2): >>46235406 #>>46235471 #
6. armenarmen ◴[] No.46235128{3}[source]
I’d read it!
7. catigula ◴[] No.46235143[source]
The irony is that Deepseek is still running with a distilled 4o model.
replies(1): >>46235772 #
8. brokencode ◴[] No.46235181[source]
Typically I think, but you could pre-train your previous model on new data too.

I don’t think it’s publicly known for sure how different the models really are. You can improve a lot just by improving the post-training set.

9. barrenko ◴[] No.46235237{3}[source]
Monsieur, if I may offer a vaaaguely similar story on how things may progress https://www.owlposting.com/p/a-body-most-amenable-to-experim...
10. ijl ◴[] No.46235299[source]
What kind of issues could prevent a company with such resources from that?
replies(1): >>46235522 #
11. verdverm ◴[] No.46235306{3}[source]
You can ask 2025 Ai to write such a book, it's happy to comply and may or may not actually write the book

https://www.pcgamer.com/software/ai/i-have-been-fooled-reddi...

12. WhyOhWhyQ ◴[] No.46235386{3}[source]
There's a scifi short story about a janitor who knows how to do basic arithmetic and becomes the most important person in the world when some disaster happens. Of course after things get set up again due to his expertise, he becomes low status again.
replies(1): >>46236959 #
13. SparkyMcUnicorn ◴[] No.46235406[source]
If the pretraining rumors are true, they're probably using continued pretraining on the older weights. Right?
replies(1): >>46236213 #
14. ssl-3 ◴[] No.46235429{3}[source]
Sounds good.

Might sell better with the protagonist learning iron age leatherworking, with hides tanned from cows that were grown within earshot, as part of a process of finding the real root of the reason for why any of us ever came to be in the first place. This realization process culminates in the formation of a global, unified steampunk BDSM movement and a wealth of new diseases, and then: Zombies.

(That's the end. Zombies are always the end.)

replies(2): >>46238931 #>>46239250 #
15. georgefrowny ◴[] No.46235455{3}[source]
An software version of Asimov's Holmes-Ginsbook device? https://sfwritersworkshop.org/node/1232

I feel like there was a similar one about software, but it might have been mathematics (also Asimov: The Feeling of Power)

16. FergusArgyll ◴[] No.46235471[source]
> This stands in sharp contrast to rivals: OpenAI’s leading researchers have not completed a successful full-scale pre-training run that was broadly deployed for a new frontier model since GPT-4o in May 2024, highlighting the significant technical hurdle that Google’s TPU fleet has managed to overcome.

- https://newsletter.semianalysis.com/p/tpuv7-google-takes-a-s...

It's also plainly obvious from using it. The "Broadly deployed" qualifier is presumably referring to 4.5

17. verdverm ◴[] No.46235522{3}[source]
Drama if I had to pick the symptom most visible from the outside.

A lot of talent left OpenAI around that time, most notably in this regard would be Ilya in May '24. Remember that time Ilya and the board ousted Sam only to reverse it almost immediately?

https://arstechnica.com/information-technology/2024/05/chief...

18. blovescoffee ◴[] No.46235772[source]
Source?
19. rockinghigh ◴[] No.46236200[source]
They add new data to the existing base model via continuous pre-training. You save on pre-training, the next token prediction task, but still have to re-run mid and post training stages like context length extension, supervised fine tuning, reinforcement learning, safety alignment ...
replies(1): >>46239272 #
20. ◴[] No.46236213{3}[source]
21. bradfitz ◴[] No.46236959{4}[source]
I had to go look that up! I assume that's https://en.wikipedia.org/wiki/The_Feeling_of_Power ? (Not a janitor, but "a low grade Technician"?)
replies(1): >>46237121 #
22. WhyOhWhyQ ◴[] No.46237121{5}[source]
Hmm it could be a false memory, since this was almost 15 years ago, but I really do remember it differently than the text of 'Feeling of Power'.
replies(1): >>46242285 #
23. wafflemaker ◴[] No.46238931{4}[source]
Sorry, but compared with the parent, my money is in you ssl-3. Do you get better results from prompting by being more poetic?
replies(1): >>46242247 #
24. astrange ◴[] No.46239250{4}[source]
This is somewhat similar to a Piers Anthony series that I suspect noone has ever read except for me.

What was with that guy anyway.

25. astrange ◴[] No.46239272{3}[source]
Continuous pretraining has issues because it starts forgetting the older stuff. There is some research into other approaches.
26. ssl-3 ◴[] No.46242247{5}[source]
> Do you get better results from prompting by being more poetic?

Is that yet-another accusation of having used the bot?

I don't use the bot to write English prose. If something I write seems particularly great or poetic or something, then that's just me: I was in the right mood, at the right time, with the right idea -- and with the right audience.

When it's bad or fucked-up, then that's also just me. I most-assuredly fuck up plenty.

They can't all be zingers. I'm fine with that.

---

I do use the hell out of the bot for translating my ideas (and the words that I use to express them) into languages that I can't speak well, like Python, C, and C++. But that's very different. (And at least so far I haven't shared any of those bot outputs with the world at all, either.)

So to take your question very literally: No, I don't get better results from prompting being more poetic. The responses to my prompts don't improve by those prompts being articulate or poetic.

Instead, I've found that I get the best results from the bot fastest by carrying a big stick, and using that stick to hammer and welt it into compliance.

Things can get rather irreverent in my interactions with the bot. Poeticism is pretty far removed from any of that business.

replies(1): >>46242933 #
27. wafflemaker ◴[] No.46242933{6}[source]
No. I just genuinely liked your style, and didn't notice previous posts by you. I haven't yet learned to look at names on hn, it's mostly anonymous posts for me. No snark here. And was also genuinely curious if better writing style yields better results.

I've observed that using proper grammar gives slightly better answers. And using more "literacy"(?) kind of language in prompts sometimes gives better answers and sometimes just more interesting ones, when bots try to follow my style.

Sorry for using the word poetic, I'm travelling and sleep deprived and couldn't find the proper word, but didn't want to just use "nice" instead either.