←back to thread

GPT-5.2

(openai.com)
1094 points atgctg | 1 comments | | HN request time: 0.518s | source
Show context
svara ◴[] No.46241936[source]
In my experience, the best models are already nearly as good as you can be for a large fraction of what I personally use them for, which is basically as a more efficient search engine.

The thing that would now make the biggest difference isn't "more intelligence", whatever that might mean, but better grounding.

It's still a big issue that the models will make up plausible sounding but wrong or misleading explanations for things, and verifying their claims ends up taking time. And if it's a topic you don't care about enough, you might just end up misinformed.

I think Google/Gemini realize this, since their "verify" feature is designed to address exactly this. Unfortunately it hasn't worked very well for me so far.

But to me it's very clear that the product that gets this right will be the one I use.

replies(14): >>46241987 #>>46242107 #>>46242173 #>>46242280 #>>46242317 #>>46242483 #>>46242537 #>>46242589 #>>46243494 #>>46243567 #>>46243680 #>>46244002 #>>46244904 #>>46245168 #
phorkyas82 ◴[] No.46241987[source]
Isn't that what no LLM can provide: being free of hallucinations?
replies(5): >>46242091 #>>46242093 #>>46242230 #>>46243681 #>>46244023 #
kyletns ◴[] No.46242093[source]
For the record, brains are also not free of hallucinations.
replies(3): >>46242289 #>>46242311 #>>46244746 #
rimeice ◴[] No.46242311[source]
I still don’t really get this argument/excuse for why it’s acceptable that LLMs hallucinate. These tools are meant to support us, but we end up with two parties who are, as you say, prone to “hallucination” and it becomes a situation of the blind leading the blind. Ideally in these scenarios there’s at least one party with a definitive or deterministic view so the other party (i.e. us) at least has some trust in the information they’re receiving and any decisions they make off the back of it.
replies(4): >>46242664 #>>46242733 #>>46242790 #>>46243300 #
1. Libidinalecon ◴[] No.46243300[source]
"The airplane wing broke and fell off during flight"

"Well humans break their leg too!"

It is just a mindlessly stupid response and a giant category error.

The way an airplane wing and a human limb is not at all the same category.

There is even another layer to this that comparing LLMs to the brain might be wrong because the mereological fallacy is attributing the brain "thinks" vs the person/system as a whole thinks.