Most active commenters
  • el_benhameen(3)

←back to thread

An LLM is a lossy encyclopedia

(simonwillison.net)
509 points tosh | 24 comments | | HN request time: 0.001s | source | bottom

(the referenced HN thread starts at https://news.ycombinator.com/item?id=45060519)
Show context
quincepie ◴[] No.45101219[source]
I totally agree with the author. Sadly, I feel like that's not what the majority of LLM users tend to view LLMs. And it's definitely not what AI companies marketing.

> The key thing is to develop an intuition for questions it can usefully answer vs questions that are at a level of detail where the lossiness matters

the problem is that in order to develop an intuition for questions that LLMs can answer, the user will at least need to know something about the topic beforehand. I believe that this lack of initial understanding of the user input is what can lead to taking LLM output as factual. If one side of the exchange knows nothing about the subject, the other side can use jargon and even present random facts or lossy facts which can almost guarantee to impress the other side.

> The way to solve this particular problem is to make a correct example available to it.

My question is how much effort would it take to make a correct example available for the LLM before it can output quality and useful data? If the effort I put in is more than what I would get in return, then I feel like it's best to write and reason it myself.

replies(7): >>45102038 #>>45102286 #>>45103159 #>>45103931 #>>45104349 #>>45105150 #>>45116121 #
cj ◴[] No.45103159[source]
> the user will at least need to know something about the topic beforehand.

I used ChatGPT 5 over the weekend to double check dosing guidelines for a specific medication. "Provide dosage guidelines for medication [insert here]"

It spit back dosing guidelines that were an order of magnitude wrong (suggested 100mcg instead of 1mg). When I saw 100mcg, I was suspicious and said "I don't think that's right" and it quickly corrected itself and provided the correct dosing guidelines.

These are the kind of innocent errors that can be dangerous if users trust it blindly.

The main challenge is LLMs aren't able to gauge confidence in its answers, so it can't adjust how confidently it communicates information back to you. It's like compressing a photo, and a photographer wrongly saying "here's the best quality image I have!" - do you trust the photographer at their word, or do you challenge him to find a better quality image?

replies(12): >>45103322 #>>45103346 #>>45103459 #>>45103642 #>>45106112 #>>45106634 #>>45108321 #>>45108605 #>>45109136 #>>45110008 #>>45110773 #>>45112140 #
1. dncornholio ◴[] No.45103322[source]
Using a LLM for medical research is just as dangerous as Googling it. Always ask your doctors!
replies(7): >>45103478 #>>45103604 #>>45103966 #>>45104534 #>>45105609 #>>45106236 #>>45108475 #
2. yojo ◴[] No.45103478[source]
This is the terrifying part: doctors do this too! I have an MD friend that told me she uses ChatGPT to retrieve dosing info. I asked her to please, please not do that.
replies(2): >>45103580 #>>45105843 #
3. ozgrakkurt ◴[] No.45103580[source]
Find good doctors. A solution doesn’t have to be perfect. A doctor doing better than regular joe with a computer is much higher as you can see in research around this topic
replies(1): >>45111188 #
4. djrj477dhsnv ◴[] No.45103604[source]
I disagree. I'd wager that state of the art LLMs can beat out of the average doctor at diagnosis given a detailed list of symptoms, especially for conditions the doctor doesn't see on a regular basis.
replies(2): >>45104305 #>>45105870 #
5. el_benhameen ◴[] No.45103966[source]
I don’t disagree that you should use your doctor as your primary source for medical decision making, but I also think this is kind of an unrealistic take. I should also say that I’m not an AI hype bro. I think we’re a long ways off from true functional AGI and robot doctors.

I have good insurance and have a primary care doctor with whom I have good rapport. But I can’t talk to her every time I have a medical question—it can take weeks to just get a phone call! If I manage to get an appointment, it’s a 15 minute slot, and I have to try to remember all of the relevant info as we speed through possible diagnoses.

Using an llm not for diagnosis but to shape my knowledge means that my questions are better and more pointed, and I have a baseline understanding of the terminology. They’ll steer you wrong on the fine points, but they’ll also steer you _right_ on the general stuff in a way that Dr. Google doesn’t.

One other anecdote. My daughter went to the ER earlier this year with some concerning symptoms. The first panel of doctors dismissed it as normal childhood stuff and sent her home. It took 24 hours, a second visit, and an ambulance ride to a children’s hospital to get to the real cause. Meanwhile, I gave a comprehensive description of her symptoms and history to an llm to try to get a handle on what I should be asking the doctors, and it gave me some possible diagnoses—including a very rare one that turned out to be the cause. (Kid is doing great now). I’m still gonna take my kids to the doctor when they’re sick, of course, but I’m also going to use whatever tools I can to get a better sense of how to manage our health and how to interact with the medical system.

replies(3): >>45106306 #>>45106491 #>>45109272 #
6. rafterydj ◴[] No.45104305[source]
"Given a detailed list of symptoms" is sure holding a lot of weight in that statement. There's way too much information that doctors tacitly understand from interactions with patients that you really cannot rely on those patients supplying in a "detailed list". Could it diagnose correctly, some of the time? Sure. But the false positive rate would be huge given LLMs suggestible nature. See the half dozen news stories covering AI induced psychosis for reference.

Regardless, it's diagnostic capability is distinct from the dangers it presents, which is what the parent comment was mentioning.

7. yujzgzc ◴[] No.45104534[source]
Plot twist, your doctor is looking it up on WebMD themselves
8. gmac ◴[] No.45105609[source]
Not really: it's arguably quite a lot worse. Because you can judge the trustworthiness of the source when you follow a link from Google (e.g. I will place quite a lot of faith in pages at an .nhs.uk URL), but nobody knows exactly how that specific LLM response got generated.
replies(1): >>45117763 #
9. nsriv ◴[] No.45105843[source]
I have a hunch that the whole "chat" interface is a brilliant but somewhat unintentional product design choice that has created this faux trust in LLM's to give back accurate information that others can get from drugs.com or Medline with a text search. This is a terrifying example, and please get her to test it out by second guessing the LLM and watching it flip flop.
10. nsriv ◴[] No.45105870[source]
What you're describing, especially with the amount of water "given a detailed list of symptoms" is carrying, is essentially a compute-intensive flowchart with no concept of diagnostic parsimony.
11. jrm4 ◴[] No.45106236[source]
Almost certainly more I would think, precisely because of magnitude errors.

The ol' "What weighs more, a pound of feathers or two pounds of bricks" trick explains this perfectly to me.

12. parpfish ◴[] No.45106306[source]
I always thought “ask your doctor” was included for liability reasons and not a thing that people actually could do.

I also have good insurance and a PCP. The idea that I could call them up just to ask “should I start doing this new exercise” or “how much aspirin for this sprained ankle?” is completely divorced from reality.

replies(3): >>45106405 #>>45106721 #>>45138619 #
13. rkomorn ◴[] No.45106405{3}[source]
I think "ask your doctor" is for prescription meds since only said doctor can write prescriptions.

And "your doctor" is actually "any doctor that is willing to write you a prescription for our medicine".

replies(1): >>45109555 #
14. shrx ◴[] No.45106491[source]
> it can take weeks to just get a phone call

> If I manage to get an appointment, it’s a 15 minute slot

I'm sorry that this is what "good insurance" gets you.

replies(1): >>45109298 #
15. el_benhameen ◴[] No.45106721{3}[source]
Yes, exactly this. I am an anxious, detail-focused person. I could call or message for every health-related question that comes to mind, but that would not be a good use of anyone’s time. My doctor is great, but she does not care about the minutiae of my health like I do, nor do I expect her to.
16. wtbdbrrr ◴[] No.45108475[source]
your doctor can have a bad day. and or be an asshole.

In 40 years, only one of my doctors had the decency to correct his mistake after I pointed it out.

He prescribed the wrong Antibiotics, which I only knew because I did something dumb and wondered if the prescribed antibiotics cover a specific strain, which they didn't, which I knew because I asked an LLM and then superficially double-checked via trustworthy official, government sources.

He then prescribed the correct antibiotics. In all other cases where I pointed out a mistake, back in the day researched without LLMs, doctors justified their logic, sometimes siding with a colleague or "the team" before evaluating the facts themselves, instead of having an independent opinion, which, AFAIK, especially in a field like medicine, is _absolutely_ imperative.

17. lurking_swe ◴[] No.45109272[source]
I live in the U.S. and my doctor is very responsive on MyChart. A few times a year i’ll send a message and I almost always get a reply within a day! From my PCP directly, or from her assistant.

I’d encourage you to find another doctor.

replies(1): >>45109987 #
18. lurking_swe ◴[] No.45109298{3}[source]
no, that’s what happens when you pick a busy doctor or a practice that’s overbooked in general. All too common these days! :(

This probably varies by locale. For example my doctor responds within 1 day on MyChart for quick questions. I can set up an in person or video appointment with her within a week, easily booked on MyChart as well.

19. parpfish ◴[] No.45109555{4}[source]
"ask your doctor" is more widespread than tthat. if you look up any diet or exercise advice, there's always an "ask your doctor before starting any new exercise program".

i'm not going to call my doctor to ask "is it okay if I try doing kettlebell squats?"

replies(1): >>45109676 #
20. rkomorn ◴[] No.45109676{5}[source]
Yes, I totally got out of context and said something a bit senseless.

But also, maybe calling your doctor would be wise (eg if you have back problems) before you start doing kettlebell squats.

I'd say that the audience for a lot of health related content skews towards people who should probably be seeing a doctor anyway.

The cynic in me also thinks some of the "ask your doctor" statements are just slapped on to artificially give credence to whatever the article is talking about (eg "this is serious exercise/diet/etc).

Edit: I guess what I meant is: I don't think it's just "liability", but genuine advice/best practice/wisdom for a sizable chunk of audiences.

21. el_benhameen ◴[] No.45109987{3}[source]
My doctor is usually pretty good at responding to messages too, but there’s still a difference between a high-certainty/high-latency reply and a medium-certainty/low-latency reply. With the llm I can ask quick follow ups or provide clarification in a way that allows me to narrow in on a solution without feeling like I’m wasting someone else’s time. But yes, if it’s bleeding, hurting, or growing, I’m definitely going to the real person.
22. SequoiaHope ◴[] No.45111188{3}[source]
I have noticed that my doctor is getting busier and busier lately. I worry that cost cutting will have doctors so frantic that they are forced to rely on things like ChatGPT, and “find good doctors” will be an option only for an elite few.
23. naasking ◴[] No.45117763[source]
Many of the big LLMs do RAG and will provide links to sources, eg. Bing/ChatGPT, Gemini Pro 2.5, etc.
24. easyThrowaway ◴[] No.45138619{3}[source]
I am constantly terrified by the American healthcare system.

That's exactly what I (and most people I know) routinely do both in Italy and France. Like, "when in doubt, call the doc". I wouldn't know where to start if I had to handle this kind of stuff exclusively by myself.