←back to thread

177 points ohjeez | 8 comments | | HN request time: 0.892s | source | bottom
1. pcrh ◴[] No.44474775[source]
How is an LLM supposed to review an original manuscript?

At their core (and as far as I understand), LLMs are based on pre-existing texts, and use statistical algorithms to stitch together text that is consistent with these.

An original research manuscript will not have formed part of any LLMs training dataset, so there is no conceivable way that it can evaluate it, regardless of claims that LLMs "understand" anything or not.

Reviewers who use LLMs are likely deluding themselves that they are now more productive due to use of AI, when in fact they are just polluting science through their own ignorance of epistemology.

replies(3): >>44474852 #>>44474964 #>>44475084 #
2. calebkaiser ◴[] No.44474852[source]
You might be interested in work around mechanistic interpretability! In particular, if you're interested in how models handle out-of-distribution information and apply in-context learning, research around so-called "circuits" might be up your alley: https://www.transformer-circuits.pub/2022/mech-interp-essay
replies(1): >>44474956 #
3. pcrh ◴[] No.44474956[source]
After a brief scan, I'm not competent to evaluate the essay by Chris Olah you posted.

I probably could get an LLM to do so, but I won't....

replies(1): >>44475123 #
4. jeroenhd ◴[] No.44474964[source]
LLMs can find problems in logic, conclusions based on circumstantial evidence, common mistakes made in other rejected papers, and other suspect language, even if it hasn't seen the exact sentence structures used in its input. You'll catch plenty of improvements to scientific preprints that way because humans aren't all that good at writing down long, complicated documents as we might think we are.

Sometimes it'll claim that a noun can only be used as a verb and will think you're Santa. LLMs can't be relied to be accurate or truthful of course.

I can imagine the non-computer science people (and unfortunately some computer science people) believe LLMs are close to infallibe. What's a biologist or a geographist going to know about the limits of ChatGPT? All they know is that the LLM did a great job spotting the grammatical issues in the paragraph they had it check so it seems pretty legit right?

replies(1): >>44474978 #
5. pcrh ◴[] No.44474978[source]
I don't doubt that LLMs can improve grammar. However, an original research paper should not be evaluated on the basis of the quality of the writing, unless this is so bad as to make the claims impenetrable.
replies(1): >>44475954 #
6. analog31 ◴[] No.44475084[source]
It's like anybody else managing their workload. Professors assign the papers to their grad students to review. Overworked grad student feeds it into the LLM. It doesn't matter if the work is novel, only that it produces something that looks like a review.
7. qingcharles ◴[] No.44475123{3}[source]
I ran it through an LLM it said the paper was absolutely outstanding and perhaps the best paper of all time.
8. jeroenhd ◴[] No.44475954{3}[source]
I totally agree, but I kind of doubt the people using LLMs to review their papers were ever interested in rigorously verifying the science in the first place.