←back to thread

Getting 50% (SoTA) on Arc-AGI with GPT-4o

(redwoodresearch.substack.com)
394 points tomduncalf | 5 comments | | HN request time: 0.203s | source
Show context
rgbrgb ◴[] No.40712154[source]
> 50% accuracy on the public test set for ARC-AGI by having GPT-4o

Isn't the public test set public on github and therefore GPT-4o trained on it?

replies(2): >>40712401 #>>40712472 #
1. bongodongobob ◴[] No.40712401[source]
I keep seeing this comment all over the place. Just because something exists 1 time in the training data doesn't mean it can just regurgitate that. That's not how training works. An LLM is not a knowledge database.
replies(2): >>40712453 #>>40713177 #
2. adroniser ◴[] No.40712453[source]
And yet it doesn't rule out that it can't. See new york times lawsuit
replies(1): >>40712544 #
3. bongodongobob ◴[] No.40712544[source]
From old pieces of articles that are quoted all over the internet? That's not surprising.
replies(1): >>40714639 #
4. spencerchubb ◴[] No.40713177[source]
It could exist many times. People can fork and clone the repo. People are likely to copy the examples and share them online.
5. ben_w ◴[] No.40714639{3}[source]
That's still sufficient for both The Times and for it to be a potential problem in this case.