←back to thread

688 points crescit_eundo | 1 comments | | HN request time: 0s | source
Show context
swiftcoder ◴[] No.42144784[source]
I feel like the article neglects one obvious possibility: that OpenAI decided that chess was a benchmark worth "winning", special-cases chess within gpt-3.5-turbo-instruct, and then neglected to add that special-case to follow-up models since it wasn't generating sustained press coverage.
replies(8): >>42145306 #>>42145352 #>>42145619 #>>42145811 #>>42145883 #>>42146777 #>>42148148 #>>42151081 #
scott_w ◴[] No.42145811[source]
I suspect the same thing. Rather than LLMs “learning to play chess,” they “learnt” to recognise a chess game and hand over instructions to a chess engine. If that’s the case, I don’t feel impressed at all.
replies(5): >>42146086 #>>42146152 #>>42146383 #>>42146415 #>>42156785 #
fires10 ◴[] No.42146086[source]
Recognize and hand over to a specialist engine? That might be useful for AI. Maybe I am missing something.
replies(5): >>42146145 #>>42146293 #>>42146329 #>>42147558 #>>42151536 #
generic92034 ◴[] No.42146293[source]
How do you think AI are (correctly) solving simple mathematical questions which they have not trained for directly? They hand it over to a specialist maths engine.
replies(1): >>42149781 #
internetter ◴[] No.42149781[source]
This is a relatively recent development (<3 months), at least for OpenAI, where the model will generate code to solve math and use the response
replies(1): >>42151065 #
1. cruffle_duffle ◴[] No.42151065[source]
They’ve been doing that a lot longer than three months. ChatGPT has been handing stuff off to python for a very long time. At least for my paid account anyway.