←back to thread

60 points QueensGambit | 1 comments | | HN request time: 0.199s | source
Show context
daxfohl ◴[] No.45686184[source]
I've seen research that shows that starting with reasoning models, and fine-tuning to slowly remove the reasoning steps, allows you to bake the reasoning directly into the model weights in a strong sense. Here's a recent example, and you can see the digits get baked into a pentagonal prism in the weights, allowing accurate multi-digit multiplication without needing notes: https://arxiv.org/abs/2510.00184. So, reasoning and tool use could be the first step, to collect a ton of training data to do something like this fine-tuning process.
replies(3): >>45686449 #>>45687650 #>>45688521 #
photonthug ◴[] No.45688521[source]
Glad to see the pentagonal multiplication prism is just as weird as the addition helix https://arxiv.org/abs/2502.00873
replies(1): >>45689060 #
1. daxfohl ◴[] No.45689060[source]
Yeah, I have to imagine animal brains are just giant Fourier transform engines under the hood, and humans brains have just evolved to make some frequencies more precise.