Most active commenters
  • bloomingkales(3)
  • zoogeny(3)

←back to thread

S1: A $6 R1 competitor?

(timkellogg.me)
851 points tkellogg | 14 comments | | HN request time: 1.359s | source | bottom
1. bloomingkales ◴[] No.42949616[source]
If an LLM output is like a sculpture, then we have to sculpt it. I never did sculpting, but I do know they first get the clay spinning on a plate.

Whatever you want to call this “reasoning” step, ultimately it really is just throwing the model into a game loop. We want to interact with it on each tick (spin the clay), and sculpt every second until it looks right.

You will need to loop against an LLM to do just about anything and everything, forever - this is the default workflow.

Those who think we will quell our thirst for compute have another thing coming, we’re going to be insatiable with how much LLM brute force looping we will do.

replies(3): >>42955281 #>>42955806 #>>42956482 #
2. MrLeap ◴[] No.42955281[source]
This is a fantastic insight and really has my gears spinning.

We need to cluster the AI's insights on a spatial grid hash, give it a minimap with the ability to zoom in and out, and give it the agency to try and find its way to an answer and build up confidence and tests for that answer.

coarse -> fine, refine, test, loop.

Maybe a parallel model that handles the visualization stuff. I imagine its training would look more like computer vision. Mind palace generation.

If you're stuck or your confidence is low, wander the palace and see what questions bubble up.

Bringing my current context back through the web is how I think deeply about things. The context has the authority to reorder the web if it's "epiphany grade".

I wonder if the final epiphany at the end of what we're creating is closer to "compassion for self and others" or "eat everything."

3. zoogeny ◴[] No.42955806[source]
I can't believe this hasn't been done yet, perhaps it is a cost issue.

My literal first thought about AI was wondering why we couldn't just put it in a loop. Heck, one update per day, or one update per hour would even be a start. You have a running "context", the output is the next context (or a set of transformations on a context that is a bit larger than the output window). Then ramp that up ... one loop per minute, one per second, millisecond, microsecond.

replies(2): >>42955958 #>>42956117 #
4. layer8 ◴[] No.42955958[source]
Same. And the next step is that it must feed back into training, to form long-term memory and to continually learn.
replies(1): >>42955988 #
5. zoogeny ◴[] No.42955988{3}[source]
I analogize this with sleep. Perhaps that is what is needed, 6 hours offline per day to LoRa the base model on some accumulated context from the day.
replies(1): >>42965779 #
6. int_19h ◴[] No.42956117[source]
The hard part is coming up with a good way to grade results. Which you need to update the weights based on the outcome, otherwise the model will not actually learn anything.
replies(1): >>42956298 #
7. zoogeny ◴[] No.42956298{3}[source]
For the "looping" I'm talking about you don't need to update the weights. It is simply, old context in, new context out, new context in, new-new context out, etc.

Of course, keeping that coherent over numerous loops isn't going to be easy. No doubt there is a chance it goes off the rails. So you might have a section of context that is constantly stable, a section of context that updates each loop, etc.

In the other response to my comment someone mentioned eventually updating the weights (e.g. daily) and you would in that case have to have some kind of loss function.

replies(3): >>42957903 #>>42958106 #>>42958279 #
8. gyomu ◴[] No.42956482[source]
> If an LLM output is like a sculpture, then we have to sculpt it. I never did sculpting, but I do know they first get the clay spinning on a plate.

That’s pottery, not sculpture. Traditionally in sculpture you start from a block of marble or wood, but you can also make sculptures of cast bronze or welded steel (or clay, but you don’t use a spinning plate).

replies(1): >>42957923 #
9. ◴[] No.42957903{4}[source]
10. bloomingkales ◴[] No.42957923[source]
Thank you for the clarification. I wanted to use some kind of visual to show the model in a loop. Otherwise, I’d just have to say explicitly that the sculptor is the one in the loop, as in the person will not stop chiseling. It’s in this infinite chiseling that we get our answers (same thing as finding a limit in calculus as it approaches infinity, we will never get the discrete answer, but we will get infinitely close enough to label a discrete point confidently).

In other words, we fly as close to the sun as possible and get our measurements :)

11. int_19h ◴[] No.42958106{4}[source]
Then I'm not quite sure what benefit you expect to derive from it? Making e.g. QwQ-32 loop isn't hard - it often does it all by itself, even. But it doesn't translate to improvements on every iteration; it just goes in circles.
replies(1): >>42962817 #
12. bloomingkales ◴[] No.42958279{4}[source]
When you look out your eyes, that rectangular viewport is all your context. For example, we cannot fit Jupiter into your viewport.

So, if we can never fit large concepts like Jupiter into your viewport (your context window), does it follow that you will never understand Jupiter? Or is there no way to take a picture of Jupiter and make it smaller so it fits into your viewport?

See, coherence is just resizing or reimagining things so it fits into context. Context can never hold it all because we have the capacity to always imagine infinitely large things.

So I don’t fully know if it’s old context out, new context in, but could be just the right context, just enough of it, and just the right looking context so we can assemble the right worldview. This process would have to go on infinitely, and that is the world as you see and understand it.

13. ◴[] No.42962817{5}[source]
14. dev0p ◴[] No.42965779{4}[source]
LLMs need to sleep too. Do they dream of electric sheep?