Most active commenters
  • slashdave(3)

←back to thread

625 points lukebennett | 40 comments | | HN request time: 0.551s | source | bottom
1. iandanforth ◴[] No.42139410[source]
A few important things to remember here:

The best engineering minds have been focused on scaling transformer pre and post training for the last three years because they had good reason to believe it would work, and it has up until now.

Progress has been measured against benchmarks which are / were largely solvable with scale.

There is another emerging paradigm which is still small(er) scale but showing remarkable results. That's full multi-modal training with embodied agents (aka robots). 1x, Figure, Physical Intelligence, Tesla are all making rapid progress on functionality which is definitely beyond frontier LLMs because it is distinctly different.

OpenAI/Google/Anthropic are not ignorant of this trend and are also reviving or investing in robots or robot-like research.

So while Orion and Claude 3.5 opus may not be another shocking giant leap forward, that does not mean that there arn't giant shocking leaps forward coming from slightly different directions.

replies(9): >>42139779 #>>42139984 #>>42140069 #>>42140194 #>>42140421 #>>42141563 #>>42142249 #>>42142983 #>>42143148 #
2. joe_the_user ◴[] No.42139779[source]
Tesla are all making rapid progress on functionality which is definitely beyond frontier LLMs because it is distinctly different

Sure, that's tautologically true but that doesn't imply that beyondness will lead to significant leaps that offer notable utility like LLMs. Deep Learning overall has been a way around the problem that intelligent behavior is very hard to code and no wants to hire many, many coders needed to do this (and no one actually how to get a mass of programmers to actually be useful beyond a certain of project complexity, to boot). People take the "bitter lesson" to mean data can do anything but I'd say a second bitter lesson is that data-things are the low hanging fruit.

Moreover, robot behavior is especially to fake. Impressive robot demos have been happening for decades without said robots getting the ability to act effectively in the complex, ad-hoc environment that human live in, IE, work with people or even cheaply emulate human behavior (but they can do choreographed/puppeteered kung fu on stage).

replies(2): >>42139926 #>>42143654 #
3. hobs ◴[] No.42139926[source]
And worth noting that Tesla faked a ton of its robot footage already, they might be making progress but their physical human robotics does not seem advanced at the moment.
replies(1): >>42140939 #
4. knicholes ◴[] No.42139984[source]
Once we've scraped the internet of its data, we need more data. Robots can take in video/audio data 24/7 and can be placed in your house to record this data by offering services like cooking/cleaning/folding laundry. Yeah, I'll pay $20k to have you record everything that happens in my house if I can stop doing dishes for five years!
replies(5): >>42140130 #>>42140146 #>>42140263 #>>42141123 #>>42142935 #
5. eli_gottlieb ◴[] No.42140069[source]
>The best engineering minds have been focused on scaling transformer pre and post training for the last three years

The best minds don't follow the herd.

replies(1): >>42149743 #
6. triyambakam ◴[] No.42140130[source]
Or get a dishwashing machine?
7. hartator ◴[] No.42140146[source]
Why 5 years?
replies(5): >>42140215 #>>42140216 #>>42140220 #>>42140608 #>>42141793 #
8. demosthanos ◴[] No.42140194[source]
> that does not mean that there arn't giant shocking leaps forward coming from slightly different directions.

Nor does it mean that there are! We've gotten into this habit of assuming that we're owed giant shocking leaps forward every year or so, and this wave of AI startups raised money accordingly, but that's never how any innovation has worked. We've always followed the same pattern: there's a breakthrough which causes a major shift in what's possible, followed by a few years of rapid growth as engineers pick up where the scientists left off, followed by a plateau while we all get used to the new normal.

We ought to be expecting a plateau, but Sam Altman and company have done their work well and have convinced many of us that this time it's different. This time it's the singularity, and we're going to see exponential growth from here on out. People want to believe it, so they do, and Altman is milking that belief for all it's worth.

But make no mistake: Altman has been telegraphing that he's eyeing the exit, and you don't eye the exit when you own a company that's set to continue exponentially increasing in value.

replies(2): >>42140584 #>>42140605 #
9. bredren ◴[] No.42140215{3}[source]
Because whatever org fills this space will be working on ARR.
10. exe34 ◴[] No.42140216{3}[source]
that's when the robot takes his job and he can't afford the robot anymore.
11. fifilura ◴[] No.42140220{3}[source]
Five years, that's all we've got.

https://en.m.wikipedia.org/wiki/Five_Years_(David_Bowie_song...

12. fldskfjdslkfj ◴[] No.42140263[source]
There's plenty of video content being uploaded and streamed everyday, i find it hard to believe the more data will really change something, excluding very specialized tasks.
replies(1): >>42140567 #
13. sincerecook ◴[] No.42140421[source]
> That's full multi-modal training with embodied agents (aka robots). 1x, Figure, Physical Intelligence, Tesla are all making rapid progress on functionality which is definitely beyond frontier LLMs because it is distinctly different.

Cool, but we already have robots doing this in 2d space (aka self driving cars) that struggle not to kill people. How is adding a third dimension going to help? People are just refusing to accept the fact that machine learning is not intelligence.

replies(4): >>42141572 #>>42141776 #>>42142802 #>>42143184 #
14. nuancebydefault ◴[] No.42140567{3}[source]
The difference with the bot is that there is a fast feedback loop between action and content. No tagging required, real physics is the playground.
15. ◴[] No.42140584[source]
16. lcnPylGDnU4H9OF ◴[] No.42140605[source]
> Altman has been telegraphing that he's eyeing the exit

Can you think of any specific examples? Not trying to express disbelief, just curious given that this is obviously not what he's intending to communicate so it would be interesting to examine what seemed to communicate it.

replies(1): >>42148862 #
17. twelve40 ◴[] No.42140608{3}[source]
> OpenAI has announced a plan to achieve artificial general intelligence (AGI) within five years, an ambitious goal as the company works to design systems that outperform humans.
18. ben_w ◴[] No.42140939{3}[source]
Indeed.

Even assuming the recent robot demo was entirely AI, the only single thing they demonstrated that would have been noteworthy was isolating one voice in a noisy crowd well enough to respond; everything else I saw Optimus do, has already been demonstrated by others.

What makes the uncertainty extra sad, is that a remote controllable humanoid robot is already directly useful for work in hazardous environments, and we know they've got at least that… but Musk would rather it be about the AI.

19. fragmede ◴[] No.42141123[source]
People go and live in a house to get recorded 24/7, to be on tv, for far more asnine situations, for way less money.
20. rafaelmn ◴[] No.42141563[source]
>There is another emerging paradigm which is still small(er) scale but showing remarkable results. That's full multi-modal training with embodied agents (aka robots). 1x, Figure, Physical Intelligence, Tesla are all making rapid progress on functionality which is definitely beyond frontier LLMs because it is distinctly different.

Tesla is selling this view for almost a decade now in self-driving - how their car fleet feeding training data is going to make them leaders in the area. I don't find it convincing anymore

replies(2): >>42143183 #>>42144438 #
21. warkdarrior ◴[] No.42141572[source]
> Cool, but we already have robots doing this in 2d space (aka self driving cars) that struggle not to kill people. How is adding a third dimension going to help?

If we have robots that operate in 3D, they'll be able to kill you not only from behind or from the side, but also from above. So that's progress!

22. akomtu ◴[] No.42141776[source]
My understanding is that machine learning today is a lot like interpolation of examples in the dataset. The breakthrough of LLMs is due to the idea that interpolation in a 1024-dimensional space works much better than in a 2d space, if we naively interpolated English letters. All the modern transformers stuff is basically an advanced interpolation method that uses a large local neighborhood than just few nearest examples. It's like the Lanczos interpolation kernel, using a 1d analogy. Increasing the size of the kernel won't bring any gains, because the current kernel already nearly perfectly approximates an ideal interpolation (a full dataset DFT).

However interpolation isn't reasoning. If we want to understand the motion of planets, we would start with a dataset of (x, y, z, t) coordinates and try to derive the law of motion. Imagine if someone simply interpolated the dataset and presented the law of gravity as an array of million coefficients (aka weights)? Our minds have to work with a very small operating memory that can hardly fit 10 coefficients. This constraint forces us to develop intelligence that compacts the entire dataset into one small differential equation. Btw, English grammar is the differential equation of English in a lot of ways: it tells what the local rules are of valid trajectories of words that we call sentences.

23. knicholes ◴[] No.42141793{3}[source]
No real reason. I just made it up. But that's kind of my reasonable expectation of longevity of a machine like a robotic lawnmower and battery life.
24. mvdtnz ◴[] No.42142249[source]
> The best engineering minds have been focused on scaling transformer pre and post training for the last three years because they had good reason to believe it would work, and it has up until now.

Or because the people running companies who have fooled investors into believing it will work can afford to pay said engineers life-changing amounts of money.

replies(1): >>42149769 #
25. tick_tock_tick ◴[] No.42142802[source]
I ride in self driving cares basically once a week in SF (Waymo). It's always felt safer then a Uber and makes ways less risky maneuvers.
replies(1): >>42144609 #
26. BobaFloutist ◴[] No.42142935[source]
There already exists a robot that does the dishes, it's called a dishwasher.
replies(1): >>42145208 #
27. slashdave ◴[] No.42142983[source]
> Tesla are all making rapid progress on functionality

The lack of progress with self driving seems to indicate that Tesla has a serious problem with scaling. The investment in enormous compute resources is another red flag (if you run out of ideas, just use brute force). This points to a fundamental flaw in model architecture.

28. airstrike ◴[] No.42143148[source]
The gap from the virtual world of software and the brutally uncompromising nature of physical reality is wider than most people seem to accept.

It's almost like saying "we've already visited every place on Earth, surely Mars is just around the corner now"

29. torguyvg46787 ◴[] No.42143183[source]
The approaches are very limited, and it's essentially artificial artificial AI (and need a lot of human teleop demos).

At CoRL last week, the progress has noticeably plateaued. Roboticists notably were pessimistic that scaling laws will apply to robotics because of the embodiment issues.

30. soheil ◴[] No.42143184[source]
How is self-driving a 2D problem when you navigate a 3D world? (please do visit hilly San Francisco sometime) not to mention additional dimensions like depth, velocity vectors among others.
replies(1): >>42144401 #
31. hereme888 ◴[] No.42143654[source]
Are we humans so different? Why do you wear what you wear? People emulate their older siblings, and so learn behavior. LLMs can create new programs, after having initially learned similar examples from others. Likewise for AI media.
32. physicsguy ◴[] No.42144401{3}[source]
The visual input and sensory input to the self driving function are of the 3D world but the car is still constrained to move along a 2D topological surface, it’s not moving up and down other than by following the curvature of that
replies(1): >>42144706 #
33. Dunedan ◴[] No.42144438[source]
While one could argue whether Tesla or another company is the leader in this space, don't all promising self-driving approaches rely on this paradigm?
34. n_ary ◴[] No.42144609{3}[source]
Could be because Uber or Taxi is trying to make most trips and maximize day earning while Waymo do not have that rush and can take things slow…

Of course Waymo needs money but if the car made fewer trips compared to Uber/Taxi, it is not suffering the same consequences.

We need to consider human factor and the severe lacking of that in these robot/self driving/LLM and drawing parallels is not a direction I am feeling comfortable.

End of the day, Tesla also sold half baked self drive that killed people, we should not forget.

35. soheil ◴[] No.42144706{4}[source]
So based on your argument they actually operate in 1D since roads go in one direction and lanes and intersections are constrained to a predetermined curly line.
replies(1): >>42145327 #
36. ogogmad ◴[] No.42145208{3}[source]
You still need to load it.
37. physicsguy ◴[] No.42145327{5}[source]
The point is clearly that they don’t have a vertical axis of control, they can’t make the car fly up in the air unless they’re driving crazy taxi style
38. tim333 ◴[] No.42148862{3}[source]
Yeah, listening to him last week he seemed very unlike that https://www.youtube.com/watch?v=xXCBz_8hM9w&t=2324s
39. slashdave ◴[] No.42149743[source]
I hear what you are saying, but "innovation" is also often used to excuse some rather badly engineered concepts
40. slashdave ◴[] No.42149769[source]
The improvements in transformer implementation (e.g. "Flash Attention") have saved gobs of money on training and inference, I am guessing most likely more than the salary of those researchers.