Most active commenters
  • anotherd1p(7)
  • HarHarVeryFunny(6)
  • re-thc(5)
  • gtech1(5)
  • astrange(4)
  • (3)
  • skeeter2020(3)
  • fooker(3)
  • HDThoreaun(3)

←back to thread

765 points MindBreaker2605 | 107 comments | | HN request time: 1.163s | source | bottom
Show context
sebmellen ◴[] No.45897467[source]
Making LeCun report to Wang was the most boneheaded move imaginable. But… I suppose Zuckerberg knows what he wants, which is AI slopware and not truly groundbreaking foundation models.
replies(20): >>45897481 #>>45897498 #>>45897518 #>>45897885 #>>45897970 #>>45897978 #>>45898040 #>>45898053 #>>45898092 #>>45898108 #>>45898186 #>>45898539 #>>45898651 #>>45898727 #>>45899160 #>>45899375 #>>45900884 #>>45900885 #>>45901421 #>>45903451 #
1. xuancanh ◴[] No.45897885[source]
In industry research, someone in a chief position like LeCun should know how to balance long-term research with short-term projects. However, for whatever reason, he consistently shows hostility toward LLMs and engineering projects, even though Llama and PyTorch are two of the most influential projects from Meta AI. His attitude doesn’t really match what is expected from a Chief position at a product company like Facebook. When Llama 4 got criticized, he distanced himself from the project, stating that he only leads FAIR and that the project falls under a different organization. That kind of attitude doesn’t seem suitable for the face of AI at the company. It's not a surprise that Zuck tried to demote him.
replies(13): >>45897942 #>>45898142 #>>45898331 #>>45898661 #>>45898893 #>>45899157 #>>45899354 #>>45900094 #>>45900130 #>>45900230 #>>45901443 #>>45901631 #>>45902275 #
2. throwaw12 ◴[] No.45897942[source]
I would pose a question differently, under his leadership did Meta achieve good outcome?

If the answer is yes, then better to keep him, because he has already proved himself and you can win in the long-term. With Meta's pockets, you can always create a new department specifically for short-term projects.

If the answer is no, then nothing to discuss here.

replies(5): >>45897962 #>>45898150 #>>45898191 #>>45899393 #>>45900070 #
3. rw2 ◴[] No.45897962[source]
I believe that the fact that Chinese models are beating the crap of of Llama means it's a huge no.
replies(1): >>45898163 #
4. rapsey ◴[] No.45898142[source]
Yann was never a good fit for Meta.
replies(1): >>45899588 #
5. xuancanh ◴[] No.45898150[source]
Meta did exactly that, kept him but reduced his scope. Did the broader research community benefit from his research? Absolutely. But did Meta achieve a good outcome? Probably not.

If you follow LeCun on social media, you can see that the way FAIR’s results are assessed is very narrow-minded and still follows the academic mindset. He mentioned that his research is evaluated by: "Research evaluation is a difficult task because the product impact may occur years (sometimes decades) after the work. For that reason, evaluation must often rely on the collective opinion of the research community through proxies such as publications, citations, invited talks, awards, etc."

But as an industry researcher, he should know how his research fits with the company vision and be able to assess that easily. If the company's vision is to be the leader in AI, then as of now, he seems to have failed that objective, even though he has been at Meta for more than 10 years.

replies(2): >>45898292 #>>45898602 #
6. amelius ◴[] No.45898163{3}[source]
Why? The Chinese are very capable. Most DL papers have at least one Chinese name on it. That doesn't mean they are Chinese but it's telling.
replies(2): >>45898196 #>>45898335 #
7. UrineSqueegee ◴[] No.45898196{4}[source]
is an american model chinese because chinese people were in the team?
replies(2): >>45898372 #>>45898389 #
8. nsonha ◴[] No.45898292{3}[source]
Also he always sounds like "I know this will not work". Dude are you a researcher? You're supposed to experiment and follow the results. That's what separates you from oracles and freaking philosophers or whatever.
replies(4): >>45898333 #>>45898783 #>>45899067 #>>45899161 #
9. rob_c ◴[] No.45898331[source]
tbf, transformers from more of a developmental perspective are hugely wasteful. they're long-range stable sure, but the whole training process requires so much power/data compared to even slightly simpler model designs I can see why people are drawn to alternative complex model designs down-playing the reliance on pure attention.
10. yawnxyz ◴[] No.45898333{4}[source]
he probably predicted the asymptote everyone is approaching right now
replies(1): >>45899048 #
11. rob_c ◴[] No.45898335{4}[source]
most papers are also written in the same language, what's your point?
12. rat9988 ◴[] No.45898372{5}[source]
What are these chinese labs made of?
replies(2): >>45898419 #>>45898574 #
13. danielbln ◴[] No.45898389{5}[source]
There is no need for that tone here.
14. 4ggr0 ◴[] No.45898419{6}[source]
500 remote indian workers (/s)
15. ◴[] No.45898574{6}[source]
16. ◴[] No.45898602{3}[source]
17. blutoot ◴[] No.45898661[source]
These are the types that want academic freedom in a cut-throat industry setup and conversely never fit into academia because their profiles and growth ambitions far exceed what an academic research lab can afford (barring some marquee names). It's an unfortunate paradox.
replies(3): >>45898951 #>>45899099 #>>45902308 #
18. uoaei ◴[] No.45898783{4}[source]
He's speaking to the entire feedforward Transformer-based paradigm. He sees little point in continuing to try to squeeze more blood out of that stone and instead move on to more appropriate ways to model ontologies per se rather than the crude-for-what-we-use-them-for embedding-based methods that are popular today.

I really resonate with his view due to my background in physics and information theory. I for one welcome his new experimentation in other realms while so many still hack away at their LLMs in pursuit of SOTA benchmarks.

replies(1): >>45898917 #
19. sharmajai ◴[] No.45898893[source]
Product companies with deprioritized R&D wings are the first ones to die.
replies(4): >>45899020 #>>45899574 #>>45900235 #>>45901707 #
20. fhd2 ◴[] No.45898917{5}[source]
If the LLM hype doesn't cool down fast, we're probably looking at another AI winter. Appears to me like he's just trying to ensure he'll have funding for chasing the global maximum going forward.
replies(1): >>45899138 #
21. sigbottle ◴[] No.45898951[source]
Maybe it's time for Bell Labs 2?

I guess everyone is racing towards AGI in a few years or whatever so it's kind of impossible to cultivate that environment.

replies(13): >>45899122 #>>45899204 #>>45899373 #>>45899504 #>>45899663 #>>45899866 #>>45900147 #>>45900934 #>>45900995 #>>45901066 #>>45902188 #>>45902731 #>>45905111 #
22. ◴[] No.45899020[source]
23. brazukadev ◴[] No.45899048{5}[source]
So did I after trying llama/Meta AI
24. lukan ◴[] No.45899067{4}[source]
Philosophers are usually more aware of their not knowing than you seem to give them credit for. (And oracles are famously vague, too).
25. kamaal ◴[] No.45899099[source]
More importantly even if you do want it, and there are business situations that support your ambitions. You still have to do get into the managerial powerplay, which quite honestly takes a separate kind of skill set, time and effort. Which Im guessing the academia oriented people aren't willing to do.

Its pretty much dog eat dog at top management positions.

Its not exactly a space for free thinking timelines.

replies(2): >>45899200 #>>45900241 #
26. belter ◴[] No.45899122{3}[source]
> I guess everyone is racing towards AGI in a few years

A pipe dream sustaining the biggest stock market bubble in history. Smart investors are jumping to the next bubble already...Quantum...

replies(1): >>45899178 #
27. re-thc ◴[] No.45899138{6}[source]
> If the LLM hype doesn't cool down fast, we're probably looking at another AI winter.

Is the real bubble ignorance? Maybe you'll cool down but the rest of the world? There will just be more DeepSeek and more advances until the US loses its standing.

replies(1): >>45904742 #
28. Grimblewald ◴[] No.45899157[source]
LLM hostility was warrented. The overhype/downright charlartan nature of ai hype and marketing threatens another AI winter. It happened to cybernetics, it'll happen to us too. The finance folks will be fine, they'll move to the next big thing to overhype, it is the researchers who suffer the fall-out. I am considered anti LLM (transformers anyway) for this reason, i like the the architecture, it is cool amd rather capable at its problem set, which is a unique set, but, it isnt going to deliver any of what has been promised, any more than a plain DNN or a CNN will.
replies(1): >>45903202 #
29. teleforce ◴[] No.45899161{4}[source]
Do you know that all formally trained researchers have Doctor of Philosophy or PhD to their name? [1]

[1] Doctor of Philosophy:

https://en.wikipedia.org/wiki/Doctor_of_Philosophy

replies(1): >>45900115 #
30. re-thc ◴[] No.45899178{4}[source]
> A pipe dream sustaining the biggest stock market bubble in history

This is why we're losing innovation.

Look at electric cars, batteries, solar panels, rare earths and many more. Bubble or struggle for survival? Right, because if US has no AI the world will have no AI? That's the real bubble - being stuck in an ancient world view.

Meta's stock has already tanked for "over" investing in AI. Bubble, where?

replies(1): >>45899194 #
31. belter ◴[] No.45899194{5}[source]
2 Trillion dollars in Capex to get code generators with hallucinations, that run at a loss, and you ask where is the Bubble?
replies(1): >>45899264 #
32. ptero ◴[] No.45899200{3}[source]
It is not a free thinking paradise in academia either. Different groups fighting for hiring, promotions and influence exist there, too. And it tends to be more pronounced: it is much easier in industry to find a comparable job to escape a toxic environment, so a lot of problems in academia settings steam forever.

But the skill sets to avoid and survive personnel issues in academia is different from industry. My 2c.

33. ryukoposting ◴[] No.45899204{3}[source]
The Bell Labs we look back on was only the result of government intervention in the telecom monopoly. The 1956 consent decree forced Bell to license thousands of its patents, royalty free, to anyone who wanted to use them. Any patent not listed in the consent decree was to be licensed at "reasonable and nondiscriminatory rates."

The US government basically forced AT&T to use revenue from its monopoly to do fundamental research for the public good. Could the government do the same thing to our modern megacorps? Absolutely! Will it? I doubt it.

https://www.nytimes.com/1956/01/25/archives/att-settles-anti...

replies(1): >>45899620 #
34. re-thc ◴[] No.45899264{6}[source]
> 2 Trillion dollars in Capex to get code generators with hallucinations

You assume that's the only use of it.

And are people not using these code generators?

Is this an issue with a lost generation that forgot what Capex is? We've moved from Capex to Opex and now the notion is lost, is it? You can hire an army of software developers but can't build hardware.

Is it better when everyone buys DeepSeek or a non-US version? Well then you don't need to spend Capex but you won't have revenue either.

replies(1): >>45899333 #
35. littlestymaar ◴[] No.45899333{7}[source]
Deepseek somehow didn't need $2T to happen.
replies(3): >>45899379 #>>45900047 #>>45900088 #
36. HarHarVeryFunny ◴[] No.45899354[source]
Meta had a two prong AI approach - product-focused group working on LLMs, and blue-sky research (FAIR) working on alternate approaches, such as LeCun's JEPA.

It seems they've given up on the research and are now doubling down on LLMs.

37. HarHarVeryFunny ◴[] No.45899373{3}[source]
It seems DeepMind is the closest thing to a well funded blue-sky AI research group, even despite the merger with Google Brain and now more of a product focus.
38. re-thc ◴[] No.45899379{8}[source]
Because you know how much they spent.

And that $2T you're referring to includes infrastructure like energy, data centers, servers and many things. DeepSeek rents from others. Someone is paying.

39. HarHarVeryFunny ◴[] No.45899393[source]
LeCun was always part of FAIR, doing research, not part of the LLM/product group, who reported to someone else.
replies(1): >>45901322 #
40. StilesCrisis ◴[] No.45899574[source]
None of Meta's revenue has anything to do with AI at all. (Other than GenAI slop in old people's feeds.) Meta is in the strange position of investing very heavily in multiple fields where they have no successful product: VR, hardware devices, and now AI. Ad revenue funds it all.
replies(2): >>45900403 #>>45900635 #
41. runeblaze ◴[] No.45899588[source]
Agreed, I am surprised he is happy to stay this long. He would have been on paper a far better match at a place like pre-Gemini-era Google
42. aatd86 ◴[] No.45899620{4}[source]
Used to be a Google X. Not sure at what scale it was. But if any state/central bank was clever they would subsidize this. That's a better trickle down strategy. Until we get to agi and all new discoveries are autonomously led by AI that is :p
replies(1): >>45904500 #
43. sllabres ◴[] No.45899628{4}[source]
If you are (obviously) interested in the matter you might find one of the Bell Labs articles discussed on HN:

"Why Bell Labs Worked" [1]

"The Influence of Bell Labs" [2]

"Bringing back the golden days of Bell Labs" [3]

"Remembering Bell Labs as legendary idea factory prepares to leave N.J. home" [4] or

"Innovation and the Bell Labs Miracle" [5]

interesting too.

[1] https://news.ycombinator.com/item?id=43957010 [2] https://news.ycombinator.com/item?id=42275944 [3] https://news.ycombinator.com/item?id=32352584 [4] https://news.ycombinator.com/item?id=39077867 [5] https://news.ycombinator.com/item?id=3635489

replies(1): >>45901536 #
44. gtech1 ◴[] No.45899663{3}[source]
This sounds crazy. We don't even know/can't define what human intelligence is or how it works , but we're trying to replicate it with AGI ?
replies(5): >>45899845 #>>45899912 #>>45899913 #>>45899981 #>>45900436 #
45. Obscurity4340 ◴[] No.45899845{4}[source]
If an LLM can pass a bar exam, isn't that at least a decent proof of concept or working model?
replies(3): >>45900030 #>>45900196 #>>45900397 #
46. diego_sandoval ◴[] No.45899866{3}[source]
The fact that people invest on the architecture that keeps getting increasingly better results is a feature, not a bug.

If LLMs actually hit a plateau, then investment will flow towards other architectures.

replies(1): >>45900215 #
47. cantor_S_drug ◴[] No.45899912{4}[source]
Intelligence and human health can't be defined neatly. They are what we call suitcase words. If there exists a physiological tradeoff between medical research about whether to live till 500 years or to be able to lift 1000kg when a person is in youth, those are different dimensions / directions across we can make progress. Same happens for intelligence. I think we are on right track.
48. afthonos ◴[] No.45899913{4}[source]
Man, why did no one tell the people who invented bronze that they weren’t allowed to do it until they had a correct definition for metals and understood how they worked? I guess the person saying something can’t be done should stay out of the way of the people doing it.
replies(2): >>45899989 #>>45900146 #
49. anotherd1p ◴[] No.45899954{4}[source]
I always take a bird's eye kind of view on things like that, because however close I get, it always loops around to make no sense.

> is massively monopolistic and have unbounded discretionary research budget

that is the case for most megacorps. if you look at all the financial instruments.

modern monopolies are not equal to single corporation domination. modern monopolies are portfolios who do business using the same methods and strategies.

the problem is that private interests strive mostly for control, not money or progress. if they have to spend a lot of money to stay in control of (their (share of the)) segments, they will do that, which is why stuff like the current graph of investments of, by and for AI companies and the industries works.

A modern equivalent and "breadth" of a Bell Labs (et. al) kind of R&D speed could not be controlled and would 100% result in actual Artificial Intelligence vs all those white labelababbebel (sry) AI toys we get now.

Post WW I and II "business psychology" have build a culture that cannot thrive in a free world (free as in undisturbed and left to all devices available) for a variety of reasons, but mostly because of elements with a medieval/dark-age kind of aggressive tendency to come to power and maintain it that way.

In other words: not having a Bell Labs kind of setup anymore ensures that the variety of approaches taken on large scales aka industry-wide or systemic, remains narrow enough.

50. anotherd1p ◴[] No.45899981{4}[source]
stretching the infinite game is exactly that, yes, "This is the way"
51. gtech1 ◴[] No.45899989{5}[source]
I'm not sure what 'inventing bronze' is supposed to be. 'Inventing' AGI is pretty much equivalent to creating new life, from scratch. And we don't have an idea on how to do that either, or how life came to be.
52. anotherd1p ◴[] No.45900030{5}[source]
I love this application of AI the most but as many have stated elsewhere: mathematical precision in law won't work, or rather, won't be tolerated.
53. anotherd1p ◴[] No.45900047{8}[source]
all that led up to Deepseek needed more. don't forget where it all comes from.
54. anotherd1p ◴[] No.45900070[source]
then we should ask: will Meta come close enough to the fulfillment of the promises made, or will it keep achieving good enough outcomes?
55. matt3D ◴[] No.45900088{8}[source]
I think the argument can be made that Deepseek is a state sponsored needle looking to pop another states bubble.

If Deepseek is free it undermines the value of LLMs, so the value of these US companies is mainly speculation/FOMO over AGI.

replies(1): >>45900589 #
56. _the_inflator ◴[] No.45900094[source]
I totally agree. He appeared to act against his employer and actively undermined Meta's effort to attract talent by his behavior visible on X.

And I stopped reading him, since he - in my opinion - trashed on autopilot everything 99% did - and these 99% were already beyond the two standard deviation of greatness.

It is even more highly problematic if you have absolutely no results eg products to back your claims.

57. anotherd1p ◴[] No.45900115{5}[source]
If academia is in question, then so are their titles. When I see "PhD", I read "we decided that he was at least good enough for the cause" PhD, or PhD (he fulfilled the criteria).
58. hbarka ◴[] No.45900130[source]
LeCun truly believes the future is in world models. He’s not alone. Good for him to now be in the position he’s always wanted and hopefully prove out what he constantly talks about.
replies(1): >>45901728 #
59. skeeter2020 ◴[] No.45900146{5}[source]
>> I guess the person saying something can’t be done should stay out of the way of the people doing it.

I'll happily step out of the way once someone simply tells me what it is you're trying to accomplish. Until you can actually define it, you can't do "it".

replies(2): >>45900270 #>>45900419 #
60. blueboo ◴[] No.45900147{3}[source]
We call it “legacy DeepMind”
61. skeeter2020 ◴[] No.45900196{5}[source]
Or does this just prove lawyers are artificially intelligent?

yes, a glib response, but think about it: we define an intelligence test for humans, which by definition is an artificial construct. If we then get a computer to do well on the test we haven't proved it's on par with human intelligence, just that both meet some of the markers that the test makers are using as rough proxies for human intelligence. Maybe this helps signal or judge if AI is a useful tool for specific problems, but it doesn't mean AGI

62. esafak ◴[] No.45900215{4}[source]
At which point companies that had the foresight to investigate those architectures earlier on will have the lead.
63. nailer ◴[] No.45900230[source]
Lecun has also consistently tried to redefine open source away from the open source definition.
64. skeeter2020 ◴[] No.45900235[source]
Hasn't happened to Google yet
replies(1): >>45900798 #
65. anotherd1p ◴[] No.45900241{3}[source]
> Its not exactly a space for free thinking timelines.

Same goes for academia. People's visions compete for other people's financial budgets, time and other resources. Some dogs get to eat, study, train at the frontier and with top tools in top environments while the others hope to find a good enough shelter.

66. gtech1 ◴[] No.45900270{6}[source]
no bro, others have done 'it' without even knowing what they were doing!
67. staticman2 ◴[] No.45900397{5}[source]
I don't think the bar exam is scientifically designed to measure intelligence so that was an odd example. Citing the bar exam is like saying it passes the "Game of thrones trivia" exam so it must be intelligent.

As for IQ tests and the like, to the extent they are "scientific" they are designed based on empirical observations of humans. It is not designed to measure the intelligence of a statistical system containing a compressed version of the internet.

68. nxor ◴[] No.45900403{3}[source]
Underrated comment
69. afthonos ◴[] No.45900419{6}[source]
The big tech companies are trying to make machines that replace all human labor. They call it artificial intelligence. Feel free to argue about definitions.
replies(1): >>45900868 #
70. meindnoch ◴[] No.45900436{4}[source]
Hi there! :) Just wanted to gently flag that one of the terms (beginning with the letter "r") in your comment isn't really aligned with the kind of inclusive language we try to encourage across the community. Totally understand it was likely unintentional - happens to all of us! Going forward, it'd be great to keep things phrased in a way that ensures everyone feels welcome and respected. Thanks so much for taking the time to share your thoughts here!
replies(1): >>45900822 #
71. re-thc ◴[] No.45900589{9}[source]
> the argument can be made that Deepseek is a state sponsored needle looking to pop another states bubble

Who says they don't make money? Same with open source software that offer a hosted version.

> If Deepseek is free it undermines the value of LLMs, so the value of these US companies is mainly speculation/FOMO over AGI

Freemium, open source and other models all exist. Does it undermine the value of e.g. Salesforce?

72. jpadkins ◴[] No.45900635{3}[source]
LLMs help ads efficiency a lot. policy labels, targeting, adaptive creatives, landing page evals, etc.
73. anshumankmr ◴[] No.45900798{3}[source]
Has Google depriortized R&D?
74. gtech1 ◴[] No.45900822{5}[source]
My apologies, I have edited my comment.
75. gtech1 ◴[] No.45900868{7}[source]
No no, let's define labor (labour?) first.
replies(1): >>45905921 #
76. ambicapter ◴[] No.45900934{3}[source]
Why would Bell Labs be a good fit? It was famous for embedding engineers with the scientists to direct research in a more results-oriented fashion.
77. musebox35 ◴[] No.45900995{3}[source]
Google Deepmind is the closest lab to that idea because Google is the only entity that is big enough to get close to the scale of AT&T. I was skeptical that the Deepmind and Google Brain merge would be successful but it seems to have worked surprisingly well. They are killing it with LLMs and image editing models. They are also backing the fastest growing cloud business in the world and collecting Nobel prizes along the way.
78. ximeng ◴[] No.45901066{3}[source]
https://www.startuphub.ai/ai-news/ai-research/2025/sam-altma...

Like the new spin out Episteme from OpenAI?

79. rockinghigh ◴[] No.45901322{3}[source]
Wasn't the original LLaMA developed by FAIR Paris?
replies(1): >>45902034 #
80. mi_lk ◴[] No.45901443[source]
This is the right take. He is obviously a pioneer and much more knowledgeable than Wang in the field, but if you don't have the product mind to serve company's business interest in short term and long term capacity anymore, you may as well stay in academia and be your own research director, let alone a chief executive in one of the largest public companies
81. mysfi ◴[] No.45901536{5}[source]
I became interested in the matter reading this thread and vaguely remember reading a couple of the articles. Saved them all in NotebookLM to get an audio overview and to read later. Thanks!
82. Nimitz14 ◴[] No.45901631[source]
Yann was in charge of FAIR which has nothing to do with llama4 or the product focussed AI orgs. In general your comment is filled with misrepresentations. Sad.
replies(1): >>45903247 #
83. astrange ◴[] No.45901707[source]
Apple doesn't have an "R&D wing". It's a bad idea to split your company into the cool part and the boring part.
replies(1): >>45903111 #
84. astrange ◴[] No.45901728[source]
He seems stuck in the GOFAI development philosophy where they just decide humans have something called a "world model" because they said so, and then decide that if they just develop some random thing and call it a "world model" it'll create intelligence because it has the same name as the thing they made up.

And of course it doesn't work. Humans don't have world models. There's no such thing as a world model!

replies(1): >>45905064 #
85. HarHarVeryFunny ◴[] No.45902034{4}[source]
I hadn't heard that, but he was heavily involved in a cancelled project called Galactica that was an LLM for scientific knowledge.
replies(1): >>45903099 #
86. meekaaku ◴[] No.45902188{3}[source]
I am of the opinion that splitting AT&T and hence Bell Labs was a net negative for America and rest of the world.

We are yet to create lab as foundational as Bell Labs.

87. whiplash451 ◴[] No.45902275[source]
It's very hard (and almost irreconcilable) to lead both Applied Research -- that optimizes for product/business outcomes -- and Fundamental Research -- that optimizes for novel ideas -- especially at the scale of Meta.

LeCun had chosen to focus on the latter. He can't be blamed for not having taken the second hat.

replies(1): >>45903160 #
88. whiplash451 ◴[] No.45902308[source]
Meta has the financial oomph to run multiple Bell Labs within its organization.

Why they decided not to do that is kind of a puzzle.

89. red2awn ◴[] No.45902731{3}[source]
I'd argue SSI and Thinking Machines Lab seem to that environment you are thinking about. Industry labs that focuses on research without immediate product requirement.
90. fooker ◴[] No.45903099{5}[source]
Yeah that stuff generated embarrassingly wrong scientific 'facts' and citations.

That kind of hallucination is somewhat acceptable for something marketed as a chatbot, less so for an assistant helping you with scientific knowledge and research.

replies(1): >>45905233 #
91. fooker ◴[] No.45903111{3}[source]
Isn't that why Siri is worse today than it was thirteen years ago?
replies(2): >>45903194 #>>45904132 #
92. HDThoreaun ◴[] No.45903160[source]
Yes he can. If he wanted to focus on fundamental research he shouldn’t have accepted a leadership position at a product company. He knew going in that releasing products was part of his job and largely blew it.
93. mi_lk ◴[] No.45903194{4}[source]
And apparently that doesn't stop people from buying their products
replies(1): >>45903559 #
94. HDThoreaun ◴[] No.45903202[source]
Meta is in last place among the big tech companies making an AI push because of lecun’s llm hostility. Refusing to properly invest in the biggest product breakthrough this century was not even a little bit warranted. He had more than enough resources available to do the research he wanted and create a fantastic open source llm.
replies(1): >>45905388 #
95. HDThoreaun ◴[] No.45903247[source]
FAIR having shit for products is the whole reason he is being demoted/fired. Yes, he had nothing to do with applied research, that was the problem.
96. fooker ◴[] No.45903559{5}[source]
It doesn't.

Apple makes the best hardware, period.

It makes sense that people are willing to overlook subpar software for top notch hardware.

97. astrange ◴[] No.45904132{4}[source]
It's better in ways you don't think about.

(It works offline, it works in other languages, the TTS is much better.)

98. williamDafoe ◴[] No.45904500{5}[source]
Google X is a complete failure. Maybe they had fei-fei on staff for a short while but most of her work was done elsewhere.
replies(1): >>45905604 #
99. uoaei ◴[] No.45904742{7}[source]
How is it a foregone conclusion that squeezing the stone will continue to produce blood?
100. HarHarVeryFunny ◴[] No.45905064{3}[source]
I don't think the focus is really on world models, rather than on animal intelligence based around predicting the real world, but to predict it you need to model it in some sense.
replies(1): >>45905460 #
101. stocksinsmocks ◴[] No.45905111{3}[source]
I thought that was Google. Regulators pretend not to notice their monopoly, they probably get large government contracts for social engineering and surveillance laundered through advertising, and the “don’t be evil” part is they make some open source contributions
102. baobabKoodaa ◴[] No.45905233{6}[source]
I thought it was weird at the time how much hate Galactica got for its hallucinations compared to hallucinations of competing models. I get your point and it partially explains things. But it's not a fully satisfying explanation.
103. Grimblewald ◴[] No.45905388{3}[source]
Meta has made some fantastic llm's publically avliable many of which continue to outperform all but the qwen series in real world applications.

LLMs cannot do any of the major claims made for them, so competing at the current frontier is a massive resource waste.

Right now a locally running 8b model with large context window (10k tokens+) beat google/openAI models easily on any task you like.

why would anyone then pay for something that is possible to run on consumer hardware with higher token/second throughput and better performance? What exactly have the billions invested given google/oai in return? Nothing more than an existensial crisis I'd say.

Companies aren't trying to force AI costs into their subscription models in dishonest ways because they've got a winning product.

104. astrange ◴[] No.45905460{4}[source]
IMO the issue is that animals can't have a specific "world model" system, because if you create a model ahead of time you will mostly waste energy because most of the model is not used.

And animals' main concern is energy conservation, so they must be doing something else.

replies(1): >>45906280 #
105. aatd86 ◴[] No.45905604{6}[source]
Didn't the current LLMs stem from this...? Or it might be Google Brain instead. For Google X, there is Waymo? I know a lot of stuff didn't pan out. This is expected. These were 'moonshots'.

But the principle is there. I think that when a company sits on a load of cash, that's what they should do. Either that or become a kind of alternative investments allocator. These are risky bets. But they should be incentivized to take those risks. From a fiscal policy standpoint for instance. Well it probably is the case already via lower taxation of capital gains and so on. But there should probably exist a more streamlined framework to make sure incentives are aligned.

And/or assigned government projects? Besides implementing their Cloud infrastructure that is...

106. CamperBob2 ◴[] No.45905921{8}[source]
Whatever you're doing for money that you wouldn't do if you didn't need money.
107. HarHarVeryFunny ◴[] No.45906280{5}[source]
There are many factors playing into "survival of the fittest", and energy conservation is only one. Animals build mental models to predict the world because this superpower of seeing into the future is critical to survival - predict where the water is in a drought, where the food is, and how to catch it, etc, etc.

The animal learns as it encounters learning signals - prediction failure - which is the only way to do it. Of course you need to learn/remember something before you can use that in the future, so in that sense it's "ahead of time", but the reason it's done that way because evolution has found that learning patterns will ultimately prove beneficial.