Skip to content(if available)orjump to list(if available)

Yann LeCun to depart Meta and launch AI startup focused on 'world models'

sebmellen

Making LeCun report to Wang was the most boneheaded move imaginable. But… I suppose Zuckerberg knows what he wants, which is AI slopware and not truly groundbreaking foundation models.

xuancanh

In industry research, someone in a chief position like LeCun should know how to balance long-term research with short-term projects. However, for whatever reason, he consistently shows hostility toward LLMs and engineering projects, even though Llama and PyTorch are two of the most influential projects from Meta AI. His attitude doesn’t really match what is expected from a Chief position at a product company like Facebook. When Llama 4 got criticized, he distanced himself from the project, stating that he only leads FAIR and that the project falls under a different organization. That kind of attitude doesn’t seem suitable for the face of AI at the company. It's not a surprise that Zuck tried to demote him.

blutoot

These are the types that want academic freedom in a cut-throat industry setup and conversely never fit into academia because their profiles and growth ambitions far exceed what an academic research lab can afford (barring some marquee names). It's an unfortunate paradox.

sigbottle

Maybe it's time for Bell Labs 2?

I guess everyone is racing towards AGI in a few years or whatever so it's kind of impossible to cultivate that environment.

kamaal

More importantly even if you do want it, and there are business situations that support your ambitions. You still have to do get into the managerial powerplay, which quite honestly takes a separate kind of skill set, time and effort. Which Im guessing the academia oriented people aren't willing to do.

Its pretty much dog eat dog at top management positions.

Its not exactly a space for free thinking timelines.

throwaw12

I would pose a question differently, under his leadership did Meta achieve good outcome?

If the answer is yes, then better to keep him, because he has already proved himself and you can win in the long-term. With Meta's pockets, you can always create a new department specifically for short-term projects.

If the answer is no, then nothing to discuss here.

xuancanh

Meta did exactly that, kept him but reduced his scope. Did the broader research community benefit from his research? Absolutely. But did Meta achieve a good outcome? Probably not.

If you follow LeCun on social media, you can see that the way FAIR’s results are assessed is very narrow-minded and still follows the academic mindset. He mentioned that his research is evaluated by: "Research evaluation is a difficult task because the product impact may occur years (sometimes decades) after the work. For that reason, evaluation must often rely on the collective opinion of the research community through proxies such as publications, citations, invited talks, awards, etc."

But as an industry researcher, he should know how his research fits with the company vision and be able to assess that easily. If the company's vision is to be the leader in AI, then as of now, he seems to have failed that objective, even though he has been at Meta for more than 10 years.

HarHarVeryFunny

LeCun was always part of FAIR, doing research, not part of the LLM/product group, who reported to someone else.

rw2

I believe that the fact that Chinese models are beating the crap of of Llama means it's a huge no.

null

[deleted]

HarHarVeryFunny

Meta had a two prong AI approach - product-focused group working on LLMs, and blue-sky research (FAIR) working on alternate approaches, such as LeCun's JEPA.

It seems they've given up on the research and are now doubling down on LLMs.

sharmajai

Product companies with deprioritized R&D wings are the first ones to die.

StilesCrisis

None of Meta's revenue has anything to do with AI at all. (Other than GenAI slop in old people's feeds.) Meta is in the strange position of investing very heavily in multiple fields where they have no successful product: VR, hardware devices, and now AI. Ad revenue funds it all.

null

[deleted]

Grimblewald

LLM hostility was warrented. The overhype/downright charlartan nature of ai hype and marketing threatens another AI winter. It happened to cybernetics, it'll happen to us too. The finance folks will be fine, they'll move to the next big thing to overhype, it is the researchers who suffer the fall-out. I am considered anti LLM (transformers anyway) for this reason, i like the the architecture, it is cool amd rather capable at its problem set, which is a unique set, but, it isnt going to deliver any of what has been promised, any more than a plain DNN or a CNN will.

rapsey

Yann was never a good fit for Meta.

runeblaze

Agreed, I am surprised he is happy to stay this long. He would have been on paper a far better match at a place like pre-Gemini-era Google

rob_c

tbf, transformers from more of a developmental perspective are hugely wasteful. they're long-range stable sure, but the whole training process requires so much power/data compared to even slightly simpler model designs I can see why people are drawn to alternative complex model designs down-playing the reliance on pure attention.

gnaman

He is also not very interested in LLMs, and that seems to be Zuck's top priority.

tinco

Yeah I think LeCun is underestimating the impact that LLM's and Diffusion models are going to have, even considering the huge impact they're already having. That's no problem as I'm sure whatever LeCun is working on is going to be amazing as well, but an enterprise like Facebook can't have their top researcher work on risky things when there's surefire paths to success still available.

jll29

I politely disagree - it is exactly an industry researcher's purpose to do the risky things that may not work, simply because the rest of the corporation cannot take such risks but must walk on more well-trodden paths.

Corporate R&D teams are there to absorb risk, innovate, disrupt, create new fields, not for doing small incremental improvements. "If we know it works, it's not research." (Albert Einstein)

I also agree with LeCun that LLMs in their current form - are a dead end. Note that this does not mean that I think we have already exploited LLMs to the limit, we are still at the beginning. We also need to create an ecosystem in which they can operate well: for instance, to combine LLMs with Web agents better we need a scalable "C2B2C" (customer delegated to business to business) micropayment infrastructure, because as these systems have already begun talking to each other, in the longer run nobody would offer their APIs for free.

I work on spatial/geographic models, inter alia, which by coincident is one of the direction mentioned in the LeCun article. I do not know what his reasoning is, but mine was/is: LMs are language models, and should (only) be used as such. We need other models - in particular a knowledge model (KM/KB) to cleanly separate knowledge from text generation - it looks to me right now that only that will solve hallucination.

fxtentacle

LLMs and Diffusion solve a completely different problem than world models.

If you want to predict future text, you use an LLM. If you want to predict future frames in a video, you go with Diffusion. But what both of them lack is object permanence. If a car isn't visible in the input frame, it won't be visible in the output. But in the real world, there are A LOT of things that are invisible (image) or not mentioned but only implied (text) that still strongly affect the future. Every kid knows that when you roll a marble behind your hand, it'll come out on the other side. But LLMs and Diffusion models routinely fail to predict that, as for them the object disappears when it stops being visible.

Based on what I heard from others, world models are considered the missing ingredient for useful robots and self-driving cars. If that's halfway accurate, it would make sense to pour A LOT of money into world models, because they will unlock high-value products.

qmr

> but an enterprise like Facebook can't have their top researcher work on risky things when there's surefire paths to success still available.

Bell Labs

netdevphoenix

>the huge impact they're already having

In the software development world yes, outside of that, virtually none. Yes, you can transcribe a video call in Office, yes, but that's not ground breaking. I dare you to list 10 impacts on different fields, excluding tech and including at least half blue collar fields and at least half white collar fields , at different levels from the lowest to the highest in the company hierarchy, that LLM/Diffusion models are having. Impact here specifically means a significant reduction of costs or a significant increase of revenue. Go on

OJFord

He's quoted in OP as calling them 'useful but fundamentally limited'; that seems correct, and not at all like he's denying their utility.

hodgehog11

Unless I've missed a few updates, much of the JEPA stuff didn't really bear a lot of fruit in the end.

sebmellen

While I agree with your point, “Superintelligence” is a far cry from what Meta will end up delivering with Wang in charge. I suppose that, at the end of the day, it’s all marketing. What else should we expect from an ads company :?

StopDisinfo910

Hard to tell.

The last time LeCun disagreed with the AI mainstream was when he kept working on neural net when everyone thought it was a dead end. He might be entirely right in his LLM scepticism. It's hardly a surefire path. He didn't prevent Meta from working on LLM anyway.

The issue is more than his position is not compatible with short term investors expectations and that's fatal in a company like Meta at the position LeCun occupies.

gdiamos

The role of basic research is to get off the beaten path.

LLMs aren’t basic research when they have 1 billion users

FartyMcFarter

> But… I suppose Zuckerberg knows what he wants, which is AI slopware and not truly groundbreaking foundation models.

When did they make groundbreaking foundation models though? DeepMind and OpenAI have done plenty of revolutionary things, what did Meta AI do while being led by LeCun?

ergocoder

LeCun is great and smart, of course. But he had his chance. It didn't go that well. Now Zuck wants somebody else to try.

Messi is the best footballer of our era. It doesn't mean he would play well in any team.

jamesblonde

I don't think Messi could do it on a wet night in Stoke. Ronaldo could, though.

/s

ekjhgkejhgk

> slopware

Damn did you just invent that? That's really catchy.

torginus

What does Meta even want with AI?

I suppose they could solve superintelligence and cure cancer and build fusion reactors with it, but that's 100% outside their comfort zone - if they manage to build synthethic conversation partners and synthethic content generators as good or better than the real thing the value of having every other human on the planet registered to one of their social network goes to zero.

Which is impossible anyway - I facebook to maintain real human connections and keep up with people who I care about, not to consume infinite content.

zamadatix

At 1.6T market cap it's very hard to 10x or greater the company anymore doing what's in their comfort zone and they've got a lot of money to play with to find easier to grow opportunities. If Zuckerberg was convinced he could do that by selling toothpicks they'd have a go at the toothpick business. They went after the "metaverse" first, then AI. Both are just very fast growth options which happen to be tech focused because that's the only way you generate new comparable value as a company (unless you're sitting on a lot of state owned oil) in the current markets.

bbarnett

You missed an opportunity to use paperclips instead of toothpicks, as your example.

Would be very inline with the AI angle.

breppp

they are out for your clicks and attention minutes

if OpenAI can build a "social" network of completely generated content, that can kill Meta. Even today I venture to guess that most of the engagements in their platforms is not driven by real friends, so an AI driven platform won't be too different, or it might make content generation be so easy as to make your friends engage again.

Apart from it the ludicrous vision of the metaverse seems much more plausible with highly realistic world models

drexlspivey

How do LLMs help with clicks and attention minutes? Why do they spend $100+B a year in AI capex, more than Google and Microsoft that actually rent AI compute to clients? What are they going to do with all that compute? It’s all so confusing

pandemic_region

Sad to hear it has come to attention minutes, used to be seconds.

ACCount37

That was obviously him getting sidelined. And it's easy to see why.

LLMs get results. None of the Yann LeCun's pet projects do. He had ample time to prove that his approach is promising, and he didn't.

chaoz_

I agree. I never understood LeCun's statement that we need to pivot toward the visual aspects of things because the bitrate of text is low while visual input through the eye is high.

Text and languages contain structured information and encode a lot of real-world complexity (or it's "modelling" that).

Not saying we won't pivot to visual data or world simulations, but he was clearly not the type of person to compete with other LLM research labs, nor did he propose any alternative that could be used to create something interesting for end-users.

ACCount37

If LeCun's research has made Meta a powerhouse of video generation or general purpose robotics - the two promising directions that benefit from working with visual I/O and world modeling as LeCun sees it - it could have been a justified detour.

But that sure didn't happen.

camillomiller

LLMs get results is quite the bold statement. If they get results, they should be getting adopted, and they should be making money. This is all built on hazy promises. If you had marketable results, you wouldn't have to hide 20+ billion dollars of debt financing into an obscure SPV. LLMs are the most baffling piece of tech. They are incredible, and yet marred by their non-deterministic hallucinatory nature, and bound to fail in adoption unless you convince everyone that they don't need precision and accuracy, but they can do their business at 75% quality, just with less human overhead. It's quite the thing to convince people of, and that's why it needs the spend it's needing. A lot of we-need-to-stay-in-the-loop CEOs and bigwigs got infatuated with the idea, and most probably they just had their companies get addicted to the tech equivalent of crack cocaine. A reckoning is coming.

ACCount37

LLMs get results, yes. They are getting adopted, and they are making money.

Frontier models are all profitable. Inference is sold with a damn good margin, and the amounts of inference AI companies sell keeps rising. This necessitates putting more and more money into infrastructure. AI R&D is extremely expensive too, and this necessitates even more spending.

A mistake I see people make over and over again is keeping track of the spending but overlooking the revenue altogether. Which sure is weird: you don't get from $0B in revenue to $12B in revenue in a few years by not having a product anyone wants to buy.

And I find all the talk of "non-deterministic hallucinatory nature" to be overrated. Because humans suffer from all of that too, just less severely. On top of a number of other issues current AIs don't suffer from.

Nonetheless, we use human labor for things. All AI has to do is provide a "good enough" alternative, and it often does.

miohtama

OpenAI and Anthropic are making north of 4B/year revenue so some companies have figured out the money making part. ChatGPT has some 800M users according to some calculations. Whether it's enough money today, enough money tomorrow, is of course a question but there is a lot of money. Users would not use them in a scale if they do not solve their problems.

null

[deleted]

dude250711

There is someone else at Facebook who's pet projects do not get results...

ergocoder

If you hire a house cleaner to clean your house, and the cleaner didn't do well, would you eject yourself out of the house? You would not. You would change to a new cleaner.

ACCount37

Sure, but that "someone else" is the man writing the checks. If the roles were reversed, he'd be the one being fired now.

jb1991

Who are you referring to?

motbus3

Zuck hired John Carmack and got nothing of it On the other hand, it was only lecunn avoiding meta to go 100p evil creepy mode too

lofaszvanitt

And Carmack complained about the bureaucracy hell that is Facebook.

Tepix

Carmack laid the foundation for the all-in-one VR headsets.

blitzar

Hopefully one day, in a galaxy far far away, someone builds something on those foundations.

llamasushi

LeCun, who's been saying LLMs are a dead end for years, is finally putting his money where his mouth is. Watch for LeCun to raise an absolutely massive VC round.

conradfr

So not his money ;)

qwertox

But his responsability.

zwnow

What is responsibility if you can afford good lawyers?

numpy-thagoras

Good. The world model is absolutely the right play in my opinion.

AI Agents like LLMs make great use of pre-computed information. Providing a comprehensive but efficient world model (one where more detail is available wherever one is paying more attention given a specific task) will definitely eke out new autonomous agents.

Swarms of these, acting in concert or with some hive mind, could be how we get to AGI.

I wish I could help, world models are something I am very passionate about.

sebmellen

Can you explain this “world model” concept to me? How do you actually interface with a model like this?

curiouscube

One theory of how humans work is the so called predictive coding approach. Basically the theory assumes that human brains work similar to a kalman filter, that is, we have an internal model of the world that does a prediction of the world and then checks if the prediction is congruent with the observed changes in reality. Learning then comes down to minimizing the error between this internal model and the actual observations, this is sometimes called the free energy principle. Specifically when researchers are talking about world models they tend to refer to internal models that model the actual external world, that is they can predict what happens next based on input streams like vision.

Why is this idea of a world model helpful? Because it allows multiple interesting things, like predict what happens next, model counterfactuals (what would happen if I do X or don't do X) and many other things that tend to be needed for actual principled reasoning.

HarHarVeryFunny

Learning from the real world, including how it responds to your own actions, is the only way to achieve real-world competency, intelligence, reasoning and creativity, including going beyond human intelligence.

The capabilities of LLMs are limited by what's in their training data. You can use all the tricks in the book to squeeze the most out of that - RL, synthetic data, agentic loops, tools, etc, but at the end of the day their core intelligence and understanding is limited by that data and their auto-regressive training. They are built for mimicry, not creativity and intelligence.

sgt

So... that seems like possible path towards AGI. Doesn't it?

natch

He is one of these people who think that humans have a direct experience of reality not mediated by as Alan Kay put it three pounds of oatmeal. So he thinks a language model can not be a world model. Despite our own contact with reality being mediated through a myriad of filters and fun house mirror distortions. Our vision transposes left and right and delivers images to our nerves upside down, for gawd’s sake. He imagines none of that is the case and that if only he can build computers more like us then they will be in direct contact with the world and then he can (he thinks) make a model that is better at understanding the world

BoxOfRain

Isn't this idea demonstrably false due to the existence of various sensory disorders too?

I have a disorder characterised by the brain failing to filter own its own sensory noise, my vision is full of analogue TV-like distortion and other artefacts. Sometimes when it's bad I can see my brain constructing an image in real time rather than this perception happening instantaneously, particularly when I'm out walking. A deer becomes a bundle of sticks becomes a muddy pile of rocks (what it actually is) for example over the space of seconds. This to me is pretty strong evidence we do not experience reality directly, and instead construct our perceptions predictively from whatever is to hand.

Gooblebrai

> humans have a direct experience of reality not mediated by as Alan Kay put it three pounds of oatmeal

Is he advocating for philosophical idealism of the mind or does he has an alternate physicalist theory?

trhway

That way he may get a very good lizard. Getting Einstein though takes layers of abstraction.

My thinking is that such world models should be integrated with LLM like the lower levels of perception are integrated with higher brain function.

Hendrikto

Great strawman.

koolala

Ouija board would work for text.

kittikitti

I think moving on from LLM's is slightly arrogant. It might just be my understanding, but I feel like there is still much to be discovered. I was hoping for development in spiking neural networks but it might be skipped over. Perhaps I need to dive even deeper and the research is truly well understood and "done" but I can't help but constantly learn something new about language models and neural networks.

Best of luck to LeCun. I hope by World Model's he means embodied AI or humanoid robots. We'll have to wait and see.

monkeydust

He needs a patient investor and realized Zuck is not that. As someone who delivers product and works a lot with researchers I get the constant tension that might exist with competing priorities. Very curious to see how he does, imho the outcome will be either of the extremes - one of the fastest growing companies by valuation ever or a total flop. Either way this move might advance us to whatever end state we are heading towards with AI.

qwertox

It would have been just as interesting to read that he moved over to Google, where the real brains and resources are located at.

Meta is now just competing against giants like OpenAI, Anthropic and Google, plus all the new Chinese companies; I see no real chance for them to offer a popular chat model, but rather to market their AI as a bundled product for companies which want to advertise, where the images and videos will be automatically generated by Meta.

sidcool

I think it was a plan by Mark to move LeCun out of Meta. And they cannot fire him without bad PR, so they got Wang to lead him. It was only a matter of time before LeCun moved out.

theanonymousone

Isn't putting Wang as leading him a worse PR compared to just letting him go?

fxtentacle

Working under LeCun but outside of Zuckerberg's sphere of influence sure sounds like a dream job.

fastball

Really? From where I'm standing LeCun is a pompous researcher who had early success in his career, and has been capitalizing on that ever since. Have you read any of his papers from the last 20 years? 90% of his citations are to his own previous papers. From there, he missed the boat on LLMs and is now pretending everyone else is wrong so that he can feel better about it.

MrScruff

His research group have introduced some pretty impactful research and open source models.

https://ai.meta.com/research/

fastball

For the same reason I don't attribute those successes to Zuckerberg I don't attribute them to LeCun either.

bn-l

It’s probably better for the world that LeCun is not at Meta. I mean if his direction is the likeliest approach to AGI meta is the last place where you want it.

energy123

It's better that he's not working on LLMs. There's enough people working on it already.

null

[deleted]

Jackson__

From the outside, it always looked like they gave LeCun just barely enough compute for small scale experiments. They'd publish a promising new paper, show it works at a small scale, then not use it at all for any of their large AI runs.

I would have loved to see a VLM utilizing JEPA for example, but it simply never happened.

sakex

I'd be surprised if they didn't scale it up.

tucnak

The obvious explanation is they have scaled it up, but it turned out to be total shite, like most new architectures.

albertzeyer

I wonder, what LeCun wants to do is more fundamental research, i.e. where the timeline to being useful is much longer, maybe 5-10 years at least, and also much more uncertain.

How does this fit together with a startup? Would investors happily invest into this knowing not to expect anything in return for at least the next 5-10 years?

Hendrikto

> Would investors happily invest into this knowing not to expect anything in return for at least the next 5-10 years?

Oh, you mean like OpenAI, Anthropic, Gemini, and xAI? None of them are profitable.

Amadiro

That's a quite different thing, OpenAI has billions of USD/year cash flow, and when you have that there's many many potential way to achieve profitability on different time horizons. It's not a situation of chance but a situation of choice.

Anyway, how much that matters for an investor is hard to form a clear answer to - investors are after all not directly looking for profitability as such, but for valuation growth. The two are linked but not the same -- any investor in OpenAI today probably also places themselves into a game of chance, betting on OpenAI making more breakthroughs and increasing the cash flow even more -- not just becoming profitable at the same rate of cash flow. So there's still some of the same risk baked into this investment.

But with a new startup like LeCun's is going to be, it's 100% on the risk side and 0% on the optionality side. The path to profitability for a startup would be something like 1) a breakthrough is made 2) that breakthrough is utilized in a way that generates cash flow 3) the company becomes profitable (and at this point hopefully the valuation is good.)

There's a lot of things that can go wrong at every step here (aside from the obvious), including e.g. making a breakthrough that doesn't represent a defensible mote for your startup, failing to build the structure of the business necessary to generate cashflow, ... OpenAI et al already have a lot of that behind them, and while that doesn't mean that they don't face upcoming risks and challenges, the huge amount of cashflow they have available helps them overcome these issues far more easily than a startup, which will stop solving problems if you stop feeding money into it.