skip to content
Stephen Van Tran
Table of Contents

Somewhere in a Paris office, a Turing Award winner is staking his reputation — and roughly $586 million in fresh capital — on the claim that the most celebrated technology of the decade is a dead end. Yann LeCun, the former chief scientist of Meta’s FAIR lab, launched AMI Labs in January 2026 with a single, incendiary thesis: large language models cannot think, and no amount of scaling will fix the problem. His alternative? World models — AI systems that learn how the physical world works by watching it, not by reading about it.

LeCun is not alone. In the span of ten weeks, Fei-Fei Li’s World Labs closed a $1 billion round anchored by Autodesk and AMD at a $5 billion valuation. Runway raised $315 million at $5.3 billion to pivot from video generation into world simulation. Google DeepMind shipped Genie 3 to paying subscribers, generating navigable 3D environments in real time. And Nvidia’s open-source Cosmos platform, built to train robots and autonomous vehicles in simulated physics, has been downloaded more than two million times. Add the startup valuations together and the world-model insurgency is sitting on a combined market cap north of $13.8 billion — with nearly $2 billion in venture money deployed since January alone.

This is no longer a research curiosity. It is a coordinated bet that the next leap in artificial intelligence will come not from predicting the next token in a sentence, but from simulating the next state of a physical environment. The question is whether LeCun’s rebellion can deliver before the LLM establishment proves him wrong.

The godfather bet $3.5 billion that your chatbot is a dead end

LeCun has been making this argument for years, but until November 2025 he was making it from inside the house. As Meta’s chief AI scientist, he oversaw FAIR, the fundamental research lab he founded in 2013 that produced breakthroughs in self-supervised learning and the original JEPA architecture. Then Mark Zuckerberg reorganized Meta’s AI apparatus around Superintelligence Labs, hired former Scale AI CEO Alexandr Wang as chief AI officer, and pivoted the company toward building commercial LLM products to compete with OpenAI and Google. LeCun found himself reporting to Wang.

He walked. In December 2025 he confirmed the creation of AMI Labs, a Paris-based startup seeking a valuation of $3.5 billion before writing a single line of production code. The CEO is Alex LeBrun, formerly the founder of health AI startup Nabla. Investors circling the deal reportedly include Cathay Innovation, Greycroft, and Hiro Capital. Meta, in an unusual twist, is expected to be a strategic partner.

LeCun’s core critique of LLMs is blunt and technical. Autoregressive text models predict the next word in a sequence, but they operate in the discrete space of language tokens. They do not know that objects persist when out of sight, they have no intuitive model of gravity, and they cannot plan a sequence of physical actions by simulating their outcomes. An LLM can describe a ball falling because it has ingested millions of sentences about falling balls. It cannot predict the trajectory of a specific ball thrown at a specific angle, because it has never experienced physics — only prose about physics.

AMI Labs intends to solve this with V-JEPA, the Joint Embedding Predictive Architecture that LeCun’s team built at Meta and published as V-JEPA 2 — a 1.2-billion-parameter model that learns by predicting missing information in an abstract representation space rather than pixel by pixel. Where GPT predicts the next word, V-JEPA predicts the next state of the world. The training data is not scraped text but high-bandwidth sensory input: video, audio, lidar, robot arm telemetry. The result, if it works, is an AI that can plan before it acts, reason about cause and effect, and maintain a persistent model of its environment across time.

The implications ripple far beyond one startup. Consider the timeline: LeCun spent a decade building the intellectual foundation at Meta, published the core architecture as open research, then left to commercialize it the moment his employer chose a different path. This is not a pivot — it is a schism. The man who helped invent convolutional neural networks, who shared the 2018 Turing Award with Geoffrey Hinton and Yoshua Bengio, is telling the industry that its most profitable product category is a technological cul-de-sac.

The stakes for the broader industry are enormous. If LeCun is right — if LLMs are architecturally incapable of achieving robust physical reasoning — then the $690 billion that hyperscalers are pouring into GPU-centric data centers this year is partially misallocated. If he is wrong, AMI Labs is a very expensive vanity project. Either way, the debate is no longer academic. It is priced into term sheets.

A field guide to the physics-first insurgents

The world-model landscape in early 2026 is fracturing into distinct schools of thought, each backed by serious capital and led by researchers who fundamentally disagree on how to build AI that understands reality.

Fei-Fei Li’s World Labs occupies the spatial-intelligence lane. The Stanford professor coined the term to describe AI systems that can perceive, generate, and interact with three-dimensional environments. Her commercial product, Marble, turns text prompts, photos, or video into editable, downloadable 3D worlds. Users can block out spatial structures in a hybrid editor, then let the AI fill in photorealistic detail. Marble already works on Vision Pro and Quest 3 headsets, and the February 2026 funding round — led by a $200 million check from Autodesk alongside AMD, Fidelity, and Nvidia — reveals the bet: spatial intelligence will reshape design, architecture, and entertainment before it touches robotics.

Google DeepMind’s approach is real-time generative simulation. Genie 3 does not render a static 3D snapshot; it generates the path ahead as you move through it, running at 24 frames per second at 720p resolution. The catch is that coherence degrades after roughly sixty seconds, and DeepMind is limiting sessions accordingly due to compute constraints. The product shipped to Google AI Ultra subscribers on January 29, positioning it as a consumer novelty today and a training-data engine for robotics tomorrow.

Runway represents the media-to-simulation pipeline. The company built its reputation on Gen-4 and Gen-4.5, the top-rated AI video models on the Artificial Analysis benchmark with 1,247 Elo points. Its $315 million Series E, led by General Atlantic with participation from Nvidia and Adobe Ventures, is explicitly earmarked for expanding beyond video into world models — AI that builds internal maps of environments so it can plan future actions. Runway’s General World Model, GWM-1, launched in December 2025, is the bridge.

Then there is the Chinese front. Zhipu AI’s GLM-5 release in February — 744 billion parameters, trained entirely on Huawei Ascend chips — proves that frontier-model development no longer requires Nvidia silicon. But GLM-5 is a language model. The question hanging over Beijing’s AI ecosystem is whether Chinese labs will replicate the world-model push with the same speed they matched LLM scaling. If they do, the fragmentation of AI research along geopolitical lines will deepen further, creating parallel stacks for simulating physical reality — one trained on American sensor data, another on Chinese.

Nvidia sits at the infrastructure layer. Rather than building a single world model, it is supplying the tools for everyone else to build theirs. The Cosmos platform, which includes open-weight Cosmos-Predict2.5 and Cosmos-Transfer2.5, specializes in simulating physical environments for robotics and autonomous-vehicle training. When Jensen Huang talks about physical AI at CES, this is the stack he means. Two million downloads suggest the developer community agrees.

Here is the proprietary calculation that no single press release contains: if you sum the venture capital raised by AMI Labs (~$586 million), World Labs ($1 billion), and Runway ($315 million) since the start of the year, the total exceeds $1.9 billion — and that excludes Google’s internal R&D budget and Nvidia’s platform investment. Divide by the combined employee headcount of these three startups (conservatively estimated at 350 to 450 people) and the implied capital intensity is between $4.2 million and $5.4 million per head. These are not lean operations. They are infrastructure-heavy bets that require massive compute to train models on video and sensor data at scales that dwarf text corpora.

The trillion-dollar question nobody wants to answer

For all its intellectual appeal, the world-model thesis rests on assumptions that have not been tested at commercial scale. The most obvious: nobody has yet demonstrated a world model that outperforms a well-tuned LLM at a task customers will pay for.

The revenue asymmetry is staggering. OpenAI reportedly generated $11.6 billion in annualized revenue in late 2025, almost entirely from text-based products. Anthropic, Google, and Meta earn billions from LLM-powered search, coding assistants, and enterprise chatbots. World-model startups, by contrast, are pre-revenue or generating token amounts from creative-tool subscriptions. World Labs charges for Marble access; Runway sells Gen-4.5 seats; AMI Labs has shipped nothing. The combined revenue of the three leading world-model startups is likely less than one percent of OpenAI’s monthly run rate.

The technical skeptics have a point, too. A vigorous counterargument on LessWrong details why autoregressive architectures may not be the dead end LeCun claims. Multimodal LLMs like GPT-5.3 and Gemini 3.1 Pro already process video, audio, and images alongside text, embedding rudimentary physical reasoning into architectures that have proven distribution advantages. Google DeepMind’s Gemini 3.1 Pro scores 77.1% on ARC-AGI-2, a benchmark designed to test generalized reasoning. These are not systems trapped in the prison of text.

There is also the integration problem. LLMs are general-purpose enough to slot into virtually any workflow — customer support, legal review, code generation, medical triage. World models, by definition, excel in spatial and physical domains but lack the linguistic fluency that makes LLMs commercially versatile. A hospital deploying an AI system cares more about summarizing discharge notes than simulating how a patient walks down a hallway. A law firm wants contract analysis, not 3D environment generation. The total addressable market for world models may be enormous in theory (robotics, autonomous vehicles, gaming, industrial simulation) but narrow in the near term.

The fragmentation risk is equally real. AMI Labs bets on V-JEPA embeddings. World Labs bets on spatial intelligence from images. Runway bets on generative video as a bridge. Google bets on real-time simulation. Nvidia bets on being the pickaxe seller. These are not compatible approaches converging on a single standard — they are divergent architectures competing for mindshare, talent, and compute. The early days of LLMs saw a similar fragmentation (remember BERT versus GPT-2 versus T5?), but the transformer architecture eventually won and enabled an ecosystem to coalesce. No equivalent convergence is visible in world models yet.

The talent wars add another layer of risk. LeCun recruited heavily from Meta FAIR when he left. Fei-Fei Li draws from Stanford’s vision lab and the broader computer-graphics community. Runway poaches from film VFX and gaming studios. Google DeepMind has the deepest bench, but its researchers are spread across dozens of projects. The world-model field is competing for a small pool of people who understand both deep learning and 3D physics simulation — a Venn diagram that, until recently, described maybe a few hundred researchers worldwide. Hiring wars at this scale distort compensation, slow progress, and create single points of failure when key engineers leave.

Finally, compute economics may throttle ambition. Training on video and sensor data is orders of magnitude more expensive than training on text. A minute of 720p video at 24 fps contains roughly 2,000 frames, each carrying far more information than a sentence. AMI Labs and World Labs need either proprietary data pipelines or access to the same hyperscaler infrastructure that LLM companies already monopolize — which means competing for GPU allocation with the very companies whose paradigm they claim is obsolete.

Where the smart money is watching — and what to build next

The world-model rebellion is real, well-funded, and led by people with the credentials to back their claims. It is also early, fragmented, and commercially unproven. For operators, investors, and builders trying to make sense of the next twelve months, the signal is in the convergence points — the places where world models and LLMs stop competing and start complementing each other.

Gartner’s inclusion of physical AI in its top ten strategic technology trends for 2026 is not a prediction; it is a lagging indicator. Companies like Tesla, Boston Dynamics, Figure, and Unitree are already deploying humanoid robots in factory settings, and every one of those deployments requires a world model — an internal representation of the environment that the robot can use to plan actions safely. Nvidia’s Cosmos downloads suggest thousands of robotics teams are building on this stack today. If you are in manufacturing, logistics, or warehousing, the relevant question is not whether world models matter but when they become a procurement decision.

The autonomous-vehicle pipeline is a parallel story. Mobileye, Waymo, and the Chinese autonomous-driving ecosystem all rely on simulation environments to train and validate self-driving systems. World models turn simulation from a data-augmentation tool into a training-data factory — generating infinite variations of road conditions, weather, pedestrian behavior, and edge cases. Nvidia’s Omniverse and Cosmos stack is purpose-built for this workflow.

The creative and design industries will feel the impact sooner than most observers expect. Autodesk’s $200 million investment in World Labs is not philanthropic; it is a bet that spatial AI will reshape architectural visualization, product design, and film preproduction within two to three years. Runway’s pivot from video to world models follows the same logic: the creators who use Gen-4.5 today to generate two-dimensional clips will demand immersive, interactive, three-dimensional outputs tomorrow.

For builders, the operator checklist is concrete:

  • Track the convergence of LLMs and world models. The winning architecture will almost certainly be hybrid — language understanding layered on top of physical reasoning. Watch for announcements where LLM companies (OpenAI, Anthropic, Google) acquire or partner with world-model startups.
  • Monitor Nvidia’s Cosmos adoption metrics. Two million downloads is a leading indicator. If Cosmos-Predict2.5 becomes the de facto training layer for physical AI, Nvidia’s moat deepens beyond hardware into the simulation stack.
  • Benchmark world-model startups on revenue, not valuation. AMI Labs at $3.5 billion and World Labs at $5 billion are priced on vision. The first company to convert a world model into a repeatable, scalable enterprise product — whether in robotics, AV, or design — will define the category.
  • Follow the talent pipeline. LeCun recruited from Meta FAIR. World Labs draws from Stanford’s vision lab. Runway poaches from film and gaming. The talent clusters tell you where the breakthroughs will come from — and which cities (Paris, San Francisco, New York) will host the next wave of AI infrastructure.
  • Assess your own data assets through a spatial lens. If your organization generates video, sensor, lidar, CAD, or 3D scan data, you are sitting on training material that world-model companies will pay for or partner to access. The content-licensing playbook that publishers are running with LLM companies today will repeat in the world-model domain within eighteen months.

The history of AI is a history of paradigm shifts arriving faster than incumbents expect. Symbolic AI gave way to statistical learning. Statistical learning gave way to deep learning. Deep learning gave way to transformers. Each transition looked premature until it did not. LeCun is betting that transformers-for-text is the paradigm about to break, and that the replacement will be trained not on the written record of human knowledge but on the raw sensory feed of physical reality.

The LLM revolution gave machines the ability to read, write, and talk. The world-model rebellion is an attempt to give them the ability to see, touch, and plan. Both capabilities will be necessary for the AI systems that matter most in the next decade — the ones that operate in hospitals, factories, roads, and homes. The $1.9 billion deployed in the first ten weeks of 2026 is the opening bid, not the final offer. Whether LeCun’s insurgency topples the LLM consensus or merges with it, the architects who understand both paradigms will build what comes next.


In other news:

Alphabet hands Sundar Pichai a $692 million pay package — The board approved a three-year compensation plan for Google’s CEO on March 4, tying the bulk of the payout to stock performance, Waymo milestones, and total shareholder return versus the S&P 100. Only $84 million vests on tenure alone; the rest requires Alphabet to outperform its mega-cap peers (TechCrunch).

Meta signs $50 million-a-year content deal with News Corp — The three-year agreement gives Meta access to Wall Street Journal and other News Corp titles for training AI models and powering Meta AI chatbot responses. It follows similar deals with CNN, Fox News, and USA Today, signaling that publisher content licensing is now a standard line item in big-tech AI budgets (Engadget).

Mustafa Suleyman proposes the $100K-to-$1M intelligence test — Microsoft’s AI CEO argued in January that the real benchmark for AGI is not mimicking conversation but autonomously and legally turning $100,000 into $1 million, reframing the Turing Test around economic capability rather than linguistic imitation (Yahoo Finance).

MWC 2026 goes all-in on wearable AI — Qualcomm announced Snapdragon Wear Elite, a chip targeting AI pendants, pins, and display-free smart glasses, with first commercial devices expected within months. Samsung showcased agentic AI on the Galaxy S26, and Deutsche Telekom unveiled an in-network AI call assistant built with ElevenLabs (Techloy).

Zhipu AI ships GLM-5, a 744B-parameter frontier model trained entirely on Huawei chips — The Chinese lab released its largest model under the MIT license with 40 billion active parameters, scoring 77.8% on SWE-bench Verified. Trained on 100,000 Huawei Ascend 910B processors and no Nvidia silicon, GLM-5 is the clearest evidence yet that export controls have not stopped China from reaching the frontier (NxCode).