The shift in three lines

// The gap
Today's AI learned about the world by reading what humans wrote about it — not by experiencing it directly.
// The shift
World models enable machines to simulate outcomes before acting — a qualitative leap beyond pattern recognition.
// The question
If world models develop alongside LLMs rather than replacing them, are current AI infrastructure forecasts large enough?

A decade of pattern recognition — and what comes next

For a decade, the frontier of artificial intelligence was defined by a single organising principle: train a system on enough human-generated data, optimise it to predict what comes next, scale it up — and watch capabilities emerge that routinely surprise even the researchers who built the system.

The results have been genuinely astonishing. Large language models can reason, converse, write, and generate code at levels that have reshaped entire industries and triggered a multi-trillion-dollar infrastructure investment wave. The commercial momentum is real, the capability advances are substantial, and the enterprise adoption curve is still early.

And yet, beneath the surface of this achievement lies a structural limitation that the field has been slow to confront openly. These systems are powerful at completing patterns, but they lack an internal sense of the world those patterns describe. They understand how our world works through a form of second-order interpretation — absorbing descriptions of reality without ever encountering reality itself. They have no first-principles grasp of physics, motion, cause and effect, or consequence.

A language model can explain, in fluent prose, that a glass will shatter if dropped from a table. It has no internal representation of the weight, the trajectory, the contact, or the outcome. That distinction barely registers when AI is asked to summarise a document or draft a communication. It becomes a hard constraint the moment AI is asked to navigate an unstructured physical environment, coordinate a complex organisation in real time, or reason about how a strategic decision will cascade through a live system.

// The structural limitation of pattern-based AI

The transition now under way is from systems that recognise and reproduce patterns to systems that build internal models of the world — and use those models to plan, simulate, and act with consequence-awareness.

// Current paradigm — pattern completion

Systems trained on human text learn the statistical structure of language. They predict what word, sentence, or argument typically follows another. They are extraordinarily capable within this frame — and bounded by it. They have no causal model of the world; they have a very sophisticated map of how humans have described it.

// Emerging paradigm — world modelling

Systems that build internal simulators can ask, repeatedly: if I take this action, what happens next? They can rehearse in digital environments where failure is cheap. They can reason about physical constraints, social dynamics, and causal chains before committing to a course of action in the real world.

This analysis explores what world models are, why the field's most credentialed researchers are converging on them, what they imply for enterprise AI deployment, and whether current AI infrastructure projections are adequately sized for a world in which they succeed.

// Section 01 of 05

What world models are — and why they matter

A world model, at its core, is an internal simulator. It allows a system to answer a simple question, repeatedly and quickly: if I do this, what happens next? Humans rely on this faculty constantly — we picture a glass tipping before it falls; we rehearse a difficult conversation before having it. Until recently, machines could not do this well.

Teaching a robot to recognise a cup is straightforward. Teaching it to pick one up without shattering it is another matter entirely. The physical world is unforgiving: objects have weight and inertia; surfaces have friction; forces compound in ways that text descriptions cannot fully capture. World models offer a path to machines that have internalised these constraints rather than merely read about them.

There are two distinct categories of world model, each pointing toward a different frontier — and each with significant implications for how organisations compete and make decisions.

// Two frontiers of world modelling
Physical and social world models address different constraints — and together suggest a deeper shift in the nature of machine intelligence
// Track one
Physical world models

These systems learn the governing logic of the material world: gravity, friction, thermodynamics, fluid dynamics, structural mechanics. Rather than learning purely through real-world trial and error — slow, expensive, and potentially dangerous — they absorb the rules of physics through simulation, practising in digital environments where failure is fast and cheap.

Gravity & inertia Friction & contact Thermodynamics Fluid dynamics Structural forces
  • Warehouse robots navigating dynamic, unstructured spaces with fewer collisions
  • Autonomous vehicles rehearsing low-probability edge cases before encountering them on real roads
  • Industrial process optimisation where physical constraints bound every decision
// The combined system

The most consequential decisions will ultimately draw on both capabilities simultaneously. Physical world models guide the robots moving goods through a port; social world models simulate the demand shock, labour response, and geopolitical disruption that changed the routing. Decisions in one domain inform actions in the other — and planning becomes continuous rather than episodic.

What makes social world models especially powerful is that they offer something beyond faster analysis: a higher-fidelity operating environment for decision-making. Enterprises already spend enormous effort attempting to anticipate how others will respond — how competitors will move, how markets will interpret signals, how boards will react under pressure. Today, those judgements rely on experience, static analysis, and intuition. Multi-agent simulation offers something closer to a living model of human systems.

Critically, world models are not forecasting tools in the conventional sense. They do not predict the future. They reveal plausible futures — ranges, paths, and feedback loops that conventional single-point forecasts cannot surface. Forecasting assumes a single correct outcome. World models reveal the shape of possibility.

// Section 02 of 05

The researchers building the next paradigm

What distinguishes the current world model moment from prior cycles of AI research enthusiasm is the calibre and credibility of the researchers converging on it. These are not peripheral voices staking out contrarian positions. They are, in several cases, the architects of the current AI paradigm — and they are arguing, in their own research and ventures, that the paradigm is incomplete.

// Physical intelligence · JEPA architecture

Yann LeCun

Founder, AMI Labs · Former Chief AI Scientist, Meta

LeCun spent years as one of the most influential figures inside a major AI lab before departing to found AMI Labs with world models as its explicit foundation. His Joint-Embedding Predictive Architecture — JEPA — is designed to build machines that develop internal models of the world through observation, in the way humans do, rather than through text prediction. LeCun has been publicly and persistently critical of the thesis that scaling language models alone will reach general intelligence. World models are his alternative.

JEPA framework Observation-based learning Abstract prediction Physical AI
// Spatial intelligence

Fei-Fei Li

Founder, World Labs · Professor, Stanford University

Li's ImageNet dataset helped ignite the deep learning revolution that produced today's dominant AI systems. She founded World Labs around a related but distinct idea: spatial intelligence. The premise is that genuine intelligence requires not just recognising objects in images but understanding how those objects exist in space, interact with each other, and change over time. Li's bet is that machines need to inhabit a model of three-dimensional reality — not merely classify it at a frame level.

Spatial intelligence 3D scene understanding Object interaction Embodied AI

The significance of this convergence is hard to overstate. These are the researchers whose earlier work produced the current AI era — and they are now directing their energy toward what they believe the current era is missing. Their departure from leading research positions to build dedicated ventures around world models signals something more than academic interest: it signals strategic conviction.

The people who built today's AI are quietly telling us it isn't enough. That signal deserves more attention than it has received from investors and enterprise strategists.
// Section 03 of 05

Not a competition — an integration

Much of the current discourse frames LLMs and world models as competing paradigms, asking which one will prevail. This framing fundamentally misunderstands how complex intelligence emerges — and how the most capable AI systems of the near future are likely to be structured.

Complex intelligence does not arise from a single dominant approach. It arises from the orchestration of many. Just as the human brain integrates specialised modules into coherent thought — perception, memory, language, motor control, social cognition — advanced AI architectures are likely to combine large language models, physical simulators, and social reasoning engines into unified systems.

// Integrated architecture — the probable outcome
Language as the interface · world models as the engine of planning

The most capable AI system of the foreseeable future will likely use language models as the interface for instruction, explanation, and communication — while relying on world models for planning, consequence modelling, and action. Physical laws constrain motion; social rules constrain behaviour. In both cases, intelligence emerges from understanding how local actions ripple outward through a system.

// Language layer (LLM)
Fluency & interface
Instruction, explanation, communication, reasoning across domains
// World model layer
Simulation & planning
Consequence-aware action, physical constraint, social dynamics, causal inference
// Emergent capability
Situational awareness
AI that not only generates answers but understands what those answers mean in the real world

The analogy that holds here is instructive: language gives AI fluency; world models give it situational awareness. For much of the recent history of enterprise AI, we have treated these systems primarily as tools that produce answers. World models suggest something more ambitious — systems that understand the context in which those answers will be acted upon.

// Where LLMs remain dominant

High-value financial, legal, and knowledge tasks that are fundamentally about language — document analysis, contract review, communications, code generation, structured reasoning over text — remain squarely in LLM territory. The transition to world models does not displace this; it extends beyond it.

// Where world models create new leverage

Tasks requiring consequence-aware planning, physical coordination, or the simulation of complex adaptive systems — logistics, manufacturing, organisational strategy, market dynamics, risk modelling — are precisely where world models extend the frontier of what AI can usefully do.

// Section 04 of 05

Enterprise and financial applications

World models are not a distant research project. The early applications are already visible across logistics, manufacturing, financial services, and strategic planning — with the most consequential use cases only beginning to be defined.

// Illustrative world-model applications by domain
Mapping physical and social world-model capabilities to enterprise use cases
Domain Model type Application What changes
Logistics & supply chain Combined Physical robots navigate warehouses and ports; virtual models simulate demand shocks, labour responses, and routing disruptions in parallel. Planning becomes continuous rather than episodic — physical and social intelligence inform each other in real time.
Insurance & reinsurance Physical High-fidelity simulation of hurricane seasons, flood events, or wildfire dynamics to stress-test insured-loss distributions across complex portfolios. Risk modelling moves from actuarial tables to dynamic physical simulation — probability distributions become richer and more realistic.
Financial markets Social Multi-agent environments where each agent reflects a different market participant with distinct incentives, information sets, and behavioural constraints. Strategy rehearsal replaces static scenario analysis — decisions are tested against adaptive adversaries, not fixed assumptions.
Policy & macroeconomics Social Simulation of how a policy shock — a rate change, a tariff, a regulatory shift — cascades through heterogeneous populations with different response functions. Policy evaluation gains a dynamic component: second- and third-order effects become visible before implementation.
Manufacturing Physical Digital twins of production facilities where physical constraints, equipment states, and process parameters are modelled at high fidelity. Process optimisation and fault prediction become physics-grounded rather than purely statistical — the model understands why, not just when.
Corporate strategy Social Board-level governance simulations where strategic options are tested against simulated competitor responses, regulator behaviour, and market interpretations. High-stakes decisions are rehearsed against adaptive, realistic counterparties rather than static scenario trees.

The common thread across these applications is leverage. In domains where mistakes are expensive and foresight creates disproportionate advantage, the value of reliable simulation compounds faster than its cost. The barrier has historically been computational — high-fidelity simulation of physical dynamics or complex social systems is expensive. That cost calculation is now shifting, and the shift is consequential.

Infrastructure implications

// The question current forecasts have not asked

The entire AI infrastructure build-out — chip projections, data centre construction timelines, energy capacity planning — has been sized around a single assumption: that the future of AI is larger language models running on more compute. The question world models raise is whether those projections are measuring the right thing.

// Partial overlap with existing infrastructure
The same GPU clusters and inference platforms that serve large language models can also be used to train and run world models. The infrastructure story is not one of replacement — it is one of extension. Existing compute assets retain their value.
// New requirements with no current equivalent
Simulation environments require purpose-built data pipelines, synthetic data generators, and physics-based engines that go well beyond text corpora. This infrastructure does not yet exist at scale — and building it will require investment categories that current forecasts do not capture.
// Compute intensity is structurally higher
High-fidelity physical simulation, multi-agent interaction, and continuous planning consume far more compute than predicting the next token in a sequence. But cost alone is the wrong metric: in domains where the value of foresight is high, the economics can support the compute bill.
// The critical framing for investors
If world models develop as a complementary layer — built alongside LLMs rather than replacing them — the aggregate compute requirements could substantially exceed what current forecasts anticipate. The infrastructure story is one of partial overlap, not seamless reuse.

Readers of our companion note on the AI build-out scenario framework will recognise the implication: the four supply-side variables that determine total AI capital — silicon service life, data-centre cost, chip architecture mix, build-out elongation — were calibrated against today's LLM-dominated workload mix. A world-model-augmented future raises every one of them, and adds a fifth: the cost of constructing high-fidelity simulation environments at scale.

// Section 05 of 05

The competitive dimension

If the transition to world models develops as described here, it will not merely be a technical refinement. It will reshape the competitive landscape of the AI industry itself — and the enterprises that depend on AI as a source of advantage.

In the current paradigm, competitive advantage in AI accrues primarily to those with the largest models, the most data, and the most compute. These advantages are real but also scale with capital — and capital, at sufficient scale, is accessible to a broad range of well-resourced actors.

Competitive advantage may soon depend as much on who builds the most faithful simulation of reality — physical, social, and economic — as on who trains the largest model.

World models introduce a different dimension of differentiation: the fidelity and accuracy of the simulated reality. A superior world model is not simply a function of compute spend — it reflects the quality of the underlying training environment, the richness of the synthetic data, the sophistication of the physics engine, and the depth of the behavioural modelling. These advantages are harder to replicate through capital alone.

For enterprise decision-makers, the implication is that the relevant question is no longer only "how much AI can we afford?" but "how accurately can our AI represent the systems we operate within?" Organisations that develop or access world models of their own competitive environments — their markets, their supply chains, their regulatory contexts — will gain a form of strategic foresight that has no precedent in conventional analysis.

// Early movers

Organisations that invest in world-model capabilities now — building simulation environments, synthetic data pipelines, and multi-agent modelling infrastructure — gain the advantage of early calibration. World models improve with use: the organisation that has been running simulations for two years will have materially better models than one starting from scratch.

// The risk of delay

The risk of waiting for world models to mature before investing is asymmetric. If the transition develops as the field's leading researchers expect, early movers will have accumulated both proprietary model quality and organisational capability that is difficult and slow to replicate. Competitive position in AI is increasingly path-dependent.

From producing answers to understanding the world

The large language model era gave AI the ability to communicate about the world with remarkable sophistication. It produced systems that can summarise, argue, translate, code, and converse at levels that have genuinely surprised their creators and reshaped industry after industry.

What it did not produce is systems with an internal sense of what the world actually is — how its physical constraints work, how its social dynamics unfold, how local actions propagate through complex systems over time. That gap has been the quiet limitation of every deployed AI system of the past decade.

World models are the field's most serious attempt to close it. The convergence of the AI era's most credentialed researchers on this problem — departing leading institutions to build dedicated ventures around it — is a signal that deserves to be taken seriously by investors, operators, and strategists.

This transition is not imminent in the sense of being months away. The commercial deployment of sophisticated world models is a multi-year journey, and large language models will remain the dominant paradigm for enterprise AI for the foreseeable future. But the trajectory is now clearly established, the researchers are now clearly committed, and the infrastructure investment required — which current forecasts have not yet adequately priced — is now clearly materialising.

Organisations that recognise this shift early will be better positioned not just to adopt AI, but to deploy it where it matters most: in decisions that shape real systems, in real time, with real consequences.

If large language models gave AI fluency, world models will give it situational awareness. For much of its recent history, we have treated artificial intelligence as a system that produces answers. World models suggest something more fundamental: machines that understand what it means to be inside the world they are describing.

// Important disclosures

This report has been prepared by Lualdi Advisors for informational and educational purposes only. It draws on publicly available research, academic literature, industry commentary, and Lualdi Advisors' proprietary analytical framework. References to named researchers and organisations are purely illustrative and informational; they do not imply endorsement, affiliation, or any commercial relationship with Lualdi Advisors. This material does not constitute investment, legal, tax, or financial advice and should not be used as the basis for any investment decision. All forward-looking statements and scenario analyses are inherently uncertain; actual outcomes may differ materially. Lualdi Advisors makes no representations or warranties regarding the accuracy or completeness of the information herein. Past performance is not indicative of future results.