Adeia Blog

All Blogs

December 11, 2025

Does AI Scale from Here — or Stall? In Search of a New Architecture

Does AI Scale from Here — or Stall? In Search of a New Architecture

The trillion-dollar question for AI infrastructure

2025 has been a dizzying year for artificial intelligence. Around the world, technology companies, banks, private equity, sovereign wealth funds and governments are making unprecedented AI infrastructure investments — on the order of trillions of dollars in data centers, land, fiber networks and power expansions. OpenAI alone has outlined plans for well over a trillion dollars in long-term compute and energy capacity.

Beneath all this activity lies a single assumption: that the current AI technical architecture, centered on transformer-based models, such as large language models (LLMs), will keep scaling indefinitely.

But can it? Or are we nearing the edge of what this design can deliver — technically, physically and economically?

At Adeia, we have seen moments like this before. In early 2022, our analysis of conversational AI, published before ChatGPT’s launch, helped analysts and investors anticipate a major inflection point. Today, we are once again asking the hard question that many in the industry are circling around: what comes after the transformer architecture?

The transformer: the architecture that built modern AI

The transformer architecture, introduced in 2017, became the foundation of nearly every large language model in use today. Its encoder-decoder structure converts input sequences into numerical representations and reconstructs outputs through a mechanism called self‑attention.

Self‑attention allows the model to consider relationships among all tokens in an input simultaneously, capturing long-range dependencies that earlier recurrent or convolutional networks could not.

A further innovation, multi-headed attention, lets transformers focus on different parts of a sequence at once, improving learning and generalization. Just as important, the design aligned almost perfectly with GPU architectures, which excel at matrix operations and parallel computation.

Together, these advances made it possible to train enormous models on vast datasets and to discover an empirical rule of thumb that has defined the last decade of AI.

Riding — and bending — the AI scale curve

Over the past half-decade, AI progress has largely followed this scaling law. Models have been trained on nearly all public internet text, fine-tuned on curated data, reinforced through human feedback, and expanded with synthetic datasets.

To extend the life of this paradigm, researchers have added:

  • Diffusion models for image and media generation
  • Mixture-of-experts (MoE) architectures to route tokens through specialized subnetworks
  • State-space models to better handle long sequences
  • Chain-of-thought prompting and reasoning techniques to extract more from the same base models

These innovations have squeezed more performance from the same foundation. But the approach is hitting natural limits.

  • Hardware performance — measured in throughput per watt and per die area — is no longer keeping pace with the computational demands of frontier models.
  • Training state-of-the-art systems now consumes hundreds of megawatts of power and requires vast physical footprints.
  • Even with optimized chips, custom interconnects, and aggressive parallelization, cost curves are bending upward faster than efficiency gains, for training the latest and greatest next model.

The transformer, while extraordinarily capable, is nearly a decade old. What propelled us here may not be enough to take us further.

Beyond the text era: why transformers may not be enough

Despite the rapid progress in generative AI, today’s most widely deployed systems remain largely text-bound. In many ways, this resembles the command-line era of computing — powerful, but abstracted from the real world.

The next generation of AI will need to integrate multiple modalities:

  • Visual and spatial data
  • Physical and sensor data
  • Biological and environmental signals

This is the foundation for world models — architectures capable of understanding context, causality, and sensory input together, not just predicting the next token in a sequence.

Here is the core challenge:

  • The transformer’s roots in sequence prediction make it inherently optimized for text and token streams.
  • Multimodal, causal, and interactive tasks — such as robotics, real-time decision-making, and digital twins — demand architectures that can process time, space, and interaction as fluently as current models process language.

To move beyond the text era, we will need AI architectures that are natively multimodal and causal, not just extended transformers with additional adapters.

The S-curve of AI innovation

Every technology evolves along an S-curve: slow beginnings, exponential acceleration, then a plateau before the next paradigm shift.

AI’s transformer era is approaching that inflection point. Much of today’s investment focuses on scaling and optimizing rather than reinventing. Efficiency tweaks, improved training pipelines, and larger GPU clusters extend performance, but they do not change the underlying physics or economics.

At Adeia, we believe the next phase of AI growth will require a fundamental architectural rethink — a new foundation that:

  • Sustains exponential improvement
  • Remains compatible with existing software stacks and developer workflows
  • Breaks the linear relationship between performance and cost

This shift must go beyond brute-force transistor scaling. It will depend on:

  • Hardware-algorithm co-design, not just faster chips
  • Innovation at the intersection of compute, memory, and model efficiency
  • Hybrid learning methods that change how models represent knowledge and reason, not just how many parameters they have

In search of a new AI foundation

So what could that new foundation look like? Across the research community, several promising directions are emerging for post-transformer AI architectures:

  • Neuromorphic and event-driven computing
    • Mimic aspects of biological systems
    • Offer orders-of-magnitude energy savings for certain workloads
  • Hybrid symbolic–neural architectures
    • Integrate logical reasoning with deep learning’s pattern recognition
    • Combine the strengths of symbolic AI and connectionist approaches
  • World and physics-based foundation models
    • Grounded in sensor data, simulation, and digital twin environments

These are not incremental changes, and this list is not comprehensive. Each represents a potential leap — and the race is on to discover which of these approaches (or which combination) or a totally asymmetric and fundamentally different novel architecture will define the next era of AI.

Adeia’s perspective: where scaling laws stall, invention begins

At Adeia, we focus on moments when industries meet architectural walls. Whether in semiconductors, media, connectivity, or AI infrastructure, our work begins where traditional scaling laws stall. We look for the invention that unlocks the next exponential curve.

We believe the future of AI will depend not only on capital and compute, but on creativity in architecture:

  • New ways to represent knowledge
  • New models that can reason about the physical world, not just text
  • New bridges between hardware and software innovation

That is why Adeia continues to engage directly with the global research ecosystem. We recently attended NeurIPS 2025 in early December to exchange ideas, meet fellow inventors and collaborators, and explore what comes after the transformer — in AI models, in compute architectures, and in the broader infrastructure stack that will support them.

The next decade of AI starts with invention

The world has effectively staked trillions of dollars on the assumption that today’s AI architecture will keep scaling. History suggests otherwise. Each generation of technology eventually encounters limits — physical, economic, or conceptual — that demand a new approach.

As AI approaches that boundary, the challenge is clear:

  • We must invent beyond the transformer, beyond brute-force scaling, and toward architectures that combine efficiency, multimodality, and real-world reasoning.
  • We must align AI models, semiconductor design, memory architectures, and networking in a coherent, co-designed hardware and software stack.

Adeia has always thrived at these turning points. As CTO, I see this moment as an opportunity — not a ceiling. We are ready once again to help reimagine the structures that will power the next decade of intelligent systems and to partner with other industry innovators.

Adeia attended NeurIPS 2025 last week, and we enjoyed seeing old colleagues and meeting new folks in the space.

Does AI Scale from Here — or Stall? In Search of a New Architecture

MicroLED Breakthroughs: Paving the Way for Scalable, Cost-Effective AR Smart Glasses

Adeia Tech Talk Recap: Hybrid Bonding & Direct-to-Chip Cooling with Roth Capital Partners

Adeia: Leading the Future of Technology and Innovation

Serhad Doken

Chief Technology Officer

Serhad Doken is responsible for the technology research strategy and advanced R&D projects. Mr. Doken previously was the Executive, Director of Innovation & Product Realization at Verizon where he drove new 5G and mobile-edge computing powered services for consumer and enterprise businesses. Prior to Verizon, Mr. Doken was VP, Innovation Partners at InterDigital focused on technology strategy and external R&D projects and partnerships. Prior to InterDigital, Mr. Doken worked on emerging mobile technology incubation at Qualcomm. Prior to this, Mr. Doken held positions at Cisco Systems, Nortel Networks and PSI AG. Mr. Doken is an inventor on 30 issued worldwide patents with over 90 worldwide applications. Mr. Doken has a Computer Engineering degree from Bosphorus University and has completed the M&A Executive Education Program at The Wharton School and the New Ventures Executive Education Program at Harvard Business School.