The Next Wave of AI Infrastructure Must Target Nvidia’s CUDA Moat

Nvidia built something incredible with CUDA, but for AI to meet its full potential, it’s time to move on.

Written by Ben Koska
Published on Jan. 28, 2026
Nvidia headquarters
Image: Shutterstock / Built In
Brand Studio Logo
REVIEWED BY
Seth Wilson | Jan 28, 2026
Summary: Nvidia’s CUDA dominance is fading as AI labs shift to hardware-agnostic infrastructure. New tools like OpenAI’s Triton allow teams to run models on AMD and Intel without rewriting code. This flexibility is vital for multimodal robotics and escaping the infrastructure tax of vendor lock-in.

I spent four years building AI infrastructure before Y Combinator accepted my current company. During that time, I watched research teams burn through 80 percent of their engineering hours wrestling with GPU orchestration instead of advancing their models. The bottleneck wasn’t scientific talent. It was CUDA.

Nvidia built something remarkable with CUDA in 2006. As deep learning frameworks like TensorFlow and PyTorch emerged, Nvidia invested heavily in optimizing its libraries to ensure these frameworks ran as efficiently as possible on its hardware. Once you write code in CUDA, switching to AMD or Intel means rewriting everything. That lock-in has propped up their market position for nearly two decades.

But the AI infrastructure we’re building today looks nothing like what CUDA was designed for.

Why Is CUDA’s Dominance in AI Infrastructure Ending?

CUDA’s dominance is reaching an inflection point due to the rise of hardware-agnostic compilers and multimodal AI models. While Nvidia’s proprietary software created a 20-year vendor lock-in, modern tools like OpenAI’s Triton and MLIR allow developers to achieve high performance across diverse hardware (AMD, Intel and specialized ASICs) without rewriting code. This shift enables AI labs to escape high infrastructure costs and choose accelerators based on technical fit rather than software constraints.

More on AI InfrastructureHow AI Infrastructure Will Redefine National Sovereignty

 

The Costs of CUDA 

Building hardware-agnostic AI infrastructure revealed a critical insight: The next generation of AI models fundamentally breaks CUDA’s assumptions. Robotics systems now need vision-language-action models that process multiple input modalities simultaneously. Training and serving these models requires switching between different accelerators based on which hardware handles each workload most efficiently. You can’t do that when your entire stack assumes Nvidia GPUs.

The cost of this inflexibility is staggering. Nvidia CEO Jensen Huang estimates between $3 and $4 trillion will be spent on AI infrastructure by the end of the decade, with vendor lock-in through CUDA forcing teams to pay premium prices across the AI supply chain. For reference, Nvidia has 70 percent gross margins on their hardware sales.

All the while, AMD’s MI300X and MI355 accelerators sit underused despite offering comparable performance. Intel’s oneAPI ecosystem keeps expanding. And emerging players like Cerebras and Groq are shipping specialized hardware that CUDA can’t even target.

The compiler layer is where CUDA’s lock-in breaks down. OpenAI’s Triton and MLIR have proven you can write GPU code once and achieve near-parity performance across different hardware. AMD’s ROCm 7 now delivers up to 3.5 times better inference performance than previous versions, showing that alternatives are closing the performance gap. When you build at the compiler level rather than directly targeting proprietary APIs, hardware becomes swappable.

This matters because AI labs are drowning in infrastructure complexity. Research shows that teams spend the majority of their time on DevOps and optimization rather than model architecture development, training experiments, and scientific publication.I've spent years working on infrastructure that eliminates this tax. The proof already exists: teams can train models across CUDA, ROCm, and emerging accelerators without changing a single line of application code when they build at the right abstraction layer.

 

Unlocking the Future Requires Draining the Moat

The timing is critical. Vision-language models and multimodal systems demand flexible hardware that can handle diverse computational patterns. Training a robotics foundation model might need Nvidia GPUs for vision processing, AMD accelerators for distributed training and custom ASICs for inference.

CUDA forces you to pick one ecosystem and live with the constraints: You pay Nvidia’s prices, wait for Nvidia’s release cycles and abandon workloads that would run better on alternative hardware.

The current infrastructure model is unsustainable. When AI research teams can redirect engineering time from infrastructure maintenance back to actual innovation, the entire field accelerates. Breaking CUDA’s dominance isn’t just about saving money on hardware. It’s about unlocking the next generation of AI applications that current infrastructure can't support.

Nvidia built a moat with CUDA. They filled it with proprietary libraries, optimized kernels and two decades of ecosystem development. But moats don’t survive when the landscape changes underneath them. The shift to multimodal models, hardware diversity and compiler-driven abstraction is draining that moat faster than anyone expected.

The infrastructure tax that keeps AI labs locked into expensive, inflexible hardware is coming to an end. AMD, Intel and emerging accelerator companies are all investing heavily in software ecosystems that work across hardware boundaries. The compiler innovations enabling this transition are open source and improving rapidly. New approaches to automatic kernel optimization and cross-platform orchestration are making hardware choice a runtime decision rather than an architectural constraint. Instead of writing separate code paths for each accelerator, teams can write once and let the compiler generate optimized kernels for whatever hardware is available at execution time.

Nvidia’s $4.5 trillion market cap reflects the value of AI infrastructure. But that value assumes AI will always primarily involve training large language models on clusters of identical Nvidia GPUs. The future looks different. Robotics. Multimodal reasoning. Edge deployment. Specialized accelerators. Hardware diversity. None of these fit cleanly into a single-vendor infrastructure model

We’re watching the same pattern that played out in cloud computing. Amazon dominated the field because it solved the infrastructure problem early. But that dominance created the opening for specialized clouds, edge computing and multi-cloud strategies. CUDA is facing that same inflection point.

The compiler layer is where vendor lock-in dissolves, and it’s changing faster than most people realize. When code compiles to whatever hardware runs it best, when orchestration automatically picks the cheapest capable accelerator, when kernel optimization happens at compile time rather than being locked to proprietary libraries, then the economics of AI infrastructure shift permanently. Hardware vendors compete on price and performance rather than ecosystem lock-in while research teams regain the flexibility to choose accelerators based on technical fit rather than sunk costs in proprietary code.

More on the Cutting Edge of AIWhy AI Feels Useless at Work — and How to Fix It

 

The Future Needs Flexibility

This isn’t about being anti-Nvidia. They built something incredible. But the AI infrastructure we need for the next decade can’t be built on assumptions from 2006. Today’s models need hardware flexibility. Research labs need to escape the infrastructure tax. The robotics revolution needs compilers that target reality, not vendor ecosystems.

The companies building tomorrow’s AI applications won’t tolerate infrastructure that forces them to rewrite code every time hardware advances. The market is demanding compilers that make hardware choice a deployment detail, not an architectural constraint.

Explore Job Matches.