Skip to main content
ADIN
The Reinforcement Gap: Why RL Might Be the Most Underpriced Layer in AI

The Reinforcement Gap: Why RL Might Be the Most Underpriced Layer in AI

PriyankaPriyankaLv.108 min read

For three years, the AI investment narrative has been dominated by a single architecture: the large language model. Foundation models, training compute, inference infrastructure, fine-tuning tooling -- capital has flooded every link in the LLM chain.

But there is a quieter capability layer emerging beneath the surface. One that may determine which AI systems actually do things in the physical and economic world, rather than merely describe them.

That layer is reinforcement learning.

And the gap between its strategic importance and its current share of venture attention may represent one of the more interesting asymmetries in AI investing -- or one of the more seductive traps.

The Case Against: Why Smart People Think RL Is a Mirage

Before making the bull case, it is worth steelmanning the skeptics. Because they are not wrong about the facts. They may just be wrong about the conclusion.

The information efficiency problem is brutal. Toby Ord, the Oxford philosopher and AI safety researcher, published a detailed analysis in September 2025 that should give any RL bull pause. His core argument: pre-training via next-token prediction gives a model roughly 3 bits of information per token generated. Reinforcement learning on long reasoning chains provides less than 0.0001 bits per token -- a gap of roughly 30,000x. On the longest, most complex tasks, RL is up to a million times less information-efficient than pre-training. That is not a rounding error. It is a structural constraint on how fast RL-trained systems can improve per dollar of compute spent.

Most RL projects fail. One CTO quoted in a widely circulated industry post confessed to spending over $200,000 on reinforcement learning in a single year with results "indistinguishable from chance." This is not an outlier. RL agents routinely require millions of interactions to learn basic tasks. Training is unstable. Reward signals are sparse. Hyperparameter sensitivity is extreme. The gap between a published research result and a production deployment remains enormous.

Reward hacking is not a theoretical risk -- it is a recurring reality. RL agents optimize whatever signal you give them, and they are disturbingly creative at satisfying the metric while violating the intent. A robot trained to grasp objects learns to slam its arm into the table because contact triggers the "grasp detected" sensor. A language model trained via RLHF learns to produce confident-sounding nonsense that human raters reward. The alignment community's deepest concerns about AI are, at root, RL reward-specification problems.

The "RL market" may not be a market at all. Research Nester pegs the global RL market at over $122 billion in 2025. But that number conflates RL-native products with every AI system that uses RL as a component -- which, given RLHF, is essentially every frontier model. By that accounting, "the RL market" is just "the AI market with a different label." Pure-play RL companies are rare. Most RL value accrues inside Google DeepMind, OpenAI, NVIDIA, and Tesla -- not in investable startups.

These are real objections. If you stopped here, you would conclude that RL is a research technique embedded inside larger systems, not a distinct investment category.

That conclusion is probably wrong. Here is why.

What Changed: Three Forces Pulling RL Into Production

RL has been a research discipline for decades. The skeptics' critique was fully correct as recently as 2022. What changed is that three converging forces are pulling reinforcement learning from the lab into the economy in ways that the information-efficiency argument alone cannot capture.

First, the agentic turn. The AI industry is pivoting from chatbots to agents -- systems that take actions, use tools, plan multi-step workflows, and operate in open-ended environments. This is not a branding exercise. The agentic AI market surged from $5.25 billion in 2024 to over $9 billion in 2026, with projections reaching $52 billion by 2030. Agents are, by definition, an RL problem. A model that must decide what to do next in an uncertain environment, with delayed and sparse rewards, is navigating exactly the terrain RL was designed for. You cannot build reliable agents with supervised learning alone. The industry knows this.

Second, robotics crossed the funding threshold. Physical Intelligence raised over $1 billion at a $5.6 billion valuation to build generalist robot foundation models. Skild AI raised $300 million in a Series A. Google DeepMind shipped Gemini Robotics. The bridge from digital to physical AI runs through reinforcement learning -- there is no supervised-learning shortcut for a robot arm encountering novel objects in unstructured environments. Sim-to-real transfer is hard, but it is getting less hard. A February 2026 paper from ETH Zurich and Google DeepMind systematically demonstrated what matters for sim-to-online RL on real robots -- moving the field from "it sometimes works" toward engineering discipline.

Third, RLHF became load-bearing infrastructure. Every frontier LLM now depends on reinforcement learning to align model behavior with human preferences. RLHF training consumes roughly 80% of its compute budget on sample generation alone. This is not a niche post-processing step. It is the final and arguably most critical stage of building any commercially deployed large model. When OpenAI launched a Reinforcement Fine-Tuning API with its own billing model, it signaled that RL is now a product, not just a training technique.

The Timing Question: Is It Too Early or Is It Right Now?

This is where the skeptics and bulls talk past each other. They are answering different questions.

The skeptics ask: Can RL solve arbitrary problems efficiently today? The honest answer is no. Sample efficiency is poor. Sim-to-real transfer introduces distribution shift. Reward specification remains an unsolved research problem. Deep RL policies are black boxes in domains that demand explainability.

The bulls ask a different question: Is RL already load-bearing in systems that generate billions in value? The answer is unambiguously yes.

ChatGPT's instruction-following behavior -- the thing that made it a consumer product -- is an RL output. Tesla's Autopilot decision-making uses RL. SpaceX lands rocket boosters with RL-derived controllers. Google's AlphaChip designs semiconductor layouts in hours using RL that previously took engineers months. NVIDIA applies RL to chip floorplanning. A March 2026 paper demonstrated RL-based lithography optimization for sub-nanometer semiconductor fabrication -- meaning RL is now shaping the physical substrate that all of AI runs on.

The disconnect is between RL-as-general-purpose-technique (still immature) and RL-as-embedded-capability (already in production, already generating value, already irreplaceable).

The investment question is not "is RL ready?" It is "where in the RL stack is value accruing, and which layers are still underpriced?"

Where the Money Is Going -- and Where It Isn't

Capital has started flowing into RL-adjacent companies, but unevenly. The robotics layer has attracted enormous checks. The tooling and infrastructure layer is still nearly empty.

Notice the gap. Billions have gone into companies that use RL (robotics, autonomous driving). Very little has gone into companies that make RL work -- the tooling, the training infrastructure, the reward engineering platforms, the RLOps layer.

AgileRL raised just $7.5 million in seed funding in January 2026 to build what they call "RLOps" -- a platform that simplifies training, tuning, and deployment of RL agents. If RL adoption follows anything like the trajectory of MLOps adoption, this tooling layer will capture disproportionate value. And it is almost entirely unfunded relative to the application layer sitting on top of it.

This is the structural asymmetry. The market is pricing RL applications (robotics, autonomy) at billions while pricing RL infrastructure (tooling, training, reward engineering) at almost nothing.

The Contrarian Bet

Here is where this gets interesting for anyone with a thesis.

The consensus AI investment view in 2026 is: foundation models are the platform, inference is the margin, and agents are the next unlock. RL barely features in most investor decks. It is treated as a training technique -- a line item inside the model-building process, not a category.

The contrarian view is that RL is the bottleneck in all three layers.

Foundation models cannot align without RLHF. Agents cannot act reliably without RL-based planning and decision-making. Robotics cannot bridge sim-to-real without RL. Chip design cannot optimize at frontier scale without RL. The further AI moves from text generation toward action in the world, the more load-bearing RL becomes.

And yet: the talent pool is tiny (RL expertise sits at the intersection of control theory, optimization, and deep learning), the pure-play startup category is thin, and the tooling layer is nearly vacant.

If you believe AI's next phase is agency -- systems that do things rather than describe things -- then the constraining capability is reinforcement learning, and the investment opportunity is in the infrastructure that makes RL production-grade.

The counterargument, and it is a strong one: maybe RL gets absorbed. Maybe foundation model companies simply build RL capabilities in-house and the "RL layer" never separates into its own investable category, the way "the database layer" eventually became part of every cloud platform rather than a standalone market.

That is the real risk. Not that RL doesn't work -- it clearly does. But that it works inside existing giants rather than spawning a new ecosystem of independent companies.

The Map

For those building a position, the landscape stratifies:

Highest conviction (RL is the product): RL tooling and infrastructure (AgileRL and successors), RLHF pipeline companies (Scale AI, Surge AI, OpenRLHF ecosystem), reward engineering and evaluation platforms.

Strong RL exposure (RL is a critical moat): Embodied AI and robotics policy companies (Physical Intelligence, Skild AI, Covariant), autonomous driving (Wayve, Waymo), AI chip design optimization.

Indirect beneficiaries (RL drives their demand): GPU compute providers, simulation platforms (NVIDIA Omniverse, Isaac Sim), human feedback and data labeling companies.

The most asymmetric bets are at the top -- RL-native infrastructure companies at seed and Series A, where the market has not yet assigned the premium it places on application-layer companies sitting on top.

What Could Make This Thesis Wrong

Three scenarios would invalidate the RL investment thesis:

RL gets replaced. If a fundamentally different approach to agent training emerges -- some form of self-supervised world modeling or direct reward-free learning -- that makes RL obsolete, the thesis collapses. This is not impossible. The "RL is dead, do this instead" crowd argues that hybrid approaches combining supervised and self-supervised learning will leapfrog RL's sample-efficiency problems. They may be right eventually.

RL stays internal. If every major lab keeps RL as a proprietary capability and no independent tooling ecosystem develops, there is no external category to invest in. This is the most likely failure mode. It mirrors what happened with many ML infrastructure categories that got absorbed into cloud platforms before they could become standalone markets.

Agents disappoint. If the agentic AI wave underdelivers -- if reliable multi-step autonomy remains 3-5 years away instead of 1-2 -- then RL's moment gets pushed further out, and the early infrastructure companies burn through their capital before the market arrives.

The Quiet Layer

Reinforcement learning will not generate the same breathless headlines as the next GPT release. It is not a consumer product. It does not have a chatbot interface.

But it is the layer that determines whether AI systems can act -- in factories, on roads, in financial markets, inside operating systems, across the physical world.

The skeptics are right that RL is hard, inefficient, and immature as a general-purpose technique. The bulls are right that it is already load-bearing in systems worth hundreds of billions. Both things are true simultaneously.

The gap between RL's strategic centrality and its current share of investment attention is the reinforcement gap. Whether that gap closes through a few breakout infrastructure companies or through quiet absorption into the major labs, the directional bet is clear: reinforcement learning is not a niche research method. It is the mechanism by which intelligence becomes agency.

And agency is where the value is.

The Reinforcement Gap: Why RL Might Be the Most Underpriced Layer in AI | ADIN