After researching and writing The Intelligence Advantage Report—26 chapters, 7 parts, four technology revolutions spanning 140 years—certain patterns become impossible to ignore. The data doesn’t just suggest what’s coming. It demands it.
These thirteen predictions are grounded in the report’s core thesis: when infrastructure commoditizes, value migrates to the application layer. It happened with electricity. It happened with telecom. It happened with silicon. Artificial intelligence is next—and it’s moving faster than any of them.
The cost per million tokens is falling from $60 in 2020 to an estimated $0.12 by 2030—a 99.8% reduction. Within four years, ninety-five percent of cognitive tasks will approach near-zero marginal cost. But the real story is not the cost collapse itself. It is what happens next.
As intelligence commoditizes, value migrates from the infrastructure that produces it to the applications that use it. We have seen this pattern before—with electricity, with telecom, with silicon. Today, 91% of AI value sits in the infrastructure layer. By roughly 2031, the inversion point arrives: applications will capture more value than infrastructure for the first time. By 2035, the split reaches 80/20 in favor of applications, and the application layer grows 116 times—from $13 billion to $1.5 trillion. The companies that build on intelligence, not the ones that generate it, will capture the dominant share of value.
Ch 1–2: Intelligence Yield & Cost Curve · Ch 5–6: Value Migration
This is a prediction that few are willing to make, but the structural evidence is overwhelming. Google is the only company on Earth that controls the full stack simultaneously—custom silicon (TPUs, backed by $185 billion in capital expenditure in 2026 alone), frontier models (Gemini), and the largest application ecosystem ever assembled: Search, YouTube, Android, Chrome, Workspace, Cloud, Maps, and Waymo.
The report’s central thesis is that value migrates from infrastructure to applications. Consider the competitive landscape through that lens. OpenAI has models but no distribution. Meta has distribution but no cloud. Microsoft has cloud but no models of its own. Google has all three layers. With 4.3 billion Chrome users, 2 billion Android devices, and 8.5 billion daily searches, they can embed AI into every interaction on Earth without acquiring a single new customer. DeepMind remains the world’s strongest AI research lab. Google is not just playing the AI game—they own the board.
Ch 5–6: Value Migration · Ch 16–18: Infrastructure
One hundred and three billion dollars in SaaS revenue will be displaced by AI agents by 2030. Customer support hits near-total disruption by 2027. CRM and Sales face $25 billion at risk. The per-seat pricing model—the foundation of the SaaS era—is already dying: it declined from 21% to 15% of pricing models in just twelve months, while outcome-based pricing surged from 27% to 41%.
The early casualties are already visible. Jasper, once a darling of AI-powered content generation, saw a 54% revenue decline as free alternatives emerged. The underlying math is existential: when a $0.99 AI agent resolves a support ticket more effectively than a $39-per-seat-per-month SaaS tool, the incumbents don’t have a pricing problem. They have a survival problem.
Ch 19–20: Enterprise Disruption & SaaS Timeline
By 2028, the provider cost per user drops to $0.10–$0.50, but users are still paying $20—a 97 to 99% margin that competition will inevitably destroy. No technology market has sustained margins like that once alternatives emerge, and alternatives are emerging fast.
The $20 price point may survive, but only if it delivers radically more value. Think broadband, not dial-up. When home internet cost $50 a month in the late 1990s, it bought you email and slow web pages. Today the same $50 buys unlimited streaming, cloud gaming, video calls, and terabytes of storage. AI subscriptions will follow the same trajectory: to justify the price, providers will need to bundle dozens of agents, unlimited generation, priority access, and premium orchestration across an entire workflow. A single chatbot behind a paywall will not cut it.
Those who fail to expand the value stack will see their subscriptions collapse to $3–5 per month or migrate to per-outcome pricing—$0.99 per legal brief, per resolved ticket, per generated campaign. The Klarna model, which saved $60 million annually by replacing 853 human agents, is becoming the standard. The era of charging $240 a year for intelligence that costs pennies to deliver is ending.
Ch 2: Cost Curve · Ch 23: Agentic Economy
The performance gap between frontier and open-source models has all but vanished. On the MMLU benchmark, the gap collapsed from 17.5 points to just 0.3 points. DeepSeek trained a model to GPT-4-level performance for $5.6 million—99.5% cheaper than what frontier labs spent to reach the same capability.
This is not an accident. It is strategy. Meta has committed $72 billion and Alibaba $53 billion—both are open-sourcing their models as strategic weapons to commoditize the model layer. When the model itself becomes a commodity, the competitive moat moves upstream to proprietary data and downstream to orchestration and workflow. By 2030, asking which model a company uses will be like asking which database engine powers their website. The answer won’t matter. What matters is what you build on top of it.
Ch 3–4: Convergence · Ch 15: Open vs Closed
The agentic AI market reaches $50 billion by 2030 and $183 billion by 2033—a 46% compound annual growth rate. The transition is already underway and follows a clear sequence: copilots assisting humans (2023–25), agents executing tasks autonomously (2025–27), and fully autonomous workflows operating without human intervention (2027–30).
The early proof points are striking. Salesforce Agentforce has reached $800 million in annual recurring revenue, executing 2 billion autonomous actions per month. Cursor, an AI-native code editor, hit $2 billion in ARR. At 296 trillion tokens processed per day by 2030, the browser-and-click interface gives way to agent-and-intent. You will not visit a website to book a flight. Your agent will do it while you sleep, cross-referencing your preferences, loyalty programs, and calendar in a single pass.
Ch 22: Routing · Ch 23: Agentic Economy
The West’s chip export controls were designed to slow China’s AI progress by restricting access to the most advanced GPUs. The unintended consequence has been the opposite: the constraints have accelerated Chinese innovation in algorithmic efficiency—the art of doing more with less silicon.
DeepSeek V3 reached GPT-4-level performance for $5.6 million in training compute, while Western labs spent over $200 million to reach the same point. Huawei’s Ascend 910B delivers 50–60% of H100 performance on domestic silicon. Alibaba has committed $53 billion over three years. Meanwhile, 90% of the world’s advanced chips still flow through TSMC in Taiwan—a single point of geopolitical failure that cuts both ways. Export controls may have bought time, but they have not bought advantage. By 2030, capability parity is not a question of if, but of when.
Ch 25: Geopolitics · Ch 15: Open vs Closed
A twelve-agent corporate intelligence system costs approximately $500,000 per year by its third year of operation—and that cost is declining 40–50% annually. Fine-tuned 7-billion-parameter models already beat GPT-4 on 85% of targeted tasks at $0.03 per million tokens. The economics are irresistible.
The key architectural insight is the 70/20/10 routing model: 70% of tasks handled by tiny, specialized agents; 20% by mid-tier models; and only 10% escalated to frontier reasoning. This architecture delivers a 5–10x cost advantage over routing everything through a single large model. By 2030, a mid-size company will run hundreds of specialized agents handling customer service, finance, HR, legal, and marketing—for less than the cost of ten employees. The question will not be whether to deploy AI agents, but how many and how fast.
Ch 12: Small Models · Ch 22: Routing
Three hundred million jobs are exposed to AI globally. The headline is alarming, but the data tells a more nuanced story. The World Economic Forum projects 170 million new roles created versus 92 million displaced—a net gain of 78 million jobs. The issue is not the quantity of work. It is the quality of the transition.
Fifty-nine percent of the global workforce—1.2 billion workers—needs reskilling. The disruption gradient is clear: routine cognitive tasks are 85% automatable, but complex judgment is only 12% automatable. And crucially, the Dallas Federal Reserve found that wages in AI-exposed industries actually rose 8.5–16.7%, suggesting augmentation rather than replacement. The real risk is not mass unemployment. It is a $102.8 billion skills gap that no government or corporation is funding fast enough.
The GPU shortage that defined 2023 and 2024 is resolving itself. TSMC’s CoWoS advanced packaging capacity has expanded from 15,000 to 130,000 wafers per month. The silicon constraint is easing. But a far larger constraint is emerging in its place: electricity.
Energy demand for AI infrastructure is growing 15–20% annually. A single GB200 NVL72 rack consumes 72 kilowatts. By 2030, AI data centers will need 286 terawatt-hours per year—equivalent to 62% of current global data center electricity consumption. The next trillion-dollar constraint is not the chip on the board but the power plant behind the wall. The companies that secure energy supplies—not compute capacity—will ultimately control the AI supply chain.
Ch 16–18: GPU, NVIDIA & Inference
Healthcare AI is projected to grow from $36.7 billion to $187.7 billion by 2030—a 38.6% CAGR, the highest of any sector. The catalyst is not technological ambition but demographic necessity: the world faces a shortage of more than 10 million doctors by 2030. AI is not optional in healthcare. It is existential.
Fine-tuned specialist models already dominate the sector, accounting for 45% of healthcare AI deployments. Finance and healthcare combined represent 62% of the entire $607 billion enterprise AI market. AI-powered diagnostics, drug discovery pipelines that compress timelines from years to months, and personalized treatment plans generated from a patient’s complete medical history—these advances will save more lives than any medical breakthrough since antibiotics.
The era of “one giant model for everything” is ending. Fine-tuned 7-billion-parameter models already beat GPT-4 on 85% of targeted tasks at $0.03 per million tokens—roughly one hundredth the cost of frontier inference. This is not a compromise. For most real-world applications, the specialist outperforms the generalist.
As intelligence encoding matures—the science of compressing frontier capabilities into smaller, purpose-built architectures—AI moves off the cloud and onto the device. Phones, cars, medical instruments, factory floors, retail kiosks. Edge AI eliminates latency, removes cloud dependency, and enables privacy-first deployments in regulated industries. By 2030, most AI interactions will not touch a data center. They will run locally, on billions of devices, invisible to the user.
The implications for the industry are profound. Frontier models evolve into “teachers” that train armies of specialized students. The competitive moat shifts from model size to fine-tuning expertise, proprietary training data, and the orchestration of hundreds of purpose-built models across the 70/20/10 routing architecture. The company that masters intelligence encoding—not the one that trains the largest model—wins.
Ch 12: Small Models · Ch 3–4: Convergence · Ch 22: Routing
This is the prediction I most hope will be proven wrong. But the trajectory is clear, and the data is unforgiving.
The EU AI Act creates compliance costs that slow innovation velocity at exactly the moment when the United States and China are accelerating. The US fragments regulation across states but moves fast. China governs with state alignment but invests at staggering scale—Saudi Arabia’s Project Transcendence alone represents $100 billion. Europe, by contrast, regulates first and innovates second.
By 2030, zero European companies will appear in the global AI top 20. The regulatory arbitrage window—the period during which European startups could relocate to more permissive jurisdictions—closes by 2028. But the innovation gap compounds year over year, making it exponentially harder to close with each passing quarter.
This is not an argument against regulation. It is an argument against regulation as a substitute for strategy. Europe must choose between regulating AI and building AI. Right now, it is choosing regulation. That choice, if unchanged, will be permanent.
Ch 24: Regulation · Ch 25: Geopolitics
These predictions are not speculation. They are extrapolations from patterns that have repeated across every major technology revolution of the past 140 years. The question is not whether these shifts will happen, but whether we will be ready when they do.
The full data, methodology, and interactive dashboards behind each of these predictions are available in The Intelligence Advantage Report.
26 chapters of original research with interactive dashboards and data.
Read The Intelligence Advantage Report →