Geopolitics of Intelligence

Sovereign AI and the Compute Arms Race

Two AI stacks are emerging — US-allied and China-aligned. Export controls, sovereign infrastructure, and supply chain concentration are fragmenting the global AI ecosystem. Enterprises must choose which stack to build on.

90%
Advanced chips from one company (TSMC)
$53B
China AI investment (Alibaba 3yr)
3 Rounds
US export control escalation
$100B+
Sovereign AI commitments globally
Part VII — The Road Ahead
Chapter 25: The Geopolitics of Intelligence

The AI revolution is not unfolding on a single, global playing field. It is fracturing along geopolitical fault lines that will determine which enterprises can access which compute, which models, and which markets for the next decade.

Throughout this report, we have treated the AI stack as a unified global system: NVIDIA designs chips, TSMC fabricates them, hyperscalers deploy them, and enterprises everywhere consume the resulting intelligence. That framing is increasingly incomplete. Three rounds of US export controls (October 2022, October 2023, January 2025) have systematically restricted China’s access to advanced AI chips. China has responded by investing billions in domestic alternatives, accelerating open-source model development (Chapter 15), and building parallel compute infrastructure. The EU, Middle East, and India are launching sovereign AI initiatives to reduce dependence on both US and Chinese technology stacks.

The result is an emerging bifurcation: a US-allied AI ecosystem (NVIDIA, AMD, TSMC, US hyperscalers) and a China-aligned ecosystem (Huawei Ascend, SMIC, domestic clouds). For multinational enterprises, this bifurcation creates strategic choices that cannot be deferred: which chips to deploy, which models to use, which data sovereignty rules to comply with, and which markets to serve from which infrastructure. The wrong choice locks you into a stack that may lose access to key markets. The right choice — or more precisely, the right dual-stack strategy — positions you for both ecosystems.

1. The Export Control Escalation

Three rounds of increasingly restrictive US controls have reshaped the global AI chip market.

US AI Chip Export Control Timeline

Each round broadened the scope and closed loopholes exploited after the previous round.
October 2022 — Round 1
Initial restrictions on advanced chips to China
Banned export of chips with >300 TOPS (int8) or >4800 TOPS (FP16) interconnect bandwidth. Targeted NVIDIA A100 and H100. Also restricted semiconductor equipment (ASML EUV lithography). NVIDIA responded with China-specific A800 and H800 variants with reduced interconnect speeds.
October 2023 — Round 2
Closed loopholes, expanded country coverage
Lowered performance thresholds to capture China-specific variants (A800, H800). Extended restrictions to 40+ countries. Added chip design software (EDA tools) restrictions. NVIDIA created further downgraded H20 chip for China market.
January 2025 — Round 3
Framework-based controls with global reach
Introduced three-tier country classification system. Tier 1 (allies): unrestricted. Tier 2 (most countries): capped at 50,000 GPUs per entity. Tier 3 (arms-embargoed + China/Russia): near-total ban. Also restricted advanced model weights and closed-source AI system exports to Tier 3.

Impact on NVIDIA’s China Business

Before export controls, China represented approximately 25% of NVIDIA’s data centre revenue. The A100 and H100 were the standard training accelerators for Baidu, Alibaba, Tencent, and ByteDance. After Round 1, NVIDIA created the A800 and H800 with reduced interconnect bandwidth — a temporary workaround that generated billions in revenue before Round 2 closed the loophole. After Round 2, NVIDIA introduced the H20, a further downgraded chip with roughly 20% of the H100’s AI training performance but priced competitively for inference workloads.

The financial impact has been significant but not catastrophic. NVIDIA’s China data centre revenue declined from an estimated $10–12 billion annually to approximately $5–7 billion, largely through H20 sales. The company has compensated with accelerating demand from US hyperscalers and sovereign AI projects globally. But the strategic impact is deeper than revenue: every chip denied to China accelerates China’s investment in domestic alternatives.

The Paradox of Export Controls: Each restriction accelerates China’s self-sufficiency. Before October 2022, Chinese labs had little incentive to switch from NVIDIA. By 2025, Huawei’s Ascend 910B is in production, SMIC is pushing 7nm process boundaries, and DeepSeek demonstrated that algorithmic innovation can partially compensate for hardware limitations. Export controls may delay China’s AI capability by 2–3 years, but they also guarantee that China builds a fully sovereign AI stack.

2. China’s Response: Building the Parallel Stack

A comprehensive national effort to achieve AI self-sufficiency across chips, models, and infrastructure.

Domestic Chip Ecosystem

Huawei Ascend 910B/C
Most advanced Chinese AI accelerator. Manufactured by SMIC on 7nm process. Performance estimated at 50–60% of NVIDIA H100 for training. Deployed across Baidu, China Telecom, and government projects. Ascend 910C aims to close the gap further.

SMIC
China’s leading foundry, producing 7nm chips without EUV lithography (using multi-patterning DUV). Yield rates reportedly improving. 5nm remains out of reach without EUV access.

Cambricon, Biren, Enflame
Second-tier Chinese chip designers. Training-focused (Cambricon) and inference-focused (others). Collectively addressing different segments of the compute stack.

The DeepSeek Model

Algorithmic Compensation
DeepSeek V3 achieved GPT-4-level performance with a reported $5.6M training compute cost (Chapter 15). Whether trained on restricted NVIDIA GPUs stockpiled before export controls or on domestic hardware, the key insight is that algorithmic innovation can partially substitute for hardware superiority.

Open-Source Strategy
By releasing models open-weight, DeepSeek ensures global adoption regardless of geopolitical alignment. Their models run on any hardware, creating a parallel model ecosystem that is not dependent on US companies.

Full-Stack Ambition
DeepSeek, backed by High-Flyer quantitative fund ($8B AUM), represents a model where AI research is funded by trading profits rather than venture capital or government grants — a sustainable alternative that is not dependent on external capital markets.

3. The Global Sovereign AI Landscape

Beyond the US-China axis, every major region is investing in AI independence.

Sovereign AI Investment Commitments

National and regional AI infrastructure investments announced or committed as of early 2026.

European Union

EU Chips Act: €43B ($47B) investment in semiconductor manufacturing and R&D through 2030.

Mistral: French AI champion. Raised €600M+ in 2024, valued at $6B. Close ties to French government. Open-weight models (Mixtral, Mistral Large).

Aleph Alpha: German AI company focused on sovereign, privacy-preserving AI for European enterprises and governments.

EuroHPC: Pan-European supercomputing infrastructure. JUPITER (Jülich) among world’s most powerful.

Middle East

Saudi Arabia: $100B+ AI and tech commitment through Vision 2030. NEOM as AI-integrated city. Project Transcendence ($100B AI fund announced 2025).

UAE: Technology Innovation Institute (TII) created Falcon series. G42 partnership with Microsoft. Positioned as AI hub between US and China ecosystems.

Qatar: AI-driven World Cup infrastructure. Qatar Computing Research Institute (QCRI).

Asia-Pacific

India: IndiaAI Mission ($1.2B budget). 10,000 GPU compute infrastructure. Focus on local language AI and public services.

Japan: $13B AI investment plan. ABCI supercomputer expansion. RIKEN research institute.

South Korea: $7B AI semiconductor initiative. Samsung foundry advanced nodes.

Singapore: National AI Strategy 2.0. AI Verify governance framework.

Compute Sovereignty vs Model Sovereignty

The sovereign AI movement reveals a strategic distinction that many governments have not yet resolved: the difference between compute sovereignty (controlling the physical infrastructure — chips, data centres, power) and model sovereignty (controlling the AI models — training data, weights, deployment).

Compute sovereignty requires semiconductor fabrication capability, which only Taiwan (TSMC), South Korea (Samsung), the US (Intel), and to a limited extent China (SMIC) and the EU possess. For most countries, compute sovereignty means securing long-term supply agreements and building domestic data centres, not fabricating chips. Model sovereignty is more accessible: any country with sufficient compute (purchased or leased) and AI talent can train models on local data in local languages. The open-source ecosystem (Chapter 15) has dramatically lowered the barrier to model sovereignty — fine-tuning a Llama or Qwen base model for a specific language or domain costs orders of magnitude less than training from scratch.

The strategic implication for enterprises: compute sovereignty is a government problem; model sovereignty is an enterprise opportunity. Companies that build domain-specific, locally-compliant AI systems on top of open-source base models can serve sovereign AI markets without owning chip fabrication facilities. This is exactly the application-layer value capture thesis of this entire report, applied to geopolitics.

4. The TSMC Concentration Risk

The entire AI revolution depends on one company on one island.

Advanced Semiconductor Manufacturing Concentration

TSMC produces over 90% of the world’s most advanced chips (<7nm). This concentration creates the single largest supply chain risk in the AI ecosystem.

The arithmetic is stark. NVIDIA designs AI chips. TSMC fabricates them. Approximately 90% of chips at the most advanced process nodes (<7nm) come from TSMC’s fabs in Taiwan. The entire AI infrastructure buildout described in Chapters 16–18 — the $1.7 trillion compute stack, the 72 million H100-equivalents by 2030 — flows through TSMC’s fabrication facilities.

Taiwan sits 130 kilometres from mainland China in one of the world’s most geopolitically sensitive zones. Any disruption to TSMC’s operations — whether from military conflict, natural disaster, or political coercion — would create a global AI chip shortage that no amount of demand-side adjustment could mitigate. The AI industry has a single point of failure, and it is located in a potential conflict zone.

Diversification efforts are underway but slow. TSMC is building fabs in Arizona (first chips expected 2025–2026, full production 2028), Japan (Kumamoto, operational 2024 for mature nodes), and Germany (Dresden, planned 2027). Intel is expanding domestic US fabrication. Samsung is investing in Taylor, Texas. But advanced-node fabrication cannot be relocated quickly — a new fab takes 3–5 years from groundbreaking to production, and the most advanced processes (3nm, 2nm) remain in Taiwan for now.

Enterprise Risk Mitigation: Companies cannot control TSMC concentration risk, but they can reduce their exposure by: (1) diversifying across chip vendors (NVIDIA + AMD + custom ASICs), (2) building model architectures that run efficiently on multiple hardware platforms, (3) maintaining inference capacity across geographically distributed data centres, and (4) investing in model compression and efficiency to reduce per-query chip dependency.

5. Enterprise Strategy: The Dual-Stack Imperative

For multinational enterprises, the question is not which stack to choose — it is how to operate across both.

US-Allied Stack

Chips: NVIDIA (H100/B200), AMD (MI300X), Intel Gaudi

Cloud: AWS, Azure, GCP

Models: GPT-4o/5, Claude, Gemini

Open-source: Llama (Meta), Mistral (EU)

Standards: NIST AI RMF, EU AI Act, ISO 42001

Data rules: GDPR, US state privacy laws

China-Aligned Stack

Chips: Huawei Ascend 910B/C, Cambricon

Cloud: Alibaba Cloud, Tencent Cloud, Huawei Cloud

Models: Qwen (Alibaba), Ernie (Baidu), GLM (Zhipu)

Open-source: Qwen, DeepSeek, Yi (01.AI)

Standards: CAC algorithm registry, GB/T standards

Data rules: PIPL, data localisation requirements

Enterprises serving both US/EU and Chinese markets need a dual-stack architecture. This does not mean maintaining two completely separate AI systems. It means designing a modular architecture where:

  • Application logic is stack-agnostic (the routing, workflow, and business logic runs anywhere)
  • Model integration uses abstraction layers (so swapping GPT-4o for Qwen or Ernie requires configuration, not re-engineering)
  • Data handling complies with local sovereignty requirements by design (Chinese data stays in China, EU data stays in EU)
  • Compute is provisioned from local cloud providers (AWS in the US, Alibaba Cloud in China, local providers in regulated markets)

This architecture is precisely the intelligence routing framework described in Chapter 22, extended to include geopolitical routing. The router does not just select the optimal model for the task — it selects the optimal model that is legally available and data-compliant for the user’s jurisdiction.

What Comes Next

The geopolitical fragmentation of the AI stack creates winners and losers at the enterprise level. But it also creates winners and losers at the workforce level. Chapter 26 examines the other dimension of “what comes next”: the impact of AI on employment, wages, skills, and the very definition of work — building on the task-level analysis from Chapter 21 to map the human consequences of the disruption this report has documented.