Research · Financial Press

Back to sweep

Research sweep · deep · 2023 – 2026

Token Cost of Ownership

AI token pricing vs true total cost of ownership from January 2023 to 19 April 2026, with emphasis on 2025–2026 signals: lab subsidisation strategies, infrastructure economics (compute, energy, data centres, hardware, security, ops), how user-facing prices have evolved, and analyst and researcher projections for token cost trajectories through 2028.

  • financial
  • frontier
  • academic
  • vc
  • blogs

Synthesised 2026-04-19

Narrative

Bloomberg has emerged as the dominant financial outlet for AI cost and infrastructure coverage. Its February 2026 news pieces quantified combined Big Tech AI capex at $650 billion for 2026 alone and characterised the associated debt financing as a '$3 trillion' market event — establishing the sheer scale of capital that undergirds current below-cost token pricing. Bloomberg's April 2025 newsletter marked the structural pivot from training to inference spend, while its January 2025 live video coverage of DeepSeek's market shock captured the moment financial markets first confronted the possibility that frontier-equivalent performance was achievable at 5–10% of incumbent labs' cost — a signal that the subsidy model could not persist indefinitely. Bloomberg's December 2024 interview with OpenAI's CFO, targeting $2,000/month enterprise subscriptions, revealed the commercial logic: pivot high-value users to flat-rate agreements that decouple revenue from per-token cost exposure. The April 2026 reporting on OpenAI pausing its UK Stargate data centre explicitly named energy cost as the binding constraint — confirming what Bloomberg Intelligence had warned: AI queries consume up to 10x the energy of traditional searches, making power a rising structural cost floor that token price declines cannot escape. Goldman Sachs' 2024 'Gen AI: Too Much Spend, Too Little Benefit?' report, featuring head of global equity research Jim Covello's scepticism that the projected $1 trillion in AI infrastructure would generate adequate returns, became the reference document for financial market bears — its core question about cost versus benefit remains unanswered in 2026.

The picture that emerges from combining financial press coverage with high-quality analytical sources (Epoch AI, SemiAnalysis, McKinsey) is one of a market in deliberate subsidy mode. Epoch AI's empirical tracking shows token prices for comparable capability falling 40x per year for some performance tiers since January 2024 — but frontier reasoning models have not followed commodity curves. Fortune's November 2025 reporting on OpenAI's leaked financial documents revealed the starkest number: $143 billion in projected cumulative cash outflow from 2024 to 2029, with inference costs at $8.4 billion in 2025 rising to $14.1 billion in 2026 — against a gross margin held below 35%. McKinsey's $7 trillion compute analysis and SemiAnalysis's bottom-up hardware economics confirm that the hardware and energy cost stack is rising, not falling, at the infrastructure level, even as algorithmic efficiency improvements drive per-token software costs down. The financial press story, as of April 2026, is that investor capital — not unit economics — is the primary determinant of what users pay per token, and the timeline to cost normalisation remains the central unresolved question for enterprise AI budgeting.


Sources

ID Title Outlet Date Significance
f1 How Much Is Big Tech Spending on AI Computing? A Staggering $650 Billion in 2026 Bloomberg 2026-02 Definitive Bloomberg News quantification of 2026 hyperscaler AI capex at $650B, establishing the scale of infrastructure investment that underpins current token pricing subsidies.
f2 The $3 Trillion AI Data Center Build-Out Becomes All-Consuming For Debt Markets Bloomberg 2026-02 Bloomberg's deep-dive into debt market financing of AI infrastructure, revealing the financial mechanics behind how data-centre construction costs are being funded and how that cost ultimately flows through to inference economics.
f3 OpenAI Pauses Stargate UK Data Center Citing Energy Costs Bloomberg 2026-04 Illustrates that energy cost constraints are already forcing project-level decisions at the frontier lab level, confirming that power is emerging as a binding cost floor for token pricing.
f4 AI Spending Boom Shifts From Training Models to Running Them Bloomberg 2025-04 Pivotal Bloomberg newsletter piece documenting the structural shift in AI capex from model training to inference workloads, the key transition defining the 2025–2026 cost and pricing landscape.
f5 Why AI Bubble Concerns Loom as OpenAI, Microsoft, Meta Ramp Up Spending Bloomberg 2025-11 Bloomberg synthesises mounting analyst concern that AI infrastructure investment is outpacing monetisation, directly relevant to whether current token prices can ever cover true costs.
f6 OpenAI Says Spending to Rise to $115B Through 2029 Bloomberg 2025-09 Bloomberg reporting on OpenAI's internal spending roadmap, confirming that compute cost trajectories are projected to rise sharply even as token prices are cut, widening the subsidy gap.
f7 Watch AI Cost Assumptions Challenged Bloomberg 2025-01 Bloomberg live coverage on the day of DeepSeek's market shock, capturing real-time financial market reaction to a rival model achieving near-frontier performance at a fraction of the cost, directly challenging incumbent pricing assumptions.
f8 OpenAI CFO Thinks Business Users Will Pay Thousands For AI Software Bloomberg 2024-12 Direct executive commentary from OpenAI's CFO on the enterprise pricing strategy, revealing the planned shift toward high-ARPU subscription models as an alternative to per-token revenue to fund infrastructure.
f9 Microsoft Sets Expensive Price Tag for New Corporate AI Products Bloomberg 2023-07 Early Bloomberg benchmarking of enterprise AI product pricing (Microsoft Copilot at $30/user/month), providing a 2023 baseline to measure how enterprise AI pricing models have evolved.
f10 AI Inferencing at Crossroads Bloomberg Intelligence 2025 Bloomberg Intelligence analysis of inference as the critical commercial battleground, detailing how model distillation and quantisation are reducing per-token costs while demand scaling offsets margin improvements.
f11 Big Tech 2025 Capex May Hit $200 Billion as Gen-AI Demand Booms Bloomberg Intelligence 2025 Bloomberg Intelligence capex projection establishing that 2025 hyperscaler infrastructure spend — the cost base that subsidises token pricing — would reach $200B, up sharply from prior years.
f12 AI Accelerator Market Looks Set to Exceed $600 Billion by 2033 Bloomberg Intelligence 2025 Bloomberg Intelligence market-sizing of the AI accelerator chip ecosystem ($116B in 2024 to $604B by 2033), quantifying the hardware cost trajectory underlying all token pricing models.
f13 AI Is a Game Changer for Power Demand Bloomberg Intelligence 2025 Bloomberg Intelligence analysis of how AI data centres are transforming energy markets, with generative AI queries consuming up to 10x the energy of traditional searches, establishing energy as a rising structural cost component.
f14 Gen AI: Too Much Spend, Too Little Benefit? Goldman Sachs 2024-06 Goldman Sachs' most-cited AI sceptic report, with head of global equity research questioning whether $1T in AI infrastructure can generate adequate returns; a key reference point for the 'cost vs. benefit' debate in financial markets.
f15 Will the $1 Trillion of Generative AI Investment Pay Off? Goldman Sachs 2024 Goldman Sachs investment research framing the core financial question around AI infrastructure: whether the capital cycle is commercially justifiable, directly informing how analysts assess the sustainability of below-cost token pricing.
f16 Why AI Companies May Invest More Than $500 Billion in 2026 Goldman Sachs 2026 Goldman Sachs' most current projection on AI infrastructure spending, providing a 2026 financial-market perspective on whether investment momentum is sustainable and what it implies for token cost floors.
f17 The Cost of Compute: A $7 Trillion Race to Scale Data Centers McKinsey & Company 2025 McKinsey's comprehensive bottom-up analysis of data centre cost structure, projecting $5.2T required investment through 2030, and decomposing the build cost into land, power, cooling, and compute components.
f18 The New Economics of Enterprise Technology in an AI World McKinsey & Company 2025 McKinsey's enterprise-facing analysis of how AI shifts IT spending from capex to opex, with FinOps and token-level cost visibility emerging as critical for managing true AI deployment TCO beyond API sticker prices.
f19 LLM Inference Prices Have Fallen Rapidly but Unequally Across Tasks Epoch AI 2025 The most rigorous empirical tracking of token price declines across performance tiers, documenting 9x–900x annual price drops depending on task and showing that frontier reasoning models have not followed commodity price trends.
f20 Inference Economics of Language Models Epoch AI 2024 Epoch AI's foundational decomposition of what drives LLM inference costs — hardware utilisation, model size, batch size, memory bandwidth — providing the analytical framework cited by financial analysts evaluating token pricing sustainability.
f21 AI Datacenter Energy Dilemma — Race for AI Datacenter Space SemiAnalysis 2024 SemiAnalysis's deep technical analysis of data-centre power constraints as a structural cost floor for AI inference, widely cited in financial press as the authoritative bottom-up view on infrastructure economics.
f22 Groq Inference Tokenomics: Speed, But At What Cost? SemiAnalysis 2024 SemiAnalysis cost-per-token breakdown for specialised inference hardware, quantifying real economics of serving tokens and demonstrating the gap between cloud-provider pricing and actual hardware cost.
f23 OpenAI Says It Plans to Report Stunning Annual Losses Through 2028 — and Then Turn Wildly Profitable Just Two Years Later Fortune 2025-11 Fortune's reporting on leaked OpenAI financial projections showing $44B cumulative losses before 2029 profitability — the definitive document source for quantifying how much investor capital is subsidising current token prices.
f24 Perspective: AI Demand Is Inflated, and Only Anthropic Is Being Realistic CNBC 2026-04 Most recent (April 2026) financial media critique of AI demand assumptions and token consumption projections, with direct commentary on Anthropic's more conservative pricing and demand forecasting relative to OpenAI and Nvidia.
f25 AI Training Costs Are Improving at 50x the Speed of Moore's Law ARK Invest 2023 ARK Invest's Wright's Law application to AI compute, projecting that AI training and inference costs decline at 50x the pace of Moore's Law — the bullish analytical counterpoint to Goldman Sachs' scepticism on AI cost trajectories.

We use analytics cookies to understand site usage and improve the service. We do not use marketing cookies.