Research · Financial Press
Back to sweepResearch sweep · deep · 2023 – 2026
Token Cost of Ownership
AI token pricing vs true total cost of ownership from January 2023 to 19 April 2026, with emphasis on 2025–2026 signals: lab subsidisation strategies, infrastructure economics (compute, energy, data centres, hardware, security, ops), how user-facing prices have evolved, and analyst and researcher projections for token cost trajectories through 2028.
- financial
- frontier
- academic
- vc
- blogs
Synthesised 2026-04-19
Narrative
Bloomberg has emerged as the dominant financial outlet for AI cost and infrastructure coverage. Its February 2026 news pieces quantified combined Big Tech AI capex at $650 billion for 2026 alone and characterised the associated debt financing as a '$3 trillion' market event — establishing the sheer scale of capital that undergirds current below-cost token pricing. Bloomberg's April 2025 newsletter marked the structural pivot from training to inference spend, while its January 2025 live video coverage of DeepSeek's market shock captured the moment financial markets first confronted the possibility that frontier-equivalent performance was achievable at 5–10% of incumbent labs' cost — a signal that the subsidy model could not persist indefinitely. Bloomberg's December 2024 interview with OpenAI's CFO, targeting $2,000/month enterprise subscriptions, revealed the commercial logic: pivot high-value users to flat-rate agreements that decouple revenue from per-token cost exposure. The April 2026 reporting on OpenAI pausing its UK Stargate data centre explicitly named energy cost as the binding constraint — confirming what Bloomberg Intelligence had warned: AI queries consume up to 10x the energy of traditional searches, making power a rising structural cost floor that token price declines cannot escape. Goldman Sachs' 2024 'Gen AI: Too Much Spend, Too Little Benefit?' report, featuring head of global equity research Jim Covello's scepticism that the projected $1 trillion in AI infrastructure would generate adequate returns, became the reference document for financial market bears — its core question about cost versus benefit remains unanswered in 2026.
The picture that emerges from combining financial press coverage with high-quality analytical sources (Epoch AI, SemiAnalysis, McKinsey) is one of a market in deliberate subsidy mode. Epoch AI's empirical tracking shows token prices for comparable capability falling 40x per year for some performance tiers since January 2024 — but frontier reasoning models have not followed commodity curves. Fortune's November 2025 reporting on OpenAI's leaked financial documents revealed the starkest number: $143 billion in projected cumulative cash outflow from 2024 to 2029, with inference costs at $8.4 billion in 2025 rising to $14.1 billion in 2026 — against a gross margin held below 35%. McKinsey's $7 trillion compute analysis and SemiAnalysis's bottom-up hardware economics confirm that the hardware and energy cost stack is rising, not falling, at the infrastructure level, even as algorithmic efficiency improvements drive per-token software costs down. The financial press story, as of April 2026, is that investor capital — not unit economics — is the primary determinant of what users pay per token, and the timeline to cost normalisation remains the central unresolved question for enterprise AI budgeting.
Sources
| ID | Title | Outlet | Date | Significance |
|---|---|---|---|---|
| f1 | How Much Is Big Tech Spending on AI Computing? A Staggering $650 Billion in 2026 | Bloomberg | 2026-02 | Definitive Bloomberg News quantification of 2026 hyperscaler AI capex at $650B, establishing the scale of infrastructure investment that underpins current token pricing subsidies. |
| f2 | The $3 Trillion AI Data Center Build-Out Becomes All-Consuming For Debt Markets | Bloomberg | 2026-02 | Bloomberg's deep-dive into debt market financing of AI infrastructure, revealing the financial mechanics behind how data-centre construction costs are being funded and how that cost ultimately flows through to inference economics. |
| f3 | OpenAI Pauses Stargate UK Data Center Citing Energy Costs | Bloomberg | 2026-04 | Illustrates that energy cost constraints are already forcing project-level decisions at the frontier lab level, confirming that power is emerging as a binding cost floor for token pricing. |
| f4 | AI Spending Boom Shifts From Training Models to Running Them | Bloomberg | 2025-04 | Pivotal Bloomberg newsletter piece documenting the structural shift in AI capex from model training to inference workloads, the key transition defining the 2025–2026 cost and pricing landscape. |
| f5 | Why AI Bubble Concerns Loom as OpenAI, Microsoft, Meta Ramp Up Spending | Bloomberg | 2025-11 | Bloomberg synthesises mounting analyst concern that AI infrastructure investment is outpacing monetisation, directly relevant to whether current token prices can ever cover true costs. |
| f6 | OpenAI Says Spending to Rise to $115B Through 2029 | Bloomberg | 2025-09 | Bloomberg reporting on OpenAI's internal spending roadmap, confirming that compute cost trajectories are projected to rise sharply even as token prices are cut, widening the subsidy gap. |
| f7 | Watch AI Cost Assumptions Challenged | Bloomberg | 2025-01 | Bloomberg live coverage on the day of DeepSeek's market shock, capturing real-time financial market reaction to a rival model achieving near-frontier performance at a fraction of the cost, directly challenging incumbent pricing assumptions. |
| f8 | OpenAI CFO Thinks Business Users Will Pay Thousands For AI Software | Bloomberg | 2024-12 | Direct executive commentary from OpenAI's CFO on the enterprise pricing strategy, revealing the planned shift toward high-ARPU subscription models as an alternative to per-token revenue to fund infrastructure. |
| f9 | Microsoft Sets Expensive Price Tag for New Corporate AI Products | Bloomberg | 2023-07 | Early Bloomberg benchmarking of enterprise AI product pricing (Microsoft Copilot at $30/user/month), providing a 2023 baseline to measure how enterprise AI pricing models have evolved. |
| f10 | AI Inferencing at Crossroads | Bloomberg Intelligence | 2025 | Bloomberg Intelligence analysis of inference as the critical commercial battleground, detailing how model distillation and quantisation are reducing per-token costs while demand scaling offsets margin improvements. |
| f11 | Big Tech 2025 Capex May Hit $200 Billion as Gen-AI Demand Booms | Bloomberg Intelligence | 2025 | Bloomberg Intelligence capex projection establishing that 2025 hyperscaler infrastructure spend — the cost base that subsidises token pricing — would reach $200B, up sharply from prior years. |
| f12 | AI Accelerator Market Looks Set to Exceed $600 Billion by 2033 | Bloomberg Intelligence | 2025 | Bloomberg Intelligence market-sizing of the AI accelerator chip ecosystem ($116B in 2024 to $604B by 2033), quantifying the hardware cost trajectory underlying all token pricing models. |
| f13 | AI Is a Game Changer for Power Demand | Bloomberg Intelligence | 2025 | Bloomberg Intelligence analysis of how AI data centres are transforming energy markets, with generative AI queries consuming up to 10x the energy of traditional searches, establishing energy as a rising structural cost component. |
| f14 | Gen AI: Too Much Spend, Too Little Benefit? | Goldman Sachs | 2024-06 | Goldman Sachs' most-cited AI sceptic report, with head of global equity research questioning whether $1T in AI infrastructure can generate adequate returns; a key reference point for the 'cost vs. benefit' debate in financial markets. |
| f15 | Will the $1 Trillion of Generative AI Investment Pay Off? | Goldman Sachs | 2024 | Goldman Sachs investment research framing the core financial question around AI infrastructure: whether the capital cycle is commercially justifiable, directly informing how analysts assess the sustainability of below-cost token pricing. |
| f16 | Why AI Companies May Invest More Than $500 Billion in 2026 | Goldman Sachs | 2026 | Goldman Sachs' most current projection on AI infrastructure spending, providing a 2026 financial-market perspective on whether investment momentum is sustainable and what it implies for token cost floors. |
| f17 | The Cost of Compute: A $7 Trillion Race to Scale Data Centers | McKinsey & Company | 2025 | McKinsey's comprehensive bottom-up analysis of data centre cost structure, projecting $5.2T required investment through 2030, and decomposing the build cost into land, power, cooling, and compute components. |
| f18 | The New Economics of Enterprise Technology in an AI World | McKinsey & Company | 2025 | McKinsey's enterprise-facing analysis of how AI shifts IT spending from capex to opex, with FinOps and token-level cost visibility emerging as critical for managing true AI deployment TCO beyond API sticker prices. |
| f19 | LLM Inference Prices Have Fallen Rapidly but Unequally Across Tasks | Epoch AI | 2025 | The most rigorous empirical tracking of token price declines across performance tiers, documenting 9x–900x annual price drops depending on task and showing that frontier reasoning models have not followed commodity price trends. |
| f20 | Inference Economics of Language Models | Epoch AI | 2024 | Epoch AI's foundational decomposition of what drives LLM inference costs — hardware utilisation, model size, batch size, memory bandwidth — providing the analytical framework cited by financial analysts evaluating token pricing sustainability. |
| f21 | AI Datacenter Energy Dilemma — Race for AI Datacenter Space | SemiAnalysis | 2024 | SemiAnalysis's deep technical analysis of data-centre power constraints as a structural cost floor for AI inference, widely cited in financial press as the authoritative bottom-up view on infrastructure economics. |
| f22 | Groq Inference Tokenomics: Speed, But At What Cost? | SemiAnalysis | 2024 | SemiAnalysis cost-per-token breakdown for specialised inference hardware, quantifying real economics of serving tokens and demonstrating the gap between cloud-provider pricing and actual hardware cost. |
| f23 | OpenAI Says It Plans to Report Stunning Annual Losses Through 2028 — and Then Turn Wildly Profitable Just Two Years Later | Fortune | 2025-11 | Fortune's reporting on leaked OpenAI financial projections showing $44B cumulative losses before 2029 profitability — the definitive document source for quantifying how much investor capital is subsidising current token prices. |
| f24 | Perspective: AI Demand Is Inflated, and Only Anthropic Is Being Realistic | CNBC | 2026-04 | Most recent (April 2026) financial media critique of AI demand assumptions and token consumption projections, with direct commentary on Anthropic's more conservative pricing and demand forecasting relative to OpenAI and Nvidia. |
| f25 | AI Training Costs Are Improving at 50x the Speed of Moore's Law | ARK Invest | 2023 | ARK Invest's Wright's Law application to AI compute, projecting that AI training and inference costs decline at 50x the pace of Moore's Law — the bullish analytical counterpoint to Goldman Sachs' scepticism on AI cost trajectories. |