代幣海嘯:為何一筆22億美元的AGI基礎設施賭注重新定義了AI軍備競賽

May 2026
Archive: May 2026
當業界專注於模型參數量的競賽時,一場更深層的危機正在醞釀:代幣消耗量即將暴增千倍。一家AGI基礎設施公司已獲得22億美元資金,押注於一個觀點——實現AGI的瓶頸並非智慧本身,而是代幣的成本與延遲。
The article body is currently shown in English by default. You can generate the full version in this language on demand.

The AI industry's current obsession with model parameter scaling is masking a fundamental shift. The next frontier is not making models smarter in isolation, but making them usable at scale. The convergence of three trends — multimodal models that process video, audio, and text simultaneously; real-time autonomous agents that operate 24/7; and world simulators that generate continuous 3D environments — will drive token demand from billions to trillions per day. A single AGI infrastructure company, which we will refer to as 'InfraCo' for this analysis, has raised $2.2 billion to preempt this demand. Their strategy is radical vertical integration: designing custom chips, proprietary cooling systems, and novel data center architectures to push token density and efficiency to physical limits. This is not a bet on current workloads; it is a bet on a future where AGI emerges not from a single breakthrough, but from the relentless scaling of token throughput. If correct, InfraCo will become the 'water and electricity' of the AI ecosystem, with strategic value surpassing any single model company. The article dissects the technical architecture, the key players, the market dynamics, and the risks of this high-stakes wager.

Technical Deep Dive

The core insight driving InfraCo's strategy is a redefinition of the scaling laws. The original scaling laws (Kaplan et al., 2020) focused on model parameters and training compute. InfraCo's thesis is that the next scaling law is about *inference throughput* — specifically, the cost and latency of generating a single token. This shift is driven by three technical realities:

1. Multimodal Token Explosion: A single frame of 1080p video, when tokenized by a modern Vision Transformer (ViT) like those used in Sora or Gemini, requires roughly 1,000-2,000 tokens. At 30 frames per second, one minute of video generates 1.8-3.6 million tokens. Compare this to a 100,000-word novel, which is roughly 130,000 tokens. A single minute of high-quality video generation consumes 14-28 times more tokens than an entire book. As models like Meta's Movie Gen and Google's Veo 2 push towards higher resolutions and longer durations, this ratio will worsen.

2. Real-Time Agent Loops: Autonomous agents, such as those built on frameworks like LangChain or CrewAI, do not just answer a single query. They operate in loops: perceive, reason, act, observe. A single task — say, booking a complex travel itinerary — might involve 50-100 internal reasoning steps, each requiring a model call. If the agent runs continuously (e.g., a personal AI assistant monitoring your inbox), the token consumption becomes a constant stream rather than discrete bursts. Early estimates from projects like AutoGPT suggest that a moderately complex agent task can consume 10-100x more tokens than a simple chatbot interaction.

3. World Simulators: The ultimate token sink is real-time 3D world generation. A world simulator like the one proposed by Fei-Fei Li's World Labs or the underlying tech in NVIDIA's Omniverse must generate a consistent, interactive 3D environment at 60+ frames per second. Each frame is not just an image; it includes geometry, physics, lighting, and object interactions, all tokenized. A single second of a simulated world could require 10-100 million tokens. This is the domain where token demand becomes truly astronomical.

InfraCo's Vertical Integration Approach:

To meet this demand, InfraCo is pursuing a strategy of extreme vertical integration, reminiscent of Apple's approach to the iPhone but applied to AI infrastructure.

- Custom Silicon (ASICs): InfraCo has designed its own inference-optimized chip, codenamed 'TensorCore-X'. Unlike NVIDIA's H100/B200, which are general-purpose for both training and inference, TensorCore-X is a pure inference engine. It strips out unnecessary training-specific circuitry (like FP64 tensor cores) and replaces them with massive SRAM banks and a novel systolic array optimized for the sparse attention patterns common in modern LLMs. Early leaks suggest a 3x improvement in tokens-per-watt over the B200 for inference workloads.

- Memory-Centric Architecture: The biggest bottleneck in inference is memory bandwidth (the 'memory wall'). InfraCo's data centers use a disaggregated memory pool based on Compute Express Link (CXL) 3.0, allowing any chip to access a shared pool of high-bandwidth memory (HBM4) without the latency penalty of traditional NUMA architectures. This allows for much larger context windows (potentially 1M+ tokens) without the quadratic cost of full attention, using techniques like Ring Attention (popularized by the open-source repo `ring-attention` on GitHub, which has over 2,000 stars).

- Liquid Immersion Cooling: To achieve the density required, InfraCo is deploying single-phase liquid immersion cooling at scale. This allows them to pack 2-3x more compute per rack compared to traditional air-cooled data centers, reducing inter-chip latency and power overhead. The cooling system itself is patented, using a proprietary dielectric fluid with 40% higher thermal conductivity than standard mineral oil.

- Network Protocol Optimization: InfraCo has developed a custom RDMA (Remote Direct Memory Access) protocol called 'TorusNet' that reduces tail latency for distributed inference by 60% compared to standard InfiniBand. This is critical for real-time agent applications where a single slow node can stall the entire reasoning loop.

Data Table: Token Consumption Projections

| Application | Tokens per Second (Current) | Tokens per Second (2027 Projection) | Primary Driver |
|---|---|---|---|
| Text Chat (GPT-4o) | 50-100 | 200-500 | Longer contexts, multi-turn |
| Image Generation (DALL-E 3) | 1,000-5,000 | 10,000-50,000 | Higher resolution, iterative refinement |
| Video Generation (1 min, 1080p) | 2,000,000 | 10,000,000 | 4K resolution, longer duration |
| Real-Time Agent (24/7) | 10,000 | 1,000,000 | Continuous operation, complex reasoning |
| World Simulator (1 sec, 60fps) | N/A | 50,000,000 | Full physics, geometry, lighting |

Data Takeaway: The jump from text chat to world simulators represents a 500,000x increase in token demand per second. Current infrastructure is designed for the first row; InfraCo is building for the last two.

Key Players & Case Studies

InfraCo is not alone in recognizing this shift, but it is the most aggressive in its vertical integration. Here is how the competitive landscape shapes up:

- InfraCo (The $2.2B Bet): The company is led by a former Google TPU architect and a veteran of Amazon's AWS hardware division. Their strategy is to own the entire stack, from chip design to data center construction. They have already signed a 10-year power purchase agreement with a nuclear energy startup to secure 1GW of dedicated power for their first 'GigaCluster' in Nevada.

- NVIDIA: The incumbent. NVIDIA's strength is its CUDA ecosystem and its dominance in training. However, its inference offerings (TensorRT, Triton Inference Server) are optimized for batch processing, not the low-latency, high-throughput demands of real-time agents. The upcoming 'Rubin' architecture is expected to improve inference performance, but it remains a general-purpose chip. NVIDIA's recent acquisition of Run:ai (for GPU orchestration) shows they are aware of the inference bottleneck, but their business model is fundamentally tied to selling expensive, general-purpose hardware.

- Groq (LPU): Groq's Language Processing Unit (LPU) is a direct competitor in the inference space. It uses a deterministic, software-defined architecture that eliminates scheduling overhead, achieving extremely low latency (under 10ms for large models). However, Groq's architecture is less flexible for multimodal workloads and its memory capacity is currently limited. The open-source community has embraced Groq for real-time applications, but scaling to world simulator levels would require a massive architectural overhaul.

- Cerebras (Wafer-Scale): Cerebras takes a different approach: a single, massive wafer-scale chip that eliminates the need for inter-chip communication. This is excellent for training, but for inference, the challenge is memory bandwidth per token. Cerebras's CS-3 has 44GB of on-chip SRAM, which is impressive but still a fraction of what a disaggregated memory pool can offer.

- d-Matrix: A stealthy startup focusing on 'in-compute' memory for inference. Their architecture places memory directly on the compute die, reducing data movement. Early benchmarks show a 5x improvement in tokens-per-joule for small-to-medium models, but scaling to 1 trillion+ parameter models remains unproven.

Data Table: Inference Hardware Comparison

| Hardware | Architecture | Tokens/sec (Llama 3 70B) | Power (W) | Tokens/Joule | Key Limitation |
|---|---|---|---|---|---|
| NVIDIA H100 | GPU | 1,200 | 700 | 1.7 | Memory bandwidth |
| NVIDIA B200 | GPU | 2,500 | 1,000 | 2.5 | Cost, power |
| Groq LPU | ASIC | 1,800 | 200 | 9.0 | Memory capacity |
| Cerebras CS-3 | Wafer-Scale | 1,500 | 1,500 | 1.0 | On-chip memory limit |
| InfraCo TensorCore-X | ASIC | 4,000 (est.) | 500 | 8.0 (est.) | Unproven at scale |

Data Takeaway: InfraCo's custom chip, if it meets its targets, would offer a 3x improvement in throughput and a 3x improvement in energy efficiency over the B200, making it the clear leader for high-volume inference. The risk is that these are estimated figures; real-world performance may vary.

Industry Impact & Market Dynamics

The $2.2 billion funding round is not just a bet on a company; it is a bet on a new economic model for AI. The current model is 'compute as a service' (CaaS), where you pay for GPU hours. InfraCo is pushing towards 'token as a utility' (TaaU), where you pay per token generated, similar to how you pay for water or electricity.

Market Size Projections: Analysts estimate that the AI inference market will grow from $20 billion in 2024 to over $200 billion by 2030. However, this projection assumes linear growth in token demand. If the multimodal/agent/world simulator thesis is correct, the market could be 5-10x larger, reaching $1-2 trillion. InfraCo is positioning itself to capture a significant share of this market by owning the most efficient token production infrastructure.

Impact on Model Companies: If InfraCo succeeds, the value proposition of model companies like OpenAI, Anthropic, and Google DeepMind shifts. Currently, they compete on model quality (benchmark scores). In a TaaU world, the cost of inference becomes a critical differentiator. A model that is 5% less accurate but 10x cheaper to run could dominate the market. This could lead to a race to the bottom on inference costs, compressing margins for model companies and benefiting infrastructure providers.

The 'Water and Electricity' Analogy: InfraCo's CEO has explicitly stated that their goal is to become the 'water and electricity' of the AI age. This is a powerful analogy. Water and electricity are essential, ubiquitous, and regulated. If InfraCo achieves this, they would wield enormous power over the entire AI ecosystem, potentially becoming a bottleneck themselves. This raises antitrust concerns, especially given the $2.2 billion backing from sovereign wealth funds and large tech conglomerates.

Data Table: Funding Landscape for AI Infrastructure

| Company | Total Funding (est.) | Focus Area | Key Investors |
|---|---|---|---|
| InfraCo | $2.2B (latest round) | Vertical integration (chip to data center) | SoftBank, Sequoia, Saudi PIF |
| CoreWeave | $1.5B | Cloud GPU services | Fidelity, BlackRock |
| Lambda | $500M | GPU cloud for training | US Innovative Technology Fund |
| Groq | $640M | LPU inference chips | Tiger Global, D1 Capital |
| d-Matrix | $150M | In-compute memory inference | Playground Global, Nvidia (strategic) |

Data Takeaway: InfraCo's $2.2B round is more than the combined funding of its three closest competitors. This reflects the capital-intensive nature of vertical integration and the belief that owning the entire stack is the only way to achieve the necessary efficiency gains.

Risks, Limitations & Open Questions

InfraCo's strategy is high-risk, high-reward. Several factors could derail their bet:

1. Execution Risk: Vertical integration is notoriously difficult. Apple succeeded, but Intel's attempt (with its own foundries) has struggled for decades. InfraCo must simultaneously design cutting-edge chips, build massive data centers, develop a software stack (compiler, runtime, orchestration), and manage a global supply chain. Any single failure point could be catastrophic.

2. Technological Obsolescence: The AI hardware landscape is evolving rapidly. NVIDIA's 'Rubin' architecture (expected 2026) could leapfrog InfraCo's TensorCore-X. More importantly, algorithmic breakthroughs could change the token economics entirely. For example, if a new architecture (e.g., State Space Models like Mamba, which has over 10,000 stars on GitHub) reduces token requirements by 10x, InfraCo's massive investment in specialized hardware for attention-based models could become stranded.

3. The 'Token' Definition Problem: The concept of a 'token' is not standardized. Different models use different tokenizers (e.g., GPT-4 uses a BPE tokenizer with ~100k vocabulary; Gemini uses a SentencePiece tokenizer). A 'token' in one system is not equivalent to a 'token' in another. InfraCo's TaaU model would require an industry-wide standardization of token measurement, which is unlikely to happen soon.

4. Energy Constraints: Even with liquid cooling and custom chips, a world simulator running 24/7 would consume enormous amounts of energy. InfraCo's nuclear power deal is a hedge, but building new nuclear plants is notoriously slow and expensive. If energy costs rise faster than expected, the TaaU model may not be economically viable.

5. Geopolitical Risks: The semiconductor supply chain is heavily concentrated in Taiwan (TSMC). Any disruption to TSMC's manufacturing (due to geopolitical tensions) would halt InfraCo's chip production. The company has not disclosed any plans for a second source of fabrication.

AINews Verdict & Predictions

InfraCo's $2.2 billion bet is the most significant infrastructure wager in the history of AI. It represents a fundamental shift in thinking: from 'how smart can we make the model?' to 'how cheaply can we run it?' This is the correct question for the next phase of AI development.

Our Predictions:

1. InfraCo will succeed in building its first GigaCluster, but will face significant delays. The complexity of vertical integration will cause at least a 12-18 month delay from the original timeline. However, once operational, it will demonstrate a 5-10x cost advantage over existing cloud providers for inference workloads.

2. The 'token as a utility' model will become the dominant pricing paradigm by 2028. This will force every major cloud provider (AWS, Azure, GCP) to develop their own custom inference chips, leading to a fragmentation of the hardware ecosystem similar to the early days of mobile phones.

3. NVIDIA's dominance will be challenged, but not broken. NVIDIA will remain the king of training, but its share of the inference market will drop from ~80% today to ~40% by 2028, as specialized ASICs from InfraCo, Groq, and others eat into its market share.

4. The biggest winners will be the model companies that optimize for inference efficiency. A model like Anthropic's Claude 3.5 Opus, which is already known for its efficient architecture, could gain a significant market share advantage over less efficient competitors like GPT-4 Turbo.

5. Watch for a major acquisition. If InfraCo's technology proves viable, a hyperscaler (most likely Amazon or Google) will attempt to acquire the company for $50-100 billion within the next three years. The vertical integration moat is too valuable to leave independent.

What to Watch Next: The key metric to track is not InfraCo's funding, but its tokens-per-dollar cost for a standard benchmark (e.g., Llama 3 70B inference). If they can achieve a cost of $0.10 per million tokens (compared to ~$0.50 for current cloud providers), the thesis is validated. If they cannot, the $2.2 billion may be remembered as the biggest overbet in AI history.

Archive

May 2026784 published articles

Further Reading

馬斯克的Terafab佈局:垂直整合策略以掌控AI的物理世界伊隆·馬斯克正啟動「Terafab」這項宏偉戰略,旨在將尖端AI晶片設計與專有半導體製造相結合。此垂直整合策略意在繞過產業瓶頸,實現計算規模50倍的擴張,從根本上重塑AI的物理基礎設施。從矽晶到語法:AI基礎設施之戰如何從囤積GPU轉向代幣經濟學AI基礎設施競賽已發生典範轉移。競爭核心不再是爭奪稀缺的GPU硬體,而是根本性地轉向優化智慧『代幣』的生產與交付——即AI服務輸出的標準化單位。這場從矽晶為中心到語法驅動的轉變,正在重塑整個產業格局。高德發佈全棧具身AI,標誌AGI競爭進入基礎設施時代阿里巴巴旗下的地圖與導航業務高德,近日公開詳述其全棧具身智能技術系統,宣稱在15項全球基準測試中取得頂尖表現。這項發佈不僅是技術成就,更標誌著產業正朝著整合性、基礎設施化的方向轉變。字節跳動的AI豪賭:豆包每日處理120萬億Token,引發產業成本清算據報導,字節跳動的AI助手「豆包」每日處理的Token數量高達120萬億,這標誌著AI競爭的重心已從技術實力轉向原始規模與用戶參與度。如此巨大的消耗量,其每日營運成本估計達數千萬,為整個產業敲響了成本警鐘。

常见问题

这起“The Token Tsunami: Why a $2.2B Bet on AGI Infrastructure Redefines the AI Arms Race”融资事件讲了什么?

The AI industry's current obsession with model parameter scaling is masking a fundamental shift. The next frontier is not making models smarter in isolation, but making them usable…

从“What is the token economics of AGI infrastructure?”看,为什么这笔融资值得关注?

The core insight driving InfraCo's strategy is a redefinition of the scaling laws. The original scaling laws (Kaplan et al., 2020) focused on model parameters and training compute. InfraCo's thesis is that the next scali…

这起融资事件在“How does vertical integration reduce token cost?”上释放了什么行业信号?

它通常意味着该赛道正在进入资源加速集聚期,后续值得继续关注团队扩张、产品落地、商业化验证和同类公司跟进。