馬克·庫班警告:OpenAI 兆美元豪賭可能永遠無法回本

Hacker News May 2026
Source: Hacker NewsOpenAIArchive: May 2026
億萬富翁投資人馬克·庫班公開警告,OpenAI 投入的兆美元資金可能永遠無法回收。AINews 分析認為,這是 AI 產業最敏感泡沫的關鍵信號,指數級增長的運算成本與不確定的營收路徑相互碰撞,迫使市場面臨殘酷的清算。
The article body is currently shown in English by default. You can generate the full version in this language on demand.

Mark Cuban's blunt assessment that OpenAI's $1 trillion capital expenditure may never be recovered has sent shockwaves through the AI investment community. The comment cuts to the heart of a paradox: as frontier model training costs double every few months, the marginal performance gains are diminishing, while commercial adoption remains fragmented. OpenAI's valuation, pegged at over $300 billion in 2025, rests on a narrative of future dominance that is increasingly at odds with the numbers. The company is burning through cash at an estimated $7–10 billion annually, with revenue from ChatGPT subscriptions and API licensing still insufficient to cover operating expenses, let alone the massive capital outlays for data centers and GPU clusters. Cuban's critique is not isolated; it reflects a growing unease among institutional investors who see parallels to the dot-com era, where infrastructure spending vastly outpaced actual demand. The core issue is the 'scale returns' experiment: does spending ten times more on compute yield ten times the intelligence, or merely incremental improvements? Early evidence from benchmarks suggests the latter. Meanwhile, open-source alternatives like Meta's Llama 3.1 and Mistral's Mixtral are closing the gap, eroding OpenAI's moat. If Cuban is right, the industry is heading for a painful correction, where only companies with clear, near-term monetization—like Microsoft's Copilot integration or NVIDIA's hardware sales—will survive the shakeout.

Technical Deep Dive

The trillion-dollar question Cuban raises is fundamentally about the scaling laws of large language models. OpenAI's approach, from GPT-3 to GPT-4 and beyond, has relied on the assumption that increasing model size, training data, and compute yields predictable improvements in capability. This is formalized by the Chinchilla scaling law, which suggests optimal training requires balancing model parameters and data tokens. However, the cost of pushing these frontiers has become astronomical. Training GPT-4 is estimated to have cost around $100 million; training a hypothetical GPT-5 or GPT-6 could exceed $1–2 billion, factoring in the need for clusters of 100,000+ H100/B200 GPUs running for months.

| Model | Estimated Training Cost | Parameters (est.) | MMLU Score | Cost per 1M Tokens (API) |
|---|---|---|---|---|
| GPT-3 | $4.6M | 175B | 43.9 | $0.02 |
| GPT-4 | ~$100M | ~1.8T (MoE) | 86.4 | $0.03 (input) / $0.06 (output) |
| GPT-4o | ~$200M (est.) | ~200B active | 88.7 | $2.50 (input) / $10.00 (output) |
| Llama 3.1 405B | ~$60M (Meta) | 405B | 88.0 | Free (open-source) |

Data Takeaway: The cost-to-performance ratio is worsening. GPT-4o cost roughly twice as much to train as GPT-4 but only improved MMLU by 2.3 points. Meanwhile, open-source Llama 3.1 405B achieves near-parity with GPT-4o at a fraction of the training cost and zero inference cost for users. This suggests diminishing returns on scale.

From an engineering perspective, OpenAI's infrastructure relies on massive clusters of NVIDIA H100 GPUs, interconnected via InfiniBand, with custom networking and cooling solutions. The company has also developed its own AI training framework, Triton, and uses a mixture-of-experts (MoE) architecture to reduce inference costs. However, the energy consumption is staggering: a single training run for a frontier model can consume 50–100 GWh of electricity, equivalent to the annual usage of 10,000 US homes. Cuban's point is that these costs are not one-time; they recur with each new model generation, and inference costs scale with user adoption. For a product like ChatGPT, which serves hundreds of millions of users, inference compute costs alone could exceed $1 billion per year.

Key Players & Case Studies

Cuban's warning is most directly aimed at OpenAI, but the implications ripple across the entire AI ecosystem. Let's examine the key players and their strategies:

- OpenAI: The poster child for the 'scale at all costs' strategy. Under CEO Sam Altman, the company has raised over $20 billion from Microsoft and other investors, with plans to raise another $100 billion for a new data center project called 'Stargate.' The bet is that by achieving AGI (Artificial General Intelligence), the company will unlock unprecedented revenue streams. However, current revenue is heavily reliant on ChatGPT Plus ($20/month) and API sales to developers, which are price-sensitive and face competition from cheaper alternatives.
- Microsoft: The primary beneficiary of OpenAI's technology, but also a cautious partner. Microsoft has integrated GPT-4 into its Copilot products (Office 365, Azure, GitHub), generating measurable revenue from enterprise subscriptions. However, Microsoft also invests in its own models (Phi-3) and in open-source initiatives, hedging against OpenAI's potential failure.
- Google/DeepMind: Google has its own frontier model (Gemini) and a massive cloud infrastructure. Its advantage is that it can use AI to improve its core advertising business, which generates over $200 billion annually. This gives Google a natural monetization path that OpenAI lacks.
- Anthropic: Founded by ex-OpenAI employees, Anthropic focuses on safety and alignment. It has raised over $7 billion, but its Claude models are also expensive to run. The company's 'constitutional AI' approach is a differentiator, but it has not yet proven to be a commercial advantage.
- Open-Source Community (Meta, Mistral, Hugging Face): Meta's Llama 3.1, Mistral's Mixtral 8x22B, and other open models are rapidly closing the performance gap. They are free to use and modify, which puts immense pressure on OpenAI's pricing power. The open-source ecosystem also benefits from community-driven optimization, such as quantization (e.g., via llama.cpp) and fine-tuning (e.g., via Unsloth), which reduce inference costs.

| Company | Model | Strategy | Revenue Model | Key Risk |
|---|---|---|---|---|
| OpenAI | GPT-4o | Closed-source, scale-first | API, subscriptions | High cost, low moat |
| Microsoft | Copilot, Phi-3 | Integration with existing products | Enterprise subscriptions | Dependency on OpenAI |
| Google | Gemini | Integration with ads, cloud | Ad revenue, cloud | Regulatory scrutiny |
| Meta | Llama 3.1 | Open-source, ecosystem play | Indirect (ads, data) | No direct revenue |
| Anthropic | Claude 3.5 | Safety-first, enterprise | API, subscriptions | Slower innovation |

Data Takeaway: The table shows that companies with existing revenue streams (Microsoft, Google) are better positioned to absorb AI costs. Pure-play AI companies like OpenAI and Anthropic face existential risk if they cannot achieve profitability before their funding runs out.

Industry Impact & Market Dynamics

Cuban's comment is a symptom of a broader market correction. The AI industry has attracted over $150 billion in venture capital and corporate investment since 2022, with a significant portion going to compute infrastructure. This has created a 'compute bubble' where the cost of GPUs and data centers has inflated, but the demand for AI services has not kept pace.

| Year | Global AI Investment ($B) | Data Center Capex ($B) | AI Revenue (Top 10 Cos) ($B) | Ratio (Capex/Revenue) |
|---|---|---|---|---|
| 2022 | 95 | 50 | 30 | 1.67 |
| 2023 | 120 | 70 | 45 | 1.56 |
| 2024 | 150 | 100 | 60 | 1.67 |
| 2025 (est.) | 180 | 130 | 80 | 1.63 |

Data Takeaway: The ratio of capital expenditure to revenue has remained stubbornly high, around 1.6x, meaning for every dollar of AI revenue, companies are spending $1.60 on infrastructure. This is unsustainable long-term. The industry needs to either dramatically increase revenue (e.g., through killer apps) or reduce costs (e.g., through more efficient models).

The market dynamics are also shifting. Enterprise adoption of AI is growing, but it is concentrated in low-risk, high-ROI use cases like code generation (GitHub Copilot), customer service chatbots, and content summarization. These applications are price-sensitive and often use smaller, cheaper models (e.g., GPT-3.5, Llama 3 8B) rather than frontier models. The 'killer app' for frontier models—something that justifies the trillion-dollar spend—has not yet emerged. Autonomous agents, video generation (Sora), and scientific research are promising but still nascent.

Risks, Limitations & Open Questions

Cuban's warning highlights several unresolved challenges:

1. Monetization Gap: The most critical risk is that AI models, especially frontier ones, are a solution in search of a problem. While they are impressive demos, they have not yet proven they can generate the kind of revenue needed to justify their costs. The enterprise market is large but fragmented, and many companies are still in the pilot phase.
2. Energy and Environmental Costs: The compute required for training and inference has a massive carbon footprint. As energy prices rise and regulations tighten, this could become a significant liability. OpenAI's 'Stargate' project alone could require 5 GW of power, equivalent to five nuclear reactors.
3. Open-Source Disruption: The rapid progress of open-source models is eroding the competitive advantage of closed-source companies. If open models can match or exceed GPT-4o's performance within a year, OpenAI's pricing power will collapse.
4. Regulatory Risk: Governments are increasingly scrutinizing AI for safety, bias, and monopoly concerns. New regulations could limit the use of AI in certain sectors or impose liability for errors, reducing demand.
5. Technical Limitations: Current models still suffer from hallucinations, lack of reasoning, and inability to perform long-horizon tasks. The path to AGI is uncertain, and if progress stalls, the investment thesis collapses.

AINews Verdict & Predictions

Mark Cuban is right to be skeptical, but his warning may be premature. The AI industry is in a classic 'hype cycle' phase, where expectations exceed reality. However, the underlying technology is genuinely transformative, and the long-term potential remains enormous. The key question is timing.

Our editorial judgment is that the industry will undergo a 'value correction' within the next 18–24 months. This will not be a crash, but a recalibration where:

- OpenAI will be forced to cut costs by either reducing model size, adopting more efficient architectures (e.g., MoE, quantization), or pivoting to a more focused product strategy. We predict they will launch a 'GPT-4 Lite' model that is cheaper to run, targeting the mass market.
- Microsoft will acquire a controlling stake in OpenAI within 2026, providing a financial backstop and integrating the technology more deeply into its ecosystem.
- Open-source models will become the default for most enterprise use cases, with closed-source models reserved for high-stakes applications (e.g., medical diagnosis, legal analysis) where reliability is paramount.
- The 'compute bubble' will burst for non-differentiated hardware providers, but NVIDIA will continue to dominate due to its software moat (CUDA).

What to watch next: The next earnings report from Microsoft and Google, which will reveal the actual revenue from AI products. Also, watch for any signs that OpenAI is struggling to raise its next funding round. If the 'Stargate' project is delayed or scaled back, Cuban's prediction will be validated.

More from Hacker News

從影片墳場到智慧知識庫:讓內容重獲新生的WordPress外掛A new WordPress plugin, developed by an independent creator, addresses a critical blind spot in content strategy: the va免費GPT工具壓力測試創業點子:AI聯合創始人時代來臨A new free GPT-based tool is gaining traction in the startup community for its ability to rigorously pressure-test businZAYA1-8B:僅啟用7.6億參數的8B MoE模型,數學能力媲美DeepSeek-R1AINews has uncovered that ZAYA1-8B, a Mixture of Experts (MoE) model with 8 billion total parameters, activates a mere 7Open source hub3040 indexed articles from Hacker News

Related topics

OpenAI104 related articles

Archive

May 2026792 published articles

Further Reading

GPT-5.5 Instant:為何速度是AI競爭的新前沿OpenAI 發布了 GPT-5.5 Instant,這是一款專為近乎零延遲推理而設計的模型。這標誌著從原始智慧轉向推理速度的戰略轉變,目標是實現即時代理協作和高頻決策,回應時間低於200毫秒。Y Combinator 持有的 0.6% OpenAI 股份:一場改寫創投規則的隱形賭注Y Combinator 持有 OpenAI 0.6% 的股份,如今價值數十億美元。AINews 深入探討這筆早期投資——在 OpenAI 還是非營利實驗室時所下的賭注——如何重新定義 AI 時代的創投策略。OpenAI 三層架構解決語音 AI 的即時延遲問題OpenAI 以三層架構破解了即時語音 AI 的挑戰,將延遲降至無法察覺的程度。推測解碼、自適應音訊壓縮與邊緣感知路由協同運作,將語音 AI 從展示噱頭轉變為可投入生產的基礎設施。AI巨頭的課堂佈局:OpenAI、Google、微軟如何贏得下一代的心智佔有率一項由OpenAI、Google和微軟支持的美國兩黨法案,提議為K-12學校的AI素養課程和教師培訓提供聯邦補助。AINews調查發現,這項看似利他的教育倡議,實際上是對下一代認知基礎設施的精心投資。

常见问题

这次公司发布“Mark Cuban Warns OpenAI's Trillion-Dollar Bet May Never Pay Off”主要讲了什么?

Mark Cuban's blunt assessment that OpenAI's $1 trillion capital expenditure may never be recovered has sent shockwaves through the AI investment community. The comment cuts to the…

从“Mark Cuban OpenAI investment warning analysis”看,这家公司的这次发布为什么值得关注?

The trillion-dollar question Cuban raises is fundamentally about the scaling laws of large language models. OpenAI's approach, from GPT-3 to GPT-4 and beyond, has relied on the assumption that increasing model size, trai…

围绕“OpenAI trillion dollar cost recoupment”,这次发布可能带来哪些后续影响?

后续通常要继续观察用户增长、产品渗透率、生态合作、竞品应对以及资本市场和开发者社区的反馈。