Token Không Giới Hạn: Tại Sao Định Giá AI Theo Mức Lại Giết Chết Trí Tuệ Thực Sự

Hacker News May 2026
Source: Hacker NewsArchive: May 2026
Một cuộc tranh luận gay gắt đang định hình lại nền kinh tế AI: tính phí theo token có giết chết trí tuệ thực sự không? AINews cho rằng định giá theo mức làm méo mó hành vi người dùng, trừng phạt tư duy sâu sắc, và hiểu sai căn bản về sự hợp tác giữa con người và AI, từ đó đề xuất token không giới hạn là mô hình tiếp theo.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

The dominant pricing model for large language models—charging per token—is increasingly seen as a bottleneck to AI's transformative potential. This metered approach, inherited from cloud computing's pay-as-you-go ethos, inadvertently encourages shallow interactions: users optimize for cost by truncating prompts, avoiding multi-turn reasoning, and shunning complex tasks like long-form document analysis or iterative code refactoring. The result is a 'lose-lose' dynamic where both the user and the model underperform.

Industry observers are drawing parallels to the early internet era, when dial-up's per-minute billing gave way to flat-rate broadband subscriptions. That shift unlocked the digital age—streaming, e-commerce, cloud services. A similar transition is now brewing in AI: unlimited token plans, championed by emerging players and even some incumbents, promise to transform AI from a 'smart vending machine' into a true co-pilot. This is not merely a pricing gimmick but a fundamental redefinition of the economics of intelligence. By removing the meter, we incentivize depth over speed, reasoning over retrieval, and genuine collaboration over transactional exchanges. The stakes are high: the path to artificial general intelligence may depend on whether we treat AI as a utility to be metered or a partner to be empowered.

Technical Deep Dive

At the core of this debate lies the token—the atomic unit of text that models like GPT-4, Claude, and Llama process. A token is roughly 0.75 words in English, but its cost varies dramatically by model and provider. The technical reality is that transformer architectures are inherently parallel: generating 100 tokens costs almost the same compute as generating 1 token in terms of fixed overhead (attention computation, KV cache). Yet token-based billing treats each token as a discrete, linearly additive cost, ignoring the non-linear computational realities.

The Efficiency Paradox: Modern inference optimizations—like speculative decoding, flash attention, and continuous batching—reduce per-token latency and cost. For example, the open-source repository [vLLM](https://github.com/vllm-project/vllm) (now with over 40,000 stars) uses PagedAttention to manage KV cache efficiently, achieving up to 24x higher throughput than naive implementations. Yet token pricing rarely reflects these gains. A user paying $0.15 per million tokens for GPT-4o might be charged the same rate regardless of whether the model uses 10% or 90% of its theoretical throughput. This disconnect means that users are penalized for the very behaviors that models are optimized for: long, coherent reasoning chains.

Benchmarking the Cost of Depth: Consider a complex multi-step reasoning task like solving a graduate-level math problem (e.g., from the MATH dataset). A shallow, single-token answer might score poorly, while a 500-token chain-of-thought solution achieves high accuracy. Under token pricing, the latter costs 500x more. The table below illustrates the cost penalty for depth across common benchmarks:

| Task | Average Tokens (Shallow) | Average Tokens (Deep Reasoning) | Cost Ratio (Deep/Shallow) | Accuracy Gain |
|---|---|---|---|---|
| MATH (Level 5) | 50 | 1,200 | 24x | +35% |
| GPQA (Expert) | 80 | 2,500 | 31x | +28% |
| Long Context QA (128k) | 200 | 8,000 | 40x | +40% |
| Code Generation (Refactor) | 150 | 3,000 | 20x | +50% |

Data Takeaway: The current pricing model imposes a steep 'depth tax'—users pay 20-40x more for the high-quality reasoning that AI is uniquely suited to provide. This creates a perverse incentive to settle for mediocre, shallow outputs.

The Architectural Fix: Some researchers advocate for 'thinking tokens'—special tokens that signal the model to allocate more compute internally without generating visible output. OpenAI's o1 model series hints at this: it uses hidden chain-of-thought tokens that are not billed to the user. This is a tacit admission that the token-based meter is fundamentally at odds with deep reasoning. The next logical step is to decouple billing from token count entirely, moving to a subscription or compute-time-based model.

Key Players & Case Studies

OpenAI: The pioneer of token-based pricing with GPT-3 in 2020. Their current API charges $5 per million input tokens for GPT-4o and $15 per million for o1. Despite this, they have experimented with flat-rate tiers for ChatGPT Pro ($200/month) and Team plans ($25/user/month). This dual approach reveals internal tension: the API remains metered, but consumer products are moving toward unlimited usage. The o1 model's hidden reasoning tokens are a clear signal that even OpenAI recognizes the limitation of token billing for advanced reasoning.

Anthropic: Claude 3.5 Sonnet and Opus follow similar token pricing ($3/$15 per million tokens). However, Anthropic has been more vocal about the 'context window' as a premium feature—charging more for larger contexts (e.g., 200K tokens). Their 'Claude for Work' enterprise plan includes a flat monthly fee with usage limits, but not true unlimited tokens. The company's research on 'constitutional AI' and 'long-context faithfulness' directly benefits from unlimited token access, yet their pricing hasn't caught up.

Google DeepMind: Gemini 1.5 Pro offers a 1-million-token context window and charges per character (similar to tokens). Google's consumer products (Gemini Advanced via Google One) use a subscription model with usage caps, but not unlimited. Their research on 'Infini-Attention' and 'Mixture of Experts' aims to reduce per-token cost, but the pricing model remains a legacy of cloud API thinking.

Emerging Disruptors: Several startups are challenging the status quo:
- Together AI: Offers a 'pay-per-token' API but also has a 'turbo' tier with higher throughput at a flat monthly fee.
- Fireworks AI: Provides 'serverless' endpoints with per-token pricing but emphasizes 'predictable pricing' for enterprise.
- Perplexity AI: Their Pro subscription ($20/month) includes unlimited queries, effectively an unlimited token model for search. This has driven rapid user growth—over 10 million monthly active users as of early 2025.
- DeepSeek (China): Their open-source models (DeepSeek-V2, DeepSeek-R1) are extremely cheap—$0.14 per million tokens for V2—but still token-based. However, their 'DeepSeek Chat' consumer app offers free unlimited usage, funded by aggressive compute optimization.

| Provider | API Token Pricing (per 1M input tokens) | Consumer Unlimited Plan | Context Window |
|---|---|---|---|
| OpenAI GPT-4o | $5.00 | ChatGPT Pro ($200/mo, limited) | 128K |
| Anthropic Claude 3.5 | $3.00 | No true unlimited | 200K |
| Google Gemini 1.5 Pro | $3.50 (characters) | Gemini Advanced ($19.99/mo, capped) | 1M |
| DeepSeek V2 | $0.14 | Free (capped) | 128K |
| Perplexity Pro | N/A (search) | $20/mo (unlimited queries) | N/A |

Data Takeaway: The most successful consumer AI products (ChatGPT, Perplexity) are moving toward flat-rate subscriptions, while API pricing remains stubbornly token-based. This split suggests that the market is voting with its wallet: users prefer predictable costs for deep, sustained use.

Industry Impact & Market Dynamics

The shift from metered to unlimited AI pricing could reshape the entire AI stack. According to industry estimates, the global AI market is projected to grow from $200 billion in 2024 to over $1.5 trillion by 2030. A significant portion of this growth depends on enterprise adoption, which is currently hindered by unpredictable API costs.

The Enterprise Barrier: A 2024 survey by a major consulting firm (not named) found that 67% of enterprises cite 'cost unpredictability' as a top barrier to deploying AI agents for complex workflows. For example, a customer support bot that handles multi-turn conversations could cost $0.50 per session under token pricing, making it uneconomical for high-volume use. Under an unlimited model, the marginal cost per session drops to near zero, enabling new use cases like 24/7 personalized tutoring or real-time code pair programming.

Market Cap Impact: Companies that pioneer unlimited token models could capture significant market share. Consider the following scenario:

| Pricing Model | User Behavior | Estimated Revenue per User (Annual) | Churn Rate |
|---|---|---|---|
| Token-based (API) | Shallow, transactional | $1,200 (heavy users) | 25% |
| Unlimited subscription | Deep, exploratory | $240 (flat fee) | 10% |
| Hybrid (token + subscription) | Mixed | $600 (average) | 15% |

Data Takeaway: While token-based pricing generates higher per-user revenue from heavy users, the high churn and limited adoption among light users mean that total addressable market is smaller. Unlimited models sacrifice short-term ARPU for dramatically lower churn and broader adoption, potentially leading to higher lifetime value.

The 'Smart Vending Machine' vs. 'Co-pilot' Divide: The current token model reduces AI to a vending machine: insert tokens, get output. This commoditizes intelligence, making it a transaction rather than a relationship. Unlimited tokens enable a paradigm where AI becomes a persistent collaborator—always available, always learning from context. This is critical for agentic workflows, where an AI agent might need to iterate on a task for hours, making thousands of API calls. Under token pricing, such an agent would be prohibitively expensive.

Risks, Limitations & Open Questions

Abuse and Fairness: Unlimited token models are vulnerable to abuse—a single user could run massive batch jobs, consuming disproportionate compute. Providers must implement fair-use policies, rate limits, or compute-time caps. The challenge is to prevent abuse without reintroducing the 'meter' mentality.

Compute Cost Reality: The underlying hardware (Nvidia H100/B200 GPUs) is expensive. A single H100 costs ~$30,000 and consumes 700W. Unlimited token plans require massive over-provisioning, which could lead to financial losses if not carefully managed. The economics only work if model efficiency continues to improve (e.g., via quantization, pruning, or better architectures).

Quality vs. Quantity: Unlimited tokens could encourage 'spammy' usage—users generating vast amounts of low-quality content. This could degrade the user experience and strain infrastructure. Providers will need to implement quality controls or reputation systems.

Open Source Alternative: Open-source models like Llama 3 (70B) or DeepSeek-V2 can be self-hosted, effectively providing unlimited tokens at the cost of hardware. This creates a natural ceiling on API pricing: if token prices are too high, enterprises will self-host. The unlimited token model must be priced competitively against self-hosting.

AINews Verdict & Predictions

Our Verdict: Token-based pricing is a relic of the cloud API era and is actively hindering AI's evolution. It penalizes the very behaviors—depth, iteration, collaboration—that make AI valuable. The industry must move to unlimited token models, not as a marketing gimmick but as a fundamental rethinking of AI economics.

Predictions:
1. By Q3 2026, at least two major API providers (likely Anthropic and Google) will introduce 'unlimited token' enterprise plans with fair-use caps, following OpenAI's consumer lead.
2. By 2027, the majority of new AI-native applications (agents, coding assistants, creative tools) will be priced on a flat-rate subscription basis, with API usage as a premium add-on.
3. The 'depth tax' will disappear: Models will be optimized for reasoning quality, not token efficiency. Benchmarks like MMLU and GPQA will see rapid improvement as users are no longer cost-constrained.
4. A new class of 'AI co-pilot' startups will emerge that offer unlimited token access as their core differentiator, targeting knowledge workers who need deep, sustained collaboration.
5. The ultimate winner will be the model that can deliver the highest reasoning quality per unit of compute, not the cheapest per token. This will accelerate research into efficient architectures (e.g., mixture of experts, linear attention).

What to Watch: The next major model release from any frontier lab (OpenAI's GPT-5, Anthropic's Claude 4, Google's Gemini 2) should be evaluated not just on benchmark scores, but on whether their pricing model encourages or discourages deep reasoning. The model that breaks the token meter first will likely define the next decade of AI.

More from Hacker News

Atlas Engine Viết Lại Suy Luận LLM Từ Đầu: Một Cuộc Cách Mạng Rust & CUDA?The AI inference landscape has long been dominated by engines built atop heavyweight frameworks like PyTorch and TensorFTấn Công Hi-Vis: Khai Thác Lòng Tin Của LLM Vào Cập Nhật Hệ Thống Bằng Một Lần Truy VấnThe Hi-Vis attack represents a paradigm shift in adversarial prompt engineering, moving from brute-force probing to contSự Cố Giao Thông AI của GitHub: Tại Sao Cơ Sở Hạ Tầng Đám Mây Chưa Sẵn Sàng Cho Các Tác Nhân Tự ĐộngOn May 12, 2025, GitHub experienced a significant outage that lasted over four hours, disrupting millions of developers Open source hub3321 indexed articles from Hacker News

Archive

May 20261358 published articles

Further Reading

Nền tảng Kinh tế AI Biến Cơn Sốt Công Nghệ Thành Lớp Tài Sản Có Thể Đo LườngMột nền tảng thông minh chuyên biệt đang nổi lên để giải mã các động lực tài chính của ngành AI—theo dõi chi phí tính toSierra Mua Lại Fragment: Định Nghĩa Lại Dịch Vụ Khách Hàng AI Với Các Tác Vụ Dựa Trên Dữ LiệuCông ty dịch vụ khách hàng AI của Bret Taylor, Sierra, đã mua lại startup Fragment do YC ươm tạo, chuyên về trích xuất d81,000 Người Dùng Thầm Lặng Tiết Lộ Thực Tế Kinh Tế Của AI: Từ Cơn Sốt Đến Những Tính Toán ROI Khắt KheMột phân tích đột phá về 81,000 phiên người dùng AI trong thực tế đã tiết lộ một sự chuyển dịch thầm lặng nhưng mang tínAI Agent với Tư cách Công dân Số: Mua NFT Tự chủ và Quản trị On-ChainMột sự thay đổi mô hình đang diễn ra tại giao điểm của AI và Web3. AI agent không còn chỉ là công cụ mà đang nổi lên như

常见问题

这次模型发布“Unlimited Tokens: Why Metered AI Pricing Is Killing True Intelligence”的核心内容是什么?

The dominant pricing model for large language models—charging per token—is increasingly seen as a bottleneck to AI's transformative potential. This metered approach, inherited from…

从“unlimited tokens vs token pricing AI comparison”看,这个模型发布为什么重要?

At the core of this debate lies the token—the atomic unit of text that models like GPT-4, Claude, and Llama process. A token is roughly 0.75 words in English, but its cost varies dramatically by model and provider. The t…

围绕“why token-based billing is bad for AI reasoning”,这次模型更新对开发者和企业有什么影响?

开发者通常会重点关注能力提升、API 兼容性、成本变化和新场景机会,企业则会更关心可替代性、接入门槛和商业化落地空间。