Tokens ilimitados: por qué el precio por uso está matando la verdadera inteligencia

Hacker News May 2026
Source: Hacker NewsArchive: May 2026
Un intenso debate está redefiniendo la economía de la IA: ¿cobrar por token elimina la inteligencia genuina? AINews sostiene que el precio medido distorsiona el comportamiento del usuario, penaliza el pensamiento profundo y malinterpreta fundamentalmente la colaboración humano-IA, abogando por los tokens ilimitados como el próximo paradigma.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

The dominant pricing model for large language models—charging per token—is increasingly seen as a bottleneck to AI's transformative potential. This metered approach, inherited from cloud computing's pay-as-you-go ethos, inadvertently encourages shallow interactions: users optimize for cost by truncating prompts, avoiding multi-turn reasoning, and shunning complex tasks like long-form document analysis or iterative code refactoring. The result is a 'lose-lose' dynamic where both the user and the model underperform.

Industry observers are drawing parallels to the early internet era, when dial-up's per-minute billing gave way to flat-rate broadband subscriptions. That shift unlocked the digital age—streaming, e-commerce, cloud services. A similar transition is now brewing in AI: unlimited token plans, championed by emerging players and even some incumbents, promise to transform AI from a 'smart vending machine' into a true co-pilot. This is not merely a pricing gimmick but a fundamental redefinition of the economics of intelligence. By removing the meter, we incentivize depth over speed, reasoning over retrieval, and genuine collaboration over transactional exchanges. The stakes are high: the path to artificial general intelligence may depend on whether we treat AI as a utility to be metered or a partner to be empowered.

Technical Deep Dive

At the core of this debate lies the token—the atomic unit of text that models like GPT-4, Claude, and Llama process. A token is roughly 0.75 words in English, but its cost varies dramatically by model and provider. The technical reality is that transformer architectures are inherently parallel: generating 100 tokens costs almost the same compute as generating 1 token in terms of fixed overhead (attention computation, KV cache). Yet token-based billing treats each token as a discrete, linearly additive cost, ignoring the non-linear computational realities.

The Efficiency Paradox: Modern inference optimizations—like speculative decoding, flash attention, and continuous batching—reduce per-token latency and cost. For example, the open-source repository [vLLM](https://github.com/vllm-project/vllm) (now with over 40,000 stars) uses PagedAttention to manage KV cache efficiently, achieving up to 24x higher throughput than naive implementations. Yet token pricing rarely reflects these gains. A user paying $0.15 per million tokens for GPT-4o might be charged the same rate regardless of whether the model uses 10% or 90% of its theoretical throughput. This disconnect means that users are penalized for the very behaviors that models are optimized for: long, coherent reasoning chains.

Benchmarking the Cost of Depth: Consider a complex multi-step reasoning task like solving a graduate-level math problem (e.g., from the MATH dataset). A shallow, single-token answer might score poorly, while a 500-token chain-of-thought solution achieves high accuracy. Under token pricing, the latter costs 500x more. The table below illustrates the cost penalty for depth across common benchmarks:

| Task | Average Tokens (Shallow) | Average Tokens (Deep Reasoning) | Cost Ratio (Deep/Shallow) | Accuracy Gain |
|---|---|---|---|---|
| MATH (Level 5) | 50 | 1,200 | 24x | +35% |
| GPQA (Expert) | 80 | 2,500 | 31x | +28% |
| Long Context QA (128k) | 200 | 8,000 | 40x | +40% |
| Code Generation (Refactor) | 150 | 3,000 | 20x | +50% |

Data Takeaway: The current pricing model imposes a steep 'depth tax'—users pay 20-40x more for the high-quality reasoning that AI is uniquely suited to provide. This creates a perverse incentive to settle for mediocre, shallow outputs.

The Architectural Fix: Some researchers advocate for 'thinking tokens'—special tokens that signal the model to allocate more compute internally without generating visible output. OpenAI's o1 model series hints at this: it uses hidden chain-of-thought tokens that are not billed to the user. This is a tacit admission that the token-based meter is fundamentally at odds with deep reasoning. The next logical step is to decouple billing from token count entirely, moving to a subscription or compute-time-based model.

Key Players & Case Studies

OpenAI: The pioneer of token-based pricing with GPT-3 in 2020. Their current API charges $5 per million input tokens for GPT-4o and $15 per million for o1. Despite this, they have experimented with flat-rate tiers for ChatGPT Pro ($200/month) and Team plans ($25/user/month). This dual approach reveals internal tension: the API remains metered, but consumer products are moving toward unlimited usage. The o1 model's hidden reasoning tokens are a clear signal that even OpenAI recognizes the limitation of token billing for advanced reasoning.

Anthropic: Claude 3.5 Sonnet and Opus follow similar token pricing ($3/$15 per million tokens). However, Anthropic has been more vocal about the 'context window' as a premium feature—charging more for larger contexts (e.g., 200K tokens). Their 'Claude for Work' enterprise plan includes a flat monthly fee with usage limits, but not true unlimited tokens. The company's research on 'constitutional AI' and 'long-context faithfulness' directly benefits from unlimited token access, yet their pricing hasn't caught up.

Google DeepMind: Gemini 1.5 Pro offers a 1-million-token context window and charges per character (similar to tokens). Google's consumer products (Gemini Advanced via Google One) use a subscription model with usage caps, but not unlimited. Their research on 'Infini-Attention' and 'Mixture of Experts' aims to reduce per-token cost, but the pricing model remains a legacy of cloud API thinking.

Emerging Disruptors: Several startups are challenging the status quo:
- Together AI: Offers a 'pay-per-token' API but also has a 'turbo' tier with higher throughput at a flat monthly fee.
- Fireworks AI: Provides 'serverless' endpoints with per-token pricing but emphasizes 'predictable pricing' for enterprise.
- Perplexity AI: Their Pro subscription ($20/month) includes unlimited queries, effectively an unlimited token model for search. This has driven rapid user growth—over 10 million monthly active users as of early 2025.
- DeepSeek (China): Their open-source models (DeepSeek-V2, DeepSeek-R1) are extremely cheap—$0.14 per million tokens for V2—but still token-based. However, their 'DeepSeek Chat' consumer app offers free unlimited usage, funded by aggressive compute optimization.

| Provider | API Token Pricing (per 1M input tokens) | Consumer Unlimited Plan | Context Window |
|---|---|---|---|
| OpenAI GPT-4o | $5.00 | ChatGPT Pro ($200/mo, limited) | 128K |
| Anthropic Claude 3.5 | $3.00 | No true unlimited | 200K |
| Google Gemini 1.5 Pro | $3.50 (characters) | Gemini Advanced ($19.99/mo, capped) | 1M |
| DeepSeek V2 | $0.14 | Free (capped) | 128K |
| Perplexity Pro | N/A (search) | $20/mo (unlimited queries) | N/A |

Data Takeaway: The most successful consumer AI products (ChatGPT, Perplexity) are moving toward flat-rate subscriptions, while API pricing remains stubbornly token-based. This split suggests that the market is voting with its wallet: users prefer predictable costs for deep, sustained use.

Industry Impact & Market Dynamics

The shift from metered to unlimited AI pricing could reshape the entire AI stack. According to industry estimates, the global AI market is projected to grow from $200 billion in 2024 to over $1.5 trillion by 2030. A significant portion of this growth depends on enterprise adoption, which is currently hindered by unpredictable API costs.

The Enterprise Barrier: A 2024 survey by a major consulting firm (not named) found that 67% of enterprises cite 'cost unpredictability' as a top barrier to deploying AI agents for complex workflows. For example, a customer support bot that handles multi-turn conversations could cost $0.50 per session under token pricing, making it uneconomical for high-volume use. Under an unlimited model, the marginal cost per session drops to near zero, enabling new use cases like 24/7 personalized tutoring or real-time code pair programming.

Market Cap Impact: Companies that pioneer unlimited token models could capture significant market share. Consider the following scenario:

| Pricing Model | User Behavior | Estimated Revenue per User (Annual) | Churn Rate |
|---|---|---|---|
| Token-based (API) | Shallow, transactional | $1,200 (heavy users) | 25% |
| Unlimited subscription | Deep, exploratory | $240 (flat fee) | 10% |
| Hybrid (token + subscription) | Mixed | $600 (average) | 15% |

Data Takeaway: While token-based pricing generates higher per-user revenue from heavy users, the high churn and limited adoption among light users mean that total addressable market is smaller. Unlimited models sacrifice short-term ARPU for dramatically lower churn and broader adoption, potentially leading to higher lifetime value.

The 'Smart Vending Machine' vs. 'Co-pilot' Divide: The current token model reduces AI to a vending machine: insert tokens, get output. This commoditizes intelligence, making it a transaction rather than a relationship. Unlimited tokens enable a paradigm where AI becomes a persistent collaborator—always available, always learning from context. This is critical for agentic workflows, where an AI agent might need to iterate on a task for hours, making thousands of API calls. Under token pricing, such an agent would be prohibitively expensive.

Risks, Limitations & Open Questions

Abuse and Fairness: Unlimited token models are vulnerable to abuse—a single user could run massive batch jobs, consuming disproportionate compute. Providers must implement fair-use policies, rate limits, or compute-time caps. The challenge is to prevent abuse without reintroducing the 'meter' mentality.

Compute Cost Reality: The underlying hardware (Nvidia H100/B200 GPUs) is expensive. A single H100 costs ~$30,000 and consumes 700W. Unlimited token plans require massive over-provisioning, which could lead to financial losses if not carefully managed. The economics only work if model efficiency continues to improve (e.g., via quantization, pruning, or better architectures).

Quality vs. Quantity: Unlimited tokens could encourage 'spammy' usage—users generating vast amounts of low-quality content. This could degrade the user experience and strain infrastructure. Providers will need to implement quality controls or reputation systems.

Open Source Alternative: Open-source models like Llama 3 (70B) or DeepSeek-V2 can be self-hosted, effectively providing unlimited tokens at the cost of hardware. This creates a natural ceiling on API pricing: if token prices are too high, enterprises will self-host. The unlimited token model must be priced competitively against self-hosting.

AINews Verdict & Predictions

Our Verdict: Token-based pricing is a relic of the cloud API era and is actively hindering AI's evolution. It penalizes the very behaviors—depth, iteration, collaboration—that make AI valuable. The industry must move to unlimited token models, not as a marketing gimmick but as a fundamental rethinking of AI economics.

Predictions:
1. By Q3 2026, at least two major API providers (likely Anthropic and Google) will introduce 'unlimited token' enterprise plans with fair-use caps, following OpenAI's consumer lead.
2. By 2027, the majority of new AI-native applications (agents, coding assistants, creative tools) will be priced on a flat-rate subscription basis, with API usage as a premium add-on.
3. The 'depth tax' will disappear: Models will be optimized for reasoning quality, not token efficiency. Benchmarks like MMLU and GPQA will see rapid improvement as users are no longer cost-constrained.
4. A new class of 'AI co-pilot' startups will emerge that offer unlimited token access as their core differentiator, targeting knowledge workers who need deep, sustained collaboration.
5. The ultimate winner will be the model that can deliver the highest reasoning quality per unit of compute, not the cheapest per token. This will accelerate research into efficient architectures (e.g., mixture of experts, linear attention).

What to Watch: The next major model release from any frontier lab (OpenAI's GPT-5, Anthropic's Claude 4, Google's Gemini 2) should be evaluated not just on benchmark scores, but on whether their pricing model encourages or discourages deep reasoning. The model that breaks the token meter first will likely define the next decade of AI.

More from Hacker News

No hay bala de plata en la IA: los costos ocultos de la magia tecnológicaThe AI industry is awash in a narrative of magic: code generators that write entire functions from a prompt, video modelAtlas Engine reescribe la inferencia de LLM desde cero: ¿Una revolución de Rust y CUDA?The AI inference landscape has long been dominated by engines built atop heavyweight frameworks like PyTorch and TensorFAtaque Hi-Vis: El Jailbreak de un Solo Disparo que Explota la Confianza de los LLM en las Actualizaciones del SistemaThe Hi-Vis attack represents a paradigm shift in adversarial prompt engineering, moving from brute-force probing to contOpen source hub3322 indexed articles from Hacker News

Archive

May 20261361 published articles

Further Reading

La plataforma de economía de IA que convierte el hype tecnológico en una clase de activo medibleEstá surgiendo una plataforma de inteligencia especializada para descifrar la dinámica financiera de la industria de la Sierra adquiere Fragment: redefiniendo el servicio al cliente con agentes basados en datosLa empresa de servicio al cliente con IA de Bret Taylor, Sierra, ha adquirido Fragment, una startup incubada en YC espec81.000 usuarios silenciosos revelan la realidad económica de la IA: del bombo a los duros cálculos de ROIUn análisis innovador de 81.000 sesiones de usuarios reales de IA revela un cambio silencioso pero sísmico: la economía Agentes de IA como ciudadanos digitales: Compras autónomas de NFT y gobernanza on-chainUn cambio de paradigma está en marcha en la intersección de la IA y Web3. Los agentes de IA ya no son solo herramientas,

常见问题

这次模型发布“Unlimited Tokens: Why Metered AI Pricing Is Killing True Intelligence”的核心内容是什么?

The dominant pricing model for large language models—charging per token—is increasingly seen as a bottleneck to AI's transformative potential. This metered approach, inherited from…

从“unlimited tokens vs token pricing AI comparison”看,这个模型发布为什么重要?

At the core of this debate lies the token—the atomic unit of text that models like GPT-4, Claude, and Llama process. A token is roughly 0.75 words in English, but its cost varies dramatically by model and provider. The t…

围绕“why token-based billing is bad for AI reasoning”,这次模型更新对开发者和企业有什么影响?

开发者通常会重点关注能力提升、API 兼容性、成本变化和新场景机会,企业则会更关心可替代性、接入门槛和商业化落地空间。