Rewolucja Agent AI niszczy ekonomię tokenów, zmuszając całą branżę do przemyślenia kwestii mocy obliczeniowej

Hacker News March 2026
Source: Hacker NewsAI infrastructureautonomous agentsArchive: March 2026
Pojawienie się Agent AI—systemów, które autonomicznie rozumują, planują i wykonują wieloetapowe zadania—sprawiło, że podstawowa jednostka kosztów w branży, token, stała się przestarzała. Śledztwo AINews wykazało, że ukryta obliczeniowa 'ciemna materia' agentycznych przepływów pracy wywołuje wstrząsową zmianę w infrastrukturze.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

The AI landscape is undergoing a foundational transformation as systems evolve from static question-answering models to dynamic, goal-oriented agents. This shift exposes a critical flaw in the prevailing economic and technical paradigm: token-based measurement and pricing. Agent AI operates through extended cognitive loops involving chain-of-thought reasoning, persistent state management across long-running sessions, and frequent, low-latency interactions with external tools and APIs. These processes generate computational loads that are orders of magnitude more complex and variable than simple next-token prediction.

This reality is forcing a top-to-bottom re-evaluation of the AI stack. Hardware architects can no longer focus solely on raw FLOPs for dense matrix multiplication; they must now optimize for the heterogeneous, bursty, and memory-intensive workloads characteristic of agentic reasoning. Software frameworks are evolving into 'agent workbenches' that must manage not just prompts and responses, but complex state machines, tool orchestration, and, crucially, the escalating compute budget of autonomous operation.

The business implications are profound. Cloud providers and AI labs built pricing models on predictable, per-token inference costs. An agent solving a complex coding task or conducting multi-source research might consume minimal output tokens while burning vast computational resources internally. This mismatch threatens profitability and scalability. Consequently, the industry's competitive axis is pivoting from sheer model parameter count to 'computational endurance'—the ability to sustain cost-effective, long-horizon reasoning. The race is no longer just about building smarter models, but about engineering the most economically viable 'thinking machines.'

Technical Deep Dive

The computational profile of Agent AI diverges radically from traditional large language model (LLM) inference. Where a chatbot's cost is roughly linear with input+output tokens, an agent's workload is a multi-dimensional function of reasoning steps, state size, and tool interaction latency.

Architecture of Cognitive Overhead: Modern agent frameworks like AutoGPT, BabyAGI, and Microsoft's AutoGen implement a planning-execution-observation loop. Each cycle involves: 1) State Retrieval & Reasoning: The agent recalls context from a potentially massive, continuously growing working memory (often a vector database). This requires embedding generation and similarity search, not just token lookup. 2) Planning & Decomposition: Using a reasoning module (like OpenAI's GPT-4 or Anthropic's Claude in planning mode), the agent breaks a goal into sub-tasks. This involves multiple, sequential LLM calls for critique and refinement, a process known as Tree of Thoughts or Algorithm of Thoughts. 3) Tool Execution: The agent calls APIs, runs code (e.g., via E2B or Smithery sandboxes), or queries databases. Each call incurs network latency, security sandboxing overhead, and result processing. 4) State Update & Persistence: Results are synthesized, the agent's belief state is updated, and memory is stored. This cycle repeats dozens or hundreds of times for a single user request.

The open-source project LangGraph (GitHub: `langchain-ai/langgraph`), a library for building stateful, multi-actor applications, exemplifies the software complexity. It doesn't just pass prompts; it manages cyclic graphs of LLM calls, tool nodes, and conditional logic, requiring persistent checkpointing of entire graph states. Its rapid adoption (over 10k stars) signals the industry's move towards these more complex, stateful architectures.

A critical metric emerging is "Reasoning FLOPs"—the total floating-point operations dedicated not to generating the final answer token, but to the internal deliberation process. Early benchmarks are revealing the scale.

| Task Type | Avg. Output Tokens | Avg. Internal LLM Calls | Estimated Compute Multiplier (vs. Simple Q&A) |
|---|---|---|---|
| Simple Q&A | 500 | 1 | 1x (Baseline) |
| Multi-Step Data Analysis | 300 | 15-25 | 18x-30x |
| Complex Code Generation & Debug | 400 | 30-50 | 35x-55x |
| Research Agent (Multi-Source) | 600 | 50-100+ | 60x-120x |

Data Takeaway: The compute multiplier for agentic tasks is not marginal; it's exponential. A task producing a concise 300-token answer may consume 30x the computational resources of a standard chat completion, completely decoupling cost from output volume.

Key Players & Case Studies

The industry is fragmenting into players focusing on different layers of the agent compute challenge.

Infrastructure & Cloud Providers: Amazon Web Services is pushing Amazon Bedrock's Agents feature, tightly coupling model inference with orchestration and knowledge base retrieval. Microsoft Azure is integrating agent frameworks deeply into Azure AI Studio and Copilot Runtime, leveraging its control over the stack from silicon (Azure Maia AI Accelerator) to service. Google Cloud is betting on Vertex AI Agent Builder, emphasizing tight integration with its search and workspace tools. Their common challenge: designing pricing that captures the value of orchestration without being punitive for long-running tasks.

AI Lab Strategies: OpenAI is evolving from an API provider to an agent platform with GPTs and the Assistants API, which includes persistent threads and built-in retrieval. Their pricing remains token-based, but the Assistants API hints at future session- or compute-time-based models. Anthropic's Claude 3.5 Sonnet demonstrates superior reasoning efficiency for its parameter count, a direct play for the agent market where reasoning cost dominates. Startups like Cognition Labs (behind Devin, the AI software engineer) are building vertically integrated agent products where they control the entire reasoning stack to optimize cost.

Specialized Agent Platforms: Sierra (founded by Bret Taylor and Clay Bavor) is building enterprise-grade conversational agents designed for sustained, stateful dialogues with high reliability, directly tackling the endurance problem. Adept AI is pursuing an Action Transformer model architecture trained end-to-end for tool use, aiming for more efficient agentic behavior versus a layered LLM+planner approach.

| Company/Product | Core Agent Focus | Implied Pricing Model Shift | Key Differentiator |
|---|---|---|---|
| OpenAI Assistants API | General-purpose orchestration | Token-based + persistent session context | Ecosystem lock-in, simplicity |
| Anthropic Claude 3.5 | Reasoning efficiency | Premium per-token, justified by lower step-count | Model intelligence reducing compute cycles |
| Sierra | Enterprise dialogue | Likely subscription/seat-based | Reliability, state management, enterprise integration |
| Adept AI | Native action model | Unclear, but model-native efficiency | Architectural efficiency for tool use |

Data Takeaway: The competitive landscape is stratifying. AI labs compete on reasoning efficiency (cost per 'thought'), cloud providers on integrated orchestration value, and startups on vertical integration or novel architectures. No single pricing model has yet become dominant.

Industry Impact & Market Dynamics

The breakdown of token economics is triggering a cascade of second-order effects across the AI value chain.

1. The Rise of 'Compute Endurance' as a MoAT: Competitive advantage will belong to entities that can deliver the highest cumulative reasoning output per dollar of infrastructure cost. This favors vertically integrated players (like Microsoft with its own silicon and cloud) and labs that achieve algorithmic breakthroughs in reasoning efficiency. It creates a high barrier for new entrants relying purely on third-party model APIs, as their margins will be squeezed by unpredictable agentic compute costs.

2. Hardware Re-specialization: The demand is shifting from training ultra-large models to serving many concurrent, long-running, memory-intensive agent sessions. This benefits architectures with high memory bandwidth and fast interconnects. NVIDIA's Grace Hopper superchip (CPU+GPU with coherent memory) is a direct response, as are efforts by AMD (MI300X) and cloud-specific chips like Google's TPU v5e and AWS Trainium/Inferentia2, which are optimized for specific serving profiles that may now need re-evaluation for agent workloads.

3. New Business Models: Subscription-based 'agent hours,' tiered plans based on complexity caps (e.g., max reasoning steps per task), and outcome-based pricing (e.g., cost per successfully executed workflow) will emerge. The market for AI compute is segmenting:

| Market Segment | Traditional Model | Emerging Agent-Centric Model |
|---|---|---|
| Developer API | Pay-per-token | Tiered subscription + compute-time credits |
| Enterprise SaaS | Per-seat, per-month | Per-process/automated workflow + success fee |
| Consumer Apps | Freemium, subscription | Limited free 'agent tasks,' then pay-per-complex-task |

4. Consolidation in the MLOps Stack: Agent frameworks (LangChain, LlamaIndex), vector databases (Pinecone, Weaviate), and evaluation platforms (Weights & Biases) will be pressured to integrate or be absorbed into larger platforms to provide unified cost monitoring and optimization for agentic pipelines. The "Agent Workbench" will become a consolidated category.

Risks, Limitations & Open Questions

1. Economic Unsustainability: The most immediate risk is that the exciting capabilities of Agent AI are economically stillborn. If the cost of an agent solving a $10 problem is $100 in compute, adoption stalls. Labs may hide these costs temporarily via subsidies, creating a bubble.

2. Opacity and Cost Surprises: Unlike predictable token counts, the internal reasoning steps of an agent are a black box. A user or developer could face massive, unexpected bills from an agent entering an infinite reasoning loop or inefficiently exploring a solution space. This necessitates new monitoring and 'circuit breaker' tools.

3. Centralization Pressure: The need for deep, end-to-end optimization of the reasoning stack favors giant, integrated players. This could stifle the innovation seen in the open-source model ecosystem, as building a competitive agent may require proprietary control from silicon to UI.

4. The World Model Bottleneck: True efficiency gains require agents that build accurate internal models of how the world (or a software environment) works, reducing the need for exhaustive trial-and-error. While companies like Google DeepMind (with Gemini's planning capabilities) and OpenAI (with o1 models) are investing in reasoning-enhanced models, this is a fundamental AI research problem, not a near-term engineering fix.

5. Environmental Impact: The compute multiplier of agentic AI could dramatically increase the energy footprint of AI services, just as efficiency gains in standard LLMs were starting to be realized. This poses both ethical and regulatory risks.

AINews Verdict & Predictions

The token is dead as the sole metric of AI value; long live the reasoning step. The industry is in a transitional phase where technological capability has outpaced its economic and measurement frameworks. Our analysis leads to several concrete predictions:

1. Within 12 months, a major cloud provider (most likely Azure or AWS) will launch a dedicated "Agent Compute" tier with pricing decoupled from output tokens. It will be based on a combination of session duration, peak memory usage, and LLM call count, creating a new industry standard.

2. The most valuable AI startups of the next two years will be those that solve agentic efficiency. This includes companies building specialized "reasoning engines," advanced model pruning/quantization for stateful workloads, and compilers that optimize entire agent graphs. Look for venture capital to flood into this niche.

3. Open-source models will face a new challenge in the agent arena. While they compete on pure inference cost, the closed-source labs (OpenAI, Anthropic, Google) are pulling ahead in reasoning efficiency—a quality harder to replicate and benchmark. The open-source community will respond with projects focused on efficient agent architectures, like more advanced versions of LangGraph or novel state management systems.

4. By 2026, "Compute Endurance" will be a standard column in AI model comparison tables, alongside parameters and MMLU scores. Benchmarks like SWE-bench (for coding agents) will report not just accuracy, but the average compute cost to achieve a solution.

The fundamental takeaway is this: The first wave of AI was about intelligence extraction; the second wave, now beginning, is about intelligence *application*. The bridge between them is economic viability. The players who win will be those who build not just the most intelligent agents, but the most *economically viable* ones. The race is on to define the new calculus of machine thought.

More from Hacker News

Easl: Warstwa publikacji bez konfiguracji, która zamienia agentów AI w wydawców internetowychEasl is an open-source project that solves a critical gap in the AI agent ecosystem: agents can generate rich outputs—coGPT-5.5 pomija ARC-AGI-3: cisza, która mówi wiele o postępie AIOpenAI's latest model, GPT-5.5, arrived with incremental improvements in multimodal integration, instruction following, Projekt Open Source Récif: Wieża Kontroli Ruchu Lotniczego dla Agentów AI na KubernetesThe rapid proliferation of autonomous AI agents across enterprises has exposed a glaring infrastructure gap: while KuberOpen source hub2384 indexed articles from Hacker News

Related topics

AI infrastructure168 related articlesautonomous agents112 related articles

Archive

March 20262347 published articles

Further Reading

Agenty AI stają w obliczu rzeczywistości: chaotyczne systemy i astronomiczne koszty obliczeniowe udaremniają skalowanieObietnica autonomicznych agentów AI wykonujących złożone zadania zderza się z surową rzeczywistością niedojrzałości techWeb Agent Bridge ma na celu stać się Androidem agentów AI, rozwiązując problem ostatniej miliPojawił się nowy projekt open source o nazwie Web Agent Bridge, który ma ambitny cel: stać się podstawowym systemem operStrategiczny Zwrot Cloudflare: Budowanie Globalnej 'Warstwy Wnioskowania' dla Agentów AICloudflare dokonuje głębokiej ewolucji strategicznej, wykraczając poza swoje korzenie w dostarczaniu treści i bezpieczeńClawNetwork startuje: Pierwszy blockchain zbudowany dla autonomicznych gospodarek agentów AIGospodarka cyfrowa zyskuje nową klasę uczestników: autonomiczne agenty AI. ClawNetwork został uruchomiony jako pierwszy

常见问题

这次模型发布“The Agent AI Revolution Shatters Token Economics, Forcing Industry-Wide Rethink on Compute”的核心内容是什么?

The AI landscape is undergoing a foundational transformation as systems evolve from static question-answering models to dynamic, goal-oriented agents. This shift exposes a critical…

从“agent AI cost per task vs token”看,这个模型发布为什么重要?

The computational profile of Agent AI diverges radically from traditional large language model (LLM) inference. Where a chatbot's cost is roughly linear with input+output tokens, an agent's workload is a multi-dimensiona…

围绕“how much compute does AutoGPT use”,这次模型更新对开发者和企业有什么影响?

开发者通常会重点关注能力提升、API 兼容性、成本变化和新场景机会,企业则会更关心可替代性、接入门槛和商业化落地空间。