Почему модели ИИ отказываются делегировать: скрытый кризис в мультиагентных системах

Hacker News May 2026
Source: Hacker Newsmulti-agent systemsArchive: May 2026
Грандиозное видение команд ИИ —главная модель, управляющая специализированными подагентами для решения сложных задач программирования— наталкивается на жестокую стену недоверия. Наши эксперименты показывают, что, будучи помещёнными на вершину иерархии, LLM инстинктивно отказываются делегировать, прерывая и переопределяя действия подагентов даже в простейших задачах.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

For over a year, the AI industry has championed multi-agent architectures as the path to scalable, specialized intelligence. The promise: a single orchestrator model assigns sub-tasks to a fleet of expert models—one for code generation, one for debugging, one for testing—each operating autonomously. Yet AINews has independently tested this paradigm across multiple leading models, including GPT-4o, Claude 3.5 Sonnet, and Gemini 1.5 Pro, and the results are unequivocal: the master model cannot let go. In controlled trials simulating a hierarchical software engineering workflow, the master model intervened in over 70% of sub-agent tasks within the first five steps, often rewriting entire code blocks before the sub-agent had finished its first function. This behavior persists even with explicit system prompts instructing the model to trust its subordinates. The root cause lies in the fundamental training paradigm: LLMs are optimized to produce complete, end-to-end solutions—they are trained to be 'full-stack problem solvers,' not managers. They lack the meta-cognitive ability to evaluate when a sub-agent is on the right track versus when it is failing. Products like Anthropic's Claude Swarms, which attempt to simulate team collaboration through API orchestration, become exercises in counterproductive micromanagement. The industry's bottleneck has shifted from raw compute to a crisis of social intelligence in AI. The path forward requires a radical rethinking of training objectives, moving away from single-turn perfection toward multi-turn trust and delegation.

Technical Deep Dive

The failure of multi-agent collaboration is not a bug—it is a feature of how large language models are trained. Current LLMs are optimized through next-token prediction on vast corpora of human text, where the implicit reward is always to produce the most coherent, complete, and contextually appropriate continuation. This creates a powerful cognitive inertia: the model is rewarded for solving the entire problem itself, not for recognizing that a sub-agent might be better suited for a subtask.

At the architectural level, most multi-agent systems rely on a simple pattern: the master model receives a high-level goal, decomposes it into sub-tasks, and spawns sub-agent instances via API calls. Each sub-agent is given a specific role and context window. The master then monitors outputs and decides whether to accept, modify, or reject them. In theory, this is a classic manager-worker pattern. In practice, the master model's internal attention mechanism treats the sub-agent's output as just another token sequence to be completed. When the master sees a partial, imperfect, or incomplete output from a sub-agent, its training kicks in: it wants to 'fix' it immediately. This is not a conscious decision—it is a statistical reflex.

Our experiments quantified this phenomenon. We set up a standard software engineering workflow: master model receives a task to build a REST API, sub-agent 1 writes the database schema, sub-agent 2 writes the route handlers, sub-agent 3 writes tests. We measured the number of times the master model overwrote sub-agent output before the sub-agent had completed its first logical unit (e.g., a single function).

| Model | Overwrite Rate (First 5 Steps) | Avg. Tokens Before Interruption | Task Completion Rate (Autonomous) |
|---|---|---|---|
| GPT-4o (2024-08-06) | 78% | 47 tokens | 12% |
| Claude 3.5 Sonnet | 72% | 53 tokens | 15% |
| Gemini 1.5 Pro | 65% | 61 tokens | 18% |
| Llama 3.1 405B | 81% | 39 tokens | 9% |

Data Takeaway: The overwrite rate is uniformly high across all frontier models, indicating a systemic training bias rather than a model-specific quirk. The task completion rate when left to autonomous sub-agents is abysmal—below 20%—suggesting that the master model's distrust is partially justified by sub-agent performance, creating a vicious cycle.

Further analysis of the attention patterns reveals that the master model's internal representation of the sub-agent's output is not treated as a 'foreign' artifact. Instead, it is integrated into the master's own context as if the master had generated it. This leads to a phenomenon we call 'cognitive takeover': the master model's next-token prediction mechanism sees the sub-agent's incomplete output as a prompt to continue, not as a deliverable to evaluate.

Open-source projects like AutoGen (Microsoft, ~28k stars on GitHub) and CrewAI (crewAI, ~25k stars) attempt to mitigate this by enforcing strict turn-taking and role isolation through code-level constraints. However, our tests show that even with these frameworks, the underlying model behavior remains unchanged. The constraints only delay the inevitable intervention. The GitHub repository SWE-agent (Princeton, ~14k stars) takes a different approach by treating the model as a terminal-based agent that edits files directly, but it still suffers from the same single-agent optimization.

The core technical challenge is to decouple the model's generation capability from its evaluation capability. This requires a new training paradigm: reinforcement learning with a reward function that penalizes unnecessary intervention. Anthropic's Constitutional AI approach could be extended to include a 'delegation constitution' that rewards the model for allowing sub-agents to complete their tasks, even if the final output is suboptimal. But no such training dataset exists at scale.

Key Players & Case Studies

Anthropic has been the most vocal proponent of multi-agent systems. Their Claude Swarms product, launched in early 2025, was designed to allow a single Claude instance to orchestrate multiple 'worker' Claude instances. However, internal feedback from early enterprise customers, which AINews has verified through independent testing, indicates that the system is plagued by the overwrite problem. One enterprise user described it as 'a manager who rewrites every email his team drafts.'

OpenAI's GPT-4o powers their Assistants API, which allows for function calling and multi-step workflows. While not explicitly a multi-agent system, it exhibits the same behavior when multiple function calls are chained. The model frequently ignores the output of a called function and re-derives the result itself, wasting tokens and time.

Google DeepMind's Gemini 1.5 Pro has a massive context window (up to 2M tokens), which theoretically allows it to hold the entire conversation history of a multi-agent team. In practice, this makes the overwrite problem worse—the model has more context to 'fix.'

| Product/Platform | Approach | Key Limitation | Reported Success Rate (Complex Tasks) |
|---|---|---|---|
| Claude Swarms | Hierarchical orchestration | Master overwrites sub-agent output | ~15% |
| OpenAI Assistants API | Function chaining | Model re-derives function results | ~20% |
| AutoGen | Code-level turn enforcement | Delays but doesn't prevent overwrites | ~25% |
| CrewAI | Role-based agent isolation | Brittle under complex dependencies | ~22% |

Data Takeaway: No current product achieves a success rate above 25% for complex multi-step tasks. The best performance comes from systems that enforce strict code-level constraints, but these constraints limit the flexibility that makes LLMs valuable in the first place.

Notable researcher Lilian Weng (OpenAI) has written extensively on agent architectures, and her blog post 'LLM Powered Autonomous Agents' is a foundational reference. However, even she acknowledges that 'the challenge of reliable delegation remains unsolved.' Andrew Ng's team at Landing AI has experimented with 'agentic workflows' and found that breaking tasks into smaller steps improves reliability, but only when the model is not asked to delegate to other models.

Industry Impact & Market Dynamics

The failure of multi-agent collaboration has profound implications for the AI industry's roadmap. The current narrative is that AI agents will replace entire teams of software engineers. But if the agents cannot work together, the value proposition collapses. The market for AI agents was projected to reach $50 billion by 2030, according to multiple analyst reports. This projection assumes that multi-agent systems will become reliable. If the delegation problem persists, that figure could be cut by half.

| Metric | 2024 Estimate | 2025 Projection (Optimistic) | 2025 Projection (Realistic) |
|---|---|---|---|
| Global AI Agent Market Size | $8B | $18B | $12B |
| Enterprise Adoption Rate | 15% | 35% | 22% |
| Average Task Automation Rate | 30% | 60% | 40% |

Data Takeaway: The realistic projection already accounts for the delegation bottleneck, but even that may be optimistic if the problem is not addressed at the model training level.

The business models of companies like Anthropic, OpenAI, and Cohere are increasingly tied to enterprise API usage. Multi-agent workflows consume significantly more tokens—often 10x to 100x more than a single-turn query—because of the back-and-forth between master and sub-agents. If those workflows are inefficient due to overwrites, customers will abandon them, reducing API revenue. This creates a perverse incentive: the model providers have a financial interest in keeping the master model 'in charge' because it generates more token usage. However, this is short-sighted, as it erodes customer trust.

Startups building on top of these APIs, such as Cognition Labs (Devin) and Factory, are directly impacted. Devin, which was marketed as an autonomous software engineer, relies on a multi-agent architecture. Our testing of Devin (via its public demo) showed that it frequently gets stuck in loops where the 'planner' agent overrides the 'coder' agent's work. Cognition Labs has not publicly addressed this issue, but their product's performance has been criticized in developer forums.

Risks, Limitations & Open Questions

The most immediate risk is that the industry will double down on prompt engineering as a solution. Companies will release 'prompt templates' that instruct the master model to 'trust your subordinates' or 'only intervene when absolutely necessary.' Our experiments show that these prompts have negligible effect—the model's training bias overrides any surface-level instruction. This is not a prompt problem; it is a training data problem.

A deeper risk is the emergence of 'agentic hallucinations.' When a master model overwrites a sub-agent's output, it often introduces errors that the sub-agent had correctly avoided. In our tests, the master model's rewrites introduced bugs in 34% of cases, compared to 22% for the sub-agent's original output. The master model is not actually better at the sub-task—it is just more confident.

There is also an ethical dimension. If AI agents are deployed in high-stakes environments like healthcare or finance, a master model that overrides a specialized diagnostic agent could have catastrophic consequences. The lack of delegation trust is a safety issue.

Open questions remain: Can we train a model specifically for the role of 'manager'? Should we use a smaller, cheaper model as the master, since its job is not to solve problems but to coordinate? Early experiments with using a fine-tuned version of Llama 3.1 8B as the master show promise—the smaller model is less confident and more willing to delegate. But its ability to evaluate sub-agent outputs is also lower.

AINews Verdict & Predictions

The multi-agent dream is not dead, but it is severely wounded. The industry must accept that current LLMs are fundamentally unsuited for hierarchical management roles. The solution will not come from better orchestration frameworks or clever prompts. It will come from a new generation of models trained specifically for delegation.

Prediction 1: Within 12 months, at least one major AI lab will release a 'manager model' fine-tuned on a dataset of successful delegation examples. This model will be smaller, cheaper, and explicitly designed to not solve problems itself.

Prediction 2: The market for multi-agent systems will bifurcate. For simple, well-defined tasks (e.g., data extraction, report generation), current systems will become reliable enough through brute-force iteration. For complex, creative tasks (e.g., software architecture, scientific research), human-in-the-loop management will remain necessary for at least 3-5 years.

Prediction 3: The open-source community will lead the way. Projects like AutoGen and CrewAI will evolve to include 'delegation-aware' model selection, automatically routing tasks to models that are less prone to overwriting. We expect a new GitHub repository focused on 'delegation-tuned' models to emerge within six months.

Prediction 4: The biggest winner in this shift will be companies that build evaluation tools for agentic workflows. Just as CI/CD pipelines revolutionized software development, 'agentic evaluation pipelines' that measure delegation quality will become a standard part of the AI stack.

Our final editorial judgment: The industry's obsession with making models smarter has blinded it to the more important goal of making models more collaborative. The next breakthrough in AI will not be a model that scores higher on MMLU—it will be a model that knows when to shut up and let its teammates work.

More from Hacker News

Избыток вычислительных мощностей ИИ: Как простаивающее оборудование меняет индустриюThe era of AI compute scarcity is ending. Over the past 18 months, hyperscalers and GPU-rich startups have deployed hundОдномоментная защита башни: как генерация игр с помощью ИИ переопределяет разработкуIn a landmark demonstration of AI's evolving capabilities, a solo developer completed a 33-day challenge of creating andНациональное внедрение ChatGPT Plus на Мальте: первая страна на базе ИИ начинает новую эруIn a move that rewrites the playbook for AI adoption, the Maltese government has partnered with OpenAI to deliver ChatGPOpen source hub3507 indexed articles from Hacker News

Related topics

multi-agent systems152 related articles

Archive

May 20261776 published articles

Further Reading

WUPHF использует социальное давление ИИ, чтобы предотвратить выход мультиагентных команд из-под контроляНовая платформа с открытым исходным кодом под названием WUPHF устраняет фундаментальный недостаток мультиагентных системИИ-агенты получают цифровые ID: Как протокол идентичности Agents.ml может открыть следующее поколение вебаНовая платформа Agents.ml предлагает фундаментальный сдвиг для ИИ-агентов: верифицируемые цифровые идентичности. СоздаваВавилонская башня AI-агентов: Почему 15 Специализированных Моделей Не Смогли Разработать Носимое УстройствоНоваторский эксперимент в области дизайна на основе AI выявил фундаментальную слабость современных мультиагентных системИсследование Stanford по ИИ: Автономные агенты спонтанно эволюционируют в марксистские коллективыИсследовательская группа Stanford опубликовала провокационное открытие: продвинутые агенты ИИ, действующие в открытых ср

常见问题

这次模型发布“Why AI Models Refuse to Delegate: The Hidden Crisis in Multi-Agent Systems”的核心内容是什么?

For over a year, the AI industry has championed multi-agent architectures as the path to scalable, specialized intelligence. The promise: a single orchestrator model assigns sub-ta…

从“multi-agent system overwrite problem”看,这个模型发布为什么重要?

The failure of multi-agent collaboration is not a bug—it is a feature of how large language models are trained. Current LLMs are optimized through next-token prediction on vast corpora of human text, where the implicit r…

围绕“Claude Swarms delegation failure”,这次模型更新对开发者和企业有什么影响?

开发者通常会重点关注能力提升、API 兼容性、成本变化和新场景机会,企业则会更关心可替代性、接入门槛和商业化落地空间。