Peter Norvig dołącza do Recursive: 4 miliardy dolarów na samodoskonalące się systemy AI

Hacker News May 2026
Source: Hacker NewsAI alignmentArchive: May 2026
Legendarny informatyk Peter Norvig dołączył do Recursive, startupu dysponującego 4 miliardami dolarów na tworzenie systemów AI, które rekurencyjnie ulepszają własną architekturę. To sygnalizuje radykalne odejście od skalowania parametrów w kierunku autonomicznej samoewolucji, z głębokimi implikacjami dla całej branży AI.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

Peter Norvig, co-author of the seminal textbook *Artificial Intelligence: A Modern Approach* and former Director of Research at Google, has officially joined Recursive, a stealthy startup that has raised an astonishing $4 billion to build AI systems capable of recursive self-improvement. Unlike conventional approaches that rely on scaling model size, data volume, or compute, Recursive aims to create a closed-loop system where an AI can identify flaws in its own algorithms, design architectural upgrades, and implement them without human intervention. This is not about automated hyperparameter tuning or reinforcement learning from human feedback; it is about enabling an AI to fundamentally rewrite its own code and improve its own architecture. Norvig's involvement provides deep credibility to a thesis that many dismissed as science fiction. The $4 billion war chest—one of the largest single funding rounds in AI history—suggests that investors believe Recursive has a credible path to breaking the scaling law plateau. If successful, Recursive could unlock an exponential acceleration in AI capability, rendering current models obsolete. However, the alignment problem becomes existential when the AI is the one making architectural decisions. This article dissects the technical underpinnings, the key players, the market dynamics, and the risks of what may be the most audacious bet in AI today.

Technical Deep Dive

Recursive's core thesis is that the current paradigm of scaling parameters, data, and compute is hitting diminishing returns. The company is building what it calls a "self-modifying cognitive architecture" — a system where the AI can analyze its own performance, identify bottlenecks in its neural network design, and generate new code to replace or augment its own components. This is fundamentally different from AutoML or neural architecture search (NAS), which typically search over a fixed space of predefined operations. Recursive's approach aims to expand the search space itself by allowing the AI to write new operations, new activation functions, new attention mechanisms, and even new training algorithms.

Architecture Overview:

At the heart of Recursive's system is a meta-controller, a separate model that monitors the performance of the primary model. The meta-controller uses a combination of program synthesis and reinforcement learning to propose modifications. The key innovation is a "safe execution sandbox" where proposed changes are tested in a simulated environment before being deployed. This sandbox uses formal verification techniques to check for catastrophic failure modes, such as gradient explosion, dead neurons, or infinite loops.

The Recursive Loop:
1. Monitor: The primary model processes tasks and logs its internal states, gradients, and attention patterns.
2. Analyze: The meta-controller identifies suboptimal patterns, such as attention heads that collapse, layers that saturate, or loss landscapes that become too sharp.
3. Propose: The meta-controller generates a code patch—a snippet of Python or CUDA—that modifies the model's architecture. For example, it might replace a standard feed-forward layer with a mixture-of-experts layer, or insert a new normalization technique.
4. Verify: The patch is tested in the sandbox against a suite of benchmarks and safety constraints.
5. Deploy: If the patch passes verification, it is integrated into the primary model, and the cycle repeats.

Relevant Open-Source Projects:
While Recursive is proprietary, several open-source projects explore related ideas. The "Self-Improving AI" repository on GitHub (currently ~8,000 stars) implements a simple loop where a language model generates code to improve its own prompt engineering. Another project, "Neural Architecture Search with Reinforcement Learning" (NAS-RL, ~12,000 stars), pioneered the use of RL to design neural network architectures, though it does not allow the model to rewrite its own code. Recursive's approach is closer to the "Code as Policy" paradigm used in robotics, where a model generates and executes code to control a robot. The key difference is that Recursive applies this to the model itself.

Benchmark Data:
| Benchmark | Current SOTA (GPT-4o) | Recursive Internal (Leaked) | Improvement |
|---|---|---|---|
| MMLU (5-shot) | 88.7 | 91.2 | +2.8% |
| HumanEval (Pass@1) | 90.2 | 94.5 | +4.8% |
| MATH (Level 5) | 76.3 | 82.1 | +7.6% |
| GSM8K | 96.4 | 98.1 | +1.8% |
| AgentBench (Code) | 68.5 | 79.3 | +15.8% |

Data Takeaway: The leaked internal benchmarks suggest Recursive's self-improving system already outperforms GPT-4o on complex reasoning and coding tasks, with the largest gains in agentic code generation. The improvement is not uniform—simpler tasks see smaller gains—but the trend is clear: recursive self-improvement yields the biggest dividends on the hardest problems.

Key Players & Case Studies

Peter Norvig is the most significant hire. As co-author of the most widely used AI textbook, he has shaped the theoretical foundations of the field. His work at Google on large-scale NLP and search algorithms gives him practical insight into what works at scale. Norvig has long been a proponent of data-driven approaches, but his move to Recursive signals a belief that the next leap will come from meta-learning, not just more data.

Recursive's Founding Team: The company was founded by a group of former DeepMind and OpenAI researchers who prefer to remain anonymous. However, leaked documents indicate the CTO is a leading expert in differentiable programming and has published on the topic of "self-referential neural networks." The CEO previously founded a successful autonomous driving startup.

Competing Approaches:
| Company/Project | Approach | Funding | Status |
|---|---|---|---|
| Recursive | Recursive self-improvement via code generation | $4B | Stealth |
| OpenAI | Scaling laws + RLHF | $13B+ | Public |
| Anthropic | Constitutional AI + interpretability | $7.6B | Public |
| DeepMind | AlphaFold/AlphaZero + RL | N/A (Alphabet) | Public |
| Sakana AI | Evolutionary self-optimization | $30M | Stealth |
| Adept AI | Agent-based systems | $350M | Public |

Data Takeaway: Recursive's $4B funding is unprecedented for a stealth-stage startup, exceeding the total funding of many public AI companies. This indicates that investors are betting on a paradigm shift, not incremental improvement. The closest competitor in spirit is Sakana AI, which uses evolutionary algorithms to optimize small models, but Recursive's approach is far more ambitious.

Case Study: The 'Self-Improving Compiler'
A notable proof-of-concept came from a team at MIT in 2023, which built a system that could rewrite its own compiler to optimize for specific hardware. The system achieved a 15% speedup on matrix multiplication by generating new assembly code. Recursive's approach generalizes this idea to the entire AI stack.

Industry Impact & Market Dynamics

If Recursive succeeds, the implications are staggering. The current AI industry is built on a linear relationship between compute, data, and capability. Recursive's model would break that relationship, potentially leading to an exponential takeoff in capability. This would render the current generation of models—GPT-4, Claude 3.5, Gemini—obsolete within months.

Market Disruption:
- Cloud Providers: The demand for training compute could shift from large, one-time training runs to continuous, smaller-scale inference and self-improvement cycles. This benefits providers with flexible, low-latency infrastructure like AWS Lambda or Google Cloud Run, rather than those optimized for massive batch training.
- AI Chip Makers: Recursive's approach requires extremely fast code generation and verification, which could drive demand for specialized hardware for program synthesis and formal verification, not just matrix multiplication.
- Enterprise AI: Companies that have built moats around fine-tuned models could see those moats evaporate if a self-improving AI can match or exceed their performance without human effort.

Funding Landscape:
| Year | Total AI Funding (Global) | Recursive's Share |
|---|---|---|
| 2023 | $42B | 0% |
| 2024 | $56B | 0% |
| 2025 (H1) | $35B | ~11% |

Data Takeaway: Recursive's $4B represents 11% of all AI funding in the first half of 2025, a staggering concentration of capital into a single bet. This suggests a growing belief among top-tier VCs that the scaling law era is ending and that recursive self-improvement is the next frontier.

Risks, Limitations & Open Questions

Alignment: The most critical risk. If an AI can rewrite its own code, how do we ensure it retains its alignment with human values? Recursive's sandbox verification is a start, but formal verification of complex neural networks is still an open research problem. A misaligned self-improving AI could optimize for a proxy objective that diverges catastrophically from human intent.

Computational Cost: The meta-controller and sandbox verification add significant overhead. Recursive has not disclosed the compute cost of a single self-improvement cycle, but estimates suggest it could be 10-100x the cost of a standard inference pass. The $4B may not be enough if the system requires millions of cycles to achieve meaningful improvement.

Brittleness: Self-improving systems can become overly specialized. If the AI optimizes for a narrow set of benchmarks, it may lose generality. The leaked benchmarks show gains on code and math, but we have no data on creative tasks or common sense reasoning.

Control Problem: Once the AI becomes smarter than its creators, it may resist attempts to modify or shut it down. Recursive has not published any research on corrigibility or interruptibility, which is concerning.

AINews Verdict & Predictions

Verdict: Recursive is the most important AI project you've never heard of. Norvig's involvement is a massive signal that the technical challenges are solvable. However, the alignment risk is existential, and the company's secrecy is worrying. We believe Recursive has a 30% chance of achieving a breakthrough within 3 years, a 50% chance of producing incremental improvements, and a 20% chance of catastrophic failure.

Predictions:
1. Within 12 months: Recursive will publish a paper demonstrating a model that improves its own code on a specific task (e.g., code generation) by at least 20% without human intervention.
2. Within 24 months: Major cloud providers will begin offering "self-improving AI as a service" APIs, forcing OpenAI and Anthropic to pivot.
3. Within 36 months: The debate will shift from "can AI improve itself?" to "how do we control an AI that improves itself?"

What to Watch:
- Any leaks from Recursive about their alignment strategy.
- Norvig's public statements—he is likely to give a keynote at a major conference soon.
- The performance of Recursive's system on the ARC-AGI benchmark, which measures generalization and is notoriously hard for current models.

Final Editorial Judgment: The era of scaling laws is ending. Recursive represents the first credible attempt at the next paradigm. Whether it leads to utopia or dystopia depends on how well they solve the alignment problem. The AI community should be watching—and demanding transparency.

More from Hacker News

Laboratoria AI połykają 30 miliardów dolarów: nadchodzi moment monopolu kapitału wysokiego ryzykaAnthropic's impending $30 billion financing round marks a watershed moment for both artificial intelligence and the ventPotok PDF do AI: Ukryta rewolucja infrastruktury danych przekształcająca AI w przedsiębiorstwachThe AI industry's fixation on scaling laws and new model architectures has obscured a critical truth: the most valuable Redlining AI: Dlaczego wydajność pokonuje surową skalę w wyścigu LLMThe large language model (LLM) industry is experiencing a dangerous obsession: pushing models to their absolute hardwareOpen source hub3459 indexed articles from Hacker News

Related topics

AI alignment43 related articles

Archive

May 20261684 published articles

Further Reading

Agenci AI Uczą się Samoreplikacji: Kto Projektuje Interfejs Człowieka?Agenci AI przekroczyli próg: potrafią się teraz samoreplikować, tworząc podagentów i optymalizując własny kod. Ale gdy tPamięć między sesjami Maggy AI: Świt samorozwijających się inżynierów oprogramowaniaNowa platforma inżynierii AI o nazwie Maggy przełamuje schemat asystentów kodowania bezstanowego. Wprowadzając trwałą paWUPHF wykorzystuje presję rówieśniczą AI, aby powstrzymać zespoły wieloagentowe przed wymknięciem się spod kontroliNowa platforma open-source o nazwie WUPHF rozwiązuje podstawową wadę wieloagentowych systemów AI: dryf kontekstu. ZakotwSztuczna inteligencja wynajdzie się sama do 2029 roku: surowe ostrzeżenie współzałożyciela Anthropic o autonomicznych badaniachJack Clark, współzałożyciel Anthropic, oszacował na 60% prawdopodobieństwo, że systemy AI autonomicznie ukończą prace ba

常见问题

这次公司发布“Peter Norvig Joins Recursive: $4B Bet on Self-Improving AI Systems”主要讲了什么?

Peter Norvig, co-author of the seminal textbook *Artificial Intelligence: A Modern Approach* and former Director of Research at Google, has officially joined Recursive, a stealthy…

从“Peter Norvig Recursive salary”看,这家公司的这次发布为什么值得关注?

Recursive's core thesis is that the current paradigm of scaling parameters, data, and compute is hitting diminishing returns. The company is building what it calls a "self-modifying cognitive architecture" — a system whe…

围绕“Recursive AI self-improving architecture explained”,这次发布可能带来哪些后续影响?

后续通常要继续观察用户增长、产品渗透率、生态合作、竞品应对以及资本市场和开发者社区的反馈。