OpenMythos與循環Transformer的崛起:重新定義超越注意力機制的AI架構

Hacker News April 2026
Source: Hacker NewsAI architecturelong-context AIArchive: April 2026
開源專案OpenMythos正在挑戰現代AI的一項基本原則:Transformer的前饋架構。透過提出「循環Transformer」設計,它旨在解決長上下文處理與計算效率的核心限制。這標誌著一個關鍵轉變,可能重塑AI模型的基礎。
The article body is currently shown in English by default. You can generate the full version in this language on demand.

A quiet revolution is brewing in the open-source AI community, centered on a project called OpenMythos. Rather than fine-tuning existing large language models (LLMs), its contributors are attempting a foundational reimagining of the core AI architecture itself. The project's central thesis is that the standard Transformer, while revolutionary, is inherently inefficient for tasks requiring persistent state, long-term memory, and iterative, multi-step reasoning. Its proposed solution is a 'Recurrent Transformer'—a hybrid architecture that integrates recurrent neural network (RNN) principles, specifically stateful loops, into the Transformer block. This is not merely an incremental tweak but a philosophical departure from the stateless, attention-heavy processing that defines models like GPT-4, Claude, and Llama.

The significance of OpenMythos lies not in its immediate production readiness—it remains largely a research framework—but in what it represents. It is a concrete manifestation of the growing consensus that the path to more capable, efficient, and agentic AI systems may require moving beyond the pure Transformer. The project directly addresses critical bottlenecks: the quadratic computational complexity of attention with sequence length, the model's 'amnesiac' nature between forward passes, and the difficulty of maintaining coherent internal state across extended interactions or documents. By enabling iterative refinement of a hidden state within a fixed computational budget, OpenMythos's architecture hints at a future where AI can 'think' more like a process than a single, massive prediction. This has profound implications for the development of advanced AI agents, persistent world models, and real-time interactive systems, potentially lowering the computational barrier to sophisticated AI and shifting competitive advantage from sheer scale to architectural ingenuity.

Technical Deep Dive

The OpenMythos architecture proposes a fundamental re-engineering of the Transformer block. In a standard Transformer, each layer processes an input sequence through self-attention and a feed-forward network, passing the entire transformed sequence to the next layer. Information flow is strictly feed-forward within a single forward pass. OpenMythos introduces a recurrent loop *within* the block or across a small group of blocks.

At its core, the proposed Recurrent Transformer block maintains a latent state vector that is updated iteratively. For a given input (or a chunk of a long sequence), the block processes it once, updates its internal state, and can then optionally process the same input again with the new context provided by the updated state. This allows for iterative refinement. A simplified conceptual update for a block's hidden state `h_t` at step `t` within a processing cycle might look like: `h_t = LayerNorm(Attention(Concat(x, h_{t-1})) + FFN(Concat(x, h_{t-1})))`, where `x` is the input. The key is that the parameters of the Attention and FFN layers are shared across these recurrent steps, dramatically increasing the effective 'depth' of processing without adding parameters.

This design draws inspiration from several research threads. The Universal Transformers (Dehghani et al., 2018) introduced adaptive computation time and recurrence across layers. More recently, models like RWKV (an RNN-style architecture with Transformer-level performance) and Mamba (a state space model) have demonstrated the viability of non-attention-based, stateful models for sequences. OpenMythos appears to be a synthesis, attempting to retain the expressive power of attention while grafting on the memory and efficiency of recurrence.

The primary GitHub repository, `open-mythos/arch`, provides a PyTorch implementation of the core building blocks. While still experimental, it has garnered significant interest, with over 2.8k stars and active forks exploring integration with existing model frameworks like Hugging Face's Transformers library. Early, non-peer-reviewed benchmarks shared by contributors on synthetic tasks show promising results on algorithmic tasks that require memory, such as copying long sequences or performing iterative arithmetic.

| Architecture | Key Mechanism | Context Window Scaling | Stateful Memory | Inference Cost (Relative) |
|---|---|---|---|---|
| Standard Transformer | Global Self-Attention | O(N²) memory, O(N²) time | No (Context Window Only) | 1.0 (Baseline) |
| Transformer + RoPE/ALiBi | Positional Encoding | Linear/Log Attention | No | ~0.9-1.0 |
| Mamba (SSM) | Selective State Space | Linear | Yes (Implicit) | ~0.3-0.5 |
| OpenMythos (Proposed) | Recurrent + Local Attention | Linear per step, O(kN) for k steps | Yes (Explicit Latent State) | ~0.4-0.7 (est.) |

Data Takeaway: The table highlights the trade-off space. Standard Transformers pay a heavy price for long context. While efficient alternatives like Mamba exist, OpenMythos seeks a middle ground, offering explicit statefulness with a more moderate estimated efficiency gain, betting that the retained attention mechanism is worth the overhead for certain reasoning tasks.

Key Players & Case Studies

The development of OpenMythos is not happening in a vacuum. It reflects a broader strategic pivot by both research institutions and corporations who are hedging against the limitations of the Transformer monopoly.

Research Pioneers: The project's conceptual debt is clear. Albert Gu's work on Mamba at Carnegie Mellon and Stanford has been a watershed, proving that state space models can compete with Transformers on language. Similarly, the RWKV project, led by Bo Peng, has built a large community around its attention-free RNN architecture. OpenMythos contributors are explicitly trying to bridge these worlds. Notably, researchers from Meta's FAIR lab have published on Infini-Transformer, which introduces a compressive memory for unbounded context, a complementary approach to the recurrence problem. DeepMind's Gemini models reportedly use a mixture of experts (MoE) for efficiency, but not fundamental architectural recurrence.

Corporate Strategic Moves: While not directly linked to OpenMythos, corporate R&D shows where the wind is blowing. Google DeepMind has long invested in memory-augmented networks. Anthropic's Claude and its 100K+ context window rely on sophisticated positional encodings and caching, not architectural change, pushing the current paradigm to its limit. xAI's Grok-1 is a standard Transformer MoE model. However, the most telling case is Microsoft Research. Their recent LongNet and RetNet papers propose dilated attention and retention mechanisms, respectively, explicitly aiming for efficient long sequences. RetNet's 'retention' mechanism is a parallelizable, recurrent-like structure—a close cousin to OpenMythos's goals. Microsoft's deep integration of AI into Windows and its pursuit of persistent, agentic Copilots creates a direct need for the stateful, efficient architectures OpenMythos explores.

| Entity | Project/Model | Approach to Long Context/State | Open Source? | Strategic Goal |
|---|---|---|---|---|
| OpenMythos Community | Recurrent Transformer | Architectural Recurrence | Yes (Core) | Prove hybrid architecture viability |
| Microsoft Research | RetNet | Retention (Recurrent + Parallel) | Yes (Paper/Code) | Enable efficient OS-level agents |
| Carnegie Mellon/Stanford | Mamba | State Space Models (SSM) | Yes | Replace Attention |
| RWKV | Eagle (7B) | RNN with Transformer-like FFN | Yes | Create full RNN ecosystem |
| Anthropic | Claude 3 | Scaling & Advanced Positional Encoding | No | Maximize current Transformer utility |
| Google DeepMind | Gemini 1.5 | MoE + Efficient Attention | No (API) | Push multimodal context limits |

Data Takeaway: The competitive landscape reveals a split. Large API providers (Anthropic, Google) are optimizing the Transformer to its breaking point. Meanwhile, Microsoft and open-source collectives are actively prototyping successors. OpenMythos's fully open approach positions it as a testbed for ideas that may be too radical for corporate roadmaps but could define the next era.

Industry Impact & Market Dynamics

If successful, architectures inspired by OpenMythos could trigger a significant market realignment. The current AI economy is built on the Transformer's hunger for compute: training costs in the hundreds of millions, inference costs scaling with context length, and a high barrier to entry for creating cutting-edge models. A shift to more efficient, stateful architectures would disrupt this calculus.

First-Order Impact: Cost and Accessibility. The primary promise is a drastic reduction in inference latency and cost for long-context and multi-turn tasks. This would make advanced AI agent features—which require maintaining session state over hours or days—economically viable for millions of applications, not just tech giants. Startups could deploy sophisticated, persistent assistants without prohibitive cloud bills. The market for edge AI, where compute and memory are constrained, would expand dramatically.

Second-Order Impact: New Application Paradigms. The ability to maintain a compact, evolving internal state enables new product categories:
1. True Persistent AI Agents: Agents that remember user preferences, learn from ongoing interaction, and execute complex, multi-day workflows without constant re-prompting or expensive context re-injection.
2. Dynamic World Models for Simulation: Games, virtual environments, and robotics simulators could use a single, efficient model to maintain the state of complex systems, predicting outcomes over long time horizons.
3. Evolutive Content Creation: Tools for long-form narrative generation, music composition, or codebase evolution where the AI's 'vision' iteratively refines, akin to a human writer's developing plot.

This would shift value from pure model scale (parameter count) to architectural design and data efficiency. Companies like Nvidia might see a change in demand patterns, with less emphasis on raw FLOPs for inference and more on memory bandwidth and latency for recurrent state updates. Cloud providers (AWS, Azure, GCP) would need to optimize their hardware stacks for these new workloads.

| Market Segment | Current Transformer-Based Challenge | Potential Impact of Recurrent Architectures | Projected Growth Catalyst |
|---|---|---|---|
| AI Agent Platforms | High cost per session, context window limits | 50-70% lower inference cost for long sessions | Enables mass-market B2B agent deployment |
| Consumer AI Assistants | Stateless, repetitive interactions | Persistent personality & memory across months | Increases user retention & perceived intelligence |
| Scientific AI / Drug Discovery | Difficulty modeling long causal chains in molecular dynamics | Efficient simulation of iterative processes | Accelerates discovery cycles, reduces compute cost |
| Edge AI (Devices, Cars) | Transformer models too large/heavy for low-power chips | Smaller, stateful models enable on-device reasoning | Unlocks real-time AI in IoT and mobile without cloud dependency |

Data Takeaway: The market impact is systemic, moving from a compute-intensive, cloud-centric model to one that favors efficiency and persistence. This could democratize advanced AI capabilities and create winners in hardware and software tailored for recurrent computation.

Risks, Limitations & Open Questions

Despite its promise, the OpenMythos path is fraught with technical and practical hurdles.

Technical Hurdles:
1. Training Instability: Recurrent networks are notoriously difficult to train due to vanishing/exploding gradients. While modern techniques like gradient clipping help, stabilizing a deep, recurrent Transformer across billions of parameters is an unsolved challenge. The shared parameters across recurrent steps could lead to degenerate learning dynamics.
2. Parallelization Loss: The core advantage of Transformers is massive parallel training. Introducing sequential recurrence breaks this, potentially increasing training time dramatically. The project must prove its proposed techniques for partial parallelization (e.g., scanning operations) are effective at scale.
3. The Benchmark Gap: Current LLM benchmarks (MMLU, GSM8K, HumanEval) are designed for feed-forward models excelling at next-token prediction. They may not adequately measure the benefits of stateful, iterative reasoning. New evaluation suites for persistent reasoning and long-horizon tasks are needed, which creates a chicken-and-egg problem for adoption.

Practical & Ecosystem Risks:
1. Ecosystem Inertia: The entire AI software stack—libraries (PyTorch, TensorFlow), compilers (TensorRT, OpenAI Triton), and hardware (GPU tensor cores)—is hyper-optimized for the Transformer's attention patterns. A new architecture faces a steep adoption cliff without equivalent optimization.
2. The Scaling Law Unknown: Transformers have predictable scaling laws. It is completely unknown if Recurrent Transformers scale as gracefully with data and parameters. The community may be reluctant to invest hundreds of millions in training a model on an unproven scaling curve.
3. Interpretability & Control: An explicit, evolving internal state is a double-edged sword. While potentially more interpretable than a static attention pattern, controlling what the model 'remembers' and how its state evolves raises new alignment and safety concerns. Ensuring a model's state doesn't become corrupted or biased over long interactions is a novel problem.

AINews Verdict & Predictions

The OpenMythos project is a bellwether, not a finished product. It signals that the architectural stagnation following the Transformer's 2017 debut is over. Our editorial judgment is that the core insight—that future AI requires efficient, stateful reasoning—is correct and inevitable. However, the winning implementation may not be a direct hybrid like OpenMythos, but could be a more radical departure like Mamba or an as-yet-uninvented paradigm.

Predictions:
1. Hybrid Architectures Will Dominate Within 3 Years: Within the next three years, we predict that the majority of new, frontier-scale models from leading labs will incorporate some form of explicit recurrent or stateful mechanism, whether via a retention layer, a dedicated memory network, or a recurrent block. Pure, vanilla Transformers will become legacy architecture for all but the most basic tasks.
2. The First "Stateful LLM" Breakthrough Will Come from Open Source: Given the risk-averse nature of large corporations with billion-dollar training budgets, the first demonstrably superior stateful model at scale (e.g., outperforming a same-parameter Transformer on a suite of long-reasoning tasks) will likely emerge from an open-source collective like OpenMythos or the RWKV community, leveraging distributed compute and collaborative development.
3. Microsoft Will Be the First Major Adopter: Microsoft, with its dual strengths in research (RetNet) and product (Copilot), is best positioned to integrate a stateful architecture into a mainstream product. We predict that within 2 years, a future Windows Copilot will be powered by an architecture directly descended from this research thread, offering a persistently helpful assistant that remembers user context across reboots and applications.
4. A New Benchmark War is Coming: The next competitive battleground will not be parameter counts or standard academic benchmarks, but on new metrics for 'reasoning depth,' 'state consistency,' and 'long-horizon task completion.' Organizations that define these benchmarks will shape the direction of the field.

What to Watch Next: Monitor the `open-mythos/arch` repository for the first medium-scale (1-7B parameter) pretrained model release. Its performance on modified, long-reasoning tasks versus a comparable Transformer will be the first real validation or refutation of its approach. Simultaneously, watch for any whisper of a 'Mamba-2' or 'RetNet-2' from corporate labs, which would signal an acceleration of this trend into the mainstream. The architecture of AI is entering its most creatively volatile period since the advent of the Transformer, and OpenMythos is one of the first maps to the territory beyond.

More from Hacker News

Agensi 與 AI 技能市場的崛起:智能體能力如何成為新經濟層The launch of Agensi represents a pivotal maturation in the AI agent landscape, transitioning the paradigm from monolithGPT Image 2 問世:原生多模態圖像生成的靜默革命The generative AI landscape is witnessing a subtle but profound architectural evolution with the emergence of GPT Image AgentSearch 推出自託管搜尋 API,挑戰 AI 代理對商業服務的依賴The development of sophisticated AI agents capable of autonomous action has been consistently hampered by a critical depOpen source hub2250 indexed articles from Hacker News

Related topics

AI architecture19 related articleslong-context AI15 related articles

Archive

April 20261936 published articles

Further Reading

Autoloom極簡AI代理框架,挑戰產業對複雜性的迷思全新開源AI代理框架Autoloom問世,其理念與業界追求更大、更複雜系統的趨勢背道而馳。它基於確定性的tinyloom庫構建,優先考慮簡潔性、可預測性和低計算開銷,為開發者提供了一種更輕量、可控的選擇。SigMap實現97%上下文壓縮,重新定義AI經濟學,終結暴力擴展上下文視窗的時代一個名為SigMap的新開源框架,正在挑戰現代AI開發的核心經濟假設:即更多上下文必然導致成本指數級增長。它通過智能壓縮和優先處理程式碼上下文,實現了高達97%的token使用量削減,有望大幅降低AI運算成本。AgentSearch 推出自託管搜尋 API,挑戰 AI 代理對商業服務的依賴一款名為 AgentSearch 的新工具,有望重新定義 AI 代理存取網路的方式。它提供一個無需商業金鑰、可自託管且容器化的搜尋 API,直接解決了長期以來限制自主代理發展的成本、隱私和控制問題。這項創新將為開發者帶來更大的自主權。Comrade AI 工作區:開源、安全至上的設計如何挑戰代理程式現狀開源專案 Comrade 已成為對當前主流 AI 開發與團隊工作區 SaaS 模式的直接挑戰。它將精緻的使用者介面與嚴格的本地優先、安全至上的理念相結合,為組織提供了一條利用先進 AI 代理程式的新途徑。

常见问题

GitHub 热点“OpenMythos and the Rise of Recurrent Transformers: Redefining AI Architecture Beyond Attention”主要讲了什么?

A quiet revolution is brewing in the open-source AI community, centered on a project called OpenMythos. Rather than fine-tuning existing large language models (LLMs), its contribut…

这个 GitHub 项目在“OpenMythos vs Mamba performance comparison”上为什么会引发关注?

The OpenMythos architecture proposes a fundamental re-engineering of the Transformer block. In a standard Transformer, each layer processes an input sequence through self-attention and a feed-forward network, passing the…

从“how to implement recurrent transformer PyTorch”看,这个 GitHub 项目的热度表现如何?

当前相关 GitHub 项目总星标约为 0,近一日增长约为 0,这说明它在开源社区具有较强讨论度和扩散能力。