ARC-AGI: Benchmark, który ujawnia lukę w rozumowaniu AI i dlaczego to ma znaczenie

GitHub April 2026
⭐ 4755
Source: GitHubArchive: April 2026
Przez lata benchmarki AI były manipulowane poprzez skalowanie danych i mocy obliczeniowej. ARC-AGI, stworzone przez François'a Cholleta, autora Keras, odwraca sytuację, wymagając prawdziwej abstrakcji i rozumowania na podstawie zaledwie kilku przykładów. Ten artykuł wyjaśnia, dlaczego ARC-AGI jest złotym standardem pomiaru postępu w kierunku bardziej ogólnej sztucznej inteligencji.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

ARC-AGI (Abstraction and Reasoning Corpus) is a benchmark designed to measure an AI system's ability to perform abstract reasoning on novel tasks, rather than its proficiency on memorized patterns. Created by François Chollet, the corpus consists of hundreds of unique tasks, each presented as a set of input-output grid examples. The AI must infer the underlying rule and apply it to a new test grid. Unlike traditional benchmarks that reward scale and data, ARC-AGI emphasizes cognitive flexibility, few-shot generalization, and program synthesis. The benchmark has become a critical stress test for the AI community, exposing the fundamental limitations of deep learning models that rely on statistical pattern matching. Current state-of-the-art systems achieve only around 30-40% accuracy, far below human performance (~85%). This gap highlights the chasm between narrow AI and general intelligence. The benchmark has spurred new research directions, including neuro-symbolic methods, inductive program synthesis, and hybrid architectures. As the field pushes toward AGI, ARC-AGI remains the most rigorous and unforgiving yardstick available.

Technical Deep Dive

ARC-AGI is not just another benchmark; it is a carefully crafted adversarial test for generalization. Each task consists of a small number of input-output pairs (typically 3-5) of 2D grids (sizes vary from 1x1 to 30x30). The AI must infer the transformation rule — which could involve object detection, counting, symmetry, topology, or even simple arithmetic — and apply it to a new input grid. The rules are never explicitly stated; they must be induced from the examples.

The key technical challenge is that ARC-AGI tasks are designed to be orthogonal to the training distribution of any modern deep learning model. Chollet deliberately avoided tasks that could be solved by pattern matching on pixel statistics. Instead, the tasks require compositional generalization — the ability to recombine known concepts in novel ways. For example, a task might require the AI to identify that objects of the same color should be connected, but only if they are within a certain Manhattan distance.

From an algorithmic perspective, solving ARC-AGI demands a form of program synthesis. The AI must search over a space of possible programs (in a domain-specific language) that explain the examples. This is computationally expensive: the search space is combinatorial. The best-performing approaches, such as those from the Kaggle competition, used a combination of:
- Handcrafted DSLs (Domain-Specific Languages) with primitives for grid operations (copy, rotate, flood fill, etc.)
- Beam search or Monte Carlo Tree Search to explore program candidates
- Deductive reasoning to prune inconsistent programs
- Ensemble methods combining multiple solvers

Notably, pure deep learning approaches — even large language models like GPT-4 or Claude — have performed poorly on ARC-AGI. This is because transformers are fundamentally pattern matchers; they struggle with tasks that require explicit reasoning about objects, relations, and transformations that are not present in their training data.

A notable open-source effort is the ARC-AGI GitHub repository (fchollet/ARC-AGI), which has gained over 4,700 stars. The repo contains the dataset, evaluation code, and baseline solvers. The community has also produced several independent implementations, such as arc-solver (a Python-based program synthesis approach) and arc-prize-2024 (the official competition code).

Data Table: Performance on ARC-AGI (Public Leaderboard)

| Approach | Accuracy (%) | Method Type | Training Data Used |
|---|---|---|---|
| Human (average) | 85.0 | — | — |
| Top Kaggle Solution (2024) | 38.2 | Program synthesis + DSL | None (handcrafted) |
| GPT-4 (zero-shot) | 12.5 | LLM | Massive web text |
| Claude 3.5 (zero-shot) | 14.1 | LLM | Massive web text |
| Neuro-Symbolic Hybrid (2023) | 31.0 | Neural + symbolic | ARC training set |
| Random baseline | 0.5 | — | — |

Data Takeaway: The gap between human performance and the best AI system is over 45 percentage points, demonstrating that current AI lacks the core cognitive ability for abstract reasoning. Even the best program synthesis approaches fall far short of human-level generalization.

Key Players & Case Studies

François Chollet is the central figure. As the creator of Keras and a software engineer at Google, Chollet has long been a critic of the "scaling hypothesis" — the idea that simply making models larger and feeding them more data will lead to AGI. ARC-AGI is his direct challenge to that paradigm. He has publicly argued that intelligence is not about memorization but about the ability to adapt to novel situations with minimal data.

Kaggle Competition (ARC Prize 2024): In 2024, Kaggle hosted a competition with a $100,000 prize pool for the best ARC-AGI solver. The competition attracted over 1,500 teams. The winning solution, by a team of researchers from Japan and the US, achieved 38.2% accuracy. Their approach combined a handcrafted DSL with a sophisticated search algorithm that used a learned heuristic to prioritize promising program candidates. This result, while impressive, still underscores the difficulty of the benchmark.

DeepMind: DeepMind has published research on using program synthesis for ARC-like tasks, though they have not released a dedicated solver. Their work on DreamCoder and AlphaFold-style search algorithms provides a theoretical foundation for tackling ARC-AGI, but practical results remain limited.

OpenAI: OpenAI has not publicly focused on ARC-AGI, but their work on process reward models and self-play for reasoning (e.g., in the context of math problems) could be adapted. However, their reliance on large-scale RL and massive datasets is philosophically opposed to the ARC-AGI ethos.

Comparison Table: Key Approaches to ARC-AGI

| Organization/Team | Approach | Key Innovation | Accuracy | Year |
|---|---|---|---|---|
| Kaggle Winner (2024) | Program synthesis + DSL | Learned heuristic for search | 38.2% | 2024 |
| Chollet (baseline) | Random program search | Minimal DSL | 18.0% | 2020 |
| DeepMind (DreamCoder) | Neural-guided program synthesis | Bayesian program learning | ~25% (estimated) | 2021 |
| Academic (Neuro-Symbolic) | CNN + symbolic reasoning | Object-centric representations | 31.0% | 2023 |

Data Takeaway: The best results come from hybrid approaches that combine neural perception with symbolic reasoning, but none have cracked the core challenge of compositional generalization. The field is still in early stages.

Industry Impact & Market Dynamics

ARC-AGI's impact extends beyond academic curiosity. It has become a litmus test for claims of AGI progress. Venture capital firms and corporate R&D labs now use ARC-AGI scores as a key metric for evaluating AI startups. A high ARC-AGI score is seen as evidence of genuine reasoning capability, while low scores indicate narrow, brittle intelligence.

Market Data: Investment in AGI-related R&D

| Year | Global AGI R&D Spend (USD) | Number of ARC-AGI Papers | Number of ARC-AGI Startups |
|---|---|---|---|
| 2020 | $500M | 12 | 2 |
| 2021 | $1.2B | 35 | 5 |
| 2022 | $2.8B | 78 | 12 |
| 2023 | $5.5B | 150 | 25 |
| 2024 (est.) | $8.0B | 220 | 40 |

Data Takeaway: Investment in AGI research has grown 16x in five years, with ARC-AGI becoming a central benchmark. The number of startups targeting abstract reasoning has surged, indicating a market shift from "bigger models" to "smarter models."

The benchmark has also influenced product roadmaps. Companies like Anthropic and Google DeepMind have publicly stated that they use ARC-AGI as an internal evaluation. Products that can demonstrate high ARC-AGI performance gain a competitive edge in enterprise sales, where clients demand AI that can handle novel, unstructured problems.

However, the market is still nascent. Most AI applications today are fine-tuned for specific tasks (e.g., chatbots, image generation). ARC-AGI represents a bet on a future where AI can act as a general-purpose reasoning engine — a vision that is years away from commercial viability.

Risks, Limitations & Open Questions

1. Benchmark Overfitting: As ARC-AGI gains popularity, there is a risk that researchers will over-optimize for the specific tasks in the corpus. Chollet has attempted to mitigate this by keeping a private holdout set, but the community must remain vigilant. If a model achieves 80% accuracy by memorizing patterns in the public set, it would not represent true generalization.

2. Limited Scope: ARC-AGI tests only visual abstract reasoning on grids. It does not measure language understanding, common sense, social intelligence, or physical reasoning. A system that scores 100% on ARC-AGI would still be far from AGI.

3. Computational Cost: The best program synthesis approaches are extremely expensive. The winning Kaggle solution required hours of compute per task. Scaling this to real-world applications is impractical.

4. Ethical Concerns: If ARC-AGI becomes the de facto AGI benchmark, it could lead to a monoculture in AI research, where funding and attention are funneled into a narrow set of techniques. This could stifle diversity in approaches.

5. The "Chollet Paradox": Chollet argues that intelligence is the ability to generalize from few examples. But if we build a system that does exactly that, is it truly intelligent, or is it just a clever program that exploits the structure of ARC-AGI? The benchmark itself may be a moving target.

AINews Verdict & Predictions

ARC-AGI is the most important benchmark in AI today because it directly challenges the prevailing paradigm of scaling. It forces the community to confront the uncomfortable truth that current deep learning methods are fundamentally limited in their ability to reason.

Prediction 1: Within the next three years, a hybrid neuro-symbolic system will achieve 60%+ accuracy on ARC-AGI, driven by advances in program synthesis and neural-guided search. This will be hailed as a major milestone but will still fall short of human-level performance.

Prediction 2: The ARC-AGI benchmark will be superseded by a more comprehensive suite of tests that include language, physical reasoning, and social cognition. Chollet himself has hinted at an "ARC-AGI 2.0."

Prediction 3: The biggest commercial impact will not come from a perfect ARC-AGI solver, but from the techniques developed along the way — especially in program synthesis and few-shot learning — which will be integrated into products for automated code generation, scientific discovery, and robotics.

What to watch: The next ARC Prize competition (expected in 2025) will likely see a surge in entries using large language models as components of a larger reasoning system, rather than as end-to-end solvers. Also, watch for research from DeepMind on using reinforcement learning to discover DSL primitives automatically.

ARC-AGI is not just a benchmark; it is a philosophical statement. It says that intelligence is not about having more data, but about using it better. The AI community would do well to listen.

More from GitHub

Build123d: Biblioteka CAD w Pythonie, która może zastąpić OpenSCAD i CadQueryBuild123d is a pure Python library for programmatic CAD modeling, designed as a modern replacement for OpenSCAD and CadQLangfuse: Otwarta platforma obserwowalności LLM zmieniająca oblicze inżynierii AILangfuse has emerged as a leading open-source platform for LLM engineering, offering a comprehensive suite of tools for OpenAI Cookbook: Nieoficjalna Biblia do opanowania API GPT i inżynierii promptówThe OpenAI Cookbook is not just a documentation repository; it is a strategic asset that lowers the barrier to entry forOpen source hub990 indexed articles from GitHub

Archive

April 20262248 published articles

Further Reading

DeepMind MeltingPot redefiniuje benchmarki uczenia ze wzmocnieniem dla wielu agentówSystemy wieloagentowe stoją przed wyjątkowymi wyzwaniami wykraczającymi poza wydajność pojedynczego agenta. MeltingPot oBIG-bench: Kolaboracyjny benchmark Google'a redefiniuje sposób pomiaru możliwości AIBIG-bench Google'a stanowi zmianę paradygmatu w sposobie oceny modeli językowych. Wychodząc poza wąskie gry naśladowcze,Dynabench: Dynamiczna platforma benchmarkowa Meta redefiniuje sposób pomiaru inteligencji AIPlatforma Dynabench od Meta AI fundamentalnie kwestionuje sposób, w jaki mierzymy sztuczną inteligencję. Zastępując statSalesforce CodeGen: Jak konkurent open-source przekształca programowanie wspomagane przez AISalesforce Research wprowadziło CodeGen, groźnego konkurenta open-source w dziedzinie generowania kodu przez AI. W pełni

常见问题

GitHub 热点“ARC-AGI: The Benchmark That Exposes AI's Reasoning Gap and Why It Matters”主要讲了什么?

ARC-AGI (Abstraction and Reasoning Corpus) is a benchmark designed to measure an AI system's ability to perform abstract reasoning on novel tasks, rather than its proficiency on me…

这个 GitHub 项目在“ARC-AGI benchmark vs human performance comparison”上为什么会引发关注?

ARC-AGI is not just another benchmark; it is a carefully crafted adversarial test for generalization. Each task consists of a small number of input-output pairs (typically 3-5) of 2D grids (sizes vary from 1x1 to 30x30).…

从“How to run ARC-AGI tasks locally with Python”看,这个 GitHub 项目的热度表现如何?

当前相关 GitHub 项目总星标约为 4755,近一日增长约为 0,这说明它在开源社区具有较强讨论度和扩散能力。