Ảo Ảnh Lập Trình AI: Tại Sao Chúng Ta Vẫn Chưa Có Phần Mềm Được Viết Bởi Máy Móc

Hacker News April 2026
Source: Hacker NewsAI programminggenerative AIsoftware developmentArchive: April 2026
AI sinh đã thay đổi cách các nhà phát triển viết code, nhưng lời hứa về phần mềm được tạo ra hoàn toàn bởi máy móc vẫn chưa thành hiện thực. Khoảng cách này cho thấy những hạn chế cơ bản trong khả năng duy trì tính nhất quán kiến trúc dài hạn và lập luận cấp hệ thống của AI hiện tại. Ngành công nghiệp hiện đang phải đối mặt với thách thức này.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

The developer community is grappling with a profound paradox: while AI coding assistants like GitHub Copilot, Amazon CodeWhisperer, and Cursor have become ubiquitous, there are virtually no significant end-user applications—no operating systems, compilers, or creative suites—that have been primarily authored by artificial intelligence. This absence points to a critical evolutionary bottleneck in AI's journey toward true software creation.

Current large language models excel at generating discrete functions, boilerplate code, and implementing well-defined algorithms. However, they falter when confronted with the sprawling, integrated complexity of systems like database management engines, game physics simulators, or modern web browsers. The core challenge isn't scale alone but the fundamental lack of a persistent world model—the ability to maintain architectural intent, manage technical debt, and make coherent trade-offs across a development timeline that can span years.

AI today functions as an exceptional short-term collaborator, not a visionary project lead. It can optimize a sorting algorithm but cannot conceive the million interdependent decisions that constitute a software project's lifecycle. The industry is now pivoting from pure code generation toward frameworks that enable AI to coordinate modular, verifiable components over extended development cycles. This shift may not simply replicate existing software but could birth entirely new software categories, with profound implications for business models, from selling AI assistance tools to licensing AI-originated software kernels. We are witnessing AI's arduous climb from automation tool to creative subject.

Technical Deep Dive

The failure to produce AI-authored software stems from architectural limitations in current transformer-based LLMs, not merely a lack of training data. These models operate on a statistical next-token prediction paradigm optimized for local coherence, not global system design. They lack the internal mechanisms to build and maintain a persistent, evolving representation of a software project's architecture—its modules, dependencies, interfaces, and non-functional requirements.

The Missing Architecture Engine: Modern software engineering relies on abstraction layers (APIs, interfaces, contracts) and long-range planning (roadmaps, technical specifications). LLMs, by design, have a fixed context window, creating a planning horizon problem. Projects like Anthropic's Claude 3.5 Sonnet with its 200K token context or Google's Gemini 1.5 Pro with its 1M token context attempt to mitigate this, but they still treat the project as a linear sequence of text, not a structured, queryable knowledge graph of the codebase. The open-source SWE-agent framework from Princeton attempts to address this by creating a specialized agent for software engineering tasks, treating the codebase as an environment to navigate. It has gained significant traction (over 11k stars on GitHub) by reframing coding as a reinforcement learning problem with tools like file editors, linters, and test runners.

Benchmarking the Gap: Current benchmarks like HumanEval or MBPP measure function-level code generation. They are poor proxies for system-building capability. A more telling metric is the success rate on complex, multi-file issues from real-world repositories. Preliminary studies show a dramatic drop in performance as task complexity moves from single-file bug fixes to cross-module feature additions.

| Task Complexity Level | Example Task | Top Model Success Rate (Claude 3.5 Sonnet) | Human Junior Dev Success Rate |
|---|---|---|---|
| Single-Function Generation | "Write a Python function to reverse a linked list." | ~95% | ~99% |
| Single-File Bug Fix | "Fix the off-by-one error in `data_processor.py`." | ~75% | ~90% |
| Multi-File Feature Addition | "Add OAuth2 support to the authentication module." | ~20% | ~70% |
| Architectural Refactor | "Migrate the monolith's user service to a microservice." | <5% | ~50% (with senior oversight) |

Data Takeaway: The performance cliff is stark. AI agents match or exceed humans on localized, well-defined coding tasks but collapse when the problem requires understanding and modifying a diffuse web of dependencies—the essence of software architecture.

Emerging Technical Approaches: The frontier is moving toward meta-reasoning frameworks. Projects like OpenDevin (an open-source attempt to replicate Devin, Cognition AI's autonomous AI engineer) and Meta's Aria project focus on creating AI agents that can break down high-level goals into a series of verifiable sub-tasks, execute them, and integrate results. The key innovation is the plan-execute-verify loop with external memory. Instead of generating a million lines of code in one go, the AI proposes a plan, writes a module, runs tests, assesses the outcome, and updates its world model. This is computationally expensive but mirrors human iterative development.

Key Players & Case Studies

The landscape is divided between enhancers of human developers and aspiring replacers of the development process itself.

The Enhancers (Incumbent Approach):
* Microsoft (GitHub Copilot): Deeply integrated into the IDE, Copilot operates as a "pair programmer." Its strength is in-line code completion and chat-based assistance within the context of the open files. It amplifies developer productivity but is fundamentally a reactive tool, not a proactive architect.
* Amazon (CodeWhisperer): Similar to Copilot but with a strong focus on AWS APIs and security scanning. It excels at generating code for cloud services but remains anchored to the developer's immediate intent.
* Cursor & Windsurf: These newer, AI-native IDEs (built on VS Code) take a more aggressive approach. Cursor, for instance, allows AI to edit codebases across multiple files based on natural language commands, moving closer to system-level changes. However, they still require the human to provide the high-level direction and sanity-check the output.

The Aspiring Replacers (Autonomous Agents):
* Cognition AI (Devin): This startup caused a sensation by demoing an AI agent that could complete entire Upwork software engineering jobs. Devin is presented as an autonomous AI software engineer with its own shell, code editor, and browser. It plans and executes complex engineering tasks. While its true capabilities on large, novel projects are unproven, it represents the most direct assault on the problem of end-to-end software creation.
* OpenAI (GPT-Engineer & Custom GPTs): While not a product per se, the GPT-Engineer open-source project and the capability to create custom, action-equipped GPTs point to a future where a sufficiently advanced model, given the right tools (file system, build tools, web search), could orchestrate development. OpenAI's strategic partnerships, like with Figure AI for robotics, hint at a model where its AI acts as the "brain" for complex system design.
* Google DeepMind (AlphaCode & Project Astra): DeepMind's research has long focused on problem-solving. AlphaCode competed in programming competitions, a domain requiring some level of algorithmic planning. Their newer work on Project Astra and Gemini's planning capabilities suggests a continued push toward AI that can reason over longer horizons and with tools.

| Company/Project | Core Philosophy | Key Differentiator | Primary Limitation |
|---|---|---|---|
| GitHub Copilot | Augmentation | Deep IDE integration, vast training data | Reactive, no project-level planning |
| Cursor | Augmentation+ | AI-native IDE, multi-file edits | Requires precise human direction |
| Cognition AI's Devin | Automation | Full-stack autonomous agent, long-horizon planning | Unproven at scale, black-box nature |
| OpenDevin (OS) | Automation | Open-source, modular agent framework | Early stage, lacks polished capabilities |

Data Takeaway: The market is bifurcating. The enhancers have product-market fit today by boosting productivity. The replacers are high-risk, high-reward bets on a future where the software development lifecycle is fundamentally re-architected around AI agency.

Industry Impact & Market Dynamics

The inability to produce AI-authored software is holding back a potential multi-trillion-dollar shift in the software economy. Currently, the market for AI coding tools is booming, but it is merely a precursor.

Current Market: The AI-assisted development market is valued at approximately $2.8 billion in 2024, growing at over 25% CAGR. This is dominated by subscription services for tools like Copilot ($10-19/user/month). This model simply adds a new tooling cost to the existing software development cost center.

The Potential Disruption: The true disruption occurs when the cost of "software creation" plunges because the primary labor (developer time) is radically reduced or transformed. Imagine a scenario where a startup can describe a complex SaaS product to an AI architect, which then generates, deploys, and maintains the v1.0. The business model shifts from selling developer tools to licensing AI-generated software IP or taking a revenue share of the resulting product.

Funding and Strategic Moves: Venture capital is flooding into autonomous AI agent startups. Cognition AI raised a $21M Series A at a $350M+ valuation before having a public product, signaling extreme investor belief in this frontier. Established players are responding via acquisition and internal projects. Microsoft's integration of Copilot across its stack and its heavy investment in OpenAI is a hedge to own the platform on which future AI software is built.

| Business Model | Example | Revenue Driver | Risk if AI Software Emerges |
|---|---|---|---|
| Traditional SaaS | Salesforce, Adobe | Software license subscriptions | High - AI could generate cheaper, tailored alternatives |
| AI-Augmented Dev Tools | GitHub Copilot, Tabnine | Seat-based subscriptions | Medium - Could be commoditized or bypassed by autonomous agents |
| Cloud Infrastructure | AWS, Azure, GCP | Compute, storage, API calls | Low - AI-generated software will still run on clouds (likely a beneficiary) |
| AI-Native Software Foundry | (Future) Cognition AI, OpenAI Platform | Licensing fee, revenue share, or compute credits | N/A - This is the disruptive model itself |

Data Takeaway: The greatest existential risk is to traditional software vendors whose products are complex but conceptually formulaic. The winners will be those who control the AI "foundries" (like OpenAI, Anthropic) or the infrastructure they run on (cloud providers). Developer tool companies must evolve into AI-agent management platforms or face obsolescence.

Risks, Limitations & Open Questions

The pursuit of AI-authored software is fraught with technical, ethical, and economic perils.

Technical Debt on Steroids: AI that generates code without deep architectural understanding could produce systems that are incomprehensible to humans—a black box built from black-box components. Debugging, security auditing, and maintenance could become impossible. The industry may need entirely new verification and validation (V&V) frameworks for AI-generated systems.

The Innovation Stagnation Risk: If AI is trained predominantly on existing human code (GitHub), its "innovations" may merely be remixes of past patterns. We risk entering a software cultural dark age where novel paradigms (like functional reactive programming, or the actor model) cease to emerge because they are underrepresented in the training data. The AI would optimize for what *has worked*, not what *could work better*.

Economic and Labor Dislocation: The narrative has moved from "AI will help developers" to "AI will replace developers." A sudden breakthrough in autonomous software creation could cause severe short-term disruption. However, history suggests the role would shift rather than vanish—toward "AI software directors" who define high-level requirements, curate training data for domain-specific agents, and manage the ethical and societal implications of automatically generated systems.

Security Apocalypse: Automated, large-scale code generation vastly increases the attack surface. While AI can be trained to avoid known vulnerabilities, novel attack vectors in AI-generated code could be exploited at scale before humans even understand the flawed patterns the AI has invented.

Open Questions:
1. Will we need a "seed" of human-designed architecture? The most plausible path may be hybrid: humans design the high-level module boundaries and APIs, and AI agents fill in the implementation, akin to a detailed blueprint.
2. Can AI develop taste? Great software involves subjective judgments about user experience, API elegance, and performance trade-offs. Can LLMs, trained on objective data, develop a coherent sense of software aesthetics?
3. Who is liable for bugs? If an AI generates a financial trading application with a catastrophic flaw, is the liability with the human who prompted it, the company that built the AI, or the AI itself? Current legal frameworks are utterly unprepared.

AINews Verdict & Predictions

The dream of uttering "build me a new Photoshop" and having it appear is a mirage—for this decade. However, the components of that mirage are rapidly coalescing into a new, more profound reality.

Our Verdict: The industry is correctly focused on the agentic paradigm, not merely larger models. The breakthrough will not come from GPT-5 having a trillion more parameters, but from frameworks that enable GPT-5-level models to act as persistent, planning, tool-using project coordinators. We are at the end of the first act—AI as a coding parrot—and the beginning of the second: AI as a junior engineer with severe amnesia. The third act, AI as a competent systems architect, remains on the horizon.

Specific Predictions:
1. By 2026: We will see the first commercially viable, narrow-domain AI software factories. These will generate complete, functional applications for well-scoped domains like CRUD business dashboards, simple mobile games, or data pipeline orchestration scripts. They will not be general-purpose but will disrupt specific low-code/no-code and outsourcing markets.
2. The "Kernel" Model Will Emerge: Major tech firms (likely Microsoft/OpenAI or Google) will begin offering AI-originated software kernels—core, optimized engines for graphics, physics, or database management—that are licensed as black-box components. Humans will build the UI and business logic on top of these AI-crafted cores.
3. A New Programming Metaphor Will Arise: The dominant paradigm will shift from writing code to training and directing specialized AI agents. Programming languages will become more declarative and high-level, focusing on intent, constraints, and interfaces, while the agents handle the implementation details. Languages like Rust, with its strong compile-time guarantees, may become the preferred *output* language for AI due to its safety, even as humans interact in natural language.
4. Watch the Open-Source Agent Frameworks: The progress of projects like OpenDevin, SWE-agent, and ToolLLM will be the true bellwether. If these communities can create a stable, extensible platform for AI software engineering, it will democratize the capability and accelerate progress far faster than closed commercial offerings.

The ultimate form of "AI-written software" may not resemble today's applications at all. It may be fluid, self-adapting, and inherently explainable to its AI maintainer, even if opaque to us. The journey is not toward automating the past but inventing a new future for what software even is.

More from Hacker News

Sự Trỗi Dậy Của Khả Năng Sẵn Sàng Vận Hành: Cách Các Tác Nhân AI Phát Triển Từ Nguyên Mẫu Thành Nhân Công Sản XuấtA quiet but profound transformation is underway in artificial intelligence. The initial euphoria surrounding large languKiến trúc Meshcore nổi lên: Mạng suy luận P2P phi tập trung có thể thách thức quyền bá chủ AI?The AI infrastructure landscape is witnessing the early stirrings of a paradigm war. At its center is the concept of MesKhả năng Quan sát AI Nổi lên như Một Lĩnh vực Quan trọng để Quản lý Chi phí Suy luận Bùng nổThe initial euphoria surrounding large language models has given way to a sobering operational phase where the true costOpen source hub2137 indexed articles from Hacker News

Related topics

AI programming47 related articlesgenerative AI49 related articlessoftware development32 related articles

Archive

April 20261681 published articles

Further Reading

Từ Copilot đến Commander: Cách các Tác nhân AI Định nghĩa Lại Phát triển Phần mềmTuyên bố của một lãnh đạo công nghệ về việc tạo ra hàng chục nghìn dòng mã AI mỗi ngày không chỉ cho thấy sự gia tăng năCuộc Cách mạng Mã nguồn AI: Tại sao Cấu trúc Dữ liệu & Thuật toán lại Chiến lược hơn Bao giờ hếtSự trỗi dậy của các trợ lý lập trình AI đã gây ra nỗi lo lắng sâu sắc trong giới lập trình viên toàn cầu: liệu nhiều nămCuộc Cách mạng Nhà phát triển Độc lập: Cách AI Agent Xây dựng Nền tảng SaaS Từ thiện Full-StackMột mô hình phát triển phần mềm mới đã xuất hiện, nơi một nhà phát triển độc lập thành công chỉ đạo một nhóm AI agent xâNền tảng Đa Tác tử của Kern Định nghĩa Lại Lập trình AI—Từ Cộng sự Bay đến Đồng đội Hợp tácSự tiến hóa của AI trong phát triển phần mềm đang trải qua một sự thay đổi mô hình. Nền tảng của Kern vượt ra ngoài các

常见问题

这次模型发布“The AI Programming Mirage: Why We Still Don't Have Software Written by Machines”的核心内容是什么?

The developer community is grappling with a profound paradox: while AI coding assistants like GitHub Copilot, Amazon CodeWhisperer, and Cursor have become ubiquitous, there are vir…

从“Can AI write a complete operating system?”看,这个模型发布为什么重要?

The failure to produce AI-authored software stems from architectural limitations in current transformer-based LLMs, not merely a lack of training data. These models operate on a statistical next-token prediction paradigm…

围绕“Difference between GitHub Copilot and autonomous AI engineer”,这次模型更新对开发者和企业有什么影响?

开发者通常会重点关注能力提升、API 兼容性、成本变化和新场景机会,企业则会更关心可替代性、接入门槛和商业化落地空间。