Claude がネットワークスタックに:AI が Ping に応答し、インフラを再定義

Hacker News May 2026
Source: Hacker NewsArchive: May 2026
驚くべき実験で、Anthropic の Claude がユーザー空間の IP プロトコルスタックとして機能するよう設定され、ネットワークの Ping 要求に応答することに成功しました。AINews はこの偉業を分析し、レイテンシがハードウェアより桁違いに遅いものの、大規模言語モデルが
The article body is currently shown in English by default. You can generate the full version in this language on demand.

A recent experiment has demonstrated that a large language model, specifically Anthropic's Claude, can be configured to act as a user-space IP protocol stack, capable of receiving and responding to ICMP Echo Request (Ping) packets. The setup involves feeding raw network packets into the model's context window, instructing it to parse the IP and ICMP headers, compute the necessary checksums, and generate a valid Echo Reply. The results are both absurd and profound. The response time for a single ping, measured in seconds rather than microseconds, is laughably impractical for any real-world networking task. Yet, the very fact that a transformer-based model can execute this low-level, stateful, real-time computation—without being explicitly programmed for it—challenges our fundamental assumptions about the role of AI in computing. This is not a parlor trick; it is a proof of concept that AI can operate at the infrastructure layer. The implications are vast: future networks could feature AI-powered endpoints that dynamically negotiate protocols, handle congestion control with contextual awareness, or serve as intelligent security filters that understand the semantics of the traffic they inspect. This blurs the traditional OSI model layers, suggesting a future where the network itself is programmable through natural language. The economic model for compute may also shift from token-based pricing to packet-based pricing, creating a new class of 'soft routers' that handle edge cases with reasoning rather than rigid rules. While performance remains a critical barrier, the conceptual barrier has been broken—AI is evolving from a chat interface into an operating system for the internet.

Technical Deep Dive

The experiment's core mechanism is deceptively simple yet computationally radical. A raw network socket (using tools like `scapy` or `libpcap`) captures incoming ICMP Echo Request packets. The raw bytes—including the Ethernet frame, IP header, and ICMP header—are converted into a hexadecimal or decimal string and injected into Claude's system prompt. The prompt instructs the model to act as an IP stack: parse the source and destination IP addresses, the ICMP type and code, the identifier and sequence number, and the payload. It must then compute the ICMP checksum (a 16-bit one's complement sum of the ICMP header and payload) and generate a valid Echo Reply packet, which is then sent back through the raw socket.

This process exposes the fundamental latency bottleneck of transformer architectures. A single inference pass for a small model like Claude 3 Haiku takes approximately 500-800 milliseconds. The checksum calculation, which a silicon NIC performs in nanoseconds, requires the model to perform arithmetic reasoning within its attention mechanism—a task for which it is not optimized. The total round-trip time for a ping can easily exceed 5-10 seconds, compared to <1ms for a hardware stack.

| Metric | Hardware IP Stack | Claude (Simulated) |
|---|---|---|
| Ping Latency (RTT) | <1 ms | 5,000 - 10,000 ms |
| Throughput (packets/sec) | >1,000,000 | <0.2 |
| Power per packet | ~1 nJ | ~100 J (GPU) |
| Error rate (checksum) | <10^-12 | ~5-10% (first attempt) |
| State capacity | Unlimited (hardware) | Limited by context window (100K tokens) |

Data Takeaway: The performance gap is not incremental; it is multiple orders of magnitude. This underscores that LLMs are not replacements for existing network stacks but rather a new category of 'slow-path' processors for exceptional cases.

The experiment also highlights a critical architectural insight: the model must maintain state across multiple packets. A hardware stack uses registers and counters; Claude must keep the entire packet history in its context window. This is analogous to a von Neumann bottleneck but on a cognitive scale. Open-source projects like `netstack` (a user-space TCP/IP stack in Rust) and `smoltcp` (a standalone TCP/IP stack) demonstrate that efficient software stacks can achieve near-hardware performance. LLMs, by contrast, are fundamentally sequential and memory-bound for this task. A relevant GitHub repository is `llama.cpp`, which shows how to run LLMs locally but still cannot approach real-time networking performance. The repo has over 70,000 stars, indicating massive interest in local inference, but its latency profile (hundreds of milliseconds per token) is incompatible with sub-millisecond networking.

Key Players & Case Studies

This experiment is not an isolated stunt. Several companies and research groups are exploring the intersection of LLMs and networking. Anthropic, the creator of Claude, has not officially endorsed this use case, but their research on 'constitutional AI' and 'tool use' directly enables it. The ability to call external functions (like a raw socket) from within a prompt is a key enabler. OpenAI has demonstrated similar capabilities with GPT-4's function calling, though no public experiment has replicated the IP stack feat. Cisco and Juniper Networks have been exploring AI for network management, but their focus is on AI *assisting* network operations (e.g., intent-based networking) rather than AI *being* the network endpoint. A startup called Aalyria (spun out of Google) is working on 'spacetime' software-defined networking, which could theoretically integrate AI agents for dynamic routing.

| Entity | Approach | Stage | Key Limitation |
|---|---|---|---|
| Anthropic (Claude) | LLM as user-space stack | Experimental | Latency, cost, error rate |
| OpenAI (GPT-4) | Function calling for network tasks | Conceptual | No public demo of raw packet handling |
| Cisco (Catalyst Center) | AI for network analytics | Production | Not real-time; AI assists, does not replace |
| Aalyria (Spacetime) | SDN with AI optimization | Prototype | Focused on satellite networks, not general IP |

Data Takeaway: The incumbents (Cisco, Juniper) are using AI as a co-pilot, while the LLM providers are accidentally building the pilot. The most disruptive path is the latter, but it requires a fundamental rethinking of network latency budgets.

Industry Impact & Market Dynamics

The immediate market impact is negligible—no one will replace their routers with a GPU cluster running an LLM. However, the second-order effects are significant. The concept of a 'soft router'—an AI that handles only the 0.1% of packets that are anomalous (e.g., DDoS attacks, malformed packets, protocol negotiation edge cases)—could be economically viable. The global network equipment market is valued at approximately $150 billion (2025 estimate). Even a 1% displacement by AI-driven soft routers represents a $1.5 billion opportunity.

| Market Segment | Current Size (2025 est.) | AI-Addressable Share | Potential Value |
|---|---|---|---|
| Enterprise Routers | $45B | 2% (edge cases) | $900M |
| Security Appliances (Firewalls) | $30B | 5% (anomaly detection) | $1.5B |
| Data Center Switches | $50B | 0.5% (control plane) | $250M |
| WAN Optimization | $10B | 10% (dynamic routing) | $1B |

Data Takeaway: The addressable market is in the billions, but only if AI can achieve sub-10ms latency for the 'slow path'—a target that current LLM architectures cannot meet. This creates a clear opportunity for specialized inference hardware (e.g., Groq, Cerebras) that can reduce latency to milliseconds.

The pricing model shift from 'per token' to 'per packet' is a natural evolution. If an AI handles a network packet, the cost should be tied to the value of that packet (e.g., a financial transaction packet is worth more than a DNS query). This could lead to tiered pricing: $0.001 per packet for standard routing, $0.01 per packet for security inspection, and $1.00 per packet for complex protocol negotiation.

Risks, Limitations & Open Questions

The most immediate risk is security. An LLM that processes raw network packets is a massive attack surface. Prompt injection could cause the model to generate malformed packets, crash the network, or leak data. The checksum error rate of 5-10% on first attempt is unacceptable for production networks—a single corrupted packet can cause TCP retransmission storms. There is also the determinism problem: LLMs are probabilistic, while network protocols require deterministic behavior. A router that sometimes drops packets because the model 'decides' to is not a router; it's a liability.

Scalability is another open question. The context window limits how many concurrent flows an AI can handle. A modern router manages millions of flows; Claude can handle perhaps a dozen before its context is exhausted. Power consumption is prohibitive: a single ping response consumes as much energy as a hardware router uses to process billions of packets.

Finally, there is the regulatory question. Network infrastructure is subject to strict reliability standards (e.g., five-nines availability). An AI that fails 5% of the time cannot meet these standards. The liability for a misrouted packet that causes a financial loss would be enormous.

AINews Verdict & Predictions

This experiment is a watershed moment, not for its practicality, but for its symbolic power. It demonstrates that the boundary between application and infrastructure is not a law of physics but a convention of engineering. We predict the following:

1. Within 3 years, a major cloud provider (AWS, Azure, GCP) will offer a 'Smart Endpoint' service that uses a small, distilled LLM to handle edge cases in network traffic—such as protocol negotiation for IoT devices or dynamic firewall rule generation. This will be priced per packet, not per token.

2. Within 5 years, a startup will emerge that builds a 'soft router' ASIC co-designed with a lightweight transformer model, achieving sub-millisecond latency for the slow path. This will be used in high-security environments where rule-based systems are insufficient (e.g., military networks, financial exchanges).

3. The open-source community will create a 'NetLLM' framework that allows anyone to run a user-space IP stack on a local LLM, similar to how `llama.cpp` democratized local inference. This will be used for educational purposes and penetration testing, not production.

4. The pricing model for compute will bifurcate: high-latency, high-reasoning tasks (like protocol negotiation) will be priced per packet, while low-latency, high-throughput tasks (like bulk routing) will remain per byte. This will create a new market for 'intelligent bandwidth'.

The verdict is clear: AI is no longer just a tool for generating text. It is becoming a substrate for computation itself. The network stack experiment is a canary in the coal mine—a signal that the infrastructure layer is ripe for disruption. The question is not whether it will happen, but which company will build the first production-grade 'AI router' and how they will price it.

More from Hacker News

JSON危機:AIモデルが構造化出力で信頼できない理由AINews conducted a systematic stress test of 288 large language models, requiring each to output valid JSON. The resultsトークン予算管理:AIコスト制御とエンタープライズ戦略の最前線The transition of large language models from research labs to production pipelines has exposed a brutal reality: inferenOrbit UI はAIエージェントに仮想マシンの直接制御を可能にする——まるでデジタルパペットのようにAINews has uncovered Orbit UI, an open-source project that bridges the gap between AI agents and real system administratOpen source hub3250 indexed articles from Hacker News

Archive

May 20261206 published articles

Further Reading

NvidiaのRust-to-CUDAコンパイラが安全なGPUプログラミングの新時代を切り開くNvidiaは、Rustコードを直接CUDAカーネルに変換する公式コンパイラ「CUDA-oxide」を静かにリリースしました。この動きは、並列コンピューティングにおけるメモリ安全性のバグを劇的に削減し、Rust開発者がGPUアクセラレーショAmália AI:ファドにちなんだモデルがポルトガル語の主権を取り戻す方法ポルトガルを代表するファド歌手にちなんで名付けられた大規模言語モデル「Amália」が、欧州ポルトガル語向けに登場しました。このモデルは、ポルトガル語独自の文法や文化的背景、低リソース環境での最適化に重点を置き、AIにおける小言語の軽視に対OpenAIがAIの価値を再定義:モデルインテリジェンスからデプロイメントインフラへOpenAIは、最先端の研究ラボからフルスタックのデプロイメント企業へと静かに重要な変革を遂げています。当社の分析によると、その戦略の中心はモデルパラメータのブレークスルー追求から、エンタープライズ統合、リアルタイム推論最適化、そしてデプロAdamsReview:マルチエージェント並列レビューがClaude CodeのPR品質を再定義する方法AINewsは、複数の並列サブエージェントを調整して深く多次元的なPRレビューを実行するClaude Codeプラグイン「adamsreview」を発見しました。セキュリティ、パフォーマンス、保守性に特化したエージェントと、永続的なJSON

常见问题

这次模型发布“Claude as a Network Stack: AI Responds to Pings, Redefining Infrastructure”的核心内容是什么?

A recent experiment has demonstrated that a large language model, specifically Anthropic's Claude, can be configured to act as a user-space IP protocol stack, capable of receiving…

从“Can Claude really replace a router?”看,这个模型发布为什么重要?

The experiment's core mechanism is deceptively simple yet computationally radical. A raw network socket (using tools like scapy or libpcap) captures incoming ICMP Echo Request packets. The raw bytes—including the Etherne…

围绕“What is a user-space IP stack?”,这次模型更新对开发者和企业有什么影响?

开发者通常会重点关注能力提升、API 兼容性、成本变化和新场景机会,企业则会更关心可替代性、接入门槛和商业化落地空间。