Nomos 실행 방화벽: 안전한 AI 에이전트 배포를 위한 결정적으로 누락된 계층

Hacker News March 2026
Source: Hacker NewsAI agent securityAI infrastructureArchive: March 2026
대화형 챗봇에서 복잡한 작업을 실행할 수 있는 자율 에이전트로 AI가 빠르게 진화하면서 위험한 보안 공백이 드러났습니다. 오픈소스 프로젝트 Nomos는 해결책을 선도하고 있습니다: 실행 전 제안된 모든 작업을 가로채고, 분석하며, 승인하는 '실행 방화벽'입니다.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

The AI landscape is undergoing a profound transformation as language models evolve from passive text generators into proactive, tool-using agents. Systems like AutoGPT, LangChain agents, and proprietary platforms from OpenAI and Anthropic can now write code, send emails, manipulate files, and call APIs. This autonomy, while powerful, introduces unprecedented risks. A single errant instruction from an agent could delete critical data, initiate fraudulent transactions, or expose sensitive information. Current safety approaches—primarily reliant on prompt engineering and post-training alignment—are proving insufficient for this new paradigm of runtime execution.

Nomos enters this breach not as another agent framework, but as a fundamental governance infrastructure. Its core innovation is the deliberate decoupling of an agent's 'planning' phase from its 'execution' phase. Every action an agent proposes—whether sending an email, executing a shell command, or making an API call—is intercepted by the Nomos firewall. The firewall then evaluates the action against a configurable policy engine before granting or denying authorization. This architecture introduces mandatory checkpoints for safety, compliance, and operational guardrails.

The immediate application is clear: enabling the trusted deployment of AI agents in sensitive domains like financial services, IT operations, and personal assistance, where a single unauthorized action could be catastrophic. The long-term implication is even more significant. Nomos conceptualizes a future where 'Agent Security Operations' (ASecOps) becomes as standard as cybersecurity, creating an entirely new layer in the AI infrastructure market. Its success hinges not just on its core technology, but on its ability to foster a rich ecosystem of policy plugins, making safety as modular and adaptable as the agents it aims to govern.

Technical Deep Dive

Nomos is architected as a policy enforcement point (PEP) that sits between an AI agent's reasoning engine and the external tools or APIs it wishes to use. Its design philosophy is rooted in the principle of least privilege and explicit authorization. Technically, it operates as a gRPC or HTTP proxy. When an agent, built on frameworks like LangGraph or AutoGen, decides to take an action, it does not execute it directly. Instead, it sends an action request to the Nomos server. This request contains a structured description of the intended action (e.g., `{action: 'send_email', params: {to: 'client@example.com', subject: 'Invoice', body: '...'}}`).

The Nomos server's workflow is a multi-stage pipeline:
1. Interception & Parsing: The action request is captured and parsed into a normalized internal representation.
2. Policy Evaluation: The parsed action is passed to a policy decision point (PDP). This is the heart of Nomos. Policies are written in a domain-specific language (DSL) or via a graphical interface, defining rules based on actor identity, action type, resource parameters, and contextual state (e.g., time of day, previous actions in a session).
3. External Enrichment (Optional): The PDP can query external systems—a compliance database, a security information and event management (SIEM) system, or even a secondary AI model for nuanced judgment—to enrich its decision context.
4. Authorization & Logging: The PDP returns an `ALLOW`, `DENY`, or `MODIFY` decision. Crucially, all decisions and the full context of the request are immutably logged to an audit trail, providing forensic capability.
5. Execution Proxy: If allowed, Nomos proxies the action to the real tool/API. If modified, it adjusts the parameters before proxying.

A key technical differentiator is its focus on session-aware policies. Unlike simple API gateways, Nomos maintains the context of an agent's entire task session. This allows for policies like "an agent can only modify a database record if it first successfully read that record in the same session" or "total monetary value of transactions in a session must not exceed $10,000."

The project is open-source, with its core repository (`nomos-system/nomos-core`) on GitHub. Recent commits show active development around a plugin architecture for policy engines, with early adapters integrating Open Policy Agent (OPA) and custom logic. The repository has garnered significant attention, reflecting the acute market need.

| Security Layer | Stage of Intervention | Primary Mechanism | Limitation for Agents |
|---|---|---|---|
| Training Alignment | Pre-deployment | RLHF, Constitutional AI | Cannot anticipate novel runtime scenarios; 'alignment tax' on capabilities |
| Prompt Engineering | Inference-time | System prompts, in-context instructions | Brittle, can be jailbroken or ignored; no hard enforcement |
| Tool-Level Permissions | Execution-time | OS/user-level access controls | Coarse-grained; lacks agent-specific context and session awareness |
| Nomos Execution Firewall | Execution-time | Interception, policy evaluation, audit | Introduces latency; requires policy definition overhead |

Data Takeaway: The table highlights the layered security model for AI. Nomos occupies a unique and necessary niche, providing fine-grained, context-aware, enforceable policies at the precise moment of action—a gap left by both pre-deployment alignment and static system permissions.

Key Players & Case Studies

The drive for agent safety is creating a new competitive axis. While Nomos is an open-source pioneer, several companies are approaching the same problem from different angles.

Open-Source Frameworks Integrating Governance: LangChain has introduced the concept of "tools" with permission flags, and projects like `langchain-community` offer rudimentary safety chains. However, these are library-level checks, not a standalone, auditable enforcement layer. Microsoft's AutoGen framework includes conversation constraints but lacks a unified policy engine for all agent actions.

Proprietary Cloud Platforms: Major cloud providers are baking agent safety into their managed services. Google Cloud's Vertex AI Agent Builder includes safety settings that filter harmful content and can restrict certain Google Search actions. Amazon Bedrock's Agents for Amazon Bedrock allow for guardrail policies on topics and content. However, these are walled-garden solutions, specific to their own tool ecosystems and opaque in their decision-making.

Enterprise Security Vendors: Established cybersecurity firms are beginning to pivot. Palo Alto Networks and CrowdStrike have research initiatives exploring "AI Security Posture Management" that could extend to agent runtime monitoring. Their strength lies in enterprise integration but they lack the native AI-centric design of Nomos.

Direct Conceptual Competitors: A project like `OpenAI's"Model Context Protocol"` (MCP) aims to standardize how models access tools and data, which could include security schemas. Another emerging concept is the "Agent Kernel," a secure sandbox for execution, as seen in research from Anthropic on sovereign AI. Nomos's distinct position is its agnosticism; it is designed to work with any agent framework and any tool, positioning it as a potential universal standard.

| Solution | Approach | Strengths | Weaknesses | Best For |
|---|---|---|---|---|
| Nomos (Open-Source) | Universal Execution Firewall | Framework-agnostic, transparent, auditable, session-aware | DIY deployment, nascent ecosystem | Tech-forward enterprises, researchers, compliance-heavy sectors |
| Cloud Platform Guardrails (e.g., Bedrock Agents) | Integrated, Managed Service | Easy setup, tightly coupled with proprietary tool suite | Vendor lock-in, limited customization, opaque | Companies all-in on a specific cloud AI stack |
| Agent Framework Safety (e.g., LangChain Tools) | Library-Level Checks | Simple, developer-friendly | Non-enforceable, no central audit, bypassable | Simple prototypes, low-risk internal tasks |
| Traditional API Gateways | Network-Level Control | Mature, high-performance | Lack AI/agent semantic understanding, no session context | Regulating API access, not agent intent |

Data Takeaway: The competitive landscape is fragmented between integrated but locked-in cloud services and flexible but complex open-source solutions. Nomos's open, agnostic architecture gives it a strategic advantage for multi-cloud or hybrid deployments and for setting a *de facto* standard, but it faces the classic open-source challenge of commercialization and enterprise support.

Industry Impact & Market Dynamics

The emergence of execution firewalls like Nomos is not merely a technical feature—it is an enabling technology that will reshape the adoption curve and business models for autonomous AI agents.

Unlocking Regulated Industries: The primary immediate impact is the potential to open floodgates for agent deployment in finance, healthcare, and legal sectors. A major bank cannot deploy an agent to handle customer account inquiries without a system like Nomos providing an immutable audit log of every proposed action and a guarantee that it cannot, under any prompt, initiate an unauthorized wire transfer. Nomos transforms the agent from a black-box liability into a governable system. We predict the first large-scale production deployments will be in IT operations (AI-driven incident response) and financial compliance (automated transaction monitoring and reporting), where the cost of error is high and regulatory requirements are strict.

Creating the "ASecOps" Market: Just as DevOps gave rise to DevSecOps, the agent economy will spawn Agent Security Operations. This will create a new vendor category. Startups will emerge to offer managed Nomos deployments, premium policy packs (e.g., "HIPAA Compliance Policy Pack for Healthcare Agents"), and specialized audit services. Established security companies will acquire or build competing solutions. The total addressable market aligns with the projected growth of the AI agent platform market itself.

| Market Segment | 2024 Estimated Size | 2027 Projected Size | CAGR | Key Driver |
|---|---|---|---|---|
| AI Agent Platforms | $3.2 Billion | $19.6 Billion | ~83% | Automation of complex knowledge work |
| AI Security & Governance | $1.5 Billion | $8.8 Billion | ~80% | Regulatory pressure & high-profile failures |
| Sub-segment: Runtime Agent Security | ~$50 Million (emergent) | ~$2.1 Billion | ~250%* | Deployment of agents in sensitive production environments |
*Note: High CAGR due to starting from near-zero base.*

Data Takeaway: The runtime agent security sub-segment is poised for explosive growth, significantly outpacing the broader AI security market. This reflects the acute, unmet need that Nomos addresses. Its growth is directly tied to the maturation and enterprise adoption of AI agents.

Shifting Value Capture: In the AI stack, value has concentrated at the model layer (OpenAI, Anthropic) and the application layer. Nomos represents a bid to create and dominate a critical *infrastructure* layer. The entity that controls the universal policy layer for AI agents holds a position analogous to a cloud identity provider (like Okta) or an API gateway (like Kong). The business model will likely follow open-source core with commercial enterprise features: advanced analytics, centralized policy management for distributed agents, and SOC2-compliant audit trails.

Risks, Limitations & Open Questions

Despite its promise, the Nomos approach faces significant hurdles.

The Policy Definition Problem: Nomos moves the safety challenge from "aligning the model" to "defining the perfect policy." Writing comprehensive, conflict-free policies for complex, multi-step agent workflows is a monumental task. An overly permissive policy is unsafe; an overly restrictive policy cripples the agent's utility. This could lead to a new form of "policy engineering" that is as complex as prompt engineering. The solution—a community-driven ecosystem of policy templates—is promising but unproven.

Performance & Latency Overhead: Introducing a network hop and a policy evaluation for every single agent action adds latency. For an agent performing a task requiring hundreds of micro-actions (e.g., researching and writing a report), this overhead could become prohibitive. The engineering challenge is to make the policy engine extremely fast and support batched or asynchronous evaluations.

The "Malicious Principal" Problem: Nomos secures the *tools* from the *agent*. But what if the human or system instructing the agent is malicious? Nomos can prevent the agent from sending a phishing email, but it cannot judge if a legally-worded but predatory sales email drafted by a human-controlled agent is "ethical." This is a fundamental limit of execution-layer security.

Adoption Chicken-and-Egg: For Nomos to become a standard, major agent frameworks need to build support for it. For frameworks to prioritize it, there needs to be significant user demand. Breaking this cycle requires a flagship enterprise deployment that demonstrates clear ROI on safety and compliance.

Open Questions: Can the policy DSL be made intuitive enough for non-coders (e.g., compliance officers)? How does Nomos handle ambiguous actions proposed by a highly persuasive agent that attempts to "justify" its violation? Will a competitive, proprietary standard (e.g., from Microsoft or Google) win out over the open-source approach?

AINews Verdict & Predictions

Nomos is more than a tool; it is a necessary conceptual breakthrough for the safe industrialization of AI agents. Its architecture correctly identifies the decoupling of planning and execution as the foundational principle for trustworthy autonomy. While not the first to consider agent safety, its focused, execution-centric approach is the most pragmatic and immediately deployable solution we have seen.

Our specific predictions are:

1. Standardization within 18 Months: Within the next year and a half, a Nomos-like execution firewall will become a *de facto* standard for any serious enterprise AI agent deployment. Major cloud providers will either adopt the open standard or release fully compatible proprietary equivalents.
2. The Rise of Policy-as-Code: A new engineering role—"Agent Security Engineer"—will emerge, specializing in writing and testing policy-as-code for AI workflows. Certifications and dedicated SaaS tools for policy simulation and testing will follow.
3. Merger & Acquisition Target: The core Nomos team or a startup built on its commercial version will be acquired by a major cybersecurity firm (like Palo Alto Networks) or a cloud platform (like Google Cloud) within the next two years, for a sum reflecting its strategic position as gatekeeper to the agent economy.
4. First Major "Save" by 2025: We will see a publicly disclosed incident by late 2025 where a Nomos-style firewall prevented a major financial loss or data breach caused by a misaligned or hijacked AI agent, cementing its value proposition.

What to Watch Next: Monitor the growth of the `nomos-core` GitHub repository's contributor base and the emergence of commercial entities offering support. Watch for announcements from financial institutions or healthcare networks piloting agentic AI with explicit mention of runtime governance. The key indicator of success will not be stars on GitHub, but the signing of a Fortune 500 company to a production contract. Nomos has correctly identified the critical missing layer. The race is now on to build it, scale it, and make it indispensable.

More from Hacker News

기하학적 충돌이 밝혀지다: LLM이 망각하는 이유와 이제 제어가 가능해진 이유For years, catastrophic forgetting in large language models (LLMs) has been an empirical black box. Practitioners reliedLLM이 20년 된 분산 시스템 설계 규칙을 무너뜨리다The fundamental principle of distributed system design—strict separation of compute, storage, and networking—is being quAI 에이전트의 무제한 스캔이 운영자를 파산시키다: 비용 인식 위기In a stark demonstration of the dangers of unconstrained AI autonomy, an operator of an AI agent scanning the DN42 amateOpen source hub3370 indexed articles from Hacker News

Related topics

AI agent security104 related articlesAI infrastructure229 related articles

Archive

March 20262347 published articles

Further Reading

Wirken: 엔터프라이즈 AI 에이전트를 해제할 단일 바이너리 보안 금고Wirken이라는 새로운 오픈소스 프로젝트는 전체 보안 게이트웨이를 단일 정적 바이너리로 압축하여 AI 에이전트의 가장 깊은 신뢰 위기를 해결하고 있습니다. 암호화된 금고가 에이전트별 세분화된 키 접근을 강제함으로써Cube Sandbox, AI 에이전트 혁명의 핵심 인프라로 부상AI 에이전트가 실험적인 데모에서 신뢰할 수 있고 확장 가능한 작업자로 전환되는 것은 근본적인 인프라 격차, 즉 안전하고 성능이 뛰어난 실행 환경 때문에 지연되고 있습니다. 즉시 시작과 경량 격리를 약속하는 새로운 Nono.sh의 커널 수준 보안 모델, 중요 인프라를 위한 AI 에이전트 안전성 재정의오픈소스 프로젝트 Nono.sh는 AI 에이전트 보안에 대한 근본적인 재고를 제안합니다. 취약한 애플리케이션 계층 권한에 의존하는 대신, 커널이 강제하는 제로 트러스트 런타임 모델을 구현하여 모든 에이전트를 본질적으단일 샌드박스 보안이 AI 에이전트에 실패하는 이유와 다음 단계AI 에이전트를 보호하는 보안 모델은 근본적인 변화를 겪고 있습니다. 업계 표준인 단일 샌드박스 접근 방식은 자율적 도구 사용 시스템의 무게 아래 무너지고 있습니다. 세분화된 도구 수준 격리를 기반으로 하는 새로운

常见问题

GitHub 热点“Nomos Execution Firewall: The Critical Missing Layer for Safe AI Agent Deployment”主要讲了什么?

The AI landscape is undergoing a profound transformation as language models evolve from passive text generators into proactive, tool-using agents. Systems like AutoGPT, LangChain a…

这个 GitHub 项目在“How to install and configure Nomos firewall with LangChain”上为什么会引发关注?

Nomos is architected as a policy enforcement point (PEP) that sits between an AI agent's reasoning engine and the external tools or APIs it wishes to use. Its design philosophy is rooted in the principle of least privile…

从“Nomos vs OpenAI moderation API for agent safety”看,这个 GitHub 项目的热度表现如何?

当前相关 GitHub 项目总星标约为 0,近一日增长约为 0,这说明它在开源社区具有较强讨论度和扩散能力。