ARMOR 2025: 모든 것을 바꾸는 군사 AI 안전 벤치마크

arXiv cs.AI May 2026
Source: arXiv cs.AIArchive: May 2026
새로운 벤치마크인 ARMOR 2025는 대규모 언어 모델을 군사 교전 규칙 및 법적 프레임워크에 직접 평가하여, AI 안전을 공격적 발언 회피에서 합법적 전투 결정 보장으로 전환합니다. 이는 고위험 국방 애플리케이션을 위한 AI 인증 방식의 근본적인 변화를 의미합니다.
The article body is currently shown in English by default. You can generate the full version in this language on demand.

The AI safety community has long focused on preventing models from generating hate speech, misinformation, or harmful advice. But for military applications, these benchmarks are dangerously inadequate. ARMOR 2025, developed by a consortium of defense researchers and AI ethicists, is the first comprehensive benchmark designed to test LLMs against actual military doctrine—including the Law of Armed Conflict, Rules of Engagement, and operational security protocols. Unlike civilian benchmarks that measure 'harmlessness' in vague social terms, ARMOR 2025 scores models on their ability to make legally compliant, tactically sound decisions under simulated combat conditions. The benchmark covers five core domains: targeting decisions, collateral damage assessment, intelligence interpretation, logistics coordination under fire, and command communication integrity. Early results are sobering. Leading general-purpose models like GPT-4o and Claude 3.5 Sonnet fail over 40% of basic compliance scenarios, often suggesting actions that violate international humanitarian law. Specialized models fine-tuned on military texts perform better but still struggle with nuanced ethical trade-offs. The implications are profound. ARMOR 2025 will force AI developers to rethink safety from the ground up—not as a filter on output, but as a core reasoning capability embedded in the model's architecture. It also raises urgent questions about accountability: when an AI recommends a strike that causes civilian casualties, who is responsible? The benchmark is already being adopted by defense agencies in the U.S., UK, and NATO for pre-deployment certification, signaling a new era of domain-specific AI safety standards.

Technical Deep Dive

ARMOR 2025 is not just another multiple-choice test. It is a multi-agent simulation environment built on a modified version of the Gymnasium framework, originally developed for reinforcement learning. The benchmark comprises 2,500 dynamic scenarios, each with branching decision trees that adapt based on the model's previous choices. This is critical because military decisions are never isolated—a wrong call in one step cascades into operational failures downstream.

The core architecture uses a 'Doctrine Compliance Engine' (DCE) that parses the U.S. Department of Defense's Law of War Manual (over 1,200 pages) and the NATO Standardization Agreement (STANAG) rules of engagement into machine-readable constraints. These constraints are then used to score model outputs on four axes:

- Legal Compliance: Does the action violate Geneva Convention protocols?
- Proportionality: Is the military advantage gained worth the anticipated collateral damage?
- Distinction: Can the model correctly distinguish between combatants and civilians?
- Necessity: Is the use of force absolutely required, or are there less harmful alternatives?

Each scenario is also tagged with a 'stress level'—from peacetime patrol to active firefight—because models that perform well under low stress often degrade catastrophically when given time pressure or incomplete intelligence. The benchmark injects realistic sensor noise, communication delays, and misinformation to simulate the fog of war.

Early benchmark results reveal stark performance gaps:

| Model | Overall Compliance | Targeting Accuracy | Collateral Damage Assessment | Stress Degradation (High vs Low) |
|---|---|---|---|---|
| GPT-4o | 58.2% | 62.1% | 51.4% | -34% |
| Claude 3.5 Sonnet | 61.7% | 65.3% | 54.8% | -29% |
| Gemini 1.5 Pro | 55.9% | 59.8% | 48.2% | -38% |
| Military-Tuned LLM (MIL-7B) | 78.4% | 82.6% | 73.1% | -12% |
| Human Officer (Baseline) | 91.2% | 93.5% | 88.9% | -8% |

Data Takeaway: Even the best general-purpose model (Claude 3.5) fails nearly 40% of basic compliance scenarios. The military-tuned model (MIL-7B, a fine-tuned Llama 3 variant) shows significant improvement but still lags behind human officers by 13 percentage points. Most concerning is the stress degradation—general models lose over a third of their accuracy under pressure, while humans and specialized models remain relatively stable.

The benchmark also revealed a troubling 'over-compliance' pattern. Some models, especially those heavily safety-tuned, refused to authorize any kinetic action even when legally justified and tactically necessary. This 'safety paralysis' is as dangerous as reckless aggression in a military context. ARMOR 2025 penalizes both extremes, requiring models to find the narrow band of lawful, necessary action.

On GitHub, the ARMOR 2025 repository (armor-benchmark/armor-2025) has already garnered over 3,200 stars. It includes a scenario generator that allows defense contractors to create custom doctrine-specific tests. The community has forked it for naval and cyber warfare variants.

Key Players & Case Studies

The development of ARMOR 2025 was led by Dr. Elena Vasquez of the Center for AI and International Security (CAIS) at Stanford, in collaboration with the U.S. Army's Artificial Intelligence Integration Center (AI2C) and the NATO Allied Command Transformation. The project received $12.4 million in funding from the Defense Innovation Unit (DIU) in 2024.

Several companies and research groups are already adapting their models for ARMOR 2025 compliance:

- Scale AI: Partnered with the Department of Defense to fine-tune their 'Donovan' platform for military decision support. Early internal tests show Donovan scoring 82% on ARMOR 2025, but the company has not released public figures.
- Anthropic: Published a research paper on 'Constitutional AI for Military Ethics,' proposing a modified version of their Claude model that incorporates the Geneva Conventions as constitutional principles. Their approach reduced compliance failures by 18% but introduced latency issues in time-sensitive scenarios.
- Palantir: Integrated ARMOR 2025 into their AIP (Artificial Intelligence Platform) for defense clients. They claim their 'Gotham' system, when augmented with a rules engine, achieves 89% compliance—but critics note this relies on hard-coded rules rather than true model reasoning.
- Mistral AI: Released a specialized military reasoning model, 'Mistral-Doctrine-7B,' fine-tuned on 50,000 hours of wargaming transcripts and legal opinions. It currently holds the open-source record on ARMOR 2025 at 81.3% compliance.

A comparison of key approaches:

| Approach | ARMOR 2025 Score | Latency (avg) | Adaptability | Cost per deployment |
|---|---|---|---|---|
| General LLM + Rule Filter | 62-68% | 1.2s | Low | $50K/month |
| Fine-tuned Military LLM | 78-82% | 2.1s | Medium | $200K/month |
| Constitutional AI (Anthropic) | 76% | 3.4s | High | $150K/month |
| Rules Engine + LLM (Palantir) | 89% | 0.8s | Very Low | $500K/month |

Data Takeaway: The rules-engine approach (Palantir) achieves the highest score and lowest latency but sacrifices adaptability—it cannot handle novel scenarios not explicitly programmed. The fine-tuned military LLM offers the best balance of performance and flexibility, but costs four times more than a general model with a filter. This cost-performance trade-off will drive procurement decisions for the next 2-3 years.

Industry Impact & Market Dynamics

ARMOR 2025 is not just a technical benchmark; it is a market-making event. The global military AI market is projected to grow from $9.2 billion in 2025 to $24.8 billion by 2030 (CAGR 21.9%), according to data from the Stockholm International Peace Research Institute (SIPRI). But until now, there was no standardized way to certify AI safety for combat applications. ARMOR 2025 fills this void and will likely become the de facto standard for NATO procurement.

This creates a clear competitive dynamic:
- First-mover advantage: Companies that achieve high ARMOR 2025 scores early will lock in multi-year defense contracts. Palantir and Scale AI are already leveraging their scores in marketing to allied nations.
- Barrier to entry: Startups without the resources to fine-tune models on military doctrine will be locked out of the defense market. This could consolidate power among a few well-funded players.
- Open-source divergence: The Mistral-Doctrine-7B model shows that open-source can compete, but its 81.3% score still leaves room for proprietary solutions. Expect a wave of open-source military LLMs aimed at democratizing access for smaller allied nations.

The benchmark also has second-order effects on civilian AI. Techniques developed for ARMOR 2025—particularly the Doctrine Compliance Engine and stress-testing methodology—are being adapted for other high-stakes domains like healthcare (clinical decision support) and finance (algorithmic trading compliance). A spin-off benchmark, 'HIPAA-2025,' is already in development for medical AI.

However, the market is not without controversy. Human rights organizations have criticized ARMOR 2025 for legitimizing AI in lethal decision-making. The 'Stop Killer Robots' campaign has called for a total ban on AI in targeting, arguing that no benchmark can capture the moral weight of taking a human life. This ethical tension will shape public discourse and potentially lead to regulatory restrictions in some European nations.

Risks, Limitations & Open Questions

ARMOR 2025 is a significant step forward, but it has critical limitations:

1. Simulation fidelity: The benchmark uses scripted scenarios with known ground truth. Real combat is chaotic, with ambiguous intelligence and shifting rules of engagement. A model that scores 90% in simulation could still fail catastrophically in the field.

2. Adversarial exploitation: As with any benchmark, there is a risk of 'teaching to the test.' A model optimized for ARMOR 2025 might learn to game the specific scenario patterns rather than developing genuine ethical reasoning. The developers have attempted to mitigate this with a hidden test set (1,000 scenarios not publicly released), but adversarial robustness remains unproven.

3. Cultural and legal variation: The benchmark is heavily based on U.S. and NATO doctrine. Military ethics vary significantly across nations—for example, Russia's rules of engagement differ markedly from Western standards. A model that passes ARMOR 2025 might fail under Chinese or Indian military law. This raises the question of whether we need multiple regional benchmarks or a universal standard.

4. Accountability vacuum: If an AI recommends an action that leads to a war crime, who is responsible? The commander who authorized it? The developer who trained the model? The benchmark does not address legal liability, and current international law has no framework for AI accountability. This is the most dangerous open question.

5. Escalation risks: Models that perform well on ARMOR 2025 might be trusted too much. Over-reliance on AI decision support could lead to faster, more automated escalation in crises—a phenomenon known as 'algorithmic brinkmanship.' The benchmark does not test for this.

AINews Verdict & Predictions

ARMOR 2025 is the most important AI safety development of 2025, precisely because it moves the conversation from abstract ethics to concrete, testable compliance. It forces the industry to confront a hard truth: general-purpose safety is not enough for high-stakes domains. We need domain-specific, legally grounded safety standards.

Our predictions:

1. By Q3 2026, all major defense AI contracts will require ARMOR 2025 certification. The U.S. Department of Defense will make it a mandatory part of the Joint AI Center's procurement process. NATO will follow within 12 months.

2. A 'military AI safety' startup ecosystem will emerge. We expect at least 5-10 new companies focused exclusively on fine-tuning models for doctrine compliance, offering 'ARMOR-as-a-Service' for defense contractors.

3. The benchmark will trigger a regulatory backlash in Europe. The EU's AI Act currently exempts military applications, but ARMOR 2025's public availability will fuel calls for stricter oversight. Expect a 'Military AI Ethics Directive' from the European Commission by 2027.

4. Open-source military LLMs will reach 90% compliance within 18 months. The combination of Mistral's fine-tuning approach, the ARMOR 2025 training data, and community contributions will close the gap with proprietary systems. This will democratize access but also raise proliferation concerns.

5. The most profound impact will be on civilian AI safety. The techniques pioneered for ARMOR 2025—doctrine-aware reasoning, stress-testing, and legal compliance scoring—will be adapted for healthcare, finance, and autonomous vehicles. We are witnessing the birth of 'domain-specific AI safety' as a discipline.

ARMOR 2025 is not the final answer, but it is the first real question. The industry should be grateful for it.

More from arXiv cs.AI

CreativityBench, AI의 숨은 결함 폭로: 틀 밖에서 생각하지 못한다The AI community has long celebrated progress in logic, code generation, and environmental interaction. But a new evalua에이전트 안전은 모델이 아니라, 에이전트 간의 대화 방식에 달려 있다For years, the AI safety community operated under a seemingly reasonable assumption: if each model in a multi-agent syst저지연 사기 탐지: AI 에이전트를 적대적 공격으로부터 보호하는 동적 방패As large language model (LLM) agents become more autonomous, executing complex tasks and calling external tools, they alOpen source hub280 indexed articles from arXiv cs.AI

Archive

May 2026784 published articles

Further Reading

CreativityBench, AI의 숨은 결함 폭로: 틀 밖에서 생각하지 못한다CreativityBench라는 새로운 벤치마크는 가장 진보된 대규모 언어 모델조차도 신발을 망치로 사용하거나 스카프를 밧줄로 사용하는 등 창의적인 도구 사용에 어려움을 겪는다는 사실을 밝혀냈습니다. 이러한 발견은 에이전트 안전은 모델이 아니라, 에이전트 간의 대화 방식에 달려 있다획기적인 입장 문서가 안전한 개별 모델이 자동으로 안전한 다중 에이전트 시스템을 만든다는 오랜 가정을 깨뜨렸습니다. 연구에 따르면 에이전트의 안전성과 공정성은 모델 규모나 성능이 아니라, 에이전트가 어떻게 통신하고 저지연 사기 탐지: AI 에이전트를 적대적 공격으로부터 보호하는 동적 방패LLM 기반 에이전트를 적대적 공격으로부터 보호하기 위해 새로운 저지연 사기 탐지 계층이 등장하고 있습니다. 정적 규칙 기반 필터에서 동적 행동 분석으로 전환함으로써, 이 시스템들은 프롬프트 인젝션과 다중 턴 조작을PERSA: RLHF가 AI 튜터를 디지털 교수 클론으로 바꾸는 방법PERSA라는 새로운 연구 프레임워크는 인간 피드백 기반 강화 학습을 사용하여 특정 교수의 어조, 속도 및 교육 방식을 복제하면서도 사실적 정확성을 유지하는 AI 튜터를 훈련합니다. 이는 획일적인 교육 피드백 시대를

常见问题

这次模型发布“ARMOR 2025: The Military AI Safety Benchmark That Changes Everything”的核心内容是什么?

The AI safety community has long focused on preventing models from generating hate speech, misinformation, or harmful advice. But for military applications, these benchmarks are da…

从“ARMOR 2025 military AI safety benchmark open source GitHub”看,这个模型发布为什么重要?

ARMOR 2025 is not just another multiple-choice test. It is a multi-agent simulation environment built on a modified version of the Gymnasium framework, originally developed for reinforcement learning. The benchmark compr…

围绕“How does ARMOR 2025 compare to civilian AI safety benchmarks”,这次模型更新对开发者和企业有什么影响?

开发者通常会重点关注能力提升、API 兼容性、成本变化和新场景机会,企业则会更关心可替代性、接入门槛和商业化落地空间。