OpenAI Strikes Pentagon AI Deal

OpenAI inks a deal with the Department of War for classified AI deployments, emphasizing strict safety guardrails against surveillance and autonomous weapons.

2 min read
OpenAI logo with a stylized Pentagon building in the background.
Image credit: OpenAI News

OpenAI announced a significant agreement with the Department of War to deploy advanced AI systems within classified environments. The company stated its deal includes more robust safety guardrails than previous classified AI deployments, including those with Anthropic.

The agreement is guided by three core principles: no use of OpenAI technology for mass domestic surveillance, no direction of autonomous weapons systems, and no high-stakes automated decisions. OpenAI claims this approach offers better protection against unacceptable use compared to relying solely on usage policies.

The company detailed a multi-layered safety strategy for the deployment. This includes a cloud-only architecture, ensuring the use of OpenAI's proprietary safety stack, and requiring cleared OpenAI personnel to be involved in operations. This is in addition to existing U.S. legal protections.

Deployment Architecture and Contractual Safeguards

The deployment will be strictly cloud-based, with OpenAI managing its safety stack. The company confirmed it will not provide 'guardrails off' models or deploy on edge devices, which could facilitate the use of autonomous weapons AI. OpenAI retains the ability to independently verify that its redlines are not breached.

Contractually, the Department of War may use the AI System for lawful purposes, consistent with oversight protocols. Critically, the AI System will not be used to independently direct autonomous weapons where human control is legally required. It also prohibits assuming other high-stakes decisions requiring human approval, referencing DoD Directive 3000.09.

The contract also explicitly bars unconstrained monitoring of U.S. persons’ private information and restricts use in domestic law-enforcement activities, except as permitted by law. This framework aims to prevent scenarios that Anthropic previously raised concerns about regarding classified AI deployments.

Expert Involvement and Rationale

OpenAI will embed cleared forward-deployed engineers and have safety and alignment researchers involved. The company explained its rationale includes the U.S. military's need for advanced AI against adversaries and a desire to foster collaboration between AI labs and the government.

OpenAI requested that the same deployment terms be made available to all AI companies, aiming to de-escalate tensions and establish a more collaborative environment. The company believes its deal offers stronger guarantees than prior agreements, citing the cloud-only limitation, its control over the safety stack, and the inclusion of personnel in the loop.