- Today
- Holidays
- Birthdays
- Reminders
- Cities
- Atlanta
- Austin
- Baltimore
- Berwyn
- Beverly Hills
- Birmingham
- Boston
- Brooklyn
- Buffalo
- Charlotte
- Chicago
- Cincinnati
- Cleveland
- Columbus
- Dallas
- Denver
- Detroit
- Fort Worth
- Houston
- Indianapolis
- Knoxville
- Las Vegas
- Los Angeles
- Louisville
- Madison
- Memphis
- Miami
- Milwaukee
- Minneapolis
- Nashville
- New Orleans
- New York
- Omaha
- Orlando
- Philadelphia
- Phoenix
- Pittsburgh
- Portland
- Raleigh
- Richmond
- Rutherford
- Sacramento
- Salt Lake City
- San Antonio
- San Diego
- San Francisco
- San Jose
- Seattle
- Tampa
- Tucson
- Washington
Anthropic Denied Pentagon Blacklisting Reversal by D.C. Court
AI firm's legal battle with Department of Defense over "supply chain risk" designation continues.
Apr. 11, 2026 at 6:57am by Ben Kaplan
Got story updates? Submit your updates here. ›
As the legal battle over AI's role in national security continues, the intricate digital infrastructure powering Anthropic's technology remains a source of both innovation and caution.San Francisco TodayAnthropic, the AI company behind the powerful language model Claude, has faced a setback in its legal fight against the Pentagon's designation of the firm as a "supply chain risk". While a San Francisco court previously granted Anthropic a temporary reprieve, preventing the administration from outright banning Claude, a D.C. appeals court has now denied the company's request to halt the enforcement of this "supply chain risk" label. This means the Pentagon can continue to view Anthropic's technology with caution, even if other agencies are still able to utilize it.
Why it matters
This legal battle highlights the complex and often conflicting interpretations of risk when it comes to integrating powerful AI tools into critical infrastructure. The Pentagon's concern about supply chain risks is understandable, but applying such a broad label can stifle innovation and create uncertainty for AI developers. The outcome of this case could set important precedents for how AI companies interact with governments worldwide.
The details
The core of the issue is the Pentagon's designation of Anthropic as a "supply chain risk". This label carries significant weight, potentially impacting the company's ability to engage with a major governmental client. While a San Francisco court previously granted Anthropic a temporary reprieve, preventing the administration from outright banning Claude, the D.C. appeals court has now denied the company's request to halt the enforcement of this "supply chain risk" designation. This means the Pentagon can continue to view Anthropic's technology with caution, even if other agencies are still able to utilize it.
- In April 2026, the D.C. appeals court denied Anthropic's request to halt the enforcement of the "supply chain risk" designation.
- Previously, a San Francisco court had granted Anthropic a temporary reprieve, preventing the administration from outright banning the use of its AI model, Claude.
The players
Anthropic
An AI firm that has developed the powerful language model Claude, which is now at the center of a legal battle with the Pentagon.
Department of Defense (Pentagon)
The U.S. Department of Defense, which has designated Anthropic as a "supply chain risk", potentially impacting the company's ability to engage with the government.
What’s next
Anthropic still has significant legal battles to fight, as the preliminary injunction in San Francisco offers a temporary shield, but the D.C. ruling means the Pentagon can still operate under its "supply chain risk" assessment. This creates a peculiar situation where the company might still be used by the Pentagon for a period, but is effectively excluded from new contracts.
The takeaway
This legal battle between Anthropic and the Pentagon highlights the broader societal conversation about how to harness the incredible potential of AI while mitigating its inherent risks. The outcome of this case could set important precedents for how AI companies interact with governments worldwide, and shape the future of AI governance in sensitive sectors.
San Francisco top stories
San Francisco events
Apr. 13, 2026
Janelle JamesApr. 13, 2026
Janelle James




