
U.S. District Judge Rita Lin Blocks Pentagon Designation Of Anthropic Supply Chain Risk
Key Takeaways
- Judge Rita F. Lin issued a preliminary injunction blocking Anthropic's supply-chain risk designation.
- The injunction bars the Pentagon from branding Anthropic a national security risk and restricting contracts.
- Court findings reference potential First Amendment retaliation against Anthropic for public statements.
New development: injunction blocks designation
Newly unfolding: a federal judge granted a preliminary injunction that blocks the Pentagon from designating Anthropic as a supply chain risk and pauses President Trump's directive to cut off its work, effectively restoring the February 27 status quo.
“SAN FRANCISCO -- Artificial intelligence company Anthropic is asking a federal judge on Tuesday to temporarily halt the Pentagon's "unprecedented and stigmatizing" designation of the company as a supply chain risk”
The decision, issued by U.S. District Judge Rita Lin, also delays enforcement for one week to allow an appeal.

The ruling frames the designation as 'likely both contrary to law and arbitrary and capricious' and casts doubt on the government's asserted national security rationale.
It preserves Anthropic’s ability to work with the government under existing contracts and makes clear that punitive actions tied to public speech risk violating the First Amendment.
The move signals a potential turn in how procurement powers are wielded against AI firms, even as the DoD has relied on Claude for military tasks such as target analysis in Iran-linked operations.
What the injunction blocks and implies
What this injunction actually blocks is central: the designation of Anthropic as a 'supply chain risk' and the concurrent White House directive to sever federal ties.
The court’s action halts actions tied to a designation typically reserved for foreign actors, and preserves Anthropic’s current government access while litigation unfolds.
Government supporters argued the designation was necessary for national security, while Anthropic argued the move was punitive retaliation for its safety stance and public criticism.
The Pentagon has been using Claude for sensitive operations, including in target analysis, and the ruling explicitly preserves a path for the government to transition away or continue under regulations, rather than under the risk label alone.
Legal rights and First Amendment
Legally, the case foregrounds First Amendment and due-process questions: is punishing a company for voicing safety and policy disagreements lawful?
“A federal judge in California has blocked the Trump administration from designating Anthropic a supply chain risk to national security and cutting off the AI company’s work with federal agencies”
The judge’s framing suggests the government’s broad punitive measures against Anthropic may violate constitutional protections, a point highlighted by advocates and several outlets.
Anthropic argues that the government cannot weaponize procurement policy to punish speech, while the administration contends the actions are justified by security concerns and contract performance.
The cross-cutting fight demonstrates how AI vendors’ public stances can shape, and be shaped by, procurement power and speech rights.
Next steps and timelines
Procedural path forward remains murky but constricted: the injunction is temporary, and the court emphasized a week-long pause to allow appeals.
The directive’s ultimate fate will hinge on further rulings, potential appellate review, and how DoD contractors adjust under new constraints.

Coverage indicates the final verdict could take weeks or months, and officials may seek to maneuver the relationship with Claude under alternative regulatory or contractual arrangements.
Anthropic may begin to lean on the court’s decision to reassure customers and partners that its leadership in safety remains recognised during the dispute.
Broader implications and context
Broader implications: the Lin decision adds a high-profile constitutional check on executive power in AI procurement, and it could influence how other agencies handle vendor safety, transparency, and public critique.
“A federal judge has sided with Anthropic in its twisty legal battle with the Trump administration, awarding the tech company an injunction against the government’s recent order that labeled it a “supply chain risk,” the Wall Street Journal reports”
The ruling is being watched by tech policy observers and civil liberties groups who argue it reinforces due-process norms in the face of fast-moving AI weaponization debates.

Industry coverage suggests the decision could clear space for Anthropic to resume engagements with customers while the legal process plays out, potentially shaping the competitive landscape for AI tools in West Asia-related security contexts and beyond.
More on Technology and Science
David Sacks Exits White House AI, Crypto Role To Co-Chair PCAST
26 sources compared

Federal Judge Grants Injunction Blocking President Trump's Ban And Pentagon Supply-Chain Designation Against Anthropic
11 sources compared
Trump Orders Ban And Supply Chain Risk Label For Anthropic; Judge Temporarily Blocks Order
10 sources compared

U.S. District Judge Rita Lin Blocks Trump Directive, Pentagon Supply-Chain Designation For Seven Days
16 sources compared