Anthropic, after refusing to relax AI usage restrictions, was labeled a supply chain risk by the U.S. Department of Defense, leading to a lawsuit alleging violation of constitutional rights. The case may reshape the military's evaluation standards for AI suppliers.
Artificial intelligence company Anthropic has recently filed a lawsuit against the U.S. Department of Defense, stemming from the department's designation of the company as a "supply chain risk," which prohibits its AI technology from participating in any Department of Defense projects. This label directly cuts off Anthropic's access to military collaborations, sparking strong concerns from the company regarding the rationality of the policy.
According to disclosures, the Department of Defense made this decision on March 5, citing that Anthropic had set usage restrictions for its AI model Claude, such as prohibiting its use in autonomous weapons development or domestic surveillance. However, the military believes that these restrictions may hinder the technology's application in legitimate defense missions, arguing that legal compliance, rather than corporate policy, should determine the boundaries of use.
Anthropic insists that its safety mechanisms are necessary designs based on ethical responsibility and public trust, and points out that the label did not give the company an opportunity for a hearing, which it alleges violates the constitutional guarantees of due process and freedom of speech. The company has applied to the court to revoke the label and requested a suspension of all related restrictions until the case is resolved.
This dispute has been brewing for some time. In the preceding months, the two parties had engaged in multiple rounds of consultations on the ethical boundaries of AI, but Anthropic consistently refused to compromise its core safety architecture. Now that the lawsuit has escalated, it not only concerns one company's market access but may also set a key precedent for how the government will evaluate AI suppliers in the future, and whether companies can independently set the boundaries for technology use.
The outcome of this case may reshape the cognitive framework of AI ethics and safety standards in the U.S. defense technology procurement system.
0 comment A文章作者M管理员
No Comments Yet. Be the first to share what you think
❯
Profile
Search
Checking in, please wait...
Click for today's check-in bonus!
You have earned {{mission.data.mission.credit}} points today