Hol
Well-known
In a significant development for the intersection of artificial intelligence policy and national security, a federal appeals court in Washington ruled on April 8 that the Department of War may designate Anthropic as a supply-chain risk while a full judicial review plays out. The decision came after the AI company sought an emergency stay to block the controversial designation.
…
The move originated after Anthropic declined a Department of War request to alter the user policies and safety guardrails of its flagship AI model, Claude. The company refused to remove restrictions that prevent the AI from being used for mass surveillance or the development and operation of fully autonomous weapons systems. Anthropic has emphasized its commitment to “constitutional AI” principles and responsible deployment, arguing that such guardrails are essential to ethical AI use.
The Pentagon has stated publicly that it does not intend to employ Claude for those specific purposes, but it has insisted on the flexibility to use the technology for all lawful military applications. President Donald Trump weighed in on social media earlier, accusing Anthropic of trying to “strong-arm” the federal government by using its AI policies to dictate military decisions.
www.zerohedge.com
…
The move originated after Anthropic declined a Department of War request to alter the user policies and safety guardrails of its flagship AI model, Claude. The company refused to remove restrictions that prevent the AI from being used for mass surveillance or the development and operation of fully autonomous weapons systems. Anthropic has emphasized its commitment to “constitutional AI” principles and responsible deployment, arguing that such guardrails are essential to ethical AI use.
The Pentagon has stated publicly that it does not intend to employ Claude for those specific purposes, but it has insisted on the flexibility to use the technology for all lawful military applications. President Donald Trump weighed in on social media earlier, accusing Anthropic of trying to “strong-arm” the federal government by using its AI policies to dictate military decisions.
Federal Appeals Court Allows Pentagon To Designate Anthropic As A Supply-Chain Risk<!-- --> | ZeroHedge
ZeroHedge - On a long enough timeline, the survival rate for everyone drops to zero