What does the US military’s feud with Anthropic mean for AI used in war?
Tech policy professor who served in US air force explains how a feud between an AI startup and the US military illuminates ethical fault linesAnthropic’s ongoing fight with the Department of Defense over what safety restrictions it can put on its artificial intelligence models has captivated the tech industry, acting as a test of how AI may be used in war and the government’s power to coerce companies to meet its demands.The negotiations have revolved around Anthropic’s refusal to allow the federal government to use its Claude AI for domestic mass surveillance or autonomous weapons systems, but the dispute also reflects the messy nature of what happens when tech companies have their products integrated into conflict. The Pentagon this week declared Anthropic a supply chain risk for its ref
Generated by Pulse AI, Glideslope's proprietary engine for interpreting market sentiment and economic signals. For informational purposes only — not financial advice.