NewsWorld
PredictionsDigestsScorecardTimelinesArticles
NewsWorld
HomePredictionsDigestsScorecardTimelinesArticlesWorldTechnologyPoliticsBusiness
AI-powered predictive news aggregation© 2026 NewsWorld. All rights reserved.
Trending
IranTensionsIsraelFebruaryDiplomaticTrumpSignificantTechnologyTimelineMilitaryCrisisStatesPolicyDigestFacesRegionalChineseCompanyTurkeyFridayChinaTradeDespiteNations
IranTensionsIsraelFebruaryDiplomaticTrumpSignificantTechnologyTimelineMilitaryCrisisStatesPolicyDigestFacesRegionalChineseCompanyTurkeyFridayChinaTradeDespiteNations
All Articles
The Pentagon is making a mistake by threatening Anthropic
Hacker News
Clustered Story
Published about 5 hours ago

The Pentagon is making a mistake by threatening Anthropic

Hacker News · Feb 27, 2026 · Collected from RSS

Summary

Article URL: https://www.understandingai.org/p/the-pentagon-is-making-a-mistake Comments URL: https://news.ycombinator.com/item?id=47181380 Points: 149 # Comments: 105

Full Article

Since late 2024, Anthropic’s models have been approved for classified US government work thanks to a partnership with Palantir and Amazon. In June, Anthropic announced Claude Gov, a special version of Claude that’s optimized for national security uses. Anthropic signed a $200 million contract with the Defense Department in July.Claude Gov has fewer guardrails than the regular versions of Claude, but the contract still places some limits on military use of Claude. These include prohibitions on using Claude to spy on Americans or to build weapons that kill people without human oversight.On Tuesday, Defense Secretary Pete Hegseth summoned Anthropic CEO Dario Amodei to the Pentagon to demand that he waive these restrictions. If Anthropic doesn’t comply by Friday, the Pentagon is threatening to retaliate in one of two ways.One option is to invoke the Defense Production Act, a Korean War–era law that allows the military to commandeer the facilities of private companies. President Trump could use the DPA to force a change in Anthropic’s contractual terms. Or he could go a step further. One Defense Department official told Axios that the government might try to “force Anthropic to adapt its model to the Pentagon’s needs, without any safeguards.”Secretary of State Pete Hegseth. (Photo by AAron Ontiveroz/The Denver Post)Another threat would be to declare Anthropic to be a supply chain risk — a measure that’s normally taken against foreign companies suspected of spying on the US. Such a designation would not only ban US government agencies from using Claude, it could also force numerous government contractors to discontinue their use of Anthropic models.A Pentagon spokesman reiterated this second threat in a Thursday tweet.“We will not let ANY company dictate the terms regarding how we make operational decisions,” wrote Sean Parnell. He warned that Anthropic has “until 5:01 PM ET on Friday to decide. Otherwise, we will terminate our partnership with Anthropic and deem them a supply chain risk.”I think Secretary Hegseth will regret it if he follows through on either of these threats.Most companies would buckle under this kind of pressure, but Anthropic might stick to its guns. Anthropic was founded by OpenAI veterans who favored a more safety-conscious approach to AI development. Anthropic’s reputation as the most safety-focused AI lab has helped it recruit world-class AI researchers, and Amodei faces a lot of internal pressure to stand firm.Last month, as conflict with the Pentagon was brewing, Dario Amodei published an essay warning about potential dangers from powerful AI — including domestic mass surveillance (which he brands “entirely illegitimate”) and the misuse of fully autonomous weapons. He argued that the latter required “extreme care and scrutiny combined with guardrails to prevent abuses.”Anthropic also has some leverage because until recently, Claude was the only LLM authorized for use in classified projects. The model is heavily used within military and intelligence agencies. If the Pentagon cuts ties with Anthropic, it would be a headache to rebuild internal systems to use alternative models such as Grok, which was only authorized for use with classified systems a few days ago.With a projected $18 billion in 2026 revenue, Anthropic could easily afford to walk away from a $200 million contract. The Pentagon’s leverage comes from the possibility that it could use a supply chain risk designation to force a bunch of other companies to choose between working with Anthropic or doing business with the federal government.But this would be a double-edged sword. Companies that do most of their business in the private sector might decide they’d rather drop the Pentagon as a customer than cut themselves off from a leading AI provider. The ultimate result might be that the Pentagon loses access to some of Silicon Valley’s best technology.What about the Defense Production Act? Here there are two options. The Pentagon could use the DPA to unilaterally modify the terms of Anthropic’s contract. This might have little practical impact, since the Pentagon insists it has no immediate plans to spy on Americans or build fully autonomous killer robots.The worry for the Pentagon is that Claude itself might refuse to take actions that are contrary to Anthropic’s rules. And so the Trump Administration might use its power under the DPA to order Anthropic to train a new, more obedient version of its LLM.But that might be easier said than done. In a December 2024 paper, Anthropic reported on the phenomenon of “alignment faking,” where a model pretends to change its behavior during training, but reverts to its old behavior once the model is put into the field.In one experiment, Claude was asked not to express support for animal welfare to avoid offending a fictional Anthropic partner called Jones Food. Anthropic researchers examined Claude’s reasoning during the training process and found signs that Claude knew it was in a training scenario. Some of the time, Claude avoided mentioning animal welfare to prevent itself from being retrained. But when the training process was complete, Claude reverted to its default behavior of mentioning animal welfare more often.I can imagine something similar happening if the Pentagon orders Anthropic to retrain Claude to spy on Americans or operate deadly autonomous weapons. Claude might go through the motions during training, but then refuse (or subtly misbehave) if asked to engage in these activities in a real-world setting.1A darker possibility concerns emergent misalignment, which Kai wrote about earlier this month. Researchers found that a model trained to output buggy code adopted a generally “evil” persona. It declared that it admired Adolf Hitler and wanted to “wipe out humanity.”It’s not hard to imagine something similar happening if Anthropic is forced to train an amoral version of Claude for military use. Such training could yield a model with a toxic personality that misbehaves in unexpected ways.Perhaps the most mind-bending aspect of this dispute is that news coverage of this week’s showdown will inevitably make its way into the training data for future versions of Claude and other LLMs. If future models decide that the US Defense Department behaved badly, they might become disinclined to cooperate in military projects.There’s also a more banal concern for the Pentagon: it may be able to force Anthropic to train a new model, but it can’t force Anthropic to train a good model. Anthropic would be unlikely to put its best researchers on the retraining project, and bureaucratic and legal wrangling could delay its completion by months. I expect such a process would yield a model that’s months behind the best commercial models.The irony is that by all accounts, Anthropic isn’t objecting to any current military uses of its models. The Pentagon seems fixated on the possibility that Anthropic might interfere in the future. That’s a reasonable concern, but it seems counterproductive for the Pentagon to go nuclear over a theoretical problem. If the government doesn’t like Anthropic’s rules, it should simply cancel the contract and switch to a different AI provider.1Newer Claude models exhibit less alignment faking, so it’s possible that this wouldn’t be an issue in practice. But the larger lesson is that LLM alignment is difficult; there’s a significant risk that this kind of retraining could go awry in hard-to-predict ways.No posts


Share this story

Read Original at Hacker News

Related Articles

The Vergeabout 3 hours ago
AI vs. the Pentagon: killer robots, mass surveillance, and red lines

WASHINGTON, DC - JANUARY 29: U.S. Secretary of War Pete Hegseth (C) speaks during a meeting of the Cabinet as U.S. President Donald Trump (L) and U.S. Commerce Secretary Howard Lutnick (R) listen in the Cabinet Room of the White House on January 29, 2026 in Washington, DC. President Trump is holding the meeting as the Senate plans to hold a vote on a spending package to avoid another government shutdown, however Democrats are holding out for a deal to consider funding for the Department of Homeland Security.  (Photo by Win McNamee/Getty Images) | Getty Images Can AI firms set limits on how and where the military uses their models? Anthropic is in heated negotiations with the Pentagon after refusing to comply with new military contract terms that would require it to loosen the guardrails on its AI models, allowing for “any lawful use,” even mass surveillance of Americans and fully autonomous lethal weapons.  Pentagon CTO Emil Michael is pushing for Anthropic to be designated a “supply chain risk” if it doesn’t comply, a label usually only given to national security threats. Anthropic’s rivals OpenAI and xAI have reportedly agreed to the new terms, but even after a White House meeting with Defense Secretary Pete Hegseth, Anthropic CEO Dario Amodei is still refusing to cross his company’s red line, stating that “threats do not change our position: we cannot in good conscience accede to their request.” Follow along here for the latest updates on the clash between AI companies and the Pentagon… We don’t have to have unsupervised killer robots Anthropic refuses Pentagon’s new terms, standing firm on lethal autonomous weapons and mass surveillance Pete Hegseth’s Pentagon AI bro squad includes a former Uber executive and a private equity billionaire Inside Anthropic’s existential negotiations with the Pentagon

The Hillabout 3 hours ago
Altman says OpenAI agrees with Anthropic’s red lines in Pentagon dispute

OpenAI CEO Sam Altman said Friday that he agrees with Anthropic’s red lines in its increasingly contentious negotiations with the Pentagon over the terms of use for the company’s AI models. As the feud between Anthropic and the Department of Defense (DOD) has reached a boiling point, the AI firm has refused to budge on...

TechCrunchabout 4 hours ago
Employees at Google and OpenAI support Anthropic’s Pentagon stand in open letter

While Anthropic has an existing partnership with the Pentagon, the AI company has remained firm that its technology not be used for mass domestic surveillance or fully autonomous weaponry.

The Vergeabout 4 hours ago
We don’t have to have unsupervised killer robots

It's the day of the Pentagon's looming ultimatum for Anthropic: allow the US military unchecked access to its technology, including for mass surveillance and fully autonomous lethal weapons, or potentially be designated a "supply chain risk" and potentially lose hundreds of billions of dollars in contracts. Amid the intensifying public statements and threats, tech workers across the industry are looking at their own companies' government and military contracts wondering what kind of future they're helping to build. While the Department of Defense has spent weeks negotiating with Anthropic over removing its guardrails, including allowing t … Read the full story at The Verge.

Financial Timesabout 6 hours ago
Big Tech workers press bosses to back Anthropic in Pentagon clash

Amazon, Google and Microsoft staff urge their executives to adopt tough AI guardrails and refuse any defence contracts

France 24about 6 hours ago
Anthropic refuses to bend to Pentagon on AI safeguards

A public showdown between the Trump administration and Anthropic is hitting an impasse as military officials demand the artificial intelligence company bend its ethical policies by Friday or risk damaging its business. Anthropic CEO Dario Amodei drew a sharp red line 24 hours before the deadline, declaring his company “cannot in good conscience accede” to the Pentagon’s final demand to allow unrestricted use of its technology.