When an AI Company Said No to $200 Million
Anthropic rejected a Pentagon contract over surveillance concerns while OpenAI rushed to sign. The market's reaction reveals what users really value in AI companies.
$200 million. That's what Anthropic walked away from when it refused to let the U.S. government use its AI to surveil American citizens or operate autonomous weapons without human oversight.
The backlash was swift and brutal. President Trump branded the company "leftwing nut jobs" on Truth Social. Defense Secretary Pete Hegseth designated Anthropic as a "Supply-Chain Risk to National Security"—a label typically reserved for hostile foreign actors like Huawei. Every federal agency received orders to immediately stop using Anthropic's products.
Meanwhile, rival OpenAI quickly swooped in to sign its own Pentagon deal.
The Unexpected Payoff of Principles
Anthropic's principled stand looked like corporate suicide. Early indicators suggest it might be genius instead.
The company's confrontation with the Pentagon proved more effective than a Super Bowl ad campaign. After this year's big game advertising push, Anthropic's AI model Claude cracked the top 10 most-downloaded free apps in America. The day after Hegseth announced the government was severing ties, Claude shot to No. 1—a position it still holds.
The numbers tell the story: daily downloads topped 1 million, and Anthropic reports it has "broken its own sign-up record every day since early last week, across every country where Claude is available."
OpenAI faced the mirror opposite. When details of its Pentagon deal emerged on February 28, ChatGPT app uninstalls spiked 295 percent. One-star reviews surged nearly 800 percent while five-star reviews dropped by half.
Silicon Valley's Solidarity Test
Perhaps more consequential than user metrics is the response from within the AI industry itself. Support letters for Anthropic are circulating among competitors' employees, with one gathering 850 signatures by Monday. Many are demanding their own companies show solidarity with Anthropic and honor the same ethical red lines. Some have reportedly threatened to quit if those demands aren't met.
The admiration extends beyond Silicon Valley. Former Republican Representative Denver Riggleman, who now leads a cybersecurity firm, was evaluating AI partners when Anthropic's stand narrowed his options to one. "Anthropic had its nonnegotiables," he told reporters, "and we have ours."
Riggleman believes Hegseth's "supply-chain risk" designation will likely be overturned in court. The U.S. government has never applied this label to an American company, and this appears to be the first time it's been used in retaliation for declining contract terms. "To say it rests on shaky legal ground," Riggleman said, "would be generous."
The Fine Print That Divided Silicon Valley
The dispute centers on contract language that reveals fundamentally different approaches to AI safety and government oversight.
compare-table
| Aspect | Anthropic's Position | OpenAI's Position |
|---|---|---|
| Core Concern | "All lawful purposes" clause too vague | Mass surveillance and autonomous weapons explicitly prohibited |
| Legal Risk | Executive orders could shift legal boundaries | Current contract prohibitions are sufficient |
| CEO Statement | "We don't want to sell something that could get our own people killed" | Deal was "definitely rushed" with poor "optics" |
| Response | Rejected entire contract | Added surveillance restrictions Monday |
compare-table
The Pentagon argued its contract contained adequate safeguards through the "all lawful purposes" limitation. Anthropic countered that this wasn't sufficient—that new executive orders or reinterpretations of statute could shift existing legal boundaries.
OpenAI maintains its subsequent Pentagon deal is safer, with explicit prohibitions on mass surveillance and autonomous weapons. But it retains the "all lawful purposes" language, making those prohibitions dependent on the Pentagon's willingness to respect legal norms. Even CEO Sam Altman conceded the deal was "definitely rushed" and acknowledged "the optics don't look good."
Trust as Currency
The standoff evokes memories from a former Navy pilot who recalls signing documents pledging never to surveil American citizens nearly three decades ago. During post-9/11 reconnaissance missions, he took for granted that somewhere in the decision chain, a human being was in the loop—that a person, not a machine, would bear the weight of life-and-death choices.
This wasn't bureaucratic procedure. It was a hard line that those in uniform were expected to hold.
In the confrontation between Anthropic and the Pentagon, a private company was forced to hold that line against its own government. The market's response suggests users noticed.
The Regulatory Reckoning
Riggleman's perspective is particularly striking given his background on a congressional AI task force focused on foreign adversaries. He once trusted his country to regulate technologies that could reshape Americans' lives. "These days," he observes, "the government is no longer creating those safeguards—it's destroying them."
The implications extend beyond any single contract. As Riggleman notes: "I don't think we appreciate yet, as a society, what it means to have private firms controlling this amount of information about citizens."
This content is AI-generated based on source articles. While we strive for accuracy, errors may occur. We recommend verifying with the original source.
Related Articles
OpenAI's contract with the Pentagon contains numerous loopholes that could allow mass surveillance of Americans, despite public assurances. Legal experts warn the 'red lines' may be blurrier than they appear.
The Pentagon blacklisted Anthropic for refusing mass surveillance uses, embracing OpenAI instead. Is America becoming the authoritarian bogeyman it warns against?
Anthropic's contract termination with Pentagon reveals the complex tensions between AI ethics and national security. Where should companies draw the line?
The Pentagon has given Anthropic until Friday evening to allow unrestricted military use of its Claude AI, setting up the biggest AI ethics showdown since 2018. What's really at stake?
Thoughts
Share your thoughts on this article
Sign in to join the conversation