Liabooks Home|PRISM News
When an AI Company Said No to $200 Million
CultureAI Analysis

When an AI Company Said No to $200 Million

4 min readSource

Anthropic rejected a Pentagon contract over surveillance concerns while OpenAI rushed to sign. The market's reaction reveals what users really value in AI companies.

$200 million. That's what Anthropic walked away from when it refused to let the U.S. government use its AI to surveil American citizens or operate autonomous weapons without human oversight.

The backlash was swift and brutal. President Trump branded the company "leftwing nut jobs" on Truth Social. Defense Secretary Pete Hegseth designated Anthropic as a "Supply-Chain Risk to National Security"—a label typically reserved for hostile foreign actors like Huawei. Every federal agency received orders to immediately stop using Anthropic's products.

Meanwhile, rival OpenAI quickly swooped in to sign its own Pentagon deal.

The Unexpected Payoff of Principles

Anthropic's principled stand looked like corporate suicide. Early indicators suggest it might be genius instead.

The company's confrontation with the Pentagon proved more effective than a Super Bowl ad campaign. After this year's big game advertising push, Anthropic's AI model Claude cracked the top 10 most-downloaded free apps in America. The day after Hegseth announced the government was severing ties, Claude shot to No. 1—a position it still holds.

The numbers tell the story: daily downloads topped 1 million, and Anthropic reports it has "broken its own sign-up record every day since early last week, across every country where Claude is available."

OpenAI faced the mirror opposite. When details of its Pentagon deal emerged on February 28, ChatGPT app uninstalls spiked 295 percent. One-star reviews surged nearly 800 percent while five-star reviews dropped by half.

Silicon Valley's Solidarity Test

Perhaps more consequential than user metrics is the response from within the AI industry itself. Support letters for Anthropic are circulating among competitors' employees, with one gathering 850 signatures by Monday. Many are demanding their own companies show solidarity with Anthropic and honor the same ethical red lines. Some have reportedly threatened to quit if those demands aren't met.

The admiration extends beyond Silicon Valley. Former Republican Representative Denver Riggleman, who now leads a cybersecurity firm, was evaluating AI partners when Anthropic's stand narrowed his options to one. "Anthropic had its nonnegotiables," he told reporters, "and we have ours."

Riggleman believes Hegseth's "supply-chain risk" designation will likely be overturned in court. The U.S. government has never applied this label to an American company, and this appears to be the first time it's been used in retaliation for declining contract terms. "To say it rests on shaky legal ground," Riggleman said, "would be generous."

The Fine Print That Divided Silicon Valley

The dispute centers on contract language that reveals fundamentally different approaches to AI safety and government oversight.

compare-table

AspectAnthropic's PositionOpenAI's Position
Core Concern"All lawful purposes" clause too vagueMass surveillance and autonomous weapons explicitly prohibited
Legal RiskExecutive orders could shift legal boundariesCurrent contract prohibitions are sufficient
CEO Statement"We don't want to sell something that could get our own people killed"Deal was "definitely rushed" with poor "optics"
ResponseRejected entire contractAdded surveillance restrictions Monday

compare-table

The Pentagon argued its contract contained adequate safeguards through the "all lawful purposes" limitation. Anthropic countered that this wasn't sufficient—that new executive orders or reinterpretations of statute could shift existing legal boundaries.

OpenAI maintains its subsequent Pentagon deal is safer, with explicit prohibitions on mass surveillance and autonomous weapons. But it retains the "all lawful purposes" language, making those prohibitions dependent on the Pentagon's willingness to respect legal norms. Even CEO Sam Altman conceded the deal was "definitely rushed" and acknowledged "the optics don't look good."

Trust as Currency

The standoff evokes memories from a former Navy pilot who recalls signing documents pledging never to surveil American citizens nearly three decades ago. During post-9/11 reconnaissance missions, he took for granted that somewhere in the decision chain, a human being was in the loop—that a person, not a machine, would bear the weight of life-and-death choices.

This wasn't bureaucratic procedure. It was a hard line that those in uniform were expected to hold.

In the confrontation between Anthropic and the Pentagon, a private company was forced to hold that line against its own government. The market's response suggests users noticed.

The Regulatory Reckoning

Riggleman's perspective is particularly striking given his background on a congressional AI task force focused on foreign adversaries. He once trusted his country to regulate technologies that could reshape Americans' lives. "These days," he observes, "the government is no longer creating those safeguards—it's destroying them."

The implications extend beyond any single contract. As Riggleman notes: "I don't think we appreciate yet, as a society, what it means to have private firms controlling this amount of information about citizens."

This content is AI-generated based on source articles. While we strive for accuracy, errors may occur. We recommend verifying with the original source.

Thoughts

Related Articles