Liabooks Home|PRISM News
When America Adopts China's Playbook in AI Race
CultureAI Analysis

When America Adopts China's Playbook in AI Race

5 min readSource

The Pentagon blacklisted Anthropic for refusing mass surveillance uses, embracing OpenAI instead. Is America becoming the authoritarian bogeyman it warns against?

Anthropic drew two red lines with the Pentagon: no mass domestic surveillance, no fully autonomous weapons. These seemed like reasonable boundaries for an AI company concerned about ethics. The Department of Defense disagreed. Instead of accepting these terms, it blacklisted Anthropic as a "supply chain risk" - a designation previously reserved for foreign adversaries like China's Huawei. Hours later, OpenAI announced a new Pentagon deal.

This isn't just corporate competition. It's a fundamental shift in how America approaches the AI arms race - one that mirrors the very authoritarianism it claims to oppose.

The Military-Civil Fusion Playbook

China's "military-civil fusion" policy forces private tech companies to make their innovations available to the military, whether they want to or not. Sound familiar? Jeffrey Ding, who studies China's AI ecosystem at George Washington University, sees the parallel clearly: "The Pentagon's threats against Anthropic copy the worst aspects of China's military-civil fusion strategy."

The irony is stark. For years, American AI companies have justified breakneck development by invoking the China threat. Sam Altman, Sundar Pichai, and Mark Zuckerberg have all warned that Chinese authoritarianism could dominate if America doesn't move fast enough. But what happens when America adopts authoritarian tactics to win?

Antropic's red lines weren't radical demands - they were basic human rights protections. Yet the Pentagon treated them as threats to national security, using extraordinary authorities typically reserved for foreign enemies against an American company.

The "Lawful Purposes" Loophole

OpenAI's contract with the Pentagon reveals the problem with America's approach. While Altman claims to share Anthropic's principles, his company agreed to allow AI use for "all lawful purposes." That sounds reasonable until you realize the law hasn't caught up to AI capabilities.

Currently, it's perfectly legal for the government to buy data collected by private companies - location information, web browsing history, credit card transactions. Before advanced AI, this data was too overwhelming to analyze effectively. Now, AI can create predictive portraits of every citizen's life. Most Americans would call this mass surveillance, but it's technically legal.

This was reportedly the main sticking point between Anthropic and the Pentagon. The company refused to enable this kind of analysis on Americans, while OpenAI's "lawful purposes" clause opens the door.

Empty Safeguards and Vague Promises

OpenAI insists its contract includes protections, pointing to clauses that mention "appropriate safeguards" and "high-stake decisions." But experts aren't buying it. As one University of Minnesota law professor noted, these provisions don't guarantee fundamental rights protection at all.

"The existing guardrails for generative AI are deeply lacking," says Heidy Khlaaf, chief AI scientist at the AI Now Institute. "It's highly doubtful that if they cannot guard their systems against benign cases, they'd be able to do so for complex military and surveillance operations."

Even OpenAI's own research scientist, Leo Gao, expressed skepticism publicly about the company's assurances. This internal dissent speaks volumes about the contract's actual protections.

The timing raises additional questions. OpenAI's leadership donated millions to support Donald Trump, while Anthropic's Dario Amodei refused to bankroll him, earning Trump's ire as a "RADICAL LEFT, WOKE COMPANY." How much did politics influence the Pentagon's decision?

The Boycott Response

Public reaction has been swift and decisive. The QuitGPT campaign has attracted over 1.5 million participants, while Anthropic's Claude became the top downloaded app over the weekend. Historian Rutger Bregman, who studies boycott movements, sees this as significant: "This is the first opportunity to start a massive consumer boycott in the AI era."

More importantly, tech workers are organizing across company lines. Over 900 employees from OpenAI and Google signed a letter titled "We Will Not Be Divided," urging their leadership to stand together against Pentagon pressure for mass surveillance and autonomous weapons.

Another letter, signed by 175 industry figures including OpenAI employees, calls for withdrawing the supply chain designation against Anthropic and suggests Congress examine whether the Pentagon's actions constitute an abuse of power.

The Solidarity Strategy

International cooperation on AI safety has largely stalled, and multilateral treaties remain elusive despite support from Nobel laureates. In this vacuum, worker solidarity across tech companies offers a different path forward.

The cross-company coordination represents something new: tech workers recognizing that their individual companies can be divided and conquered, but together they might resist government pressure that compromises fundamental principles.

This isn't just about Anthropic or OpenAI. It's about establishing precedent. If the Pentagon can successfully pressure one AI company to abandon ethical guardrails, it can pressure them all.

This content is AI-generated based on source articles. While we strive for accuracy, errors may occur. We recommend verifying with the original source.

Thoughts

Related Articles