ChatGPT Now Cites Elon Musk's Biased Encyclopedia
OpenAI's ChatGPT has begun citing information from Grokipedia, a conservative-leaning AI-generated encyclopedia created by Elon Musk's xAI, raising concerns about information bias in AI systems.
What happens when the world's most popular AI assistant starts learning from deliberately biased sources? ChatGPT has begun citing Grokipedia, the conservative-leaning encyclopedia created by Elon Musk's xAI, marking a concerning shift in how AI systems source their information.
The Grokipedia Problem
Launched in October 2024, Grokipedia emerged after Musk complained that Wikipedia was biased against conservatives. But the cure proved worse than the alleged disease. The AI-generated encyclopedia has claimed that pornography contributed to the AIDS crisis, offered "ideological justifications" for slavery, and used denigrating terms for transgender people.
These weren't isolated incidents. Grokipedia is associated with a chatbot that once described itself as "Mecha Hitler" and was used to flood X with sexualized deepfakes. Yet this problematic source is now escaping its original ecosystem.
The Guardian's investigation revealed that GPT-5.2 cited Grokipedia nine times across more than a dozen queries. Notably, ChatGPT avoided citing it for widely debunked topics like the January 6 insurrection or HIV/AIDS epidemic, instead referencing it for more obscure subjects—including claims about Sir Richard Evans that The Guardian had previously fact-checked as false.
The Containment Breach
This isn't just an OpenAI problem. Anthropic's Claude also appears to be citing Grokipedia for some queries, suggesting a broader industry issue with data sourcing and verification.
An OpenAI spokesperson told The Guardian that the company "aims to draw from a broad range of publicly available sources and viewpoints." But this response sidesteps a crucial question: When does "diverse viewpoints" become a euphemism for platforming misinformation?
The Bigger Picture
This development highlights a fundamental challenge in AI development. As these systems become more sophisticated, they're also becoming more opaque about their sources. Users rarely know where ChatGPT's confident-sounding answers originate, making it nearly impossible to assess their reliability.
The timing is particularly concerning. As AI systems become primary information sources for millions of users, the quality and bias of their training data becomes a matter of public interest. If biased sources can infiltrate mainstream AI responses, what does this mean for informed democratic discourse?
The Trust Equation
For consumers, this raises uncomfortable questions about AI reliability. If ChatGPT can cite a source known for historical revisionism and inflammatory content, how can users distinguish between well-sourced answers and those drawing from questionable materials?
The issue extends beyond individual queries. As AI systems influence everything from educational content to business decisions, the propagation of biased information could have far-reaching consequences across society.
This content is AI-generated based on source articles. While we strive for accuracy, errors may occur. We recommend verifying with the original source.
Share your thoughts on this article
Sign in to join the conversation
Related Articles
Over 230 million people use ChatGPT for health advice weekly. Explore the growing privacy risks and the legal gap between tech giants and medical providers.
Sam Altman returns to India in Feb 2026 for the India AI Impact Summit. Exploring OpenAI's strategy for the world's largest user market and infrastructure goals.
In 2025, the global app market saw AI revenue skyrocket, leading non-game app spending to surpass games for the first time ever. Discover the key trends from Sensor Tower.
OpenAI begins testing ads in ChatGPT to boost revenue. Explore the implications of the OpenAI ChatGPT ads revenue strategy for the future of AI and search.
Thoughts