Meta Faces Trial Over Alleged Child Safety Failures
Meta stands trial in New Mexico over allegations it failed to protect children from online predators on Facebook and Instagram. Multiple lawsuits could reshape social media industry accountability.
Create a fake profile of a 13-year-old girl, and watch what happens. Within hours, the account was "simply inundated with images and targeted solicitations," according to New Mexico Attorney General Raúl Torrez. That undercover operation now forms the backbone of a high-stakes trial against Meta beginning Monday.
Standing Alone in the Dock
Unlike previous tech lawsuits where companies stood together, Meta faces this battle solo. The 2023 lawsuit alleges the company "steered and connected users — including children — to sexually explicit, exploitative and child sex abuse materials and facilitated human trafficking" within New Mexico.
The timing isn't coincidental. In Los Angeles, another major trial is underway where Meta joins YouTube, TikTok, and Snap as defendants. The allegation: these companies knew their app designs harmed young users' mental health but failed to warn the public. TikTok and Snap settled before trial, leaving Meta to fight on multiple fronts.
Eighteen jurors were impaneled Friday, setting the stage for opening arguments Monday. Instagram head Adam Mosseri testifies Wednesday, followed by CEO Mark Zuckerberg a week later.
The 'Big Tobacco' Moment for Big Tech
Legal experts draw parallels to the 1990s tobacco litigation — products that allegedly harm users while companies mislead the public about risks. The comparison isn't lost on anyone watching these proceedings unfold.
Meta denies the allegations, stating it's "focused on demonstrating our longstanding commitment to supporting young people." Social media companies have long argued that Section 230 of the Communications Decency Act protects content shared on their platforms.
But here's the key shift: these lawsuits target app design and features, not just content. The claim is that tech companies deliberately built "defective apps" that cause addictive and unhealthy behaviors in teens and children.
A federal case in California's Northern District, involving Meta, TikTok, YouTube, and Snap, is slated to begin later this year with similar allegations.
Beyond the Courtroom
The financial stakes are enormous. Meta's market cap of over $1.5 trillion means even a significant settlement wouldn't threaten the company's existence. But the precedent could reshape how social media platforms operate.
Parents, educators, and policymakers worldwide are watching. If successful, these lawsuits could force platforms to prioritize user safety over engagement metrics — a fundamental shift in how social media operates.
The cases also highlight a broader question about corporate responsibility in the digital age. When algorithms decide what users see and features are designed to maximize time spent on platforms, can companies claim neutrality?
This content is AI-generated based on source articles. While we strive for accuracy, errors may occur. We recommend verifying with the original source.
Related Articles
A New Mexico jury found Meta willfully violated consumer protection law by exposing children to predators on Facebook and Instagram, ordering $375 million in damages. What does this mean for Big Tech accountability?
Meta has increased its El Paso AI data center investment more than sixfold, from $1.5B to $10B, targeting 1GW capacity by 2028. What this means for investors, competitors, and the AI infrastructure race.
Meta's second round of layoffs in 2026 hits Facebook, Reality Labs, recruiting, and sales. While slashing hundreds of jobs, the company is doubling down on AI talent and locking in top execs with aggressive stock options.
Meta is shutting down Horizon Worlds on Quest VR on June 15. After billions in losses and 1,000+ layoffs, what does the metaverse's quiet collapse mean for the next big tech bet: AI?
Thoughts
Share your thoughts on this article
Sign in to join the conversation