The Algorithm on Trial: Meta and YouTube Lose
A Los Angeles jury found Meta and YouTube liable for intentionally addicting a child to social media, awarding $6M in damages. What this landmark verdict means for Big Tech, parents, and regulators worldwide.
She was six years old when she first opened YouTube. Nine when she joined Instagram. By ten, she was experiencing anxiety and depression. The platform never once asked how old she was.
On March 26, 2026, a Los Angeles jury delivered what may be the most consequential verdict in the history of social media litigation. Meta and Google were found liable for intentionally building addictive platforms that harmed the mental health of a young woman known in court only as Kaley, now 20. The jury awarded her $6 million in damages—half of it punitive, because jurors concluded the companies had acted with "malice, oppression, or fraud."
What Actually Happened in That Courtroom
The five-week trial was, in many ways, a collision between two narratives. Meta and Google argued their platforms were designed responsibly, that teen mental health is complex, and that a single app cannot be blamed for a person's psychological struggles. Kaley's lawyers called the platforms "addiction machines" built to trap children for profit.
The evidence Kaley's team presented was damaging. Internal Meta documents suggested the company knew children under 13 were using its platforms despite official policies prohibiting it. Former Meta executives testified about growth strategies that explicitly targeted young users because they were more likely to stay on the platform longer. Mark Zuckerberg himself took the stand in February, defending the company's under-13 ban—but when confronted with internal research showing children were on the platforms anyway, he said he "always wished" for faster progress in identifying underage users.
Perhaps the moment that defined the trial came when Adam Mosseri, head of Instagram, was told that Kaley's longest single day on the platform stretched to 16 hours. He declined to call it addiction. He called it "problematic." The jury apparently saw it differently.
Meta was assigned 70% of the liability; Google the remaining 30%. Snap and TikTok had already settled with Kaley before trial for undisclosed amounts.
Why This Verdict Is Different
American courts have seen social media lawsuits before. What makes this one different is the word "intentional." Jurors didn't just find that these platforms caused harm—they found that harm was a foreseeable consequence of deliberate design choices. Infinite scroll. Algorithmic amplification. No meaningful age verification. These weren't bugs. Kaley's lawyers argued they were features.
The timing amplifies the significance. Just one day before the LA verdict, a New Mexico jury found Meta liable for exposing children to sexually explicit material and contact with predators. Back-to-back verdicts in separate jurisdictions, within 48 hours. Mike Proulx, research director at Forrester, called the twin rulings a "breaking point" between social media companies and the public.
Hundreds of similar cases are currently moving through US courts. A separate class action against Meta and other platforms is set to begin in California federal court in June. This verdict doesn't just affect Kaley—it potentially reshapes the legal landscape for all of them.
The Stakeholders React
For Meta and Google, the response was swift and predictable. Both companies announced they would appeal. Meta insisted that "teen mental health is profoundly complex and cannot be linked to a single app." Google took a different tack, arguing that YouTube is "a responsibly built streaming platform, not a social media site"—a framing that, if accepted by appeals courts, could significantly limit its future exposure.
For parents, the verdict felt like vindication. Outside the courthouse, families who had watched the trial for weeks embraced and wept. Amy Neville, whose child was not part of Kaley's case but is part of the broader wave of litigation, was among those celebrating. These parents have argued for years that their children's suffering was dismissed as parenting failure rather than corporate negligence.
For regulators and lawmakers, the verdict arrives at a moment of unusual political alignment. Australia has already banned social media for children under 16. The UK is running a pilot program exploring a similar restriction. In the US, bipartisan support for children's online safety legislation has been building—though converting that sentiment into law has proven difficult. A jury verdict, however, carries a different kind of weight than a congressional hearing.
For investors, the implications are worth watching carefully. Meta's stock dipped after the verdict. If hundreds of similar cases proceed to trial on the strength of this precedent, the financial exposure could grow substantially—and more importantly, it raises questions about the long-term viability of an advertising model that depends on maximizing engagement time, particularly among younger users.
The Arguments That Didn't Win
It's worth taking seriously the arguments Meta and Google made, not because they prevailed, but because they reflect genuine complexity. Mental health outcomes are shaped by genetics, family environment, peer relationships, socioeconomic stress, and yes, technology use. Isolating any single cause is scientifically difficult. Platforms also carry content that is educational, connective, and genuinely valuable to young people.
Meta has also pointed to real investments in teen safety features: default private accounts for users under 18, restrictions on who can message minors, and tools for parents to monitor usage. Whether these measures are sufficient—or whether they arrived too late, and were too easy to circumvent—is a question the courts will continue to examine.
There's also a legal question that appeals courts will scrutinize: Section 230 of the Communications Decency Act has historically shielded platforms from liability for user-generated content. Kaley's case focused on platform design rather than specific content, a distinction her lawyers deliberately crafted. Whether that distinction holds under appellate review will determine how far this verdict's precedent actually reaches.
This content is AI-generated based on source articles. While we strive for accuracy, errors may occur. We recommend verifying with the original source.
Related Articles
In landmark social media addiction trial, YouTube VP testified that platform's billion-hour daily viewing goal aimed to provide user value, not create harmful binge-watching habits.
BLACKPINK breaks YouTube records with 100 million subscribers, marking a new era for K-pop's global influence and redefining how we measure artistic success in the digital age.
BLACKPINK becomes first artist channel to reach 100 million YouTube subscribers. Beyond the numbers, what does this milestone reveal about the changing landscape of global entertainment?
Meta CEO Mark Zuckerberg testified that Instagram doesn't target children under 13, but internal documents revealed strategies to "bring teens in as tweens." A landmark trial examining social media's impact on youth mental health.
Thoughts
Share your thoughts on this article
Sign in to join the conversation