A Los Angeles jury delivered a landmark verdict on March 25, 2026, finding Meta and Google liable for designing platforms harmful to children and teenagers. The jury ruled the companies were negligent in their platform design and operation, ordering Meta to pay $4.2 million and Google $1.8 million in damages to the plaintiff, a 20-year-old woman identified as Kaley.
The case centered on allegations that Instagram and YouTube were deliberately engineered with addictive features like infinite scrolling and algorithm-driven recommendations to maximize user engagement at the expense of user wellbeing. Kaley testified that her addiction to these platforms began before she was a teenager, dominated her life for years, and contributed to depression, anxiety, and suicidal thoughts. The jury found both companies failed to warn users about the dangers of their platforms.
This verdict is significant because it focused on platform design rather than user-generated content, making it harder for the companies to avoid liability under Section 230 of the Communications Decency Act, which typically shields online platforms from liability for content posted by users. Legal experts note this distinction could redefine liability across the technology sector.
Both Meta and Google plan to appeal the decision. Meta stated it disagrees with the verdict and is reviewing legal options, while Google argued the ruling mischaracterized YouTube as a social media platform rather than a responsibly built streaming service. Despite the verdict, Meta's shares closed up 0.3% and Alphabet's (Google's parent) finished 0.2% higher on the day of the ruling.
The trial revealed internal company documents showing how Meta and Google worked to attract younger users. During testimony, Meta CEO Mark Zuckerberg was questioned about a decision to lift a temporary ban on beauty filters that some Meta employees had warned could harm teen girls. Zuckerberg stated he allowed it so users could express themselves.
This case is part of a broader wave of litigation against major technology companies. Snap and TikTok were originally named as defendants but settled before trial began, with terms undisclosed. More than 2,000 plaintiffs have filed similar lawsuits accusing companies of designing addictive platforms that expose children to risks.
In a separate but related development, a New Mexico jury ruled against Meta just one day before the Los Angeles verdict, finding the company violated state law in a case brought by the state's attorney general over child safety on Facebook, Instagram, and WhatsApp.
Legal analysts are drawing parallels with historic litigation against tobacco companies, suggesting the social media industry could face similar outcomes including stricter regulations and mandated design changes. The ruling comes as at least 20 US states passed laws last year related to children's social media use, though Congress has not passed any federal legislation on the issue.