- Today
- Holidays
- Birthdays
- Reminders
- Cities
- Atlanta
- Austin
- Baltimore
- Berwyn
- Beverly Hills
- Birmingham
- Boston
- Brooklyn
- Buffalo
- Charlotte
- Chicago
- Cincinnati
- Cleveland
- Columbus
- Dallas
- Denver
- Detroit
- Fort Worth
- Houston
- Indianapolis
- Knoxville
- Las Vegas
- Los Angeles
- Louisville
- Madison
- Memphis
- Miami
- Milwaukee
- Minneapolis
- Nashville
- New Orleans
- New York
- Omaha
- Orlando
- Philadelphia
- Phoenix
- Pittsburgh
- Portland
- Raleigh
- Richmond
- Rutherford
- Sacramento
- Salt Lake City
- San Antonio
- San Diego
- San Francisco
- San Jose
- Seattle
- Tampa
- Tucson
- Washington
Landmark Rulings Put Meta and YouTube Under Scrutiny Over Social Media Harms
Courts find tech giants liable for failing to protect users, especially children, from mental health and addiction risks.
Mar. 31, 2026 at 1:39am
Got story updates? Submit your updates here. ›
As social media companies face mounting legal liability for mental health harms, the intricate digital infrastructure powering these platforms is revealed in all its luminous complexity.Today in DaytonIn landmark cases, juries have ordered Meta to pay $375 million in damages and found Meta and YouTube liable for negligence over social media addiction and mental health harms, especially to young users. Experts say these "bellwether" cases set the tone for thousands of similar lawsuits and will likely lead to new state-level regulations on social media access and use.
Why it matters
These rulings mark a significant shift in how the legal system views the responsibility of social media companies to protect their users, especially vulnerable populations like children, from the documented harms of excessive social media use. The decisions could pave the way for more lawsuits, stricter regulations, and fundamental changes in how these platforms operate.
The details
In the New Mexico case, a jury found that Meta failed to provide adequate warnings to users about potential risks and did not implement sufficient safeguards to protect children from sexual predators. In California, a court determined that Meta and YouTube were aware of the mental health and addiction risks posed to young users but did not do enough to mitigate those harms. These "bellwether" cases were selected from thousands of similar lawsuits as examples that will guide the handling of the others.
- On March 24, 2026, a New Mexico jury ordered Meta to pay $375 million in damages.
- Just days later, on March 29, 2026, a California court found Meta and YouTube liable for negligence over social media addiction and mental health harms.
The players
Meta
The parent company of Facebook, Instagram, and other social media platforms.
YouTube
The popular video-sharing platform owned by Google's parent company, Alphabet.
John Hart, Esq.
Senior Counsel and Government Liaison at Cedarville University, who provided expert analysis on the legal implications of the cases.
Dr. Art Jipson
Associate Professor of Sociology and Criminology at the University of Dayton, who discussed why minors are more susceptible to the negative impacts of social media.
What they’re saying
“What did they know? When did they know it? And what did they do about it in terms of the risks,”
— John Hart, Esq., Senior Counsel and Government Liaison, Cedarville University
“Younger people are very focused on their social status, how they are perceived by others. The reluctance to accept other people's opinions that we have as adults hasn't been formed yet. In adolescence, the resilience that we have as adults Young people are still building.”
— Dr. Art Jipson, Associate Professor of Sociology and Criminology, University of Dayton
What’s next
Appeals have been filed by Meta and YouTube, but experts say these cases set the tone for how similar lawsuits will be handled moving forward. State legislators are also considering new measures to regulate social media access and use, including age verification, parental consent, and other guardrails.
The takeaway
These landmark rulings mark a significant shift in how the legal system views the responsibility of social media companies to protect their users, especially vulnerable populations like children, from the documented harms of excessive social media use. The decisions could pave the way for more lawsuits, stricter regulations, and fundamental changes in how these platforms operate.


