CASSIS, France (AP) — In the moment when her world shattered three years ago, Stephanie Mistre found her 15-year-old daughter, Marie, lifeless in the bedroom where she died by suicide.
“I went from light to darkness in a fraction of a second,” Mistre said, describing the day in September 2021 that marked the start of her fight against TikTok, the Chinese-owned video app she blames for pushing her daughter toward despair.
___
EDITOR’S NOTE — This story includes discussion of suicide. If you or someone you know needs help, the national suicide and crisis lifeline in the U.S. is available by calling or texting 988. There is also an online chat at 988lifeline.org. Helplines outside the U.S. can be found at www.iasp.info/suicidalthoughts.
___
Delving into her daughter’s phone after her death, Mistre discovered videos promoting suicide methods, tutorials and comments encouraging users to go beyond “mere suicide attempts.” She said TikTok’s algorithm had repeatedly pushed such content to her daughter.
“It was brainwashing,” said Mistre, who lives in Cassis, near Marseille, in the south of France. “They normalized depression and self-harm, turning it into a twisted sense of belonging.”
Now Mistre and six other families are suing TikTok France, accusing the platform of failing to moderate harmful content and exposing children to life-threatening material. Out of the seven families, two experienced the loss of a child.
Asked about the lawsuit, TikTok said its guidelines forbid any promotion of suicide and that it employs 40,000 trust and safety professionals worldwide — hundreds of which are French-speaking moderators — to remove dangerous posts. The company also said it refers users who search for suicide-related videos to mental health services.
Before killing herself, Marie Le Tiec made several videos to explain her decision, citing various difficulties in her life, and quoted a song by the Louisiana-based emo rap group Suicideboys, who are popular on TikTok.
Her mother also claims that her daughter was repeatedly bullied and harassed at school and online. In addition to the lawsuit, the 51-year-old mother and her husband have filed a complaint against five of Marie’s classmates and her previous high school.
Above all, Mistre blames TikTok, saying that putting the app “in the hands of an empathetic and sensitive teenager who does not know what is real from what is not is like a ticking bomb.”
Scientists have not established a clear link between social media and mental health problems or psychological harm, said Grégoire Borst, a professor of psychology and cognitive neuroscience at Paris-Cité University.
“It’s very difficult to show clear cause and effect in this area,” Borst said, citing a leading peer-reviewed study that found only 0.4% of the differences in teenagers’ well-being could be attributed to social media use.
Additionally, Borst pointed out that no current studies suggest TikTok is any more harmful than rival apps such as Snapchat, X, Facebook or Instagram.
While most teens use social media without significant harm, the real risks, Borst said, lie with those already facing challenges such as bullying or family instability.
“When teenagers already feel bad about themselves and spend time exposed to distorted images or harmful social comparisons,” it can worsen their mental state, Borst said.
Lawyer Laure Boutron-Marmion, who represents the seven families suing TikTok, said their case is based on “extensive evidence.” The company “can no longer hide behind the claim that it’s not their responsibility because they don’t create the content,” Boutron-Marmion said.
The lawsuit alleges that TikTok’s algorithm is designed to trap vulnerable users in cycles of despair for profit and seeks reparations for the families.
“Their strategy is insidious,” Mistre said. “They hook children into depressive content to keep them on the platform, turning them into lucrative re-engagement products.”
Boutron-Marmion noted that TikTok’s Chinese version, Douyin, features much stricter content controls for young users. It includes a “youth mode” mandatory for users under 14 that restricts screen time to 40 minutes a day and offers only approved content.
“It proves they can moderate content when they choose to,” Boutron-Marmion said. “The absence of these safeguards here is telling.”
A report titled “Children and Screens,” commissioned by French President Emmanuel Macron in April and to which Borst contributed, concluded that certain algorithmic features should be considered addictive and banned from any app in France. The report also called for restricting social media access for minors under 15 in France. Neither measure has been adopted.
TikTok, which faced being shut down in the U.S. until President Donald Trump suspended a ban on it, has also come under scrutiny globally.
The U.S. has seen similar legal efforts by parents. One lawsuit in Los Angeles County accuses Meta and its platforms Instagram and Facebook, as well as Snapchat and TikTok, of designing defective products that cause serious injuries. The lawsuit lists three teens who died by suicide. In another complaint, two tribal nations accuse major social media companies, including YouTube owner Alphabet, of contributing to high rates of suicide among Native youths.
Meta CEO Mark Zuckerberg apologized to parents who had lost children while testifying last year in the U.S. Senate.
In December, Australia enacted a groundbreaking law banning social media accounts for children under 16.
In France, Boutron-Marmion expects TikTok Limited Technologies, the European Union subsidiary for ByteDance — the Chinese company that owns TikTok — to answer the allegations in the first quarter of 2025. Authorities will later decide whether and when a trial would take place.
When contacted by The Associated Press, TikTok said it had not been notified about the French lawsuit, which was filed in November. It could take months for the French justice system to process the complaint and for authorities in Ireland — home to TikTok’s European headquarters — to formally notify the company, Boutron-Marmion said.
Instead, a Tiktok spokesperson highlighted company guidelines that prohibit content promoting suicide or self-harm.
Critics argue that TikTok’s claims of robust moderation fall short.
Imran Ahmed, the CEO of the Center for Countering Digital Hate, dismissed TikTok’s assertion that over 98.8% of harmful videos had been flagged and removed between April and June.
When asked about the blind spots of their moderation efforts, social media platforms claim that users are able to bypass detection by using ambiguous language or allusions that algorithms struggle to flag, Ahmed said.
The term “algospeak” has been coined to describe techniques such as using zebra or armadillo emojis to talk about cutting yourself, or the Swiss flag emoji as an allusion to suicide.
Such code words “aren’t particularly sophisticated,” Ahmed said. “The only reason TikTok can’t find them when independent researchers, journalists and others can is because they’re not looking hard enough,” Ahmed said.
Ahmed’s organization conducted a study in 2022 simulating the experience of a 13-year-old girl on TikTok.
“Within 2.5 minutes, the accounts were served self-harm content,” Ahmed said. “By eight minutes, they saw eating disorder content. On average, every 39 seconds, the algorithm pushed harmful material.”
The algorithm “knows that eating disorder and self-harm content is especially addictive” for young girls.
For Mistre, the fight is deeply personal. Sitting in her daughter’s room, where she has kept the decor untouched for the last three years, she said parents must know about the dangers of social media.
Had she known about the content being sent to her daughter, she never would have allowed her on TikTok, she said. Her voice breaks as she describes Marie as a “sunny, funny” teenager who dreamed of becoming a lawyer.
“In memory of Marie, I will fight as long as I have the strength,” she said. “Parents need to know the truth. We must confront these platforms and demand accountability.”
___
Associated Press writers Haleluya Hadero and Zen Soo contributed to this story.