TikTok

U.S. attorneys general want TikTok and Snapchat to give parents more control of kids’ DMs

The National Association of Attorneys General (NAAG) has sent a letter to TikTok and Snapchat urging them to beef up parental controls by allowing third-party monitoring apps to check what kids are up to not just in public portions of platforms, but in direct messages to other users.

Attorneys general from 44 U.S. states signed the letter, which pitches the idea that social media platforms’ community guidelines and the content moderators they hire to enforce those rules “are not always sufficient to protect children and teenagers who are particularly vulnerable to online threats, especially with regard to direct messaging.”

The letter cited a study done by parental control app Bark, which examined 3.4 billion direct messages sent across 30 social media apps in 2021.

Subscribe to get the latest creator news

Subscribe

Bark says messages revealed that 74.61% of teenagers expressed self-harm or suicidal ideation, that 90.73% of teenagers had encountered nudity and/or sexual content on social media, that 93.31% of teens “engaged in conversations surrounding drugs/alcohol,” that 94.5% of teens “expressed or experienced violent subject matter/thoughts,” and that 85% of teens “experienced bullying as a bully, victim, or witness.”

NAAG argues these findings indicate the negative impact of social media on teens (not a novel revelation) as well as the potential for them to be exposed to sexual predators and cyberbullying.

Lawmakers hammering social media platforms for failing to adequately protect young users is far from new. But this letter takes a different tack than previous warnings and investigations in two ways: one, it’s specifically pointing to direct messages; and two, it’s not necessarily asking platforms to improve their own in-house systems.

Instead, this letter urges platforms to partner with—and thus give access to—third-party parental control apps.

“Parental control apps can alert parents or schools to messages and posts on your platforms that have the potential to be harmful and dangerous,” the letter says. “On other platforms where these apps are allowed to operate appropriately parents have received notifications of millions of instances of severe bullying and hundreds of thousands of self-harm situations, showing that these apps have the potential to save lives and prevent harm to our youth.”

NAAG did not endorse a particular third-party app. It also did not publicize details about which platforms have embraced parental control apps, and does not appear to have discussed the potential child privacy and data security issues that could come with allowing third-party apps to access kids’ private communications.

Share
Published by
James Hale
Tags: snaptiktok

Recent Posts

YouTube just made a Shorts deepfake machine so creators don’t have to be in their own videos

Hey YouTubers! Do you want to be rid of the pesky chore of actually appearing…

1 day ago

Have you heard? Gaming Historian says so long, Ms. Rachel sells shoes, and TikTok ad exec moves on.

Each week, we handpick a selection of stories to give you a snapshot of trends,…

1 day ago

NAB Show wants to be the meeting ground for creators and legacy entertainment: “These two segments have so much to offer each other right now”

Back in 2024, the National Association of Broadcasters recognized the importance of content creators by…

1 day ago

Hoorae returns to Issa Rae’s web series roots with “Screen Time” microdrama

Too much screen time can be a dangerous thing, and Hoorae is taking that idea literally. The…

1 day ago

Kylie Jenner brings “star power and aura” to hydration product k2o, launched in tandem with Night

The latest product backed by Night's venture studio emerged out of a partnership between the creator…

1 day ago

Hollywood has a lot to learn from creator animators (and their IPs), YouTube says in latest Culture & Trends report

Indie animation is flourishing on YouTube. From the pop culture juggernaut that is The Amazing…

2 days ago