Addressing concerns about users under the age of 13 on TikTok, the spokesperson also described ways that TikTok has already taken action to prevent underage users from accessing the platform: “While most people understand the importance of being truthful about their age, some do not provide the correct information, which is a challenge many online services face. That’s why our commitment to enforcing our minimum age requirements does not end at the age gate, and we take a number of additional approaches to identify and remove suspected underage account holders. We have implemented methods and processes to detect and remove underage users, which have led to approximately 20 million accounts globally being removed in Q1 2022.” TikTok also expects that “Content Levels will provide an added level of protection for those community members who truthfully indicate that they are under 18.”]
TikTok’s safety features recently became the focus of a lawsuit filed by parents who claim that the app’s addictive design is responsible for the deaths of at least seven children, six of whom were too young to be on TikTok. Those parents suggested that TikTok take steps to protect young users, urging the platform to add an age verification process to restrict content or terminate the accounts of child users under the age of 13—the minimum age required to join TikTok.
That’s not the direction TikTok has decided to go, though. At least, not yet. Instead, TikTok announced on Wednesday that it is adding new safety measures for all users designed to limit exposure to harmful content and give users more control over what shows up in their feeds. That includes giving users the power to block content containing certain words, hashtags, or sounds.
Specifically focusing on improving safety measures for TikTok’s “teenage community members,” TikTok is also “working to build a new system to organize content based on thematic maturity”—essentially, creating maturity ratings for TikTok videos, like the ratings you see on movies or video games.
“In the coming weeks, we’ll begin to introduce an early version to help prevent content with overtly mature themes from reaching audiences between ages 13-17,” TikTok’s Head of Trust and Safety Cormac Keenan wrote in the blog post.
Additionally, TikTok provided an update on the previously announced steps it is taking with its algorithm to protect users from endlessly scrolling through “potentially challenging or triggering viewing experiences.”
“Last year, we began testing ways to avoid recommending a series of similar content on topics that may be fine as a single video but potentially problematic if viewed repeatedly, such as topics related to dieting, extreme fitness, sadness, and other well-being topics,” Keenan wrote. “We’ve also been testing ways to recognize if our system may inadvertently be recommending a narrower range of content to a viewer.”
Likely because TikTok community standards limit accounts to ages 13 and up, TikTok so far has only discussed the desired impact of new safety features on adult and teen users. That means TikTok still has yet to take action to address concerns that parents have about “hundreds of thousands of children as young as 6 years old,” whom the lawsuit alleges that TikTok knows “are currently using its social media product” without safety features designed just for them.
TikTok did not immediately respond to a request for comment on the new safety features or on how maturity ratings will work. It’s unclear if there are any future safety features—like age verification—planned to address growing concerns about children under 13 reportedly being harmed while using the app.
#Mourning #parents #asked #TikTok #age #verification #maturity #ratings #Updated