Categories: TikTok

TikTok Moderators Were Told To Suppress Videos Made By Marginalized Users Because They Might Be Bullied

Back in September, leaked content moderation guidelines from TikTok showed the shortform video app and its China-based parent company, Bytedance, instructing moderators to remove videos about topics the Chinese government doesn’t approve of, such as Tinanmen Square and the fight for Tibetan independence.

At the time, TikTok and Bytedance explained the guidelines were part of its “early days,” and said that back then, “we took a blunt approach to minimizing conflict on the platform, and our moderation guidelines allowed penalties to be given for things like content that promoted conflict, such as between religious sects or ethnic groups.” It added that at some unspecified point, it recognized “this was not the correct approach,” and subsequently changed its rules.

Now, the platform has issued a very similar statement in response to more leaked moderation documents that show it suppressed content made by people it deemed at high risk of bullying, including disabled, fat, and LGBTQ+ people.

Subscribe to get the latest creator news

Subscribe

German digital rights blog Netzpolitik, which uncovered the documents and spoke to a person familiar with the matter to confirm they were actively used by moderators, reports that TikTok flagged these people as “special users” and capped the reach of all their videos, regardless of what their videos were about.

TikTok apparently suppressed users’ accounts in two main ways, restricting their videos either geographically or by number of views.

In a section of its moderation guidelines called ‘Imagery Depicting A Subject Highly Vulnerable To Cyberbullying,’ the platform instructs moderators to mark people “susceptible to harassment or cyberbullying based on their physical or mental condition” as ‘Risk 4,’ a level that restricts their content to only being viewed by users within their own countries. That’s the geographic level.

There’s also even stricter suppression level, ‘Auto R,’ which moderators are told to apply to users with facial disfigurement, autism, Down syndrome, and “disabled people or people with some facial problems such as birthmark, slight squint and etc.” Auto R prevents videos from showing up on users’ For You pages — which recommend rising star videos they might be interested in — after they’ve reached a certain number of views. (Netzpolitik mentions caps of 6,000 and 10,000 views, but it’s not clear if those are the actual caps or just example viewcounts.)

Interestingly, the Auto R instructions note that if a disabled person or person with a disfigurement is “showing with positive energy and value,” they should be labeled Risk 4 rather than Auto R.

A screencap of the Auto R guidelines obtained by Neopolitik

The bottom line is that despite apparently being formulated with protective intentions, these guidelines punish marginalized people rather than the people who might target them.

A TikTok spokesperson told Netzpolitik that these guidelines were used “at the beginning” of TikTok to counteract bullying.

“This approach was never intended to be a long-term solution, and although we had a good intention, we realized that it was not the right approach,” they said, adding that the rules Netzpolitik got ahold of were changed. However, Netzpolitik reports the documents it obtained show these guidelines were in place as recently as September 2019 — which is, again, when TikTok came under fire for its other moderation policies, and again claimed they were old policies that had been changed.

TikTok would not share information about the moderation policy that replaced this one.

Share
Published by
James Hale
Tags: tiktok

Recent Posts

After cutting 15% of staff and saying goodbye to its CEO, Peloton must figure out what’s next

Peloton is dismissing a chunk of its workforce, including its top executive. Barry McCarthy announced that he is…

3 days ago

Meta is using AI to power brand and creator matchmaking on Facebook and Instagram

Meta is looking to improve creator and brand experiences on its platform by investing in AI. The…

3 days ago

Bob Does Sports cracks a cold one with new “Have a Day” tequila line

Bob Does Sports, the self-dubbed home of "brilliantly dumb sporting adventures" hosted by Robby Berger,…

3 days ago

Billion Dollar Boy launches biz dev community for creators with flagship location in London

Influencer marketing agency Billion Dollar Boy is launching a new membership community that's "dedicated to…

3 days ago

Millionaires: Giulia Amato on faith, finding her niche, and getting up at 4 a.m.

Welcome to Millionaires, where we profile creators who have recently crossed the one million follower…

3 days ago

Creators on the Rise: Celestial Sylvia reads the danger all around us

Welcome to Creators on the Rise, where we find and profile breakout creators who are…

4 days ago