Categories: YouTube

The YouTube Radicalization Pipeline Exists, And It’s Driving Users Toward Increasingly Alt-Right Content (Study)

A new study out of Cornell University has found “strong evidence for radicalization among YouTube users,” and concludes that viewers who consume “mild” radical right-wing content (it cites Joe Rogan, a YouTuber with a talk show and 6.1 million subscribers, as a creator of such “mild” content) often migrate to viewing much more radical alt-right content.

The study was prompted by ongoing concerns that YouTube offers a kind of “pipeline” that funnels viewers toward increasingly disturbing white supremacist content.

Researchers Manoel Horta Ribeiro, Raphael Ottoni, Robert West, Virgílio A. F. Almeida, and Wagner Meira looked at English-language radical content produced uploaded to YouTube from around the world, and divided that content into three types, produced by three communities. To examine how this growing amount of radical content moves on YouTube, they looked at 331,849 videos with a collective total of more than 79 million comments.

Subscribe to get the latest creator news

Subscribe

At the mildest end, there’s content produced by the Intellectual Dark Web (IDW), which is populated by people who believe they’re shut out of the mainstream media because they have “controversial” (often meaning Islamophobia, transphobic, and sexist views, the study says).

Then, in the middle, there’s the alt-lite, which the Cornell researchers class as people who “constantly flirt with ”espousing white supremacist ideals (including popularizing conspiracy theories like the anti-Semitic concept that a collective of “lizard people” secretly runs the world), but deny openly being white supremacists.

Both of these types of content are deemed “gateways” to the final, most extreme type, per the study: alt-right content. That content is made by people who are openly white supremacist, and who admit to wanting to achieve things like a state where white people live separately from people of color. The researchers note Richard Spencer, a white nationalist who’s perhaps most famous for taking a solid punch to the face, as an example of the alt-right.

All three types of content began “skyrocketing” — in number of videos uploaded, the number of views and likes received, and most importantly, the number of comments left — during the leadup to the 2016 election, the researchers say.

Processing the comments showed “that the three communities increasingly share the same user base,” the study concludes, which means it’s becoming more and more difficult to draw lines between each faction, since many of them are watching the same YouTube content and expressing the same views.

Even more crucially, though, the study showed “that users consistently migrate from milder to more extreme content,” and that “a large percentage of users who consume alt-right content now consumed alt-lite and IDW content in the past.” That backs up claims that YouTube is functioning as a “pipeline” for extremism.

But does its recommendation algorithm play a part in that pipeline? Researchers also looked at more than 2 million videos and 10,000 channels recommended by YouTube’s algorithm. They concluded that “YouTube’s recommendation algorithms frequently suggest alt-lite and IDW content,” but noted that extreme alt-right videos are not recommended by YouTube’s algorithm. Instead, alt-right content is linked to IDW and alt-lite content through channel recommendations, where viewers of a channel are recommended other channels based on similarity of content.

“This is, to our knowledge, the first large-scale quantitative audit of user radicalization on YouTube,” researchers said in the study’s write-up. “We find strong evidence for radicalization among YouTube users,” they added, and suggest “appropriate measures” need to be taken.

It’s worth noting YouTube has taken actions to suppress the spread of alt-right content, including, most recently, updating its policies to ban “videos alleging that a group is superior in order to justify discrimination, segregation or exclusion,” as well as videos that posit events like the Holocaust and the Sandy Hook shooting didn’t happen. (Those are common conspiracy theories traded among alt-righters.)

This study comes on the heels of another study published earlier this month which chronicled YouTube’s part in the radicalization of Brazilian citizens, including spreading right-wing and conspiracy theory-themed videos from the YouTuber who would later become Brazil’s current president, Jair Bolsonaro.

You can read the Cornell study in its entirety here.

Share
Published by
James Hale
Tags: YouTube

Recent Posts

Top 5 Branded Videos of the Week: The name of that man on top of our ranking? Albert Einstein.

The most-watched sponsored YouTube video of the week didn't come from MrBeast -- instead, Derek Muller's Veritasium provided…

10 hours ago

Ahead of Brandcast, YouTube’s Neal Mohan has a message: Creators should win Emmys

Neal Mohan wants the Television Academy to watch more YouTube videos. The CEO of the Google-owned platform recently penned a…

11 hours ago

Sapnap-signed Cheez-Its are now on Walmart shelves

Sapnap's face is officially a snack. The Dream SMP co-founder/NRG co-owner teamed up with Walmart…

16 hours ago

Millionaires: Rachel Chaleff says “never, ever be afraid to be your most authentic self”

Welcome to Millionaires, where we profile creators who have recently crossed the one million follower…

3 days ago

Facing Congressional pressure, an influential advocacy group ends its relationship with TikTok

TikTok's plan to challenge the new "divest-or-ban" law has encountered a setback. NetChoice, an advocacy group…

4 days ago

Is Warner Bros’ next ‘Lord of the Rings’ movie inspired by a YouTube fan film?

Peter Jackson and Andy Serkis are planning to tell the story of The Hunt For Gollum, but a…

4 days ago