Subscriptions drive views of alternative and extremist videos on YouTube and viewership of harmful content is concentrated among a small group of users, a new study shows.
As the second most popular social media platform in the world, YouTube frequently attracts criticism. In particular, critics argue that its algorithmic recommendations facilitates radicalization and extremism by sending users down “rabbit holes” of harmful content.
According to a new study published in Science Advances, however, exposure to alternative and extremist video channels on YouTube is not driven by recommendations. Instead, most consumption of these channels on the platform can be attributed to a small group of users high in gender and racial resentment who subscribe to these channels and follow links to their videos.
The authors caution that these findings do not exonerate the platform. “YouTube’s algorithms may not be recommending alternative and extremist content to nonsubscribers very often, but they are nonetheless hosting it for free and funneling it to subscribers in ways that are of great concern,” said co-author Brendan Nyhan, the James O. Freedman Presidential Professor at Dartmouth.
“The problem of potentially harmful content on YouTube is real. The challenge is understanding the nature of the problems, so we can think about how best to address it.”
Co-author Professor Jason Reifler, from the University of Exeter, said: “Social media companies make choices about what content they allow on their platforms, including YouTube. Why people consume harmful content is more complex than ‘the algorithm did it’. But companies choose policies that permit harmful content and for the algorithm to surface this content to interested users, even if the algorithm is not directly radicalizing unsuspecting people by sending them down a proverbial rabbit hole.”
In 2019, YouTube announced that changes to its algorithms had reduced watch time of harmful content by 50%, with a 70% decline in watch time by nonsubscribers. These reports had not been independently verified, so the research team set out to determine who is watching this type of content and evaluate what recommendations are offered by YouTube’s algorithm.
The research team analyzed more than 1,100 participants’ web browsing data. Participants were recruited from a general population sample of 2,000 people; a group of 1,000 people who had previously expressed high levels of racism and hostile sexism in another survey; and a sample of 1,000 people with high levels of self-reported YouTube use.
All participants who opted in provided informed consent to allow anonymized tracking of their web browsing behavior in Chrome or Firefox from July to December 2020, with various security protocols in place. The browser extension was automatically uninstalled from their computers at the conclusion of the study period.
Given the challenges of trying to characterize the content of every single video viewed, the researchers focused on the type of YouTube channels people watched. They compiled lists of channels that had been identified as alternative or extreme by journalists and academics and then examined how often a participant visited videos from those channels.
Alternative channels included men’s rights activists, anti-social justice warriors, and intellectual dark web channels, and extreme channels included white supremacist, alt-right, and extremist channels.
The results showed that exposure to alternative and extremist channels was quite rare among the study groups. Only 15% of people who opted to provide daily browser activity data visited an alternative channel video and only 6% viewed an extremist channel video.
A majority of viewers of the potentially harmful channels were subscribers to the type of channel in question: 61% subscribers for alternative channels and 55% for extremist channels. Almost all subscribed either to the channel in question or another one like it (e.g., another alternative or extremist channel): 93% for alternative channels and 85% for extremist channels.
Viewing time data showed that a tiny percentage of people were responsible for most of the time participants spent watching potentially harmful channels. Specifically, 1.7% of participants were responsible for 80% of time spent on alternative channels while only 0.6% of participants were responsible for 80% of the time spent on extremist channels.
The researchers also found that people who scored high in hostile sexism and racial resentment were more likely to visit videos from alternative and extremist channels.
“What really stands out is the correlation between content subscribers’ prior levels of hostile sexism and more time spent watching videos from alternative and extremist channels,” says Nyhan. “We interpret that relationship as suggesting that people are seeking this content out.”
By contrast, the researchers found that recommendations to alternative and extremist channel videos were very rare and that “rabbit hole”-type events were only observed a handful of times during the study period.
The findings do not speak to what was happening on YouTube prior to the changes made to the website’s algorithm in 2019; recommendations and viewing patterns during that period may have differed substantially from what the researchers observed in 2020.
Other co-authors of the research paper include: Annie Chen at CUNY Institute State & Local Governance; Ronald Robertson at Stanford University; and Christo Wilson at Northeastern University. The team conducted the research with the support of the Belfer Fellowship at the ADL’s Center for Technology and Society and a grant from the Russell Sage Foundation.