Algorithmic radicalization

Algorithmic radicalization is the concept that algorithms on popular social media sites such as YouTube and Facebook drive users toward progressively more extreme content over time, leading to them developing radicalized extremist political views. Algorithms record user interactions, from likes/dislikes to amount of time spent on posts, to generate endless media aimed to keep users engaged. Through echo chamber channels, the consumer is driven to be more polarized through preferences in media and self-confirmation.[1][2][3][4]

Algorithmic radicalization remains a controversial phenomenon as it is often not in the best interest of social media companies to remove echo chamber channels.[5][6] Though social media companies have admitted to algorithmic radicalization's existence, it remains unclear how each will manage this growing threat.

Social media echo chambers and filter bubbles

Social media platforms learn the interests and likes of the user to modify their experiences in their feed to keep them engaged and scrolling. An echo chamber is formed when users come across beliefs that magnify or reinforce their thoughts and form a group of like-minded users in a closed system.[7] The issue with echo chambers is that it spreads information without any opposing beliefs and can possibly lead to confirmation bias. According to a group polarization theory, an echo chamber can potentially lead users and groups towards more extreme radicalized positions.[8] According to the National Library of Medicine, "Users online tend to prefer information adhering to their worldviews, ignore dissenting information, and form polarized groups around shared narratives. Furthermore, when polarization is high, misinformation quickly proliferates"[9]

Facebook's algorithms

Facebook's algorithm focuses on recommending content that makes the user want to interact. They rank content by prioritizing popular posts by friends, viral content, and sometimes divisive content. Each feed is personalized to the user's specific interests which can sometimes lead users towards an echo chamber of troublesome content.[10] Users can find their list of interests the algorithm uses by going to the "Your ad Preferences page. According to a Pew Research study, 74% of Facebook users did not know that list existed until they were directed towards that page in the study.[11] It is also relatively common for Facebook to assign political labels to their users. In recent years, Facebook has started using artificial intelligence to change the content users see in their feed and what is recommended to them. A document known as The Facebook Files, has revealed that their AI system prioritizes user engagement over everything else. The Facebook Files has also demonstrated that controlling the AI systems has proven difficult to handle.[12]

Facebook's allegations

In an August 2019 internal memo leaked in 2021, Facebook has admitted that "the mechanics of our platforms are not neutral",[13][14] concluding that in order to reach maximum profits, optimization for engagement is necessary. In order to increase engagement, algorithms have found that hate, misinformation, and politics are instrumental for app activity.[15]  As referenced in the memo, "The more incendiary the material, the more it keeps users engaged, the more it is boosted by the algorithm."[13] According to a 2018 study, "false rumors spread faster and wider than true information... They found falsehoods are 70% more likely to be retweeted on Twitter than the truth, and reach their first 1,500 people six times faster. This effect is more pronounced with political news than other categories."[16]

YouTube's algorithm

YouTube has been around since 2005 and has more than 2.5 billion monthly users. YouTube discovery content systems focus on the user's personal activity (watched, favorites, likes) to direct them to recommended content. YouTube's algorithm is accountable for roughly 70% of users recommended videos and what drives people to watch certain content.[17] According to a new study, users have little power to keep unsolicited videos out of their suggested recommended content. This includes videos about hate speech, livestreams, etc. [17]

YouTube's allegations

YouTube has been identified as an influential platform for spreading radicalized content. Al-Qaeda and similar extremist groups have been linked to using YouTube for recruitment videos and engaging with international media outlets. In a research study published by the American Behavioral Scientist Journal, they researched "whether it is possible to identify a set of attributes that may help explain part of the YouTube algorithm's decision-making process" [18] The results of the study showed that YouTube's algorithm recommendations for extremism content factor into the presence of radical keywords in a video's title. In February of 2023, in the case of Gonzalez v. Google, the question at hand is whether or not Google, the parent company of YouTube, is protected from lawsuits claiming that the site's algorithms aided terrorists in recommending ISIS videos to users. Section 230 is known to generally protect online platforms from civil liability for the content posted by its users [19]

TikTok algorithms

TikTok is an app that recommends videos to your 'For You Page' (FYP), making every users' page different. With the nature of the algorithm behind the app, TikTok's FYP has been linked to showing more explicit and radical videos over time based on your previous interactions on the app.[20] Since TikTok's inception, the app has been scrutinized for misinformation and hate speech as those forms of media usually generate more interactions to the algorithm.[21]

As of 2022, TikTok's head of US Security has put out a statement that "81,518,334 videos were removed globally between April - June for violating our Community Guidelines or Terms of Service" to cut back on hate speech, harassment, and misinformation.[22]

Alt-right pipeline

Graphic of interactions between right-wing personalities on YouTube from January 2017 to April 2018. Each line indicates a shared appearance in a YouTube video, allowing audiences of one personality to discover another.[23]

The alt-right pipeline (also called the alt-right rabbit hole) is a proposed conceptual model regarding internet radicalization toward the alt-right movement. It describes a phenomenon in which consuming provocative right-wing political content, such as antifeminist or anti-SJW ideas, gradually increases exposure to the alt-right or similar far-right politics. It posits that this interaction takes place due to the interconnected nature of political commentators and online communities, allowing members of one audience or community to discover more extreme groups.[23][24] This process is most commonly associated with and has been documented on the video platform YouTube, and is largely faceted by the method in which algorithms on various social media platforms function through the process recommending content that is similar to what users engage with, but can quickly lead users down rabbit-holes.[24][25][26]

Many political movements have been associated with the pipeline concept. The intellectual dark web,[24] libertarianism,[27] the men's rights movement,[28] and the alt-lite movement[24] have all been identified as possibly introducing audiences to alt-right ideas. Audiences that seek out and are willing to accept extreme content in this fashion typically consist of young men, commonly those that experience significant loneliness and seek belonging or meaning.[29] In an attempt to find community and belonging, message boards that are often proliferated with hard right social commentary, such as 4chan and 8chan, have been well documented in their importance in the radicalization process.[30]

The alt-right pipeline may be a contributing factor to domestic terrorism.[31][32] Many social media platforms have acknowledged this path of radicalization and have taken measures to prevent it, including the removal of extremist figures and rules against hate speech and misinformation.[25][29] Left-wing movements, such as BreadTube, also oppose the alt-right pipeline and "seek to create a 'leftist pipeline' as a counterforce to the alt-right pipeline."[33]

The effects of YouTube's algorithmic bias in radicalizing users has been replicated by one study,[24][34][35][36] while two other studies have found little or no evidence of a radicalization process.[25][37][38]

Self-radicalization

An infographic from the United States Department of Homeland Security's "If You See Something, Say Something" campaign. The campaign is a national initiative to raise awareness to homegrown terrorism and terrorism-related crime.

The U.S. department of Justice defines 'Lone-wolf' (self) terrorism as "someone who acts alone in a terrorist attack without the help or encouragement of a government or a terrorist organization".[39] Through social media outlets on the internet, 'Lone-wolf' terrorism has been on the rise, being linked to algorithmic radicalization.[40] Through echo-chambers on the internet, viewpoints typically seen as radical were accepted and quickly adopted by other extremists.[41] These viewpoints are encouraged by forums, group chats, and social media to reinforce their beliefs.[42]

References in media

The Social Dilemma

"The Social Dilemma" was a 2020 docudrama about how algorithms behind social media enables addiction, while possessing abilities to manipulate people's views, emotions, and behavior to spread conspiracy theories and disinformation. The film repeatedly uses buzz words such as 'echo chambers' and 'fake news' to prove psychological manipulation on social media, therefore leading to political manipulation. In the film, we follow Ben as he falls deeper into a social media addiction as the algorithm found that his social media page has a 62.3% chance of long-term engagement. This leads into more videos on the recommended feed for Ben and he eventually becomes more immersed into propaganda and conspiracy theories, becoming more polarized with each video.

Possible solutions

Section 230

In the Communications Decency Act of 1996, section 230 states that "No provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider".[43] Section 230 protects the media from liabilities or being sued of third-party content, such as illegal activity from a user.[43]  However, this approach reduces a company's incentive to remove harmful content or misinformation.  This loophole has allowed social media companies to maximize profits through pushing radical content without legal risks.[44]

See also

References

  1. "What is a Social Media Echo Chamber? | Stan Richards School of Advertising". advertising.utexas.edu. Retrieved November 2, 2022.
  2. "The Websites Sustaining Britain's Far-Right Influencers". bellingcat. February 24, 2021. Retrieved March 10, 2021.
  3. Camargo, Chico Q. (January 21, 2020). "YouTube's algorithms might radicalise people – but the real problem is we've no idea how they work". The Conversation. Retrieved March 10, 2021.
  4. E&T editorial staff (May 27, 2020). "Facebook did not act on own evidence of algorithm-driven extremism". eandt.theiet.org. Retrieved March 10, 2021.
  5. "How Can Social Media Firms Tackle Hate Speech?". Knowledge at Wharton. Retrieved November 22, 2022.
  6. "Internet Association - We Are The Voice Of The Internet Economy. | Internet Association". December 17, 2021. Archived from the original on December 17, 2021. Retrieved November 22, 2022.
  7. "What is a Social Media Echo Chamber? | Stan Richards School of Advertising". advertising.utexas.edu. Retrieved April 12, 2023.
  8. Cinelli, Matteo; De Francisci Morales, Gianmarco; Galeazzi, Alessandro; Quattrociocchi, Walter; Starnini, Michele (March 2, 2021). "The echo chamber effect on social media". Proceedings of the National Academy of Sciences of the United States of America. 118 (9): –2023301118. Bibcode:2021PNAS..11823301C. doi:10.1073/pnas.2023301118. ISSN 0027-8424. PMC 7936330. PMID 33622786.
  9. Cinelli, Matteo; De Francisci Morales, Gianmarco; Starnini, Michele; Galeazzi, Alessandro; Quattrociocchi, Walter (January 14, 2021). "The echo chamber effect on social media". Proceedings of the National Academy of Sciences of the United States of America. 118 (9): e2023301118. Bibcode:2021PNAS..11823301C. doi:10.1073/pnas.2023301118. ISSN 0027-8424. PMC 7936330. PMID 33622786.
  10. Oremus, Will; Alcantara, Chris; Merrill, Jeremy; Galocha, Artur (October 26, 2021). "How Facebook shapes your feed". The Washington Post. Retrieved April 12, 2023.
  11. Atske, Sara (January 16, 2019). "Facebook Algorithms and Personal Data". Pew Research Center: Internet, Science & Tech. Retrieved April 12, 2023.
  12. Korinek, Anton (December 8, 2021). "Why we need a new agency to regulate advanced artificial intelligence: Lessons on AI control from the Facebook Files". Brookings. Retrieved April 12, 2023.
  13. "Disinformation, Radicalization, and Algorithmic Amplification: What Steps Can Congress Take?". Just Security. February 7, 2022. Retrieved November 2, 2022.
  14. Isaac, Mike (October 25, 2021). "Facebook Wrestles With the Features It Used to Define Social Networking". The New York Times. ISSN 0362-4331. Retrieved November 2, 2022.
  15. Little, Olivia (March 26, 2021). "TikTok is prompting users to follow far-right extremist accounts". Media Matters for America. Retrieved November 2, 2022.
  16. "Study: False news spreads faster than the truth". MIT Sloan. Retrieved November 2, 2022.
  17. "Hated that video? YouTube's algorithm might push you another just like it". MIT Technology Review. Retrieved April 11, 2023.
  18. Murthy, Dhiraj (May 1, 2021). "Evaluating Platform Accountability: Terrorist Content on YouTube". American Behavioral Scientist. 65 (6): 800–824. doi:10.1177/0002764221989774. S2CID 233449061 via JSTOR.
  19. Root, Damon (April 2023). "Scotus Considers Section 230's Scope". Reason. 54 (11): 8. ISSN 0048-6906.
  20. "TikTok's algorithm leads users from transphobic videos to far-right rabbit holes". Media Matters for America. October 5, 2021. Retrieved November 22, 2022.
  21. Little, Olivia (April 2, 2021). "Seemingly harmless conspiracy theory accounts on TikTok are pushing far-right propaganda and TikTok is prompting users to follow them". Media Matters for America. Retrieved November 22, 2022.
  22. "Our continued fight against hate and harassment". Newsroom | TikTok. August 16, 2019. Retrieved November 22, 2022.
  23. Lewis, Rebecca (September 18, 2018). Alternative Influence: Broadcasting the Reactionary Right on YouTube (Report). Data & Society. Archived from the original on May 25, 2022. Retrieved July 14, 2022.
  24. Ribeiro, Manoel Horta; Ottoni, Raphael; West, Robert; Almeida, Virgílio A. F.; Meira, Wagner (January 27, 2020). "Auditing radicalization pathways on YouTube". Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency. pp. 131–141. doi:10.1145/3351095.3372879. ISBN 9781450369367. S2CID 201316434.
  25. Ledwich, Mark; Zaitsev, Anna (February 26, 2020). "Algorithmic extremism: Examining YouTube's rabbit hole of radicalization". First Monday. arXiv:1912.11211. doi:10.5210/fm.v25i3.10419. ISSN 1396-0466. S2CID 209460683. Archived from the original on October 28, 2022. Retrieved October 28, 2022.
  26. Mozilla (July 7, 2021). "Mozilla Investigation: YouTube Algorithm Recommends Videos that Violate the Platform's Very Own Policies". Mozilla Foundation. Archived from the original on March 25, 2023. Retrieved March 25, 2023.
  27. Hermansson, Patrik; Lawrence, David; Mulhall, Joe; Murdoch, Simon (January 31, 2020). The International Alt-Right: Fascism for the 21st Century?. Routledge. pp. 57–58. ISBN 978-0-429-62709-5. Archived from the original on July 25, 2023. Retrieved September 21, 2022.
  28. Mamié, Robin; Horta Ribeiro, Manoel; West, Robert (June 21, 2021). "Are Anti-Feminist Communities Gateways to the Far Right? Evidence from Reddit and YouTube". 13th ACM Web Science Conference 2021. WebSci '21. New York, NY, USA: Association for Computing Machinery. pp. 139–147. arXiv:2102.12837. doi:10.1145/3447535.3462504. ISBN 978-1-4503-8330-1. S2CID 232045966. Archived from the original on July 25, 2023. Retrieved September 21, 2022.
  29. Roose, Kevin (June 8, 2019). "The Making of a YouTube Radical". The New York Times. ISSN 0362-4331. Archived from the original on May 17, 2023. Retrieved October 26, 2022.
  30. Hughes, Terwyn (January 26, 2021). "Canada's alt-right pipeline". The Pigeon. Archived from the original on March 25, 2023. Retrieved March 25, 2023.
  31. Piazza, James A. (January 2, 2022). "Fake news: the effects of social media disinformation on domestic terrorism". Dynamics of Asymmetric Conflict. 15 (1): 55–77. doi:10.1080/17467586.2021.1895263. ISSN 1746-7586. S2CID 233679934. Archived from the original on July 25, 2023. Retrieved November 4, 2022.
  32. Munn, Luke (June 1, 2019). "Alt-right pipeline: Individual journeys to extremism online". First Monday. doi:10.5210/fm.v24i6.10108. ISSN 1396-0466. S2CID 184483249. Archived from the original on May 24, 2022. Retrieved July 14, 2022.
  33. Cotter, Kelley (March 18, 2022). "Practical knowledge of algorithms: The case of BreadTube". New Media & Society. doi:10.1177/14614448221081802. ISSN 1461-4448. S2CID 247560346.
  34. Lomas, Natasha (January 28, 2020). "Study of YouTube comments finds evidence of radicalization effect". TechCrunch. Retrieved July 17, 2021.
  35. Newton, Casey (August 28, 2019). "YouTube may push users to more radical views over time, a new paper argues". The Verge. Retrieved July 17, 2021.
  36. Ribeiro, Manoel Horta; Ottoni, Raphael; West, Robert; Almeida, Virgílio A. F.; Meira, Wagner (August 22, 2019). "Auditing Radicalization Pathways on YouTube". arXiv:1908.08313 [cs.CY].
  37. Hosseinmardi, Homa; Ghasemian, Amir; Clauset, Aaron; Mobius, Markus; Rothschild, David M.; Watts, Duncan J. (August 2, 2021). "Examining the consumption of radical content on You Tube". Proceedings of the National Academy of Sciences. 118 (32). arXiv:2011.12843. Bibcode:2021PNAS..11801967H. doi:10.1073/pnas.2101967118. PMC 8364190. PMID 34341121.
  38. "Lone Wolf Terrorism in America | Office of Justice Programs". www.ojp.gov. Retrieved November 2, 2022.
  39. Alfano, Mark; Carter, J. Adam; Cheong, Marc (2018). "Technological Seduction and Self-Radicalization". Journal of the American Philosophical Association. 4 (3): 298–322. doi:10.1017/apa.2018.27. ISSN 2053-4477. S2CID 150119516.
  40. Dubois, Elizabeth; Blank, Grant (May 4, 2018). "The echo chamber is overstated: the moderating effect of political interest and diverse media". Information, Communication & Society. 21 (5): 729–745. doi:10.1080/1369118X.2018.1428656. ISSN 1369-118X. S2CID 149369522.
  41. Sunstein, Cass R. (May 13, 2009). Going to Extremes: How Like Minds Unite and Divide. Oxford University Press. ISBN 978-0-19-979314-3.
  42. "47 U.S. Code § 230 - Protection for private blocking and screening of offensive material". LII / Legal Information Institute. Retrieved November 2, 2022.
  43. Smith, Michael D.; Alstyne, Marshall Van (August 12, 2021). "It's Time to Update Section 230". Harvard Business Review. ISSN 0017-8012. Retrieved November 2, 2022.
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.