Social media platforms like Twitter play a pivotal role in public debates. Recent studies showed that users online tend to join groups of like-minded peers, called echo chambers, in which they frame and reinforce a shared narrative. Such a polarized configuration may trigger heated debates and foment misinformation spreading. In this work, we explore the interplay between the systematic spreading of misinformation and the emergence of toxic conversations. We perform a thorough quantitative analysis on 3.3 M comments by more than 1 M unique users from 25 K conversations involving 60 news outlets active on Twitter from January 2020 to April 2022. By tagging the news triggering the conversation with a specific reliability score provided by an independent fact-checking organization (NewsGuard), we perform a network-based analysis of the structure of toxic conversations on Twitter. We find that users using toxic language are few and evenly distributed over the entire reliability score range, showing no significant evidence for the interplay between toxic speech and misinformation spreading.
Reliability of News and Toxicity in Twitter Conversations / Quattrociocchi, A.; Etta, G.; Avalle, M.; Cinelli, M.; Quattrociocchi, W.. - 13618 LNCS:(2022), pp. 245-256. (Intervento presentato al convegno 13th International Conference on Social Informatics, SocInfo 2022 tenutosi a Glasgow) [10.1007/978-3-031-19097-1_15].
Reliability of News and Toxicity in Twitter Conversations
Quattrociocchi A.;Etta G.;Avalle M.;Cinelli M.;Quattrociocchi W.
2022
Abstract
Social media platforms like Twitter play a pivotal role in public debates. Recent studies showed that users online tend to join groups of like-minded peers, called echo chambers, in which they frame and reinforce a shared narrative. Such a polarized configuration may trigger heated debates and foment misinformation spreading. In this work, we explore the interplay between the systematic spreading of misinformation and the emergence of toxic conversations. We perform a thorough quantitative analysis on 3.3 M comments by more than 1 M unique users from 25 K conversations involving 60 news outlets active on Twitter from January 2020 to April 2022. By tagging the news triggering the conversation with a specific reliability score provided by an independent fact-checking organization (NewsGuard), we perform a network-based analysis of the structure of toxic conversations on Twitter. We find that users using toxic language are few and evenly distributed over the entire reliability score range, showing no significant evidence for the interplay between toxic speech and misinformation spreading.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.