Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Nov 11;11(1):22083.
doi: 10.1038/s41598-021-01487-w.

Dynamics of online hate and misinformation

Affiliations

Dynamics of online hate and misinformation

Matteo Cinelli et al. Sci Rep. .

Abstract

Online debates are often characterised by extreme polarisation and heated discussions among users. The presence of hate speech online is becoming increasingly problematic, making necessary the development of appropriate countermeasures. In this work, we perform hate speech detection on a corpus of more than one million comments on YouTube videos through a machine learning model, trained and fine-tuned on a large set of hand-annotated data. Our analysis shows that there is no evidence of the presence of "pure haters", meant as active users posting exclusively hateful comments. Moreover, coherently with the echo chamber hypothesis, we find that users skewed towards one of the two categories of video channels (questionable, reliable) are more prone to use inappropriate, violent, or hateful language within their opponents' community. Interestingly, users loyal to reliable sources use on average a more toxic language than their counterpart. Finally, we find that the overall toxicity of the discussion increases with its length, measured both in terms of the number of comments and time. Our results show that, coherently with Godwin's law, online debates tend to degenerate towards increasingly toxic exchanges of views.

PubMed Disclaimer

Conflict of interest statement

The authors declare no competing interests.

Figures

Figure 1
Figure 1
The distribution of the four hate speech labels in the manually annotated training (a) and evaluation (b) sets. The training set is intentionally biased to contain more hate speech while the evaluation set is unbiased.
Figure 2
Figure 2
Ranking of YouTube channels by number of comments and proportions of comment types per channel.
Figure 3
Figure 3
Proportion of the four hate speech labels in the whole dataset (a) over time (b), and for questionable (c) and reliable (d) YouTube channels. Panel b displays four dashed lines in correspondence of events of paramount relevance for the year 2020 in Italy. The first line is placed on 30/01/2020 when the first two cases of COVID-19 were detected in Italy. The second line is placed on 09/03/2020 when the Prime Minister enforced the first lockdown to the whole nation. The third line is placed on 10/04/2020 when the Prime Minister communicated to the nation an extension of the lockdown until May the 3rd. The fourth line is placed on 04/05/2020 when the “phase 2” (i.e., the suspension of the full lockdown) began. Interestingly, we note a higher share of Acceptable comments between the second and third lines, that is during the lockdown, perhaps due to positive messages and encouragement among people. Instead, as a possible consequence of the extension of the lockdown, we note a lower share of Acceptable comments right after the third line.
Figure 4
Figure 4
Distribution of comment delays in the whole dataset (a) and for questionable (b) and reliable (c) YouTube channels. The capital letters on the x-axis represent the different types of comments: acceptable (A); inappropriate (I); offensive (O); violent (V).
Figure 5
Figure 5
Users balance between different comment types. In panel (a) brighter dots indicate a higher density of users while in panel (b) brighter dots indicate a higher average activity of the users in terms of number of comments. We note that users focused on posting comments labelled as offensive and violent are almost absent in the data.
Figure 6
Figure 6
Panel (a) displays the relationship occurring between the preference of users for questionable and reliable channels (the user leaning lj) and the fraction of unacceptable comments posted by the user (a¯j) as a joint distribution. Panel (b) displays the distribution of unacceptable comments for users displaying a remarkable tendency to comment under videos posted by questionable (lj[0.75,1)) and reliable (lj(0,0.25]) channels. Panel (c) displays the distribution of unacceptable comments posted by users with leaning towards questionable channels (lj[0.75,1) indicated as q) under videos of questionable channels (dashed line indicated as q to q in the legend) and under videos of reliable channels (solid line indicated as q to r in the legend). Panel (d) displays the distribution of unacceptable comments posted by users with leaning towards reliable channels (lj(0,0.25] indicated as r) under videos of questionable channels (solid line indicated as r to q in the legend) and under videos of reliable channels (dashed line indicated as r to r in the legend).
Figure 7
Figure 7
Linear regression models for number of comments and comment delay. On the x-axis of panel (a) the comments are grouped in logarithmic bins while on the x-axis of panel (b) the comment delays are grouped in linear bins.
Figure 8
Figure 8
Transition probabilities between different comments types represented by a 4×4 transition matrix in the real (panel a) and in the random case (panel b). Brighter entries of the matrix indicate higher transition probabilities.

References

    1. Adamic, L. A., Glance, N. The political blogosphere and the 2004 us election: Divided they blog. In Proceedings of the 3rd International Workshop on Link Discovery, pp. 36–43 (2005).
    1. Flaxman S, Goel S, Rao JM. Filter bubbles, echo chambers, and online news consumption. Public Opin. Q. 2016;80(S1):298–320. doi: 10.1093/poq/nfw006. - DOI
    1. Coe K, Kenski K, Rains SA. Online and uncivil? Patterns and determinants of incivility in newspaper website comments. J. Commun. 2014;64(4):658–679. doi: 10.1111/jcom.12104. - DOI
    1. Siegel, A. A. Online hate speech. Social Media and Democracy, p. 56 (2019).
    1. Gagliardone, I., Gal, D., Alves, T. & Martinez, G. Countering Online Hate Speech (Unesco Publishing, 2015).

Publication types