From a recent Science Reports paper:
Online debates are often characterised by extreme polarisation and heated discussions among users. The presence of hate speech online is becoming increasingly problematic, making necessary the development of appropriate countermeasures. In this work, we perform hate speech detection on a corpus of more than one million comments on YouTube videos through a machine learning model, trained and fine-tuned on a large set of hand-annotated data.
Our analysis shows that there is no evidence of the presence of "pure haters", meant as active users posting exclusively hateful comments. Moreover, coherently with the echo chamber hypothesis, we find that users skewed towards one of the two categories of video channels (questionable, reliable) are more prone to use inappropriate, violent, or hateful language within their opponents' community.
Interestingly, users loyal to reliable sources use on average a more toxic language than their counterpart. Finally, we find that the overall toxicity of the discussion increases with its length, measured both in terms of the number of comments and time. Our results show that, coherently with Godwin's law, online debates tend to degenerate towards increasingly toxic exchanges of views.
Journal Reference:
M. Cinelli, A. Pelicon, I. Mozetič, et al. Dynamics of online hate and misinformation. [open] Sci Rep 11, 22083 (2021).
DOI: 10.1038/s41598-021-01487-w
(Score: 2) by Phoenix666 on Thursday November 18 2021, @08:56PM
That is an excellent test. Everyone should use it.
The people who support using the FBI to intimidate parents who attend school board meetings will really not like it when the people with ideas they don't like get control of the FBI and return the favor. (Of course, I personally think it might be time to dissolve the FBI; they got a second chance after the Hoover era and yet here they are attacking democracy again)
Washington DC delenda est.