Understanding Longitudinal Behaviors of Toxic Accounts on Reddit

@article{Kumar2022UnderstandingLB,
  title={Understanding Longitudinal Behaviors of Toxic Accounts on Reddit},
  author={Deepak Kumar and Jeffrey T. Hancock and Kurt Thomas and Zakir Durumeric},
  journal={ArXiv},
  year={2022},
  volume={abs/2209.02533}
}
Toxic comments are the top form of hate and harassment experienced online. While many studies have investigated the types of toxic comments posted online, the effects that such content has on people, and the impact of potential defenses, no study has captured the long-term behaviors of the accounts that post toxic comments or how toxic comments are operationalized. In this paper, we present a longitudinal measurement study of 929K accounts that post toxic comments on Reddit over an 18 month… 

References

SHOWING 1-10 OF 49 REFERENCES

The Structure of Toxic Conversations on Twitter

The relationship between structure and toxicity in conversations on Twitter is studied and it is demonstrated that models based on the structural characteristics of conversations can be used to detect early signs of toxicity and potentially steer conversations in a less toxic direction.

Designing Toxic Content Classification for a Diversity of Perspectives

It is demonstrated how existing classifiers for identifying toxic comments online fail to generalize to the diverse concerns of Internet users, and how current one-size-fits-all toxicity classification algorithms, like the Perspective API from Jigsaw, can improve in accuracy by 86% on average through personalized model tuning.

Using Authorship Verification to Mitigate Abuse in Online Communities

This paper expands on the similarity-based authorship verification approach, previously applied on large fanfictions, and shows that it works in open-world settings, shorter documents, and is largely topic-agnostic.

SoK: Hate, Harassment, and the Changing Landscape of Online Abuse

This work argues that existing security, privacy, and antiabuse protections fail to address the growing threat of online hate and harassment, and proposes a taxonomy of seven classes of attacks that each stem from different attacker capabilities and intents.

Trajectories of Blocked Community Members: Redemption, Recidivism and Departure

This work introduces a computational framework for studying the future behavior of blocked users on Wikipedia and reveals that these trajectories are tied to factors rooted both in the characteristics of the blocked individual and in whether they perceived the block to be fair and justified.

Exploring Antecedents and Consequences of Toxicity in Online Discussions

Through regression analyses on Reddit comments, it is found that both author propensity and toxicity in discussion context were strong positive antecedents of language toxicity; meanwhile, language toxicity significantly increased the volume and user evaluation of the discussion in some sub-communities, while toxicity elicitation showed mixed effects.

A large-scale characterization of online incitements to harassment across platforms

This work creates a taxonomy of attack strategies used by coordinated attackers, and uses this taxonomy to categorize the preferred approaches of coordinated attackers and the proportion of incitements for various types of harassment on different platforms.

POISED: Spotting Twitter Spam Off the Beaten Paths

POISED is presented, a system that leverages the differences in propagation between benign and malicious messages on social networks to identify spam and other unwanted content and significantly outperforms three state-of-the-art spam detection systems.

Anyone Can Become a Troll: Causes of Trolling Behavior in Online Discussions

A predictive model of trolling behavior reveals that mood and discussion context together can explain trolling behavior better than an individual's history of trolling, and suggests that ordinary people can, under the right circumstances, behave like trolls.

Characterizing and Detecting Hateful Users on Twitter

This work obtains a sample of Twitter's retweet graph and annotates 4,972 users as hateful or normal, and finds that a node embedding algorithm outperforms content-based approaches for detecting both hateful and suspended users.