Abstract

Online social networks serve as a convenient way to connect, share, and promote content with others. As a result, these networks can be used with malicious intent, causing disruption and harm to public debate through the sharing of misinformation. However, automatically identifying such content through its use of natural language is a significant challenge compared to our solution which uses less computational resources, language-agnostic and without the need for complex semantic analysis. Consequently alternative and complementary approaches are highly valuable. In this paper, we assess content that has the potential for misinformation and focus on patterns of user association with online social media communities (subreddits) in the popular Reddit social media platform, and generate networks of behaviour capturing user interaction with different subreddits. We examine these networks using both global and local metrics, in particular noting the presence of induced substructures (graphlets) assessing 7,876,064 posts from 96,634 users. From subreddits identified as having potential for misinformation, we note that the associated networks have strongly defined local features relating to node degree — these are evident both from analysis of dominant graphlets and degree-related global metrics. We find that these local features support high accuracy classification of subreddits that are categorised as having the potential for misinformation. Consequently we observe that induced local substructures of high degree are fundamental metrics for subreddit classification, and support automatic detection capabilities for online misinformation independent from any particular language.

Highlights

  • Misinformation is a major cause for concern with potentially dangerous ramifications for social processes, including the stability of democracy [1,2]

  • To provide a basis for comparison of potential for misinformation (PFM) subreddits, we introduce three other sets of subreddits so that we can benchmark against alternative forms of user interaction with this social media platform

  • We have used a general network representation of user association with social media content as a basis for prediction of important sub-classes of content that align with the potential for misinformation (PFM)

Read more

Summary

Introduction

Misinformation is a major cause for concern with potentially dangerous ramifications for social processes, including the stability of democracy [1,2]. We build machine learning models to determine the predictive power of graphlet and global features in distinguishing between the activity of different sets of subreddits. This gives a basis to assess the role of local features, including substructures, in capturing online behaviours aligned to potential misinformation. Once misinformation is embedded, echo chambers are known to take hold and to support engagement of misinformation, using weak ties [15,16] and lack of effective moderation [17] alongside ‘‘soft facts’’ These occur as a result of users sharing potentially misleading content without knowing the entire facts of an event [18,19,20]

Objectives
Methods
Discussion
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.