Abstract
With proliferation of user generated contents in social media platforms, establishing mechanisms to automatically identify toxic and abusive content becomes a prime concern for regulators, researchers, and society. Keeping the balance between freedom of speech and respecting each other dignity is a major concern of social media platform regulators. Although, automatic detection of offensive content using deep learning approaches seems to provide encouraging results, training deep learning-based models requires large amounts of high-quality labeled data, which is often missing. In this regard, we present in this paper a new deep learning-based method that fuses a Back Translation method, and a Paraphrasing technique for data augmentation. Our pipeline investigates different word-embedding-based architectures for classification of hate speech. The back translation technique relies on an encoder–decoder architecture pre-trained on a large corpus and mostly used for machine translation. In addition, paraphrasing exploits the transformer model and the mixture of experts to generate diverse paraphrases. Finally, LSTM, and CNN are compared to seek enhanced classification results. We evaluate our proposal on five publicly available datasets; namely, AskFm corpus, Formspring dataset, Warner and Waseem dataset, Olid, and Wikipedia toxic comments dataset. The performance of the proposal together with comparison to some related state-of-art results demonstrate the effectiveness and soundness of our proposal.
Highlights
With the exponential increase in use of social media platforms where people can freely communicate their opinions and thoughts, online hate speech has seen the emergence of appropriate ecosystem, which caused concerns to authority, researchers, and society, especially in the last decade
We evaluate our approach of cyberbullying detection on five publicly available datasets: AskFm, Formspring, OLID 2019, Warner and Waseem, and Wikipedia toxic comments dataset
Once the augmentation techniques are investigated in terms of datasets sizes, we evaluate the performance of using these augmented datasets compared to original ones for detecting hate speech and cyberbullying
Summary
With the exponential increase in use of social media platforms where people can freely communicate their opinions and thoughts, online hate speech has seen the emergence of appropriate ecosystem, which caused concerns to authority, researchers, and society, especially in the last decade. The easy access to various social media platforms as well as the anonymization schemes have boosted the scope of offensive online content and harassment cases. Some political and racist organizations have exploited such channels to propagate toxic content such as hate speeches [1]. Maintaining a balance between freedom of speech and societal protection needs while respecting cultural and gender diversity is often challenging. Hate speech can lead to serious consequences on both individual and community scale, which may trigger violence and raise public safety concerns. One predominant kind of online hate speech is cyberbullying
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.