Efficient Intelligent System for Cyberbullying Detection in English and Roman Urdu Social Media Posts

Authors

  • Muhammad Talha Jahangir Department of Computer Science, MNS University of Engineering and Technology, Multan, Pakistan.
  • Muhammad Ahmad Institute of Computing, MNS University of Agriculture Multan, Pakistan.
  • Hamna Rehman Institute of Computing, MNS University of Agriculture Multan, Pakistan.

Keywords:

Online Toxicity, Hate Speech, Abusive Language, Machine Learning Algorithms, English; Roman Urdu, Toxic Comment Classification

Abstract

The internet has revolutionized communication, offering new platforms like social media, blogs, and comment sections for people to connect. However, these platforms have also seen an uptick in abusive language, hate speech, and cyberbullying. In the more recent work, training models to identify harmful remarks across several classes is explored using algorithms. A recent study examined the efficacy of naive Bayes, logistic regression, and support vector machine as three different approaches to using an online negative feedback engine. Toxic, offensive, disparaging, hateful, and healthy (non-toxic) comments are screened out of the process. With 97.5% accuracy in English analysis and 92.9% accuracy in Urdu analysis, Support Vector Machines (SVM) performed better than the other approaches, according to the data. SVM has shown a strong capacity to identify hate speech and insults. This research is significant because it will contribute to the development of technologies that online platforms can employ to identify and eliminate unwanted information, making the internet a safer and more secure place.

Downloads

Published

2024-04-01

How to Cite

Muhammad Talha Jahangir, Muhammad Ahmad, & Hamna Rehman. (2024). Efficient Intelligent System for Cyberbullying Detection in English and Roman Urdu Social Media Posts. Journal of Computing & Biomedical Informatics. Retrieved from https://jcbi.org/index.php/Main/article/view/501