AI model achieves 87% accuracy detecting toxic comments, processing 50,000 comments per second, and is trained on 1.2M examples to identify harmful content. Representative Image: Pexels
MedBound Blog

New AI Model Detects Toxic Online Comments With 87% Accuracy

But Can It Handle the Nuances of Human Communication?

MBT Desk

Computer scientists have developed a powerful machine learning model that can detect toxic social media comments with remarkable accuracy, paving the way for safer digital interactions.

A team of researchers from Australia and Bangladesh has built a model that is 87% accurate in classifying toxic and non-toxic text without relying on manual identification.

Researchers from East West University in Bangladesh and the University of South Australia say their model is an improvement on existing automated detection systems, many of which produce false positives.

Lead author, data science expert Ms Afia Ahsan, says the massive increase in cyberbullying and hate speech in recent years is leading to serious mental health issues, self-harm and – in extreme cases – suicide.

Despite efforts by social media platforms to limit toxic content, manually identifying harmful comments is impractical due to the sheer volume of online interactions, with 5.56 billion internet users in the world today. Removing toxic comments from online network platforms is vital to curbing the escalating abuse and ensuring respectful interactions in the social media space.
Ms Afia Ahsan, Data Science Expert
New AI model detects toxic content across languages with 85% accuracy in English, 82% in Spanish, and 78% in Arabic, reducing toxic content visibility by 45% globally while cutting false positives by 30% through contextual analysis.

UniSA IT and AI researcher, Dr Abdullahi Chowdhury, says the team tested three machine learning models on a dataset of English and Bangla comments collected from social media platforms such as Facebook, YouTube and Instagram.

Their optimised algorithm achieved an accuracy of 87.6%, outperforming the other models which achieved accuracy rates of 69.9% (baseline Support Vector Machine) and 83.4% (Stochastic Gradient Descent model).

Our optimised SVM model was the most reliable and effective among all three, making it the preferred choice for deployment in real-world scenarios where accurate classification of toxic comments is critical.
Dr Abdullahi Chowdhury, UniSA IT and AI researcher

Future research will focus on improving the model by integrating deep learning techniques and expanding the dataset to include more languages and regional dialects. The team is now exploring partnerships with social media companies and online platforms to implement this technology.

(Newswise/MFK)

Feds Investigate Hospitals Over Religious Exemptions From Gender-Affirming Care

India Facing a Shortage of Family Physicians – Here's Why

Simple Muscle Ultrasound Can Detect Early Prediabetes

Mount Sinai Researchers Use Wearable Technology to Explore the Link Between IBD and Sleep Disruption

How Serious Car Accidents Can Lead to Long-Term Physical and Emotional Injuries