Login | Register

Innovative Approaches for Real-Time Toxicity Detection in Social Media Using Deep Reinforcement Learning

Title:

Innovative Approaches for Real-Time Toxicity Detection in Social Media Using Deep Reinforcement Learning

Bodaghi, Arezo (2024) Innovative Approaches for Real-Time Toxicity Detection in Social Media Using Deep Reinforcement Learning. PhD thesis, Concordia University.

[thumbnail of Bodaghi_PhD_S2025.pdf]
Preview
Text (application/pdf)
Bodaghi_PhD_S2025.pdf - Accepted Version
Available under License Spectrum Terms of Access.
7MB

Abstract

Toxic comments on social media discourage user engagement and have serious consequences for mental health and social well-being. Such negativity heightens feelings of anxiety, depression, and social isolation among users, ultimately diminishing their experience on these platforms. For businesses, these toxic interactions are detrimental as they lead to reduced user engagement, subsequently affecting advertising revenue and market share.
Creating a safe and inclusive online environment is essential for business success and social responsibility. This requires real-time detection of toxic behavior through automated methods. However, many existing toxicity detectors focus mainly on accuracy, often neglecting important factors including throughput, computational costs, and the impact of false positives and negatives on user engagement. Additionally, these methods are evaluated in controlled experimental settings (offline tests), which do not reflect the complexities of large-scale social media environments. This limitation hinders their practical applicability in real-world scenarios.
This thesis addresses these limitations by introducing a Profit-driven Simulation (PDS) framework for evaluating the real-time performance of deep learning classifiers in complex social media settings. The PDS framework integrates performance, computational efficiency, and user engagement, revealing that optimal classifier selection depends on the toxicity level of the environment. High-throughput classifiers are most effective in low- and high-toxicity scenarios, while classifiers offering moderate accuracy and throughput excel in medium-toxicity contexts.
Additionally, the thesis tackles the challenge of imbalanced datasets by introducing a novel method for augmenting toxic text data. By applying Reinforcement Learning with Human Feedback (RLHF) and Proximal Policy Optimization (PPO), this method fine-tunes Large Language Models (LLMs) to generate diverse, semantically consistent toxic data. This approach enhances classifier robustness, particularly in detecting minority class instances.
The thesis also proposes a Proximal Policy Optimization-based Cascaded Inference System (PPO-CIS), which dynamically assigns classifiers based on performance and computational costs. This system improves efficiency by using high-throughput classifiers for initial filtering and more accurate classifiers for final decisions, reducing the workload on human moderators.
Extensive evaluations on datasets such as Kaggle-Jigsaw and ToxiGen demonstrate significant improvements in processing time, detection accuracy, and overall user satisfaction, contributing to the development of scalable, cost-effective toxicity detection systems for social media platforms.

Divisions:Concordia University > Gina Cody School of Engineering and Computer Science > Concordia Institute for Information Systems Engineering
Item Type:Thesis (PhD)
Authors:Bodaghi, Arezo
Institution:Concordia University
Degree Name:Ph. D.
Program:Information and Systems Engineering
Date:6 September 2024
Thesis Supervisor(s):Schmitt, Ketra A. and Fung, Benjamin C. M.
ID Code:994801
Deposited By: Arezo Bodaghi
Deposited On:17 Jun 2025 14:06
Last Modified:17 Jun 2025 14:06
All items in Spectrum are protected by copyright, with all rights reserved. The use of items is governed by Spectrum's terms of access.

Repository Staff Only: item control page

Downloads per month over past year

Research related to the current document (at the CORE website)
- Research related to the current document (at the CORE website)
Back to top Back to top