TY - JOUR
T1 - Assessing Text Classification Methods for Cyberbullying Detection on Social Media Platforms
AU - Philipo, Adamu Gaston
AU - Sebastian Sarwatt, Doreen
AU - Ding, Jianguo
AU - Daneshmand, Mahmoud
AU - Ning, Huansheng
N1 - Publisher Copyright:
© IEEE. 2005-2012 IEEE.
PY - 2025
Y1 - 2025
N2 - Cyberbullying significantly impacts mental health by adversely affecting victims' psychological well-being. It is a prevalent issue on social media platforms, necessitating effective real-time detection systems to identify harmful content. However, current detection systems face challenges related to performance, dataset quality, time efficiency, and computational costs. This study compares existing text classification techniques for cyberbullying detection, evaluating their effectiveness on social media platforms. Large language models such as BERT, RoBERTa, XLNet, DistilBERT, and GPT-2.0 are assessed for their suitability. Results show that BERT achieves optimal performance, with 95% accuracy, precision, recall, and F1 score; a 5% error rate; 0.053 seconds inference time; 35.28 MB RAM usage; 0.4% CPU/GPU utilization; and 0.000263 kWh energy consumption. These findings highlight that while generative AI models are powerful, fine-tuned models often outperform them when adapted to specific datasets and tasks.
AB - Cyberbullying significantly impacts mental health by adversely affecting victims' psychological well-being. It is a prevalent issue on social media platforms, necessitating effective real-time detection systems to identify harmful content. However, current detection systems face challenges related to performance, dataset quality, time efficiency, and computational costs. This study compares existing text classification techniques for cyberbullying detection, evaluating their effectiveness on social media platforms. Large language models such as BERT, RoBERTa, XLNet, DistilBERT, and GPT-2.0 are assessed for their suitability. Results show that BERT achieves optimal performance, with 95% accuracy, precision, recall, and F1 score; a 5% error rate; 0.053 seconds inference time; 35.28 MB RAM usage; 0.4% CPU/GPU utilization; and 0.000263 kWh energy consumption. These findings highlight that while generative AI models are powerful, fine-tuned models often outperform them when adapted to specific datasets and tasks.
KW - Cyberbullying instances
KW - detection methods
KW - social media platforms
KW - text classification
UR - https://www.scopus.com/pages/publications/105010910942
UR - https://www.scopus.com/pages/publications/105010910942#tab=citedBy
U2 - 10.1109/TIFS.2025.3588728
DO - 10.1109/TIFS.2025.3588728
M3 - Article
AN - SCOPUS:105010910942
SN - 1556-6013
VL - 20
SP - 7602
EP - 7616
JO - IEEE Transactions on Information Forensics and Security
JF - IEEE Transactions on Information Forensics and Security
ER -