OBJECT-ORIENTED APPROACH TO NEURAL NETWORK-BASED DETECTION OF CYBERBULLYING SUBJECTS FROM MESSAGES IN A MANAGED CLOUD ENVIRONMENT
Abstract
The aim of the work is to develop and substantiate an object-oriented approach to neural network detection of cyber- bullying subjects from messages with a combination of primary detection and subsequent syntactic-semantic interpretation in a managed cloud environment. A coherent architecture is proposed in which the neural network module filters messages at the “cyberbullying / non-cyberbullying” level, after which the results undergo dependency analysis with the reconstruction of role relationships of the “subject – action – object” type. An object model, which includes classes of messages, sentences, tokens, dependencies, predicates, role triples and summary structures, is defined as the basis for ensuring transparent traceability of decisions, while a managed cloud environment ensures the reproducibility of launches and scalability of experiments. The effectiveness of the initial detection was experimentally confirmed: the BERT-based module demonstrated a metric of F1 = 0.98 in a two-class setting (“cyberbullying” / “not cyberbullying”), which indicates a sufficient level of screening out irrel- evant messages before role analysis. Consistent indicators of the quality of role identification were established on the expert-ver- ified subset: for the subject, values of 0.88, 0.86, 0.87 were obtained for Precision, Recall and F1, respectively; for the object – 0.85, 0.83, 0.84; for the verb center – 0.91, 0.89, 0.90. The exact restoration of the role triple provided a value of F1 = 0.76. The inter-expert agreement was Cohen’s coefficient of 0.82 with 87 % complete agreement, which indicates the reliability of the reference labels and the correctness of the applied evaluation procedure. In controversial cases, a third auxiliary assessment by a language model with a fixed instruction was used; the final labels were determined by majority rule. The results obtained demonstrate that the proposed approach not only allows to determine the fact of aggressive commu- nication, but also provides a structured presentation of information about its addressability, remaining reproducible and audit- able in practical conditions. Conclusions are drawn that create a basis for further integration of the approach into moderation systems and possible expansion to corpora with more detailed role markup and multilingual support.
References
2. Paul S., Saha S., Hasanuzzaman M. Identification of cyberbullying: A deep learning based multimodal approach. Multimedia Tools and Applications. 2020. https://doi.org/10.1007/s11042-020-09631-w (дата звернення: 07.11.2025).
3. Human Activity Recognition for the Identification of Bullying and Cyberbullying Using Smartphone Sensors / V. Gattulli et al. Electronics. 2023. Vol. 12, no. 2. P. 261. https://doi.org/10.3390/electronics12020261 (дата звернення: 07.11.2025).
4. Method for neural network cyberbullying detection in text content with visual analytic / I. Krak et al. CEUR Workshop Proceedings. 2025. Vol. 3917, PP. 298–309. URL: https://ceur-ws.org/Vol-3917/paper57.pdf (дата звернення: 07.11.2025).
5. Method for cyberbullying neuronetwork detection using cloud services and object-oriented model / М. Молчанова та ін. Herald of Khmelnytskyi National University. Technical sciences. 2024. Vol. 333, no. 2. P. 200–206. https://doi.org/10.31891/2307-5732-2024-333-2 (дата звернення: 07.11.2025).
6. Verma R., Kumar K., Verma H. K. Code smell prioritization in object‐oriented software systems: A systematic literature review. Journal of Software: Evolution and Process. 2023. https://doi.org/10.1002/smr.2536 (дата звернення: 07.11.2025).
7. Load Balancing in cloud Environment: A State of-the-Art Review / Y. Lohumi et al. IEEE Access. 2023. P. 1. https://doi.org/10.1109/access.2023.3337146 (дата звернення: 07.11.2025).
8. OffensEval 2023: Offensive language identification in the age of Large Language Models / M. Zampieri et al. Natural Language Engineering. 2023. Vol. 29, no. 6. P. 1416–1435. https://doi.org/10.1017/ s1351324923000517 (дата звернення: 07.11.2025).
9. SemEval-2021 Task 5: Toxic Spans Detection / J. Pavlopoulos et al. Proceedings of the 15th International Workshop on Semantic Evaluation (SemEval-2021), Online. Stroudsburg, PA, USA, 2021. https://doi.org/10.18653/ v1/2021.semeval-1.6 (дата звернення: 07.11.2025).
10. Sihab-Us-Sakib S., Rahman M. R., Forhad M. S. A., Aziz M. A. Cyberbullying detection of resource constrained language from social media using transformer-based approach. Natural Language Processing Journal. 2024. Vol. 9. Article No. 100104. https://doi.org/10.1016/j.nlp.2024.100104 (дата звернення: 07.11.2025).
11. Aliyeva Ç. O., Yağanoğlu M. Deep learning approach to detect cyberbullying on twitter. Multimedia Tools and Applications. 2024. https://doi.org/10.1007/s11042-024-19869-3 (дата звернення: 07.11.2025).
12. Yi P., Zubiaga A. Session-based cyberbullying detection in social media: A survey. Online Social Networks and Media. 2023. Vol. 36. P. 100250. https://doi.org/10.1016/j.osnem.2023.100250 (дата звернення: 07.11.2025).
13. HateXplain: A Benchmark Dataset for Explainable Hate Speech Detection / B. Mathew et al. Proceedings of the AAAI Conference on Artificial Intelligence. 2021. Vol. 35, no. 17. P. 14867–14875. https://doi.org/10.1609/ aaai.v35i17.17745 (дата звернення: 07.11.2025).
14. Hate-speech-CNERG/hatexplain · Datasets at Hugging Face. Hugging Face – The AI community building the future. URL: https://huggingface.co/datasets/Hate-speech-CNERG/hatexplain (дата звернення: 07.11.2025).
15. Cyberbullying Classification. Kaggle. URL: https://www.kaggle.com/datasets/andrewmvd/cyberbullying- classification (дата звернення: 07.11.2025).
16. Cyberbullying Detection. Kaggle. URL: https://www.kaggle.com/datasets/gbiamgaurav/cyberbullying- detection (дата звернення: 07.11.2025).
17. GPT-5. OpenAI. URL: https://openai.com/gpt-5/ (дата звернення: 07.11.2025).
18. BERT applications in natural language processing: a review / N. M. Gardazi et al. Artificial Intelligence Review. 2025. Vol. 58, no. 6. https://doi.org/10.1007/s10462-025-11162-5 (дата звернення: 07.11.2025).
ISSN 



