Cargando…

Enhancing Offensive Language Detection with Data Augmentation and Knowledge Distillation

Offensive language detection has received important attention and plays a crucial role in promoting healthy communication on social platforms, as well as promoting the safe deployment of large language models. Training data is the basis for developing detectors; however, the available offense-relate...

Descripción completa

Detalles Bibliográficos
Autores principales: Deng, Jiawen, Chen, Zhuang, Sun, Hao, Zhang, Zhexin, Wu, Jincenzi, Nakagawa, Satoshi, Ren, Fuji, Huang, Minlie
Formato: Online Artículo Texto
Lenguaje:English
Publicado: AAAS 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10506735/
https://www.ncbi.nlm.nih.gov/pubmed/37727321
http://dx.doi.org/10.34133/research.0189
_version_ 1785107166121689088
author Deng, Jiawen
Chen, Zhuang
Sun, Hao
Zhang, Zhexin
Wu, Jincenzi
Nakagawa, Satoshi
Ren, Fuji
Huang, Minlie
author_facet Deng, Jiawen
Chen, Zhuang
Sun, Hao
Zhang, Zhexin
Wu, Jincenzi
Nakagawa, Satoshi
Ren, Fuji
Huang, Minlie
author_sort Deng, Jiawen
collection PubMed
description Offensive language detection has received important attention and plays a crucial role in promoting healthy communication on social platforms, as well as promoting the safe deployment of large language models. Training data is the basis for developing detectors; however, the available offense-related dataset in Chinese is severely limited in terms of data scale and coverage when compared to English resources. This significantly affects the accuracy of Chinese offensive language detectors in practical applications, especially when dealing with hard cases or out-of-domain samples. To alleviate the limitations posed by available datasets, we introduce AugCOLD (Augmented Chinese Offensive Language Dataset), a large-scale unsupervised dataset containing 1 million samples gathered by data crawling and model generation. Furthermore, we employ a multiteacher distillation framework to enhance detection performance with unsupervised data. That is, we build multiple teachers with publicly accessible datasets and use them to assign soft labels to AugCOLD. The soft labels serve as a bridge for knowledge to be distilled from both AugCOLD and multiteacher to the student network, i.e., the final offensive detector. We conduct experiments on multiple public test sets and our well-designed hard tests, demonstrating that our proposal can effectively improve the generalization and robustness of the offensive language detector.
format Online
Article
Text
id pubmed-10506735
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher AAAS
record_format MEDLINE/PubMed
spelling pubmed-105067352023-09-19 Enhancing Offensive Language Detection with Data Augmentation and Knowledge Distillation Deng, Jiawen Chen, Zhuang Sun, Hao Zhang, Zhexin Wu, Jincenzi Nakagawa, Satoshi Ren, Fuji Huang, Minlie Research (Wash D C) Research Article Offensive language detection has received important attention and plays a crucial role in promoting healthy communication on social platforms, as well as promoting the safe deployment of large language models. Training data is the basis for developing detectors; however, the available offense-related dataset in Chinese is severely limited in terms of data scale and coverage when compared to English resources. This significantly affects the accuracy of Chinese offensive language detectors in practical applications, especially when dealing with hard cases or out-of-domain samples. To alleviate the limitations posed by available datasets, we introduce AugCOLD (Augmented Chinese Offensive Language Dataset), a large-scale unsupervised dataset containing 1 million samples gathered by data crawling and model generation. Furthermore, we employ a multiteacher distillation framework to enhance detection performance with unsupervised data. That is, we build multiple teachers with publicly accessible datasets and use them to assign soft labels to AugCOLD. The soft labels serve as a bridge for knowledge to be distilled from both AugCOLD and multiteacher to the student network, i.e., the final offensive detector. We conduct experiments on multiple public test sets and our well-designed hard tests, demonstrating that our proposal can effectively improve the generalization and robustness of the offensive language detector. AAAS 2023-09-18 /pmc/articles/PMC10506735/ /pubmed/37727321 http://dx.doi.org/10.34133/research.0189 Text en Copyright © 2023 Jiawen Deng et al. https://creativecommons.org/licenses/by/4.0/Exclusive licensee Science and Technology Review Publishing House. No claim to original U.S. Government Works. Distributed under a Creative Commons Attribution License 4.0 (CC BY 4.0) (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Research Article
Deng, Jiawen
Chen, Zhuang
Sun, Hao
Zhang, Zhexin
Wu, Jincenzi
Nakagawa, Satoshi
Ren, Fuji
Huang, Minlie
Enhancing Offensive Language Detection with Data Augmentation and Knowledge Distillation
title Enhancing Offensive Language Detection with Data Augmentation and Knowledge Distillation
title_full Enhancing Offensive Language Detection with Data Augmentation and Knowledge Distillation
title_fullStr Enhancing Offensive Language Detection with Data Augmentation and Knowledge Distillation
title_full_unstemmed Enhancing Offensive Language Detection with Data Augmentation and Knowledge Distillation
title_short Enhancing Offensive Language Detection with Data Augmentation and Knowledge Distillation
title_sort enhancing offensive language detection with data augmentation and knowledge distillation
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10506735/
https://www.ncbi.nlm.nih.gov/pubmed/37727321
http://dx.doi.org/10.34133/research.0189
work_keys_str_mv AT dengjiawen enhancingoffensivelanguagedetectionwithdataaugmentationandknowledgedistillation
AT chenzhuang enhancingoffensivelanguagedetectionwithdataaugmentationandknowledgedistillation
AT sunhao enhancingoffensivelanguagedetectionwithdataaugmentationandknowledgedistillation
AT zhangzhexin enhancingoffensivelanguagedetectionwithdataaugmentationandknowledgedistillation
AT wujincenzi enhancingoffensivelanguagedetectionwithdataaugmentationandknowledgedistillation
AT nakagawasatoshi enhancingoffensivelanguagedetectionwithdataaugmentationandknowledgedistillation
AT renfuji enhancingoffensivelanguagedetectionwithdataaugmentationandknowledgedistillation
AT huangminlie enhancingoffensivelanguagedetectionwithdataaugmentationandknowledgedistillation