Cargando…

Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings

This work presents a comprehensive approach to reduce bias in word embedding vectors and evaluate the impact on various Natural Language Processing (NLP) tasks. Two GloVe variations (840B and 50) are debiased by identifying the gender direction in the word embedding space and then removing or reduci...

Descripción completa

Detalles Bibliográficos
Autores principales: Ioannides, Georgios, Jadhav, Aishwarya, Sharma, Aditi, Navali, Samarth, Black, Alan W.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10613201/
https://www.ncbi.nlm.nih.gov/pubmed/37898713
http://dx.doi.org/10.1038/s41598-023-45677-0
_version_ 1785128778632003584
author Ioannides, Georgios
Jadhav, Aishwarya
Sharma, Aditi
Navali, Samarth
Black, Alan W.
author_facet Ioannides, Georgios
Jadhav, Aishwarya
Sharma, Aditi
Navali, Samarth
Black, Alan W.
author_sort Ioannides, Georgios
collection PubMed
description This work presents a comprehensive approach to reduce bias in word embedding vectors and evaluate the impact on various Natural Language Processing (NLP) tasks. Two GloVe variations (840B and 50) are debiased by identifying the gender direction in the word embedding space and then removing or reducing the gender component from the embeddings of target words, while preserving useful semantic information. Their gender bias is assessed through the Word Embedding Association Test. The performance of co-reference resolution and text classification models trained on both original and debiased embeddings is evaluated in terms of accuracy. A compressed co-reference resolution model is examined to gauge the effectiveness of debiasing techniques on resource-efficient models. To the best of the authors’ knowledge, this is the first attempt to apply compression techniques to debiased models. By analyzing the context preservation of debiased embeddings using a Twitter misinformation dataset, this study contributes valuable insights into the practical implications of debiasing methods for real-world applications such as person profiling.
format Online
Article
Text
id pubmed-10613201
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-106132012023-10-30 Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings Ioannides, Georgios Jadhav, Aishwarya Sharma, Aditi Navali, Samarth Black, Alan W. Sci Rep Article This work presents a comprehensive approach to reduce bias in word embedding vectors and evaluate the impact on various Natural Language Processing (NLP) tasks. Two GloVe variations (840B and 50) are debiased by identifying the gender direction in the word embedding space and then removing or reducing the gender component from the embeddings of target words, while preserving useful semantic information. Their gender bias is assessed through the Word Embedding Association Test. The performance of co-reference resolution and text classification models trained on both original and debiased embeddings is evaluated in terms of accuracy. A compressed co-reference resolution model is examined to gauge the effectiveness of debiasing techniques on resource-efficient models. To the best of the authors’ knowledge, this is the first attempt to apply compression techniques to debiased models. By analyzing the context preservation of debiased embeddings using a Twitter misinformation dataset, this study contributes valuable insights into the practical implications of debiasing methods for real-world applications such as person profiling. Nature Publishing Group UK 2023-10-28 /pmc/articles/PMC10613201/ /pubmed/37898713 http://dx.doi.org/10.1038/s41598-023-45677-0 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Ioannides, Georgios
Jadhav, Aishwarya
Sharma, Aditi
Navali, Samarth
Black, Alan W.
Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_full Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_fullStr Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_full_unstemmed Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_short Compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
title_sort compressed models for co-reference resolution: enhancing efficiency with debiased word embeddings
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10613201/
https://www.ncbi.nlm.nih.gov/pubmed/37898713
http://dx.doi.org/10.1038/s41598-023-45677-0
work_keys_str_mv AT ioannidesgeorgios compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings
AT jadhavaishwarya compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings
AT sharmaaditi compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings
AT navalisamarth compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings
AT blackalanw compressedmodelsforcoreferenceresolutionenhancingefficiencywithdebiasedwordembeddings