Cargando…

Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks

Natural language understanding (NLU) has made massive progress driven by large benchmarks, but benchmarks often leave a long tail of infrequent phenomena underrepresented. We reflect on the question: Have transfer learning methods sufficiently addressed the poor performance of benchmark-trained mode...

Descripción completa

Detalles Bibliográficos
Autores principales: Naik, Aakanksha, Lehman, Jill, Rosé, Carolyn
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MIT Press 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9590102/
https://www.ncbi.nlm.nih.gov/pubmed/36303892
http://dx.doi.org/10.1162/tacl_a_00500
_version_ 1784814442907697152
author Naik, Aakanksha
Lehman, Jill
Rosé, Carolyn
author_facet Naik, Aakanksha
Lehman, Jill
Rosé, Carolyn
author_sort Naik, Aakanksha
collection PubMed
description Natural language understanding (NLU) has made massive progress driven by large benchmarks, but benchmarks often leave a long tail of infrequent phenomena underrepresented. We reflect on the question: Have transfer learning methods sufficiently addressed the poor performance of benchmark-trained models on the long tail? We conceptualize the long tail using macro-level dimensions (underrepresented genres, topics, etc.), and perform a qualitative meta-analysis of 100 representative papers on transfer learning research for NLU. Our analysis asks three questions: (i) Which long tail dimensions do transfer learning studies target? (ii) Which properties of adaptation methods help improve performance on the long tail? (iii) Which methodological gaps have greatest negative impact on long tail performance? Our answers highlight major avenues for future research in transfer learning for the long tail. Lastly, using our meta-analysis framework, we perform a case study comparing the performance of various adaptation methods on clinical narratives, which provides interesting insights that may enable us to make progress along these future avenues.
format Online
Article
Text
id pubmed-9590102
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MIT Press
record_format MEDLINE/PubMed
spelling pubmed-95901022022-10-25 Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks Naik, Aakanksha Lehman, Jill Rosé, Carolyn Trans Assoc Comput Linguist Research Article Natural language understanding (NLU) has made massive progress driven by large benchmarks, but benchmarks often leave a long tail of infrequent phenomena underrepresented. We reflect on the question: Have transfer learning methods sufficiently addressed the poor performance of benchmark-trained models on the long tail? We conceptualize the long tail using macro-level dimensions (underrepresented genres, topics, etc.), and perform a qualitative meta-analysis of 100 representative papers on transfer learning research for NLU. Our analysis asks three questions: (i) Which long tail dimensions do transfer learning studies target? (ii) Which properties of adaptation methods help improve performance on the long tail? (iii) Which methodological gaps have greatest negative impact on long tail performance? Our answers highlight major avenues for future research in transfer learning for the long tail. Lastly, using our meta-analysis framework, we perform a case study comparing the performance of various adaptation methods on clinical narratives, which provides interesting insights that may enable us to make progress along these future avenues. MIT Press 2022-09-07 /pmc/articles/PMC9590102/ /pubmed/36303892 http://dx.doi.org/10.1162/tacl_a_00500 Text en © 2022 Association for Computational Linguistics https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. For a full description of the license, please visit https://creativecommons.org/licenses/by/4.0/.
spellingShingle Research Article
Naik, Aakanksha
Lehman, Jill
Rosé, Carolyn
Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks
title Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks
title_full Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks
title_fullStr Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks
title_full_unstemmed Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks
title_short Adapting to the Long Tail: A Meta-Analysis of Transfer Learning Research for Language Understanding Tasks
title_sort adapting to the long tail: a meta-analysis of transfer learning research for language understanding tasks
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9590102/
https://www.ncbi.nlm.nih.gov/pubmed/36303892
http://dx.doi.org/10.1162/tacl_a_00500
work_keys_str_mv AT naikaakanksha adaptingtothelongtailametaanalysisoftransferlearningresearchforlanguageunderstandingtasks
AT lehmanjill adaptingtothelongtailametaanalysisoftransferlearningresearchforlanguageunderstandingtasks
AT rosecarolyn adaptingtothelongtailametaanalysisoftransferlearningresearchforlanguageunderstandingtasks