Cargando…
Transformer-based deep neural network language models for Alzheimer’s disease risk assessment from targeted speech
BACKGROUND: We developed transformer-based deep learning models based on natural language processing for early risk assessment of Alzheimer’s disease from the picture description test. METHODS: The lack of large datasets poses the most important limitation for using complex models that do not requir...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
BioMed Central
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7971114/ https://www.ncbi.nlm.nih.gov/pubmed/33750385 http://dx.doi.org/10.1186/s12911-021-01456-3 |
_version_ | 1783666555569045504 |
---|---|
author | Roshanzamir, Alireza Aghajan, Hamid Soleymani Baghshah, Mahdieh |
author_facet | Roshanzamir, Alireza Aghajan, Hamid Soleymani Baghshah, Mahdieh |
author_sort | Roshanzamir, Alireza |
collection | PubMed |
description | BACKGROUND: We developed transformer-based deep learning models based on natural language processing for early risk assessment of Alzheimer’s disease from the picture description test. METHODS: The lack of large datasets poses the most important limitation for using complex models that do not require feature engineering. Transformer-based pre-trained deep language models have recently made a large leap in NLP research and application. These models are pre-trained on available large datasets to understand natural language texts appropriately, and are shown to subsequently perform well on classification tasks with small training sets. The overall classification model is a simple classifier on top of the pre-trained deep language model. RESULTS: The models are evaluated on picture description test transcripts of the Pitt corpus, which contains data of 170 AD patients with 257 interviews and 99 healthy controls with 243 interviews. The large bidirectional encoder representations from transformers (BERT(Large)) embedding with logistic regression classifier achieves classification accuracy of 88.08%, which improves the state-of-the-art by 2.48%. CONCLUSIONS: Using pre-trained language models can improve AD prediction. This not only solves the problem of lack of sufficiently large datasets, but also reduces the need for expert-defined features. |
format | Online Article Text |
id | pubmed-7971114 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | BioMed Central |
record_format | MEDLINE/PubMed |
spelling | pubmed-79711142021-03-19 Transformer-based deep neural network language models for Alzheimer’s disease risk assessment from targeted speech Roshanzamir, Alireza Aghajan, Hamid Soleymani Baghshah, Mahdieh BMC Med Inform Decis Mak Research Article BACKGROUND: We developed transformer-based deep learning models based on natural language processing for early risk assessment of Alzheimer’s disease from the picture description test. METHODS: The lack of large datasets poses the most important limitation for using complex models that do not require feature engineering. Transformer-based pre-trained deep language models have recently made a large leap in NLP research and application. These models are pre-trained on available large datasets to understand natural language texts appropriately, and are shown to subsequently perform well on classification tasks with small training sets. The overall classification model is a simple classifier on top of the pre-trained deep language model. RESULTS: The models are evaluated on picture description test transcripts of the Pitt corpus, which contains data of 170 AD patients with 257 interviews and 99 healthy controls with 243 interviews. The large bidirectional encoder representations from transformers (BERT(Large)) embedding with logistic regression classifier achieves classification accuracy of 88.08%, which improves the state-of-the-art by 2.48%. CONCLUSIONS: Using pre-trained language models can improve AD prediction. This not only solves the problem of lack of sufficiently large datasets, but also reduces the need for expert-defined features. BioMed Central 2021-03-09 /pmc/articles/PMC7971114/ /pubmed/33750385 http://dx.doi.org/10.1186/s12911-021-01456-3 Text en © The Author(s) 2021 Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data. |
spellingShingle | Research Article Roshanzamir, Alireza Aghajan, Hamid Soleymani Baghshah, Mahdieh Transformer-based deep neural network language models for Alzheimer’s disease risk assessment from targeted speech |
title | Transformer-based deep neural network language models for Alzheimer’s disease risk assessment from targeted speech |
title_full | Transformer-based deep neural network language models for Alzheimer’s disease risk assessment from targeted speech |
title_fullStr | Transformer-based deep neural network language models for Alzheimer’s disease risk assessment from targeted speech |
title_full_unstemmed | Transformer-based deep neural network language models for Alzheimer’s disease risk assessment from targeted speech |
title_short | Transformer-based deep neural network language models for Alzheimer’s disease risk assessment from targeted speech |
title_sort | transformer-based deep neural network language models for alzheimer’s disease risk assessment from targeted speech |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7971114/ https://www.ncbi.nlm.nih.gov/pubmed/33750385 http://dx.doi.org/10.1186/s12911-021-01456-3 |
work_keys_str_mv | AT roshanzamiralireza transformerbaseddeepneuralnetworklanguagemodelsforalzheimersdiseaseriskassessmentfromtargetedspeech AT aghajanhamid transformerbaseddeepneuralnetworklanguagemodelsforalzheimersdiseaseriskassessmentfromtargetedspeech AT soleymanibaghshahmahdieh transformerbaseddeepneuralnetworklanguagemodelsforalzheimersdiseaseriskassessmentfromtargetedspeech |