Cargando…

Pay attention and you won’t lose it: a deep learning approach to sequence imputation

In most areas of machine learning, it is assumed that data quality is fairly consistent between training and inference. Unfortunately, in real systems, data are plagued by noise, loss, and various other quality reducing factors. While a number of deep learning algorithms solve end-stage problems of...

Descripción completa

Detalles Bibliográficos
Autores principales: Sucholutsky, Ilia, Narayan, Apurva, Schonlau, Matthias, Fischmeister, Sebastian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7924680/
https://www.ncbi.nlm.nih.gov/pubmed/33816863
http://dx.doi.org/10.7717/peerj-cs.210
_version_ 1783659140033282048
author Sucholutsky, Ilia
Narayan, Apurva
Schonlau, Matthias
Fischmeister, Sebastian
author_facet Sucholutsky, Ilia
Narayan, Apurva
Schonlau, Matthias
Fischmeister, Sebastian
author_sort Sucholutsky, Ilia
collection PubMed
description In most areas of machine learning, it is assumed that data quality is fairly consistent between training and inference. Unfortunately, in real systems, data are plagued by noise, loss, and various other quality reducing factors. While a number of deep learning algorithms solve end-stage problems of prediction and classification, very few aim to solve the intermediate problems of data pre-processing, cleaning, and restoration. Long Short-Term Memory (LSTM) networks have previously been proposed as a solution for data restoration, but they suffer from a major bottleneck: a large number of sequential operations. We propose using attention mechanisms to entirely replace the recurrent components of these data-restoration networks. We demonstrate that such an approach leads to reduced model sizes by as many as two orders of magnitude, a 2-fold to 4-fold reduction in training times, and 95% accuracy for automotive data restoration. We also show in a case study that this approach improves the performance of downstream algorithms reliant on clean data.
format Online
Article
Text
id pubmed-7924680
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-79246802021-04-02 Pay attention and you won’t lose it: a deep learning approach to sequence imputation Sucholutsky, Ilia Narayan, Apurva Schonlau, Matthias Fischmeister, Sebastian PeerJ Comput Sci Algorithms and Analysis of Algorithms In most areas of machine learning, it is assumed that data quality is fairly consistent between training and inference. Unfortunately, in real systems, data are plagued by noise, loss, and various other quality reducing factors. While a number of deep learning algorithms solve end-stage problems of prediction and classification, very few aim to solve the intermediate problems of data pre-processing, cleaning, and restoration. Long Short-Term Memory (LSTM) networks have previously been proposed as a solution for data restoration, but they suffer from a major bottleneck: a large number of sequential operations. We propose using attention mechanisms to entirely replace the recurrent components of these data-restoration networks. We demonstrate that such an approach leads to reduced model sizes by as many as two orders of magnitude, a 2-fold to 4-fold reduction in training times, and 95% accuracy for automotive data restoration. We also show in a case study that this approach improves the performance of downstream algorithms reliant on clean data. PeerJ Inc. 2019-08-12 /pmc/articles/PMC7924680/ /pubmed/33816863 http://dx.doi.org/10.7717/peerj-cs.210 Text en ©2019 Sucholutsky et al. https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
spellingShingle Algorithms and Analysis of Algorithms
Sucholutsky, Ilia
Narayan, Apurva
Schonlau, Matthias
Fischmeister, Sebastian
Pay attention and you won’t lose it: a deep learning approach to sequence imputation
title Pay attention and you won’t lose it: a deep learning approach to sequence imputation
title_full Pay attention and you won’t lose it: a deep learning approach to sequence imputation
title_fullStr Pay attention and you won’t lose it: a deep learning approach to sequence imputation
title_full_unstemmed Pay attention and you won’t lose it: a deep learning approach to sequence imputation
title_short Pay attention and you won’t lose it: a deep learning approach to sequence imputation
title_sort pay attention and you won’t lose it: a deep learning approach to sequence imputation
topic Algorithms and Analysis of Algorithms
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7924680/
https://www.ncbi.nlm.nih.gov/pubmed/33816863
http://dx.doi.org/10.7717/peerj-cs.210
work_keys_str_mv AT sucholutskyilia payattentionandyouwontloseitadeeplearningapproachtosequenceimputation
AT narayanapurva payattentionandyouwontloseitadeeplearningapproachtosequenceimputation
AT schonlaumatthias payattentionandyouwontloseitadeeplearningapproachtosequenceimputation
AT fischmeistersebastian payattentionandyouwontloseitadeeplearningapproachtosequenceimputation