Cargando…

Attention-Based Sequence-to-Sequence Model for Time Series Imputation

Time series data are usually characterized by having missing values, high dimensionality, and large data volume. To solve the problem of high-dimensional time series with missing values, this paper proposes an attention-based sequence-to-sequence model to imputation missing values in time series (AS...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Yurui, Du, Mingjing, He, Sheng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9778091/
https://www.ncbi.nlm.nih.gov/pubmed/36554203
http://dx.doi.org/10.3390/e24121798
_version_ 1784856271361409024
author Li, Yurui
Du, Mingjing
He, Sheng
author_facet Li, Yurui
Du, Mingjing
He, Sheng
author_sort Li, Yurui
collection PubMed
description Time series data are usually characterized by having missing values, high dimensionality, and large data volume. To solve the problem of high-dimensional time series with missing values, this paper proposes an attention-based sequence-to-sequence model to imputation missing values in time series (ASSM), which is a sequence-to-sequence model based on the combination of feature learning and data computation. The model consists of two parts, encoder and decoder. The encoder part is a BIGRU recurrent neural network and incorporates a self-attentive mechanism to make the model more capable of handling long-range time series; The decoder part is a GRU recurrent neural network and incorporates a cross-attentive mechanism into associate with the encoder part. The relationship weights between the generated sequences in the decoder part and the known sequences in the encoder part are calculated to achieve the purpose of focusing on the sequences with a high degree of correlation. In this paper, we conduct comparison experiments with four evaluation metrics and six models on four real datasets. The experimental results show that the model proposed in this paper outperforms the six comparative missing value interpolation algorithms.
format Online
Article
Text
id pubmed-9778091
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-97780912022-12-23 Attention-Based Sequence-to-Sequence Model for Time Series Imputation Li, Yurui Du, Mingjing He, Sheng Entropy (Basel) Article Time series data are usually characterized by having missing values, high dimensionality, and large data volume. To solve the problem of high-dimensional time series with missing values, this paper proposes an attention-based sequence-to-sequence model to imputation missing values in time series (ASSM), which is a sequence-to-sequence model based on the combination of feature learning and data computation. The model consists of two parts, encoder and decoder. The encoder part is a BIGRU recurrent neural network and incorporates a self-attentive mechanism to make the model more capable of handling long-range time series; The decoder part is a GRU recurrent neural network and incorporates a cross-attentive mechanism into associate with the encoder part. The relationship weights between the generated sequences in the decoder part and the known sequences in the encoder part are calculated to achieve the purpose of focusing on the sequences with a high degree of correlation. In this paper, we conduct comparison experiments with four evaluation metrics and six models on four real datasets. The experimental results show that the model proposed in this paper outperforms the six comparative missing value interpolation algorithms. MDPI 2022-12-09 /pmc/articles/PMC9778091/ /pubmed/36554203 http://dx.doi.org/10.3390/e24121798 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Li, Yurui
Du, Mingjing
He, Sheng
Attention-Based Sequence-to-Sequence Model for Time Series Imputation
title Attention-Based Sequence-to-Sequence Model for Time Series Imputation
title_full Attention-Based Sequence-to-Sequence Model for Time Series Imputation
title_fullStr Attention-Based Sequence-to-Sequence Model for Time Series Imputation
title_full_unstemmed Attention-Based Sequence-to-Sequence Model for Time Series Imputation
title_short Attention-Based Sequence-to-Sequence Model for Time Series Imputation
title_sort attention-based sequence-to-sequence model for time series imputation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9778091/
https://www.ncbi.nlm.nih.gov/pubmed/36554203
http://dx.doi.org/10.3390/e24121798
work_keys_str_mv AT liyurui attentionbasedsequencetosequencemodelfortimeseriesimputation
AT dumingjing attentionbasedsequencetosequencemodelfortimeseriesimputation
AT hesheng attentionbasedsequencetosequencemodelfortimeseriesimputation