Cargando…

A Comparative Analysis between Efficient Attention Mechanisms for Traffic Forecasting without Structural Priors

Dot-product attention is a powerful mechanism for capturing contextual information. Models that build on top of it have acclaimed state-of-the-art performance in various domains, ranging from sequence modelling to visual tasks. However, the main bottleneck is the construction of the attention map, w...

Descripción completa

Detalles Bibliográficos
Autores principales: Rad, Andrei-Cristian, Lemnaru, Camelia, Munteanu, Adrian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9571981/
https://www.ncbi.nlm.nih.gov/pubmed/36236555
http://dx.doi.org/10.3390/s22197457
_version_ 1784810499538419712
author Rad, Andrei-Cristian
Lemnaru, Camelia
Munteanu, Adrian
author_facet Rad, Andrei-Cristian
Lemnaru, Camelia
Munteanu, Adrian
author_sort Rad, Andrei-Cristian
collection PubMed
description Dot-product attention is a powerful mechanism for capturing contextual information. Models that build on top of it have acclaimed state-of-the-art performance in various domains, ranging from sequence modelling to visual tasks. However, the main bottleneck is the construction of the attention map, which is quadratic with respect to the number of tokens in the sequence. Consequently, efficient alternatives have been developed in parallel, but it was only recently that their performances were compared and contrasted. This study performs a comparative analysis between some efficient attention mechanisms in the context of a purely attention-based spatio-temporal forecasting model used for traffic prediction. Experiments show that these methods can reduce the training times by up to 28% and the inference times by up to 31%, while the performance remains on par with the baseline.
format Online
Article
Text
id pubmed-9571981
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-95719812022-10-17 A Comparative Analysis between Efficient Attention Mechanisms for Traffic Forecasting without Structural Priors Rad, Andrei-Cristian Lemnaru, Camelia Munteanu, Adrian Sensors (Basel) Article Dot-product attention is a powerful mechanism for capturing contextual information. Models that build on top of it have acclaimed state-of-the-art performance in various domains, ranging from sequence modelling to visual tasks. However, the main bottleneck is the construction of the attention map, which is quadratic with respect to the number of tokens in the sequence. Consequently, efficient alternatives have been developed in parallel, but it was only recently that their performances were compared and contrasted. This study performs a comparative analysis between some efficient attention mechanisms in the context of a purely attention-based spatio-temporal forecasting model used for traffic prediction. Experiments show that these methods can reduce the training times by up to 28% and the inference times by up to 31%, while the performance remains on par with the baseline. MDPI 2022-10-01 /pmc/articles/PMC9571981/ /pubmed/36236555 http://dx.doi.org/10.3390/s22197457 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Rad, Andrei-Cristian
Lemnaru, Camelia
Munteanu, Adrian
A Comparative Analysis between Efficient Attention Mechanisms for Traffic Forecasting without Structural Priors
title A Comparative Analysis between Efficient Attention Mechanisms for Traffic Forecasting without Structural Priors
title_full A Comparative Analysis between Efficient Attention Mechanisms for Traffic Forecasting without Structural Priors
title_fullStr A Comparative Analysis between Efficient Attention Mechanisms for Traffic Forecasting without Structural Priors
title_full_unstemmed A Comparative Analysis between Efficient Attention Mechanisms for Traffic Forecasting without Structural Priors
title_short A Comparative Analysis between Efficient Attention Mechanisms for Traffic Forecasting without Structural Priors
title_sort comparative analysis between efficient attention mechanisms for traffic forecasting without structural priors
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9571981/
https://www.ncbi.nlm.nih.gov/pubmed/36236555
http://dx.doi.org/10.3390/s22197457
work_keys_str_mv AT radandreicristian acomparativeanalysisbetweenefficientattentionmechanismsfortrafficforecastingwithoutstructuralpriors
AT lemnarucamelia acomparativeanalysisbetweenefficientattentionmechanismsfortrafficforecastingwithoutstructuralpriors
AT munteanuadrian acomparativeanalysisbetweenefficientattentionmechanismsfortrafficforecastingwithoutstructuralpriors
AT radandreicristian comparativeanalysisbetweenefficientattentionmechanismsfortrafficforecastingwithoutstructuralpriors
AT lemnarucamelia comparativeanalysisbetweenefficientattentionmechanismsfortrafficforecastingwithoutstructuralpriors
AT munteanuadrian comparativeanalysisbetweenefficientattentionmechanismsfortrafficforecastingwithoutstructuralpriors