Cargando…

Relation is an option for processing context information

Attention mechanisms are one of the most frequently used architectures in the development of artificial intelligence because they can process contextual information efficiently. Various artificial intelligence architectures, such as Transformer for processing natural language, image data, etc., incl...

Descripción completa

Detalles Bibliográficos
Autores principales: Yamada, Kazunori D, Baladram, M. Samy, Lin, Fangzhou
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9592710/
https://www.ncbi.nlm.nih.gov/pubmed/36304959
http://dx.doi.org/10.3389/frai.2022.924688
_version_ 1784814988494372864
author Yamada, Kazunori D
Baladram, M. Samy
Lin, Fangzhou
author_facet Yamada, Kazunori D
Baladram, M. Samy
Lin, Fangzhou
author_sort Yamada, Kazunori D
collection PubMed
description Attention mechanisms are one of the most frequently used architectures in the development of artificial intelligence because they can process contextual information efficiently. Various artificial intelligence architectures, such as Transformer for processing natural language, image data, etc., include the Attention. Various improvements have been made to enhance its performance since Attention is a powerful component to realize artificial intelligence. The time complexity of Attention depends on the square of the input sequence length. Developing methods to improve the time complexity of Attention is one of the most popular research topics. Attention is a mechanism that conveys contextual information of input sequences to downstream networks. Thus, if one wants to improve the performance of processing contextual information, the focus should not be confined only on improving Attention but also on devising other similar mechanisms as possible alternatives. In this study, we devised an alternative mechanism called “Relation” that can understand the context information of sequential data. Relation is easy to implement, and its time complexity depends only on the length of the sequences; a comparison of the performance of Relation and Attention on several benchmark datasets showed that the context processing capability of Relation is comparable to that of Attention but with less computation time. Processing contextual information at high speeds would be useful because natural language processing and biological sequence processing sometimes deal with very long sequences. Hence, Relation is an ideal option for processing context information.
format Online
Article
Text
id pubmed-9592710
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-95927102022-10-26 Relation is an option for processing context information Yamada, Kazunori D Baladram, M. Samy Lin, Fangzhou Front Artif Intell Artificial Intelligence Attention mechanisms are one of the most frequently used architectures in the development of artificial intelligence because they can process contextual information efficiently. Various artificial intelligence architectures, such as Transformer for processing natural language, image data, etc., include the Attention. Various improvements have been made to enhance its performance since Attention is a powerful component to realize artificial intelligence. The time complexity of Attention depends on the square of the input sequence length. Developing methods to improve the time complexity of Attention is one of the most popular research topics. Attention is a mechanism that conveys contextual information of input sequences to downstream networks. Thus, if one wants to improve the performance of processing contextual information, the focus should not be confined only on improving Attention but also on devising other similar mechanisms as possible alternatives. In this study, we devised an alternative mechanism called “Relation” that can understand the context information of sequential data. Relation is easy to implement, and its time complexity depends only on the length of the sequences; a comparison of the performance of Relation and Attention on several benchmark datasets showed that the context processing capability of Relation is comparable to that of Attention but with less computation time. Processing contextual information at high speeds would be useful because natural language processing and biological sequence processing sometimes deal with very long sequences. Hence, Relation is an ideal option for processing context information. Frontiers Media S.A. 2022-10-11 /pmc/articles/PMC9592710/ /pubmed/36304959 http://dx.doi.org/10.3389/frai.2022.924688 Text en Copyright © 2022 Yamada, Baladram and Lin. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Yamada, Kazunori D
Baladram, M. Samy
Lin, Fangzhou
Relation is an option for processing context information
title Relation is an option for processing context information
title_full Relation is an option for processing context information
title_fullStr Relation is an option for processing context information
title_full_unstemmed Relation is an option for processing context information
title_short Relation is an option for processing context information
title_sort relation is an option for processing context information
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9592710/
https://www.ncbi.nlm.nih.gov/pubmed/36304959
http://dx.doi.org/10.3389/frai.2022.924688
work_keys_str_mv AT yamadakazunorid relationisanoptionforprocessingcontextinformation
AT baladrammsamy relationisanoptionforprocessingcontextinformation
AT linfangzhou relationisanoptionforprocessingcontextinformation