Cargando…
Relation is an option for processing context information
Attention mechanisms are one of the most frequently used architectures in the development of artificial intelligence because they can process contextual information efficiently. Various artificial intelligence architectures, such as Transformer for processing natural language, image data, etc., incl...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9592710/ https://www.ncbi.nlm.nih.gov/pubmed/36304959 http://dx.doi.org/10.3389/frai.2022.924688 |
_version_ | 1784814988494372864 |
---|---|
author | Yamada, Kazunori D Baladram, M. Samy Lin, Fangzhou |
author_facet | Yamada, Kazunori D Baladram, M. Samy Lin, Fangzhou |
author_sort | Yamada, Kazunori D |
collection | PubMed |
description | Attention mechanisms are one of the most frequently used architectures in the development of artificial intelligence because they can process contextual information efficiently. Various artificial intelligence architectures, such as Transformer for processing natural language, image data, etc., include the Attention. Various improvements have been made to enhance its performance since Attention is a powerful component to realize artificial intelligence. The time complexity of Attention depends on the square of the input sequence length. Developing methods to improve the time complexity of Attention is one of the most popular research topics. Attention is a mechanism that conveys contextual information of input sequences to downstream networks. Thus, if one wants to improve the performance of processing contextual information, the focus should not be confined only on improving Attention but also on devising other similar mechanisms as possible alternatives. In this study, we devised an alternative mechanism called “Relation” that can understand the context information of sequential data. Relation is easy to implement, and its time complexity depends only on the length of the sequences; a comparison of the performance of Relation and Attention on several benchmark datasets showed that the context processing capability of Relation is comparable to that of Attention but with less computation time. Processing contextual information at high speeds would be useful because natural language processing and biological sequence processing sometimes deal with very long sequences. Hence, Relation is an ideal option for processing context information. |
format | Online Article Text |
id | pubmed-9592710 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-95927102022-10-26 Relation is an option for processing context information Yamada, Kazunori D Baladram, M. Samy Lin, Fangzhou Front Artif Intell Artificial Intelligence Attention mechanisms are one of the most frequently used architectures in the development of artificial intelligence because they can process contextual information efficiently. Various artificial intelligence architectures, such as Transformer for processing natural language, image data, etc., include the Attention. Various improvements have been made to enhance its performance since Attention is a powerful component to realize artificial intelligence. The time complexity of Attention depends on the square of the input sequence length. Developing methods to improve the time complexity of Attention is one of the most popular research topics. Attention is a mechanism that conveys contextual information of input sequences to downstream networks. Thus, if one wants to improve the performance of processing contextual information, the focus should not be confined only on improving Attention but also on devising other similar mechanisms as possible alternatives. In this study, we devised an alternative mechanism called “Relation” that can understand the context information of sequential data. Relation is easy to implement, and its time complexity depends only on the length of the sequences; a comparison of the performance of Relation and Attention on several benchmark datasets showed that the context processing capability of Relation is comparable to that of Attention but with less computation time. Processing contextual information at high speeds would be useful because natural language processing and biological sequence processing sometimes deal with very long sequences. Hence, Relation is an ideal option for processing context information. Frontiers Media S.A. 2022-10-11 /pmc/articles/PMC9592710/ /pubmed/36304959 http://dx.doi.org/10.3389/frai.2022.924688 Text en Copyright © 2022 Yamada, Baladram and Lin. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Artificial Intelligence Yamada, Kazunori D Baladram, M. Samy Lin, Fangzhou Relation is an option for processing context information |
title | Relation is an option for processing context information |
title_full | Relation is an option for processing context information |
title_fullStr | Relation is an option for processing context information |
title_full_unstemmed | Relation is an option for processing context information |
title_short | Relation is an option for processing context information |
title_sort | relation is an option for processing context information |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9592710/ https://www.ncbi.nlm.nih.gov/pubmed/36304959 http://dx.doi.org/10.3389/frai.2022.924688 |
work_keys_str_mv | AT yamadakazunorid relationisanoptionforprocessingcontextinformation AT baladrammsamy relationisanoptionforprocessingcontextinformation AT linfangzhou relationisanoptionforprocessingcontextinformation |