Cargando…
Beyond the Transformer: A Novel Polynomial Inherent Attention (PIA) Model and Its Great Impact on Neural Machine Translation
This paper describes a novel polynomial inherent attention (PIA) model that outperforms all state-of-the-art transformer models on neural machine translation (NMT) by a wide margin. PIA is based on the simple idea that natural language sentences can be transformed into a special type of binary atten...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9519290/ https://www.ncbi.nlm.nih.gov/pubmed/36188704 http://dx.doi.org/10.1155/2022/1912750 |
_version_ | 1784799362645229568 |
---|---|
author | ELAffendi, Mohammed Alrajhi, Khawlah |
author_facet | ELAffendi, Mohammed Alrajhi, Khawlah |
author_sort | ELAffendi, Mohammed |
collection | PubMed |
description | This paper describes a novel polynomial inherent attention (PIA) model that outperforms all state-of-the-art transformer models on neural machine translation (NMT) by a wide margin. PIA is based on the simple idea that natural language sentences can be transformed into a special type of binary attention context vectors that accurately capture the semantic context and the relative dependencies between words in a sentence. The transformation is performed using a simple power-of-two polynomial transformation that maintains strict consistent positioning of words in the resulting vectors. It is shown how this transformation reduces the neural machine translation process to a simple neural polynomial regression model that provides excellent solutions to the alignment and positioning problems haunting transformer models. The test BELU scores obtained on the WMT-2014 data set are 75.07 BELU for the EN-FR data set and 66.35 BELU for the EN-DE data set—well above accuracies achieved by state-of-the-art transformer models for the same data sets. The improvements are, respectively, 65.7% and 87.42%. |
format | Online Article Text |
id | pubmed-9519290 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-95192902022-09-29 Beyond the Transformer: A Novel Polynomial Inherent Attention (PIA) Model and Its Great Impact on Neural Machine Translation ELAffendi, Mohammed Alrajhi, Khawlah Comput Intell Neurosci Research Article This paper describes a novel polynomial inherent attention (PIA) model that outperforms all state-of-the-art transformer models on neural machine translation (NMT) by a wide margin. PIA is based on the simple idea that natural language sentences can be transformed into a special type of binary attention context vectors that accurately capture the semantic context and the relative dependencies between words in a sentence. The transformation is performed using a simple power-of-two polynomial transformation that maintains strict consistent positioning of words in the resulting vectors. It is shown how this transformation reduces the neural machine translation process to a simple neural polynomial regression model that provides excellent solutions to the alignment and positioning problems haunting transformer models. The test BELU scores obtained on the WMT-2014 data set are 75.07 BELU for the EN-FR data set and 66.35 BELU for the EN-DE data set—well above accuracies achieved by state-of-the-art transformer models for the same data sets. The improvements are, respectively, 65.7% and 87.42%. Hindawi 2022-09-21 /pmc/articles/PMC9519290/ /pubmed/36188704 http://dx.doi.org/10.1155/2022/1912750 Text en Copyright © 2022 Mohammed ELAffendi and Khawlah Alrajhi. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article ELAffendi, Mohammed Alrajhi, Khawlah Beyond the Transformer: A Novel Polynomial Inherent Attention (PIA) Model and Its Great Impact on Neural Machine Translation |
title | Beyond the Transformer: A Novel Polynomial Inherent Attention (PIA) Model and Its Great Impact on Neural Machine Translation |
title_full | Beyond the Transformer: A Novel Polynomial Inherent Attention (PIA) Model and Its Great Impact on Neural Machine Translation |
title_fullStr | Beyond the Transformer: A Novel Polynomial Inherent Attention (PIA) Model and Its Great Impact on Neural Machine Translation |
title_full_unstemmed | Beyond the Transformer: A Novel Polynomial Inherent Attention (PIA) Model and Its Great Impact on Neural Machine Translation |
title_short | Beyond the Transformer: A Novel Polynomial Inherent Attention (PIA) Model and Its Great Impact on Neural Machine Translation |
title_sort | beyond the transformer: a novel polynomial inherent attention (pia) model and its great impact on neural machine translation |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9519290/ https://www.ncbi.nlm.nih.gov/pubmed/36188704 http://dx.doi.org/10.1155/2022/1912750 |
work_keys_str_mv | AT elaffendimohammed beyondthetransformeranovelpolynomialinherentattentionpiamodelanditsgreatimpactonneuralmachinetranslation AT alrajhikhawlah beyondthetransformeranovelpolynomialinherentattentionpiamodelanditsgreatimpactonneuralmachinetranslation |