Cargando…

Recognizing Semantic Relations: Attention-Based Transformers vs. Recurrent Models

Automatically recognizing an existing semantic relation (such as “is a”, “part of”, “property of”, “opposite of” etc.) between two arbitrary words (phrases, concepts, etc.) is an important task affecting many information retrieval and artificial intelligence tasks including query expansion, common-s...

Descripción completa

Detalles Bibliográficos
Autores principales: Roussinov, Dmitri, Sharoff, Serge, Puchnina, Nadezhda
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7148207/
http://dx.doi.org/10.1007/978-3-030-45439-5_37
_version_ 1783520543455051776
author Roussinov, Dmitri
Sharoff, Serge
Puchnina, Nadezhda
author_facet Roussinov, Dmitri
Sharoff, Serge
Puchnina, Nadezhda
author_sort Roussinov, Dmitri
collection PubMed
description Automatically recognizing an existing semantic relation (such as “is a”, “part of”, “property of”, “opposite of” etc.) between two arbitrary words (phrases, concepts, etc.) is an important task affecting many information retrieval and artificial intelligence tasks including query expansion, common-sense reasoning, question answering, and database federation. Currently, two classes of approaches exist to classify a relation between words (concepts) X and Y: (1) path-based and (2) distributional. While the path-based approaches look at word-paths connecting X and Y in text, the distributional approaches look at statistical properties of X and Y separately, not necessary in the proximity of each other. Here, we suggest how both types can be improved and empirically compare them using several standard benchmarking datasets. For our distributional approach, we are suggesting using an attention-based transformer. While they are known to be capable of supporting knowledge transfer between different tasks, and recently set a number of benchmarking records in various applications, we are the first to successfully apply them to the task of recognizing semantic relations. To improve a path-based approach, we are suggesting our original neural word path model that combines useful properties of convolutional and recurrent networks, and thus addressing several shortcomings from the prior path-based models. Both our models significantly outperforms the state-of-the-art within its type accordingly. Our transformer-based approach outperforms current state-of-the-art by 1–12% points on 4 out of 6 standard benchmarking datasets. This results in 15–40% error reduction and is closing the gap between the automated and human performance by up to 50%. It also needs much less training data than prior approaches. For the ease of re-producing our results, we make our source code and trained models publicly available.
format Online
Article
Text
id pubmed-7148207
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-71482072020-04-13 Recognizing Semantic Relations: Attention-Based Transformers vs. Recurrent Models Roussinov, Dmitri Sharoff, Serge Puchnina, Nadezhda Advances in Information Retrieval Article Automatically recognizing an existing semantic relation (such as “is a”, “part of”, “property of”, “opposite of” etc.) between two arbitrary words (phrases, concepts, etc.) is an important task affecting many information retrieval and artificial intelligence tasks including query expansion, common-sense reasoning, question answering, and database federation. Currently, two classes of approaches exist to classify a relation between words (concepts) X and Y: (1) path-based and (2) distributional. While the path-based approaches look at word-paths connecting X and Y in text, the distributional approaches look at statistical properties of X and Y separately, not necessary in the proximity of each other. Here, we suggest how both types can be improved and empirically compare them using several standard benchmarking datasets. For our distributional approach, we are suggesting using an attention-based transformer. While they are known to be capable of supporting knowledge transfer between different tasks, and recently set a number of benchmarking records in various applications, we are the first to successfully apply them to the task of recognizing semantic relations. To improve a path-based approach, we are suggesting our original neural word path model that combines useful properties of convolutional and recurrent networks, and thus addressing several shortcomings from the prior path-based models. Both our models significantly outperforms the state-of-the-art within its type accordingly. Our transformer-based approach outperforms current state-of-the-art by 1–12% points on 4 out of 6 standard benchmarking datasets. This results in 15–40% error reduction and is closing the gap between the automated and human performance by up to 50%. It also needs much less training data than prior approaches. For the ease of re-producing our results, we make our source code and trained models publicly available. 2020-03-17 /pmc/articles/PMC7148207/ http://dx.doi.org/10.1007/978-3-030-45439-5_37 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Roussinov, Dmitri
Sharoff, Serge
Puchnina, Nadezhda
Recognizing Semantic Relations: Attention-Based Transformers vs. Recurrent Models
title Recognizing Semantic Relations: Attention-Based Transformers vs. Recurrent Models
title_full Recognizing Semantic Relations: Attention-Based Transformers vs. Recurrent Models
title_fullStr Recognizing Semantic Relations: Attention-Based Transformers vs. Recurrent Models
title_full_unstemmed Recognizing Semantic Relations: Attention-Based Transformers vs. Recurrent Models
title_short Recognizing Semantic Relations: Attention-Based Transformers vs. Recurrent Models
title_sort recognizing semantic relations: attention-based transformers vs. recurrent models
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7148207/
http://dx.doi.org/10.1007/978-3-030-45439-5_37
work_keys_str_mv AT roussinovdmitri recognizingsemanticrelationsattentionbasedtransformersvsrecurrentmodels
AT sharoffserge recognizingsemanticrelationsattentionbasedtransformersvsrecurrentmodels
AT puchninanadezhda recognizingsemanticrelationsattentionbasedtransformersvsrecurrentmodels