Cargando…
Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model
The identification of drug–drug and chemical–protein interactions is essential for understanding unpredictable changes in the pharmacological effects of drugs and mechanisms of diseases and developing therapeutic drugs. In this study, we extract drug-related interactions from the DDI (Drug–Drug Inte...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10215465/ https://www.ncbi.nlm.nih.gov/pubmed/37237656 http://dx.doi.org/10.3390/bioengineering10050586 |
_version_ | 1785048069602017280 |
---|---|
author | Kim, Seonho Yoon, Juntae Kwon, Ohyoung |
author_facet | Kim, Seonho Yoon, Juntae Kwon, Ohyoung |
author_sort | Kim, Seonho |
collection | PubMed |
description | The identification of drug–drug and chemical–protein interactions is essential for understanding unpredictable changes in the pharmacological effects of drugs and mechanisms of diseases and developing therapeutic drugs. In this study, we extract drug-related interactions from the DDI (Drug–Drug Interaction) Extraction-2013 Shared Task dataset and the BioCreative ChemProt (Chemical–Protein) dataset using various transfer transformers. We propose BERT(GAT) that uses a graph attention network (GAT) to take into account the local structure of sentences and embedding features of nodes under the self-attention scheme and investigate whether incorporating syntactic structure can help relation extraction. In addition, we suggest T5(slim_dec), which adapts the autoregressive generation task of the T5 (text-to-text transfer transformer) to the relation classification problem by removing the self-attention layer in the decoder block. Furthermore, we evaluated the potential of biomedical relation extraction of GPT-3 (Generative Pre-trained Transformer) using GPT-3 variant models. As a result, T5(slim_dec), which is a model with a tailored decoder designed for classification problems within the T5 architecture, demonstrated very promising performances for both tasks. We achieved an accuracy of 91.15% in the DDI dataset and an accuracy of 94.29% for the CPR (Chemical–Protein Relation) class group in ChemProt dataset. However, BERT(GAT) did not show a significant performance improvement in the aspect of relation extraction. We demonstrated that transformer-based approaches focused only on relationships between words are implicitly eligible to understand language well without additional knowledge such as structural information. |
format | Online Article Text |
id | pubmed-10215465 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-102154652023-05-27 Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model Kim, Seonho Yoon, Juntae Kwon, Ohyoung Bioengineering (Basel) Article The identification of drug–drug and chemical–protein interactions is essential for understanding unpredictable changes in the pharmacological effects of drugs and mechanisms of diseases and developing therapeutic drugs. In this study, we extract drug-related interactions from the DDI (Drug–Drug Interaction) Extraction-2013 Shared Task dataset and the BioCreative ChemProt (Chemical–Protein) dataset using various transfer transformers. We propose BERT(GAT) that uses a graph attention network (GAT) to take into account the local structure of sentences and embedding features of nodes under the self-attention scheme and investigate whether incorporating syntactic structure can help relation extraction. In addition, we suggest T5(slim_dec), which adapts the autoregressive generation task of the T5 (text-to-text transfer transformer) to the relation classification problem by removing the self-attention layer in the decoder block. Furthermore, we evaluated the potential of biomedical relation extraction of GPT-3 (Generative Pre-trained Transformer) using GPT-3 variant models. As a result, T5(slim_dec), which is a model with a tailored decoder designed for classification problems within the T5 architecture, demonstrated very promising performances for both tasks. We achieved an accuracy of 91.15% in the DDI dataset and an accuracy of 94.29% for the CPR (Chemical–Protein Relation) class group in ChemProt dataset. However, BERT(GAT) did not show a significant performance improvement in the aspect of relation extraction. We demonstrated that transformer-based approaches focused only on relationships between words are implicitly eligible to understand language well without additional knowledge such as structural information. MDPI 2023-05-12 /pmc/articles/PMC10215465/ /pubmed/37237656 http://dx.doi.org/10.3390/bioengineering10050586 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Kim, Seonho Yoon, Juntae Kwon, Ohyoung Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model |
title | Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model |
title_full | Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model |
title_fullStr | Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model |
title_full_unstemmed | Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model |
title_short | Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model |
title_sort | biomedical relation extraction using dependency graph and decoder-enhanced transformer model |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10215465/ https://www.ncbi.nlm.nih.gov/pubmed/37237656 http://dx.doi.org/10.3390/bioengineering10050586 |
work_keys_str_mv | AT kimseonho biomedicalrelationextractionusingdependencygraphanddecoderenhancedtransformermodel AT yoonjuntae biomedicalrelationextractionusingdependencygraphanddecoderenhancedtransformermodel AT kwonohyoung biomedicalrelationextractionusingdependencygraphanddecoderenhancedtransformermodel |