Cargando…

Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model

The identification of drug–drug and chemical–protein interactions is essential for understanding unpredictable changes in the pharmacological effects of drugs and mechanisms of diseases and developing therapeutic drugs. In this study, we extract drug-related interactions from the DDI (Drug–Drug Inte...

Descripción completa

Detalles Bibliográficos
Autores principales: Kim, Seonho, Yoon, Juntae, Kwon, Ohyoung
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10215465/
https://www.ncbi.nlm.nih.gov/pubmed/37237656
http://dx.doi.org/10.3390/bioengineering10050586
_version_ 1785048069602017280
author Kim, Seonho
Yoon, Juntae
Kwon, Ohyoung
author_facet Kim, Seonho
Yoon, Juntae
Kwon, Ohyoung
author_sort Kim, Seonho
collection PubMed
description The identification of drug–drug and chemical–protein interactions is essential for understanding unpredictable changes in the pharmacological effects of drugs and mechanisms of diseases and developing therapeutic drugs. In this study, we extract drug-related interactions from the DDI (Drug–Drug Interaction) Extraction-2013 Shared Task dataset and the BioCreative ChemProt (Chemical–Protein) dataset using various transfer transformers. We propose BERT(GAT) that uses a graph attention network (GAT) to take into account the local structure of sentences and embedding features of nodes under the self-attention scheme and investigate whether incorporating syntactic structure can help relation extraction. In addition, we suggest T5(slim_dec), which adapts the autoregressive generation task of the T5 (text-to-text transfer transformer) to the relation classification problem by removing the self-attention layer in the decoder block. Furthermore, we evaluated the potential of biomedical relation extraction of GPT-3 (Generative Pre-trained Transformer) using GPT-3 variant models. As a result, T5(slim_dec), which is a model with a tailored decoder designed for classification problems within the T5 architecture, demonstrated very promising performances for both tasks. We achieved an accuracy of 91.15% in the DDI dataset and an accuracy of 94.29% for the CPR (Chemical–Protein Relation) class group in ChemProt dataset. However, BERT(GAT) did not show a significant performance improvement in the aspect of relation extraction. We demonstrated that transformer-based approaches focused only on relationships between words are implicitly eligible to understand language well without additional knowledge such as structural information.
format Online
Article
Text
id pubmed-10215465
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-102154652023-05-27 Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model Kim, Seonho Yoon, Juntae Kwon, Ohyoung Bioengineering (Basel) Article The identification of drug–drug and chemical–protein interactions is essential for understanding unpredictable changes in the pharmacological effects of drugs and mechanisms of diseases and developing therapeutic drugs. In this study, we extract drug-related interactions from the DDI (Drug–Drug Interaction) Extraction-2013 Shared Task dataset and the BioCreative ChemProt (Chemical–Protein) dataset using various transfer transformers. We propose BERT(GAT) that uses a graph attention network (GAT) to take into account the local structure of sentences and embedding features of nodes under the self-attention scheme and investigate whether incorporating syntactic structure can help relation extraction. In addition, we suggest T5(slim_dec), which adapts the autoregressive generation task of the T5 (text-to-text transfer transformer) to the relation classification problem by removing the self-attention layer in the decoder block. Furthermore, we evaluated the potential of biomedical relation extraction of GPT-3 (Generative Pre-trained Transformer) using GPT-3 variant models. As a result, T5(slim_dec), which is a model with a tailored decoder designed for classification problems within the T5 architecture, demonstrated very promising performances for both tasks. We achieved an accuracy of 91.15% in the DDI dataset and an accuracy of 94.29% for the CPR (Chemical–Protein Relation) class group in ChemProt dataset. However, BERT(GAT) did not show a significant performance improvement in the aspect of relation extraction. We demonstrated that transformer-based approaches focused only on relationships between words are implicitly eligible to understand language well without additional knowledge such as structural information. MDPI 2023-05-12 /pmc/articles/PMC10215465/ /pubmed/37237656 http://dx.doi.org/10.3390/bioengineering10050586 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kim, Seonho
Yoon, Juntae
Kwon, Ohyoung
Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model
title Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model
title_full Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model
title_fullStr Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model
title_full_unstemmed Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model
title_short Biomedical Relation Extraction Using Dependency Graph and Decoder-Enhanced Transformer Model
title_sort biomedical relation extraction using dependency graph and decoder-enhanced transformer model
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10215465/
https://www.ncbi.nlm.nih.gov/pubmed/37237656
http://dx.doi.org/10.3390/bioengineering10050586
work_keys_str_mv AT kimseonho biomedicalrelationextractionusingdependencygraphanddecoderenhancedtransformermodel
AT yoonjuntae biomedicalrelationextractionusingdependencygraphanddecoderenhancedtransformermodel
AT kwonohyoung biomedicalrelationextractionusingdependencygraphanddecoderenhancedtransformermodel