Cargando…
DeepCAC: a deep learning approach on DNA transcription factors classification based on multi-head self-attention and concatenate convolutional neural network
Understanding gene expression processes necessitates the accurate classification and identification of transcription factors, which is supported by high-throughput sequencing technologies. However, these techniques suffer from inherent limitations such as time consumption and high costs. To address...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
BioMed Central
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10506269/ https://www.ncbi.nlm.nih.gov/pubmed/37723425 http://dx.doi.org/10.1186/s12859-023-05469-9 |
_version_ | 1785107086862974976 |
---|---|
author | Zhang, Jidong Liu, Bo Wu, Jiahui Wang, Zhihan Li, Jianqiang |
author_facet | Zhang, Jidong Liu, Bo Wu, Jiahui Wang, Zhihan Li, Jianqiang |
author_sort | Zhang, Jidong |
collection | PubMed |
description | Understanding gene expression processes necessitates the accurate classification and identification of transcription factors, which is supported by high-throughput sequencing technologies. However, these techniques suffer from inherent limitations such as time consumption and high costs. To address these challenges, the field of bioinformatics has increasingly turned to deep learning technologies for analyzing gene sequences. Nevertheless, the pursuit of improved experimental results has led to the inclusion of numerous complex analysis function modules, resulting in models with a growing number of parameters. To overcome these limitations, it is proposed a novel approach for analyzing DNA transcription factor sequences, which is named as DeepCAC. This method leverages deep convolutional neural networks with a multi-head self-attention mechanism. By employing convolutional neural networks, it can effectively capture local hidden features in the sequences. Simultaneously, the multi-head self-attention mechanism enhances the identification of hidden features with long-distant dependencies. This approach reduces the overall number of parameters in the model while harnessing the computational power of sequence data from multi-head self-attention. Through training with labeled data, experiments demonstrate that this approach significantly improves performance while requiring fewer parameters compared to existing methods. Additionally, the effectiveness of our approach is validated in accurately predicting DNA transcription factor sequences. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1186/s12859-023-05469-9. |
format | Online Article Text |
id | pubmed-10506269 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | BioMed Central |
record_format | MEDLINE/PubMed |
spelling | pubmed-105062692023-09-19 DeepCAC: a deep learning approach on DNA transcription factors classification based on multi-head self-attention and concatenate convolutional neural network Zhang, Jidong Liu, Bo Wu, Jiahui Wang, Zhihan Li, Jianqiang BMC Bioinformatics Research Understanding gene expression processes necessitates the accurate classification and identification of transcription factors, which is supported by high-throughput sequencing technologies. However, these techniques suffer from inherent limitations such as time consumption and high costs. To address these challenges, the field of bioinformatics has increasingly turned to deep learning technologies for analyzing gene sequences. Nevertheless, the pursuit of improved experimental results has led to the inclusion of numerous complex analysis function modules, resulting in models with a growing number of parameters. To overcome these limitations, it is proposed a novel approach for analyzing DNA transcription factor sequences, which is named as DeepCAC. This method leverages deep convolutional neural networks with a multi-head self-attention mechanism. By employing convolutional neural networks, it can effectively capture local hidden features in the sequences. Simultaneously, the multi-head self-attention mechanism enhances the identification of hidden features with long-distant dependencies. This approach reduces the overall number of parameters in the model while harnessing the computational power of sequence data from multi-head self-attention. Through training with labeled data, experiments demonstrate that this approach significantly improves performance while requiring fewer parameters compared to existing methods. Additionally, the effectiveness of our approach is validated in accurately predicting DNA transcription factor sequences. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1186/s12859-023-05469-9. BioMed Central 2023-09-18 /pmc/articles/PMC10506269/ /pubmed/37723425 http://dx.doi.org/10.1186/s12859-023-05469-9 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/ (https://creativecommons.org/publicdomain/zero/1.0/) ) applies to the data made available in this article, unless otherwise stated in a credit line to the data. |
spellingShingle | Research Zhang, Jidong Liu, Bo Wu, Jiahui Wang, Zhihan Li, Jianqiang DeepCAC: a deep learning approach on DNA transcription factors classification based on multi-head self-attention and concatenate convolutional neural network |
title | DeepCAC: a deep learning approach on DNA transcription factors classification based on multi-head self-attention and concatenate convolutional neural network |
title_full | DeepCAC: a deep learning approach on DNA transcription factors classification based on multi-head self-attention and concatenate convolutional neural network |
title_fullStr | DeepCAC: a deep learning approach on DNA transcription factors classification based on multi-head self-attention and concatenate convolutional neural network |
title_full_unstemmed | DeepCAC: a deep learning approach on DNA transcription factors classification based on multi-head self-attention and concatenate convolutional neural network |
title_short | DeepCAC: a deep learning approach on DNA transcription factors classification based on multi-head self-attention and concatenate convolutional neural network |
title_sort | deepcac: a deep learning approach on dna transcription factors classification based on multi-head self-attention and concatenate convolutional neural network |
topic | Research |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10506269/ https://www.ncbi.nlm.nih.gov/pubmed/37723425 http://dx.doi.org/10.1186/s12859-023-05469-9 |
work_keys_str_mv | AT zhangjidong deepcacadeeplearningapproachondnatranscriptionfactorsclassificationbasedonmultiheadselfattentionandconcatenateconvolutionalneuralnetwork AT liubo deepcacadeeplearningapproachondnatranscriptionfactorsclassificationbasedonmultiheadselfattentionandconcatenateconvolutionalneuralnetwork AT wujiahui deepcacadeeplearningapproachondnatranscriptionfactorsclassificationbasedonmultiheadselfattentionandconcatenateconvolutionalneuralnetwork AT wangzhihan deepcacadeeplearningapproachondnatranscriptionfactorsclassificationbasedonmultiheadselfattentionandconcatenateconvolutionalneuralnetwork AT lijianqiang deepcacadeeplearningapproachondnatranscriptionfactorsclassificationbasedonmultiheadselfattentionandconcatenateconvolutionalneuralnetwork |