Cargando…
Electromagnetic Modulation Signal Classification Using Dual-Modal Feature Fusion CNN
AMC (automatic modulation classification) plays a vital role in spectrum monitoring and electromagnetic abnormal signal detection. Up to now, few studies have focused on the complementarity between features of different modalities and the importance of the feature fusion mechanism in the AMC method....
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9142120/ https://www.ncbi.nlm.nih.gov/pubmed/35626583 http://dx.doi.org/10.3390/e24050700 |
_version_ | 1784715505252171776 |
---|---|
author | Bai, Jiansheng Yao, Jinjie Qi, Juncheng Wang, Liming |
author_facet | Bai, Jiansheng Yao, Jinjie Qi, Juncheng Wang, Liming |
author_sort | Bai, Jiansheng |
collection | PubMed |
description | AMC (automatic modulation classification) plays a vital role in spectrum monitoring and electromagnetic abnormal signal detection. Up to now, few studies have focused on the complementarity between features of different modalities and the importance of the feature fusion mechanism in the AMC method. This paper proposes a dual-modal feature fusion convolutional neural network (DMFF-CNN) for AMC to use the complementarity between different modal features fully. DMFF-CNN uses the gram angular field (GAF) image coding and intelligence quotient (IQ) data combined with CNN. Firstly, the original signal is converted into images by GAF, and the GAF images are used as the input of ResNet50. Secondly, it is converted into IQ data and as the complex value network (CV-CNN) input to extract features. Furthermore, a dual-modal feature fusion mechanism (DMFF) is proposed to fuse the dual-modal features extracted by GAF-ResNet50 and CV-CNN. The fusion feature is used as the input of DMFF-CNN for model training to achieve AMC of multi-type signals. In the evaluation stage, the advantages of the DMFF mechanism proposed in this paper and the accuracy improvement compared with other feature fusion algorithms are discussed. The experiment shows that our method performs better than others, including some state-of-the-art methods, and has superior robustness at a low signal-to-noise ratio (SNR), and the average classification accuracy of the dataset signals reaches 92.1%. The DMFF-CNN proposed in this paper provides a new path for the AMC field. |
format | Online Article Text |
id | pubmed-9142120 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-91421202022-05-28 Electromagnetic Modulation Signal Classification Using Dual-Modal Feature Fusion CNN Bai, Jiansheng Yao, Jinjie Qi, Juncheng Wang, Liming Entropy (Basel) Article AMC (automatic modulation classification) plays a vital role in spectrum monitoring and electromagnetic abnormal signal detection. Up to now, few studies have focused on the complementarity between features of different modalities and the importance of the feature fusion mechanism in the AMC method. This paper proposes a dual-modal feature fusion convolutional neural network (DMFF-CNN) for AMC to use the complementarity between different modal features fully. DMFF-CNN uses the gram angular field (GAF) image coding and intelligence quotient (IQ) data combined with CNN. Firstly, the original signal is converted into images by GAF, and the GAF images are used as the input of ResNet50. Secondly, it is converted into IQ data and as the complex value network (CV-CNN) input to extract features. Furthermore, a dual-modal feature fusion mechanism (DMFF) is proposed to fuse the dual-modal features extracted by GAF-ResNet50 and CV-CNN. The fusion feature is used as the input of DMFF-CNN for model training to achieve AMC of multi-type signals. In the evaluation stage, the advantages of the DMFF mechanism proposed in this paper and the accuracy improvement compared with other feature fusion algorithms are discussed. The experiment shows that our method performs better than others, including some state-of-the-art methods, and has superior robustness at a low signal-to-noise ratio (SNR), and the average classification accuracy of the dataset signals reaches 92.1%. The DMFF-CNN proposed in this paper provides a new path for the AMC field. MDPI 2022-05-15 /pmc/articles/PMC9142120/ /pubmed/35626583 http://dx.doi.org/10.3390/e24050700 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Bai, Jiansheng Yao, Jinjie Qi, Juncheng Wang, Liming Electromagnetic Modulation Signal Classification Using Dual-Modal Feature Fusion CNN |
title | Electromagnetic Modulation Signal Classification Using Dual-Modal Feature Fusion CNN |
title_full | Electromagnetic Modulation Signal Classification Using Dual-Modal Feature Fusion CNN |
title_fullStr | Electromagnetic Modulation Signal Classification Using Dual-Modal Feature Fusion CNN |
title_full_unstemmed | Electromagnetic Modulation Signal Classification Using Dual-Modal Feature Fusion CNN |
title_short | Electromagnetic Modulation Signal Classification Using Dual-Modal Feature Fusion CNN |
title_sort | electromagnetic modulation signal classification using dual-modal feature fusion cnn |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9142120/ https://www.ncbi.nlm.nih.gov/pubmed/35626583 http://dx.doi.org/10.3390/e24050700 |
work_keys_str_mv | AT baijiansheng electromagneticmodulationsignalclassificationusingdualmodalfeaturefusioncnn AT yaojinjie electromagneticmodulationsignalclassificationusingdualmodalfeaturefusioncnn AT qijuncheng electromagneticmodulationsignalclassificationusingdualmodalfeaturefusioncnn AT wangliming electromagneticmodulationsignalclassificationusingdualmodalfeaturefusioncnn |