Cargando…

Auditory Inspired Convolutional Neural Networks for Ship Type Classification with Raw Hydrophone Data

Detecting and classifying ships based on radiated noise provide practical guidelines for the reduction of underwater noise footprint of shipping. In this paper, the detection and classification are implemented by auditory inspired convolutional neural networks trained from raw underwater acoustic si...

Descripción completa

Detalles Bibliográficos
Autores principales: Shen, Sheng, Yang, Honghui, Li, Junhao, Xu, Guanghui, Sheng, Meiping
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512589/
https://www.ncbi.nlm.nih.gov/pubmed/33266713
http://dx.doi.org/10.3390/e20120990
_version_ 1783586193012686848
author Shen, Sheng
Yang, Honghui
Li, Junhao
Xu, Guanghui
Sheng, Meiping
author_facet Shen, Sheng
Yang, Honghui
Li, Junhao
Xu, Guanghui
Sheng, Meiping
author_sort Shen, Sheng
collection PubMed
description Detecting and classifying ships based on radiated noise provide practical guidelines for the reduction of underwater noise footprint of shipping. In this paper, the detection and classification are implemented by auditory inspired convolutional neural networks trained from raw underwater acoustic signal. The proposed model includes three parts. The first part is performed by a multi-scale 1D time convolutional layer initialized by auditory filter banks. Signals are decomposed into frequency components by convolution operation. In the second part, the decomposed signals are converted into frequency domain by permute layer and energy pooling layer to form frequency distribution in auditory cortex. Then, 2D frequency convolutional layers are applied to discover spectro-temporal patterns, as well as preserve locality and reduce spectral variations in ship noise. In the third part, the whole model is optimized with an objective function of classification to obtain appropriate auditory filters and feature representations that are correlative with ship categories. The optimization reflects the plasticity of auditory system. Experiments on five ship types and background noise show that the proposed approach achieved an overall classification accuracy of 79.2%, which improved by 6% compared to conventional approaches. Auditory filter banks were adaptive in shape to improve accuracy of classification.
format Online
Article
Text
id pubmed-7512589
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75125892020-11-09 Auditory Inspired Convolutional Neural Networks for Ship Type Classification with Raw Hydrophone Data Shen, Sheng Yang, Honghui Li, Junhao Xu, Guanghui Sheng, Meiping Entropy (Basel) Article Detecting and classifying ships based on radiated noise provide practical guidelines for the reduction of underwater noise footprint of shipping. In this paper, the detection and classification are implemented by auditory inspired convolutional neural networks trained from raw underwater acoustic signal. The proposed model includes three parts. The first part is performed by a multi-scale 1D time convolutional layer initialized by auditory filter banks. Signals are decomposed into frequency components by convolution operation. In the second part, the decomposed signals are converted into frequency domain by permute layer and energy pooling layer to form frequency distribution in auditory cortex. Then, 2D frequency convolutional layers are applied to discover spectro-temporal patterns, as well as preserve locality and reduce spectral variations in ship noise. In the third part, the whole model is optimized with an objective function of classification to obtain appropriate auditory filters and feature representations that are correlative with ship categories. The optimization reflects the plasticity of auditory system. Experiments on five ship types and background noise show that the proposed approach achieved an overall classification accuracy of 79.2%, which improved by 6% compared to conventional approaches. Auditory filter banks were adaptive in shape to improve accuracy of classification. MDPI 2018-12-19 /pmc/articles/PMC7512589/ /pubmed/33266713 http://dx.doi.org/10.3390/e20120990 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Shen, Sheng
Yang, Honghui
Li, Junhao
Xu, Guanghui
Sheng, Meiping
Auditory Inspired Convolutional Neural Networks for Ship Type Classification with Raw Hydrophone Data
title Auditory Inspired Convolutional Neural Networks for Ship Type Classification with Raw Hydrophone Data
title_full Auditory Inspired Convolutional Neural Networks for Ship Type Classification with Raw Hydrophone Data
title_fullStr Auditory Inspired Convolutional Neural Networks for Ship Type Classification with Raw Hydrophone Data
title_full_unstemmed Auditory Inspired Convolutional Neural Networks for Ship Type Classification with Raw Hydrophone Data
title_short Auditory Inspired Convolutional Neural Networks for Ship Type Classification with Raw Hydrophone Data
title_sort auditory inspired convolutional neural networks for ship type classification with raw hydrophone data
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512589/
https://www.ncbi.nlm.nih.gov/pubmed/33266713
http://dx.doi.org/10.3390/e20120990
work_keys_str_mv AT shensheng auditoryinspiredconvolutionalneuralnetworksforshiptypeclassificationwithrawhydrophonedata
AT yanghonghui auditoryinspiredconvolutionalneuralnetworksforshiptypeclassificationwithrawhydrophonedata
AT lijunhao auditoryinspiredconvolutionalneuralnetworksforshiptypeclassificationwithrawhydrophonedata
AT xuguanghui auditoryinspiredconvolutionalneuralnetworksforshiptypeclassificationwithrawhydrophonedata
AT shengmeiping auditoryinspiredconvolutionalneuralnetworksforshiptypeclassificationwithrawhydrophonedata