Cargando…

Low-Pass Image Filtering to Achieve Adversarial Robustness

In this paper, we continue the research cycle on the properties of convolutional neural network-based image recognition systems and ways to improve noise immunity and robustness. Currently, a popular research area related to artificial neural networks is adversarial attacks. The adversarial attacks...

Descripción completa

Detalles Bibliográficos
Autores principales: Ziyadinov, Vadim, Tereshonok, Maxim
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10675189/
https://www.ncbi.nlm.nih.gov/pubmed/38005420
http://dx.doi.org/10.3390/s23229032
_version_ 1785141005622706176
author Ziyadinov, Vadim
Tereshonok, Maxim
author_facet Ziyadinov, Vadim
Tereshonok, Maxim
author_sort Ziyadinov, Vadim
collection PubMed
description In this paper, we continue the research cycle on the properties of convolutional neural network-based image recognition systems and ways to improve noise immunity and robustness. Currently, a popular research area related to artificial neural networks is adversarial attacks. The adversarial attacks on the image are not highly perceptible to the human eye, and they also drastically reduce the neural network’s accuracy. Image perception by a machine is highly dependent on the propagation of high frequency distortions throughout the network. At the same time, a human efficiently ignores high-frequency distortions, perceiving the shape of objects as a whole. We propose a technique to reduce the influence of high-frequency noise on the CNNs. We show that low-pass image filtering can improve the image recognition accuracy in the presence of high-frequency distortions in particular, caused by adversarial attacks. This technique is resource efficient and easy to implement. The proposed technique makes it possible to measure up the logic of an artificial neural network to that of a human, for whom high-frequency distortions are not decisive in object recognition.
format Online
Article
Text
id pubmed-10675189
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-106751892023-11-07 Low-Pass Image Filtering to Achieve Adversarial Robustness Ziyadinov, Vadim Tereshonok, Maxim Sensors (Basel) Article In this paper, we continue the research cycle on the properties of convolutional neural network-based image recognition systems and ways to improve noise immunity and robustness. Currently, a popular research area related to artificial neural networks is adversarial attacks. The adversarial attacks on the image are not highly perceptible to the human eye, and they also drastically reduce the neural network’s accuracy. Image perception by a machine is highly dependent on the propagation of high frequency distortions throughout the network. At the same time, a human efficiently ignores high-frequency distortions, perceiving the shape of objects as a whole. We propose a technique to reduce the influence of high-frequency noise on the CNNs. We show that low-pass image filtering can improve the image recognition accuracy in the presence of high-frequency distortions in particular, caused by adversarial attacks. This technique is resource efficient and easy to implement. The proposed technique makes it possible to measure up the logic of an artificial neural network to that of a human, for whom high-frequency distortions are not decisive in object recognition. MDPI 2023-11-07 /pmc/articles/PMC10675189/ /pubmed/38005420 http://dx.doi.org/10.3390/s23229032 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Ziyadinov, Vadim
Tereshonok, Maxim
Low-Pass Image Filtering to Achieve Adversarial Robustness
title Low-Pass Image Filtering to Achieve Adversarial Robustness
title_full Low-Pass Image Filtering to Achieve Adversarial Robustness
title_fullStr Low-Pass Image Filtering to Achieve Adversarial Robustness
title_full_unstemmed Low-Pass Image Filtering to Achieve Adversarial Robustness
title_short Low-Pass Image Filtering to Achieve Adversarial Robustness
title_sort low-pass image filtering to achieve adversarial robustness
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10675189/
https://www.ncbi.nlm.nih.gov/pubmed/38005420
http://dx.doi.org/10.3390/s23229032
work_keys_str_mv AT ziyadinovvadim lowpassimagefilteringtoachieveadversarialrobustness
AT tereshonokmaxim lowpassimagefilteringtoachieveadversarialrobustness