Cargando…

Fooling Examples: Another Intriguing Property of Neural Networks

Neural networks have been proven to be vulnerable to adversarial examples; these are examples that can be recognized by both humans and neural networks, although neural networks give incorrect predictions. As an intriguing property of neural networks, adversarial examples pose a serious threat to th...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Ming, Chen, Yongkang, Qian, Cheng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10383212/
https://www.ncbi.nlm.nih.gov/pubmed/37514672
http://dx.doi.org/10.3390/s23146378
_version_ 1785080852247478272
author Zhang, Ming
Chen, Yongkang
Qian, Cheng
author_facet Zhang, Ming
Chen, Yongkang
Qian, Cheng
author_sort Zhang, Ming
collection PubMed
description Neural networks have been proven to be vulnerable to adversarial examples; these are examples that can be recognized by both humans and neural networks, although neural networks give incorrect predictions. As an intriguing property of neural networks, adversarial examples pose a serious threat to the secure application of neural networks. In this article, we present another intriguing property of neural networks: the fact that well-trained models believe some examples to be recognizable objects (often with high confidence), while humans cannot recognize such examples. We refer to these as “fooling examples”. Specifically, we take inspiration from the construction of adversarial examples and develop an iterative method for generating fooling examples. The experimental results show that fooling examples can not only be easily generated, with a success rate of nearly 100% in the white-box scenario, but also exhibit strong transferability across different models in the black-box scenario. Tests on the Google Cloud Vision API show that fooling examples can also be recognized by real-world computer vision systems. Our findings reveal a new cognitive deficit of neural networks, and we hope that these potential security threats will be addressed in future neural network applications.
format Online
Article
Text
id pubmed-10383212
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-103832122023-07-30 Fooling Examples: Another Intriguing Property of Neural Networks Zhang, Ming Chen, Yongkang Qian, Cheng Sensors (Basel) Article Neural networks have been proven to be vulnerable to adversarial examples; these are examples that can be recognized by both humans and neural networks, although neural networks give incorrect predictions. As an intriguing property of neural networks, adversarial examples pose a serious threat to the secure application of neural networks. In this article, we present another intriguing property of neural networks: the fact that well-trained models believe some examples to be recognizable objects (often with high confidence), while humans cannot recognize such examples. We refer to these as “fooling examples”. Specifically, we take inspiration from the construction of adversarial examples and develop an iterative method for generating fooling examples. The experimental results show that fooling examples can not only be easily generated, with a success rate of nearly 100% in the white-box scenario, but also exhibit strong transferability across different models in the black-box scenario. Tests on the Google Cloud Vision API show that fooling examples can also be recognized by real-world computer vision systems. Our findings reveal a new cognitive deficit of neural networks, and we hope that these potential security threats will be addressed in future neural network applications. MDPI 2023-07-13 /pmc/articles/PMC10383212/ /pubmed/37514672 http://dx.doi.org/10.3390/s23146378 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhang, Ming
Chen, Yongkang
Qian, Cheng
Fooling Examples: Another Intriguing Property of Neural Networks
title Fooling Examples: Another Intriguing Property of Neural Networks
title_full Fooling Examples: Another Intriguing Property of Neural Networks
title_fullStr Fooling Examples: Another Intriguing Property of Neural Networks
title_full_unstemmed Fooling Examples: Another Intriguing Property of Neural Networks
title_short Fooling Examples: Another Intriguing Property of Neural Networks
title_sort fooling examples: another intriguing property of neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10383212/
https://www.ncbi.nlm.nih.gov/pubmed/37514672
http://dx.doi.org/10.3390/s23146378
work_keys_str_mv AT zhangming foolingexamplesanotherintriguingpropertyofneuralnetworks
AT chenyongkang foolingexamplesanotherintriguingpropertyofneuralnetworks
AT qiancheng foolingexamplesanotherintriguingpropertyofneuralnetworks