Cargando…
Fooling Examples: Another Intriguing Property of Neural Networks
Neural networks have been proven to be vulnerable to adversarial examples; these are examples that can be recognized by both humans and neural networks, although neural networks give incorrect predictions. As an intriguing property of neural networks, adversarial examples pose a serious threat to th...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10383212/ https://www.ncbi.nlm.nih.gov/pubmed/37514672 http://dx.doi.org/10.3390/s23146378 |
_version_ | 1785080852247478272 |
---|---|
author | Zhang, Ming Chen, Yongkang Qian, Cheng |
author_facet | Zhang, Ming Chen, Yongkang Qian, Cheng |
author_sort | Zhang, Ming |
collection | PubMed |
description | Neural networks have been proven to be vulnerable to adversarial examples; these are examples that can be recognized by both humans and neural networks, although neural networks give incorrect predictions. As an intriguing property of neural networks, adversarial examples pose a serious threat to the secure application of neural networks. In this article, we present another intriguing property of neural networks: the fact that well-trained models believe some examples to be recognizable objects (often with high confidence), while humans cannot recognize such examples. We refer to these as “fooling examples”. Specifically, we take inspiration from the construction of adversarial examples and develop an iterative method for generating fooling examples. The experimental results show that fooling examples can not only be easily generated, with a success rate of nearly 100% in the white-box scenario, but also exhibit strong transferability across different models in the black-box scenario. Tests on the Google Cloud Vision API show that fooling examples can also be recognized by real-world computer vision systems. Our findings reveal a new cognitive deficit of neural networks, and we hope that these potential security threats will be addressed in future neural network applications. |
format | Online Article Text |
id | pubmed-10383212 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-103832122023-07-30 Fooling Examples: Another Intriguing Property of Neural Networks Zhang, Ming Chen, Yongkang Qian, Cheng Sensors (Basel) Article Neural networks have been proven to be vulnerable to adversarial examples; these are examples that can be recognized by both humans and neural networks, although neural networks give incorrect predictions. As an intriguing property of neural networks, adversarial examples pose a serious threat to the secure application of neural networks. In this article, we present another intriguing property of neural networks: the fact that well-trained models believe some examples to be recognizable objects (often with high confidence), while humans cannot recognize such examples. We refer to these as “fooling examples”. Specifically, we take inspiration from the construction of adversarial examples and develop an iterative method for generating fooling examples. The experimental results show that fooling examples can not only be easily generated, with a success rate of nearly 100% in the white-box scenario, but also exhibit strong transferability across different models in the black-box scenario. Tests on the Google Cloud Vision API show that fooling examples can also be recognized by real-world computer vision systems. Our findings reveal a new cognitive deficit of neural networks, and we hope that these potential security threats will be addressed in future neural network applications. MDPI 2023-07-13 /pmc/articles/PMC10383212/ /pubmed/37514672 http://dx.doi.org/10.3390/s23146378 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhang, Ming Chen, Yongkang Qian, Cheng Fooling Examples: Another Intriguing Property of Neural Networks |
title | Fooling Examples: Another Intriguing Property of Neural Networks |
title_full | Fooling Examples: Another Intriguing Property of Neural Networks |
title_fullStr | Fooling Examples: Another Intriguing Property of Neural Networks |
title_full_unstemmed | Fooling Examples: Another Intriguing Property of Neural Networks |
title_short | Fooling Examples: Another Intriguing Property of Neural Networks |
title_sort | fooling examples: another intriguing property of neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10383212/ https://www.ncbi.nlm.nih.gov/pubmed/37514672 http://dx.doi.org/10.3390/s23146378 |
work_keys_str_mv | AT zhangming foolingexamplesanotherintriguingpropertyofneuralnetworks AT chenyongkang foolingexamplesanotherintriguingpropertyofneuralnetworks AT qiancheng foolingexamplesanotherintriguingpropertyofneuralnetworks |