Cargando…
Humans can decipher adversarial images
Does the human mind resemble the machine-learning systems that mirror its performance? Convolutional neural networks (CNNs) have achieved human-level benchmarks in classifying novel images. These advances support technologies such as autonomous vehicles and machine diagnosis; but beyond this, they s...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6430776/ https://www.ncbi.nlm.nih.gov/pubmed/30902973 http://dx.doi.org/10.1038/s41467-019-08931-6 |
_version_ | 1783405814242869248 |
---|---|
author | Zhou, Zhenglong Firestone, Chaz |
author_facet | Zhou, Zhenglong Firestone, Chaz |
author_sort | Zhou, Zhenglong |
collection | PubMed |
description | Does the human mind resemble the machine-learning systems that mirror its performance? Convolutional neural networks (CNNs) have achieved human-level benchmarks in classifying novel images. These advances support technologies such as autonomous vehicles and machine diagnosis; but beyond this, they serve as candidate models for human vision itself. However, unlike humans, CNNs are “fooled” by adversarial examples—nonsense patterns that machines recognize as familiar objects, or seemingly irrelevant image perturbations that nevertheless alter the machine’s classification. Such bizarre behaviors challenge the promise of these new advances; but do human and machine judgments fundamentally diverge? Here, we show that human and machine classification of adversarial images are robustly related: In 8 experiments on 5 prominent and diverse adversarial imagesets, human subjects correctly anticipated the machine’s preferred label over relevant foils—even for images described as “totally unrecognizable to human eyes”. Human intuition may be a surprisingly reliable guide to machine (mis)classification—with consequences for minds and machines alike. |
format | Online Article Text |
id | pubmed-6430776 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-64307762019-03-25 Humans can decipher adversarial images Zhou, Zhenglong Firestone, Chaz Nat Commun Article Does the human mind resemble the machine-learning systems that mirror its performance? Convolutional neural networks (CNNs) have achieved human-level benchmarks in classifying novel images. These advances support technologies such as autonomous vehicles and machine diagnosis; but beyond this, they serve as candidate models for human vision itself. However, unlike humans, CNNs are “fooled” by adversarial examples—nonsense patterns that machines recognize as familiar objects, or seemingly irrelevant image perturbations that nevertheless alter the machine’s classification. Such bizarre behaviors challenge the promise of these new advances; but do human and machine judgments fundamentally diverge? Here, we show that human and machine classification of adversarial images are robustly related: In 8 experiments on 5 prominent and diverse adversarial imagesets, human subjects correctly anticipated the machine’s preferred label over relevant foils—even for images described as “totally unrecognizable to human eyes”. Human intuition may be a surprisingly reliable guide to machine (mis)classification—with consequences for minds and machines alike. Nature Publishing Group UK 2019-03-22 /pmc/articles/PMC6430776/ /pubmed/30902973 http://dx.doi.org/10.1038/s41467-019-08931-6 Text en © The Author(s) 2019 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/. |
spellingShingle | Article Zhou, Zhenglong Firestone, Chaz Humans can decipher adversarial images |
title | Humans can decipher adversarial images |
title_full | Humans can decipher adversarial images |
title_fullStr | Humans can decipher adversarial images |
title_full_unstemmed | Humans can decipher adversarial images |
title_short | Humans can decipher adversarial images |
title_sort | humans can decipher adversarial images |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6430776/ https://www.ncbi.nlm.nih.gov/pubmed/30902973 http://dx.doi.org/10.1038/s41467-019-08931-6 |
work_keys_str_mv | AT zhouzhenglong humanscandecipheradversarialimages AT firestonechaz humanscandecipheradversarialimages |