Cargando…

Humans can decipher adversarial images

Does the human mind resemble the machine-learning systems that mirror its performance? Convolutional neural networks (CNNs) have achieved human-level benchmarks in classifying novel images. These advances support technologies such as autonomous vehicles and machine diagnosis; but beyond this, they s...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhou, Zhenglong, Firestone, Chaz
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6430776/
https://www.ncbi.nlm.nih.gov/pubmed/30902973
http://dx.doi.org/10.1038/s41467-019-08931-6
_version_ 1783405814242869248
author Zhou, Zhenglong
Firestone, Chaz
author_facet Zhou, Zhenglong
Firestone, Chaz
author_sort Zhou, Zhenglong
collection PubMed
description Does the human mind resemble the machine-learning systems that mirror its performance? Convolutional neural networks (CNNs) have achieved human-level benchmarks in classifying novel images. These advances support technologies such as autonomous vehicles and machine diagnosis; but beyond this, they serve as candidate models for human vision itself. However, unlike humans, CNNs are “fooled” by adversarial examples—nonsense patterns that machines recognize as familiar objects, or seemingly irrelevant image perturbations that nevertheless alter the machine’s classification. Such bizarre behaviors challenge the promise of these new advances; but do human and machine judgments fundamentally diverge? Here, we show that human and machine classification of adversarial images are robustly related: In 8 experiments on 5 prominent and diverse adversarial imagesets, human subjects correctly anticipated the machine’s preferred label over relevant foils—even for images described as “totally unrecognizable to human eyes”. Human intuition may be a surprisingly reliable guide to machine (mis)classification—with consequences for minds and machines alike.
format Online
Article
Text
id pubmed-6430776
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-64307762019-03-25 Humans can decipher adversarial images Zhou, Zhenglong Firestone, Chaz Nat Commun Article Does the human mind resemble the machine-learning systems that mirror its performance? Convolutional neural networks (CNNs) have achieved human-level benchmarks in classifying novel images. These advances support technologies such as autonomous vehicles and machine diagnosis; but beyond this, they serve as candidate models for human vision itself. However, unlike humans, CNNs are “fooled” by adversarial examples—nonsense patterns that machines recognize as familiar objects, or seemingly irrelevant image perturbations that nevertheless alter the machine’s classification. Such bizarre behaviors challenge the promise of these new advances; but do human and machine judgments fundamentally diverge? Here, we show that human and machine classification of adversarial images are robustly related: In 8 experiments on 5 prominent and diverse adversarial imagesets, human subjects correctly anticipated the machine’s preferred label over relevant foils—even for images described as “totally unrecognizable to human eyes”. Human intuition may be a surprisingly reliable guide to machine (mis)classification—with consequences for minds and machines alike. Nature Publishing Group UK 2019-03-22 /pmc/articles/PMC6430776/ /pubmed/30902973 http://dx.doi.org/10.1038/s41467-019-08931-6 Text en © The Author(s) 2019 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
spellingShingle Article
Zhou, Zhenglong
Firestone, Chaz
Humans can decipher adversarial images
title Humans can decipher adversarial images
title_full Humans can decipher adversarial images
title_fullStr Humans can decipher adversarial images
title_full_unstemmed Humans can decipher adversarial images
title_short Humans can decipher adversarial images
title_sort humans can decipher adversarial images
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6430776/
https://www.ncbi.nlm.nih.gov/pubmed/30902973
http://dx.doi.org/10.1038/s41467-019-08931-6
work_keys_str_mv AT zhouzhenglong humanscandecipheradversarialimages
AT firestonechaz humanscandecipheradversarialimages