Cargando…

The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks

People deploy top-down, goal-directed attention to accomplish tasks, such as finding lost keys. By tuning the visual system to relevant information sources, object recognition can become more efficient (a benefit) and more biased toward the target (a potential cost). Motivated by selective attention...

Descripción completa

Detalles Bibliográficos
Autores principales: Luo, Xiaoliang, Roads, Brett D., Love, Bradley C.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer International Publishing 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8550459/
https://www.ncbi.nlm.nih.gov/pubmed/34723095
http://dx.doi.org/10.1007/s42113-021-00098-y
_version_ 1784590965484290048
author Luo, Xiaoliang
Roads, Brett D.
Love, Bradley C.
author_facet Luo, Xiaoliang
Roads, Brett D.
Love, Bradley C.
author_sort Luo, Xiaoliang
collection PubMed
description People deploy top-down, goal-directed attention to accomplish tasks, such as finding lost keys. By tuning the visual system to relevant information sources, object recognition can become more efficient (a benefit) and more biased toward the target (a potential cost). Motivated by selective attention in categorisation models, we developed a goal-directed attention mechanism that can process naturalistic (photographic) stimuli. Our attention mechanism can be incorporated into any existing deep convolutional neural networks (DCNNs). The processing stages in DCNNs have been related to ventral visual stream. In that light, our attentional mechanism incorporates top-down influences from prefrontal cortex (PFC) to support goal-directed behaviour. Akin to how attention weights in categorisation models warp representational spaces, we introduce a layer of attention weights to the mid-level of a DCNN that amplify or attenuate activity to further a goal. We evaluated the attentional mechanism using photographic stimuli, varying the attentional target. We found that increasing goal-directed attention has benefits (increasing hit rates) and costs (increasing false alarm rates). At a moderate level, attention improves sensitivity (i.e. increases [Formula: see text] ) at only a moderate increase in bias for tasks involving standard images, blended images and natural adversarial images chosen to fool DCNNs. These results suggest that goal-directed attention can reconfigure general-purpose DCNNs to better suit the current task goal, much like PFC modulates activity along the ventral stream. In addition to being more parsimonious and brain consistent, the mid-level attention approach performed better than a standard machine learning approach for transfer learning, namely retraining the final network layer to accommodate the new task.
format Online
Article
Text
id pubmed-8550459
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Springer International Publishing
record_format MEDLINE/PubMed
spelling pubmed-85504592021-10-29 The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks Luo, Xiaoliang Roads, Brett D. Love, Bradley C. Comput Brain Behav Original Paper People deploy top-down, goal-directed attention to accomplish tasks, such as finding lost keys. By tuning the visual system to relevant information sources, object recognition can become more efficient (a benefit) and more biased toward the target (a potential cost). Motivated by selective attention in categorisation models, we developed a goal-directed attention mechanism that can process naturalistic (photographic) stimuli. Our attention mechanism can be incorporated into any existing deep convolutional neural networks (DCNNs). The processing stages in DCNNs have been related to ventral visual stream. In that light, our attentional mechanism incorporates top-down influences from prefrontal cortex (PFC) to support goal-directed behaviour. Akin to how attention weights in categorisation models warp representational spaces, we introduce a layer of attention weights to the mid-level of a DCNN that amplify or attenuate activity to further a goal. We evaluated the attentional mechanism using photographic stimuli, varying the attentional target. We found that increasing goal-directed attention has benefits (increasing hit rates) and costs (increasing false alarm rates). At a moderate level, attention improves sensitivity (i.e. increases [Formula: see text] ) at only a moderate increase in bias for tasks involving standard images, blended images and natural adversarial images chosen to fool DCNNs. These results suggest that goal-directed attention can reconfigure general-purpose DCNNs to better suit the current task goal, much like PFC modulates activity along the ventral stream. In addition to being more parsimonious and brain consistent, the mid-level attention approach performed better than a standard machine learning approach for transfer learning, namely retraining the final network layer to accommodate the new task. Springer International Publishing 2021-02-12 2021 /pmc/articles/PMC8550459/ /pubmed/34723095 http://dx.doi.org/10.1007/s42113-021-00098-y Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Original Paper
Luo, Xiaoliang
Roads, Brett D.
Love, Bradley C.
The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks
title The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks
title_full The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks
title_fullStr The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks
title_full_unstemmed The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks
title_short The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks
title_sort costs and benefits of goal-directed attention in deep convolutional neural networks
topic Original Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8550459/
https://www.ncbi.nlm.nih.gov/pubmed/34723095
http://dx.doi.org/10.1007/s42113-021-00098-y
work_keys_str_mv AT luoxiaoliang thecostsandbenefitsofgoaldirectedattentionindeepconvolutionalneuralnetworks
AT roadsbrettd thecostsandbenefitsofgoaldirectedattentionindeepconvolutionalneuralnetworks
AT lovebradleyc thecostsandbenefitsofgoaldirectedattentionindeepconvolutionalneuralnetworks
AT luoxiaoliang costsandbenefitsofgoaldirectedattentionindeepconvolutionalneuralnetworks
AT roadsbrettd costsandbenefitsofgoaldirectedattentionindeepconvolutionalneuralnetworks
AT lovebradleyc costsandbenefitsofgoaldirectedattentionindeepconvolutionalneuralnetworks