Cargando…

Classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis

Facial expressions are widely recognized as universal indicators of underlying internal states in most species of animals, thereby presenting as a non-invasive measure for assessing physical and mental conditions. Despite the advancement of artificial intelligence-assisted tools for automated analys...

Descripción completa

Detalles Bibliográficos
Autores principales: Tanaka, Yudai, Nakata, Takuto, Hibino, Hiroshi, Nishiyama, Masaaki, Ino, Daisuke
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10359012/
https://www.ncbi.nlm.nih.gov/pubmed/37471381
http://dx.doi.org/10.1371/journal.pone.0288930
_version_ 1785075790425096192
author Tanaka, Yudai
Nakata, Takuto
Hibino, Hiroshi
Nishiyama, Masaaki
Ino, Daisuke
author_facet Tanaka, Yudai
Nakata, Takuto
Hibino, Hiroshi
Nishiyama, Masaaki
Ino, Daisuke
author_sort Tanaka, Yudai
collection PubMed
description Facial expressions are widely recognized as universal indicators of underlying internal states in most species of animals, thereby presenting as a non-invasive measure for assessing physical and mental conditions. Despite the advancement of artificial intelligence-assisted tools for automated analysis of voluminous facial expression data in human subjects, the corresponding tools for mice still remain limited so far. Considering that mice are the most prevalent model animals for studying human health and diseases, a comprehensive characterization of emotion-dependent patterns of facial expressions in mice could extend our knowledge on the basis of emotions and the related disorders. Here, we present a framework for the development of a deep learning-powered tool for classifying facial expressions in head-fixed mouse. We demonstrate that our machine vision was capable of accurately classifying three different emotional states from lateral facial images in head-fixed mouse. Moreover, we objectively determined how our classifier characterized the differences among the facial images through the use of an interpretation technique called Gradient-weighted Class Activation Mapping. Importantly, our machine vision presumably discerned the data by leveraging multiple facial features. Our approach is likely to facilitate the non-invasive decoding of a variety of emotions from facial images in head-fixed mice.
format Online
Article
Text
id pubmed-10359012
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-103590122023-07-21 Classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis Tanaka, Yudai Nakata, Takuto Hibino, Hiroshi Nishiyama, Masaaki Ino, Daisuke PLoS One Research Article Facial expressions are widely recognized as universal indicators of underlying internal states in most species of animals, thereby presenting as a non-invasive measure for assessing physical and mental conditions. Despite the advancement of artificial intelligence-assisted tools for automated analysis of voluminous facial expression data in human subjects, the corresponding tools for mice still remain limited so far. Considering that mice are the most prevalent model animals for studying human health and diseases, a comprehensive characterization of emotion-dependent patterns of facial expressions in mice could extend our knowledge on the basis of emotions and the related disorders. Here, we present a framework for the development of a deep learning-powered tool for classifying facial expressions in head-fixed mouse. We demonstrate that our machine vision was capable of accurately classifying three different emotional states from lateral facial images in head-fixed mouse. Moreover, we objectively determined how our classifier characterized the differences among the facial images through the use of an interpretation technique called Gradient-weighted Class Activation Mapping. Importantly, our machine vision presumably discerned the data by leveraging multiple facial features. Our approach is likely to facilitate the non-invasive decoding of a variety of emotions from facial images in head-fixed mice. Public Library of Science 2023-07-20 /pmc/articles/PMC10359012/ /pubmed/37471381 http://dx.doi.org/10.1371/journal.pone.0288930 Text en © 2023 Tanaka et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Tanaka, Yudai
Nakata, Takuto
Hibino, Hiroshi
Nishiyama, Masaaki
Ino, Daisuke
Classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis
title Classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis
title_full Classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis
title_fullStr Classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis
title_full_unstemmed Classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis
title_short Classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis
title_sort classification of multiple emotional states from facial expressions in head-fixed mice using a deep learning-based image analysis
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10359012/
https://www.ncbi.nlm.nih.gov/pubmed/37471381
http://dx.doi.org/10.1371/journal.pone.0288930
work_keys_str_mv AT tanakayudai classificationofmultipleemotionalstatesfromfacialexpressionsinheadfixedmiceusingadeeplearningbasedimageanalysis
AT nakatatakuto classificationofmultipleemotionalstatesfromfacialexpressionsinheadfixedmiceusingadeeplearningbasedimageanalysis
AT hibinohiroshi classificationofmultipleemotionalstatesfromfacialexpressionsinheadfixedmiceusingadeeplearningbasedimageanalysis
AT nishiyamamasaaki classificationofmultipleemotionalstatesfromfacialexpressionsinheadfixedmiceusingadeeplearningbasedimageanalysis
AT inodaisuke classificationofmultipleemotionalstatesfromfacialexpressionsinheadfixedmiceusingadeeplearningbasedimageanalysis