Cargando…

Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning

As opposed to macro-expressions, micro-expressions are subtle and not easily detectable emotional expressions, often containing rich information about mental activities. The practical recognition of micro-expressions is essential in interrogation and healthcare. Neural networks are currently one of...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Yanju, Li, Yange, Yi, Xinhai, Hu, Zuojin, Zhang, Huiyu, Liu, Yanzhong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9280988/
https://www.ncbi.nlm.nih.gov/pubmed/35845761
http://dx.doi.org/10.3389/fnbot.2022.922761
_version_ 1784746776729747456
author Liu, Yanju
Li, Yange
Yi, Xinhai
Hu, Zuojin
Zhang, Huiyu
Liu, Yanzhong
author_facet Liu, Yanju
Li, Yange
Yi, Xinhai
Hu, Zuojin
Zhang, Huiyu
Liu, Yanzhong
author_sort Liu, Yanju
collection PubMed
description As opposed to macro-expressions, micro-expressions are subtle and not easily detectable emotional expressions, often containing rich information about mental activities. The practical recognition of micro-expressions is essential in interrogation and healthcare. Neural networks are currently one of the most common approaches to micro-expression recognition. Still, neural networks often increase their complexity when improving accuracy, and overly large neural networks require extremely high hardware requirements for running equipment. In recent years, vision transformers based on self-attentive mechanisms have achieved accuracy in image recognition and classification that is no less than that of neural networks. Still, the drawback is that without the image-specific biases inherent to neural networks, the cost of improving accuracy is an exponential increase in the number of parameters. This approach describes training a facial expression feature extractor by transfer learning and then fine-tuning and optimizing the MobileViT model to perform the micro-expression recognition task. First, the CASME II, SAMM, and SMIC datasets are combined into a compound dataset, and macro-expression samples are extracted from the three macro-expression datasets. Each macro-expression sample and micro-expression sample are pre-processed identically to make them similar. Second, the macro-expression samples were used to train the MobileNetV2 block in MobileViT as a facial expression feature extractor and to save the weights when the accuracy was highest. Finally, some of the hyperparameters of the MobileViT model are determined by grid search and then fed into the micro-expression samples for training. The samples are classified using an SVM classifier. In the experiments, the proposed method obtained an accuracy of 84.27%, and the time to process individual samples was only 35.4 ms. Comparative experiments show that the proposed method is comparable to state-of-the-art methods in terms of accuracy while improving recognition efficiency.
format Online
Article
Text
id pubmed-9280988
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-92809882022-07-15 Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning Liu, Yanju Li, Yange Yi, Xinhai Hu, Zuojin Zhang, Huiyu Liu, Yanzhong Front Neurorobot Neuroscience As opposed to macro-expressions, micro-expressions are subtle and not easily detectable emotional expressions, often containing rich information about mental activities. The practical recognition of micro-expressions is essential in interrogation and healthcare. Neural networks are currently one of the most common approaches to micro-expression recognition. Still, neural networks often increase their complexity when improving accuracy, and overly large neural networks require extremely high hardware requirements for running equipment. In recent years, vision transformers based on self-attentive mechanisms have achieved accuracy in image recognition and classification that is no less than that of neural networks. Still, the drawback is that without the image-specific biases inherent to neural networks, the cost of improving accuracy is an exponential increase in the number of parameters. This approach describes training a facial expression feature extractor by transfer learning and then fine-tuning and optimizing the MobileViT model to perform the micro-expression recognition task. First, the CASME II, SAMM, and SMIC datasets are combined into a compound dataset, and macro-expression samples are extracted from the three macro-expression datasets. Each macro-expression sample and micro-expression sample are pre-processed identically to make them similar. Second, the macro-expression samples were used to train the MobileNetV2 block in MobileViT as a facial expression feature extractor and to save the weights when the accuracy was highest. Finally, some of the hyperparameters of the MobileViT model are determined by grid search and then fed into the micro-expression samples for training. The samples are classified using an SVM classifier. In the experiments, the proposed method obtained an accuracy of 84.27%, and the time to process individual samples was only 35.4 ms. Comparative experiments show that the proposed method is comparable to state-of-the-art methods in terms of accuracy while improving recognition efficiency. Frontiers Media S.A. 2022-06-30 /pmc/articles/PMC9280988/ /pubmed/35845761 http://dx.doi.org/10.3389/fnbot.2022.922761 Text en Copyright © 2022 Liu, Li, Yi, Hu, Zhang and Liu. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Liu, Yanju
Li, Yange
Yi, Xinhai
Hu, Zuojin
Zhang, Huiyu
Liu, Yanzhong
Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning
title Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning
title_full Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning
title_fullStr Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning
title_full_unstemmed Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning
title_short Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning
title_sort lightweight vit model for micro-expression recognition enhanced by transfer learning
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9280988/
https://www.ncbi.nlm.nih.gov/pubmed/35845761
http://dx.doi.org/10.3389/fnbot.2022.922761
work_keys_str_mv AT liuyanju lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning
AT liyange lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning
AT yixinhai lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning
AT huzuojin lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning
AT zhanghuiyu lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning
AT liuyanzhong lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning