Cargando…
Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning
As opposed to macro-expressions, micro-expressions are subtle and not easily detectable emotional expressions, often containing rich information about mental activities. The practical recognition of micro-expressions is essential in interrogation and healthcare. Neural networks are currently one of...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9280988/ https://www.ncbi.nlm.nih.gov/pubmed/35845761 http://dx.doi.org/10.3389/fnbot.2022.922761 |
_version_ | 1784746776729747456 |
---|---|
author | Liu, Yanju Li, Yange Yi, Xinhai Hu, Zuojin Zhang, Huiyu Liu, Yanzhong |
author_facet | Liu, Yanju Li, Yange Yi, Xinhai Hu, Zuojin Zhang, Huiyu Liu, Yanzhong |
author_sort | Liu, Yanju |
collection | PubMed |
description | As opposed to macro-expressions, micro-expressions are subtle and not easily detectable emotional expressions, often containing rich information about mental activities. The practical recognition of micro-expressions is essential in interrogation and healthcare. Neural networks are currently one of the most common approaches to micro-expression recognition. Still, neural networks often increase their complexity when improving accuracy, and overly large neural networks require extremely high hardware requirements for running equipment. In recent years, vision transformers based on self-attentive mechanisms have achieved accuracy in image recognition and classification that is no less than that of neural networks. Still, the drawback is that without the image-specific biases inherent to neural networks, the cost of improving accuracy is an exponential increase in the number of parameters. This approach describes training a facial expression feature extractor by transfer learning and then fine-tuning and optimizing the MobileViT model to perform the micro-expression recognition task. First, the CASME II, SAMM, and SMIC datasets are combined into a compound dataset, and macro-expression samples are extracted from the three macro-expression datasets. Each macro-expression sample and micro-expression sample are pre-processed identically to make them similar. Second, the macro-expression samples were used to train the MobileNetV2 block in MobileViT as a facial expression feature extractor and to save the weights when the accuracy was highest. Finally, some of the hyperparameters of the MobileViT model are determined by grid search and then fed into the micro-expression samples for training. The samples are classified using an SVM classifier. In the experiments, the proposed method obtained an accuracy of 84.27%, and the time to process individual samples was only 35.4 ms. Comparative experiments show that the proposed method is comparable to state-of-the-art methods in terms of accuracy while improving recognition efficiency. |
format | Online Article Text |
id | pubmed-9280988 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-92809882022-07-15 Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning Liu, Yanju Li, Yange Yi, Xinhai Hu, Zuojin Zhang, Huiyu Liu, Yanzhong Front Neurorobot Neuroscience As opposed to macro-expressions, micro-expressions are subtle and not easily detectable emotional expressions, often containing rich information about mental activities. The practical recognition of micro-expressions is essential in interrogation and healthcare. Neural networks are currently one of the most common approaches to micro-expression recognition. Still, neural networks often increase their complexity when improving accuracy, and overly large neural networks require extremely high hardware requirements for running equipment. In recent years, vision transformers based on self-attentive mechanisms have achieved accuracy in image recognition and classification that is no less than that of neural networks. Still, the drawback is that without the image-specific biases inherent to neural networks, the cost of improving accuracy is an exponential increase in the number of parameters. This approach describes training a facial expression feature extractor by transfer learning and then fine-tuning and optimizing the MobileViT model to perform the micro-expression recognition task. First, the CASME II, SAMM, and SMIC datasets are combined into a compound dataset, and macro-expression samples are extracted from the three macro-expression datasets. Each macro-expression sample and micro-expression sample are pre-processed identically to make them similar. Second, the macro-expression samples were used to train the MobileNetV2 block in MobileViT as a facial expression feature extractor and to save the weights when the accuracy was highest. Finally, some of the hyperparameters of the MobileViT model are determined by grid search and then fed into the micro-expression samples for training. The samples are classified using an SVM classifier. In the experiments, the proposed method obtained an accuracy of 84.27%, and the time to process individual samples was only 35.4 ms. Comparative experiments show that the proposed method is comparable to state-of-the-art methods in terms of accuracy while improving recognition efficiency. Frontiers Media S.A. 2022-06-30 /pmc/articles/PMC9280988/ /pubmed/35845761 http://dx.doi.org/10.3389/fnbot.2022.922761 Text en Copyright © 2022 Liu, Li, Yi, Hu, Zhang and Liu. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Liu, Yanju Li, Yange Yi, Xinhai Hu, Zuojin Zhang, Huiyu Liu, Yanzhong Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning |
title | Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning |
title_full | Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning |
title_fullStr | Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning |
title_full_unstemmed | Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning |
title_short | Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning |
title_sort | lightweight vit model for micro-expression recognition enhanced by transfer learning |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9280988/ https://www.ncbi.nlm.nih.gov/pubmed/35845761 http://dx.doi.org/10.3389/fnbot.2022.922761 |
work_keys_str_mv | AT liuyanju lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning AT liyange lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning AT yixinhai lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning AT huzuojin lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning AT zhanghuiyu lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning AT liuyanzhong lightweightvitmodelformicroexpressionrecognitionenhancedbytransferlearning |