Cargando…
Gait-ViT: Gait Recognition with Vision Transformer
Identifying an individual based on their physical/behavioral characteristics is known as biometric recognition. Gait is one of the most reliable biometrics due to its advantages, such as being perceivable at a long distance and difficult to replicate. The existing works mostly leverage Convolutional...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9572525/ https://www.ncbi.nlm.nih.gov/pubmed/36236462 http://dx.doi.org/10.3390/s22197362 |
_version_ | 1784810636797018112 |
---|---|
author | Mogan, Jashila Nair Lee, Chin Poo Lim, Kian Ming Muthu, Kalaiarasi Sonai |
author_facet | Mogan, Jashila Nair Lee, Chin Poo Lim, Kian Ming Muthu, Kalaiarasi Sonai |
author_sort | Mogan, Jashila Nair |
collection | PubMed |
description | Identifying an individual based on their physical/behavioral characteristics is known as biometric recognition. Gait is one of the most reliable biometrics due to its advantages, such as being perceivable at a long distance and difficult to replicate. The existing works mostly leverage Convolutional Neural Networks for gait recognition. The Convolutional Neural Networks perform well in image recognition tasks; however, they lack the attention mechanism to emphasize more on the significant regions of the image. The attention mechanism encodes information in the image patches, which facilitates the model to learn the substantial features in the specific regions. In light of this, this work employs the Vision Transformer (ViT) with an attention mechanism for gait recognition, referred to as Gait-ViT. In the proposed Gait-ViT, the gait energy image is first obtained by averaging the series of images over the gait cycle. The images are then split into patches and transformed into sequences by flattening and patch embedding. Position embedding, along with patch embedding, are applied on the sequence of patches to restore the positional information of the patches. Subsequently, the sequence of vectors is fed to the Transformer encoder to produce the final gait representation. As for the classification, the first element of the sequence is sent to the multi-layer perceptron to predict the class label. The proposed method obtained 99.93% on CASIA-B, 100% on OU-ISIR D and 99.51% on OU-LP, which exhibit the ability of the Vision Transformer model to outperform the state-of-the-art methods. |
format | Online Article Text |
id | pubmed-9572525 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-95725252022-10-17 Gait-ViT: Gait Recognition with Vision Transformer Mogan, Jashila Nair Lee, Chin Poo Lim, Kian Ming Muthu, Kalaiarasi Sonai Sensors (Basel) Article Identifying an individual based on their physical/behavioral characteristics is known as biometric recognition. Gait is one of the most reliable biometrics due to its advantages, such as being perceivable at a long distance and difficult to replicate. The existing works mostly leverage Convolutional Neural Networks for gait recognition. The Convolutional Neural Networks perform well in image recognition tasks; however, they lack the attention mechanism to emphasize more on the significant regions of the image. The attention mechanism encodes information in the image patches, which facilitates the model to learn the substantial features in the specific regions. In light of this, this work employs the Vision Transformer (ViT) with an attention mechanism for gait recognition, referred to as Gait-ViT. In the proposed Gait-ViT, the gait energy image is first obtained by averaging the series of images over the gait cycle. The images are then split into patches and transformed into sequences by flattening and patch embedding. Position embedding, along with patch embedding, are applied on the sequence of patches to restore the positional information of the patches. Subsequently, the sequence of vectors is fed to the Transformer encoder to produce the final gait representation. As for the classification, the first element of the sequence is sent to the multi-layer perceptron to predict the class label. The proposed method obtained 99.93% on CASIA-B, 100% on OU-ISIR D and 99.51% on OU-LP, which exhibit the ability of the Vision Transformer model to outperform the state-of-the-art methods. MDPI 2022-09-28 /pmc/articles/PMC9572525/ /pubmed/36236462 http://dx.doi.org/10.3390/s22197362 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Mogan, Jashila Nair Lee, Chin Poo Lim, Kian Ming Muthu, Kalaiarasi Sonai Gait-ViT: Gait Recognition with Vision Transformer |
title | Gait-ViT: Gait Recognition with Vision Transformer |
title_full | Gait-ViT: Gait Recognition with Vision Transformer |
title_fullStr | Gait-ViT: Gait Recognition with Vision Transformer |
title_full_unstemmed | Gait-ViT: Gait Recognition with Vision Transformer |
title_short | Gait-ViT: Gait Recognition with Vision Transformer |
title_sort | gait-vit: gait recognition with vision transformer |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9572525/ https://www.ncbi.nlm.nih.gov/pubmed/36236462 http://dx.doi.org/10.3390/s22197362 |
work_keys_str_mv | AT moganjashilanair gaitvitgaitrecognitionwithvisiontransformer AT leechinpoo gaitvitgaitrecognitionwithvisiontransformer AT limkianming gaitvitgaitrecognitionwithvisiontransformer AT muthukalaiarasisonai gaitvitgaitrecognitionwithvisiontransformer |