Gait-ViT: Gait Recognition with Vision Transformer

Citation

Mogan, Jashila Nair and Lee, Chin Poo and Lim, Kian Ming and Sonai Muthu Anbananthen, Kalaiarasi (2022) Gait-ViT: Gait Recognition with Vision Transformer. Sensors, 22 (19). p. 7362. ISSN 1424-8220

[img] Text
36.pdf - Published Version
Restricted to Repository staff only

Download (479kB)

Abstract

Identifying an individual based on their physical/behavioral characteristics is known as biometric recognition. Gait is one of the most reliable biometrics due to its advantages, such as being perceivable at a long distance and difficult to replicate. The existing works mostly leverage Convolutional Neural Networks for gait recognition. The Convolutional Neural Networks perform well in image recognition tasks; however, they lack the attention mechanism to emphasize more on the significant regions of the image. The attention mechanism encodes information in the image patches, which facilitates the model to learn the substantial features in the specific regions. In light of this, this work employs the Vision Transformer (ViT) with an attention mechanism for gait recognition, referred to as Gait-ViT. In the proposed Gait-ViT, the gait energy image is first obtained by averaging the series of images over the gait cycle. The images are then split into patches and transformed into sequences by flattening and patch embedding. Position embedding, along with patch embedding, are applied on the sequence of patches to restore the positional information of the patches. Subsequently, the sequence of vectors is fed to the Transformer encoder to produce the final gait representation. As for the classification, the first element of the sequence is sent to the multi-layer perceptron to predict the class label. The proposed method obtained 99.93% on CASIA-B, 100% on OU-ISIR D and 99.51% on OU-LP, which exhibit the ability of the Vision Transformer model to outperform the state-of-the-art methods.

Item Type: Article
Uncontrolled Keywords: Gait recognition, deep learning, transformers, vision transformer, vit, attention
Subjects: Q Science > Q Science (General) > Q300-390 Cybernetics
Divisions: Faculty of Information Science and Technology (FIST)
Depositing User: Ms Nurul Iqtiani Ahmad
Date Deposited: 31 Oct 2022 07:48
Last Modified: 27 Apr 2023 13:17
URII: http://shdl.mmu.edu.my/id/eprint/10587

Downloads

Downloads per month over past year

View ItemEdit (login required)