Hand pose/gesture analysis using artificial intelligence

Hand gesture recognition is a crucial research topic in human-computer interaction, enabling efficient, intuitive, and natural communication between humans and computers. In short, hand gesture recognition is the task of classifying hand gestures. Despite tremendous progress, hand gesture recognitio...

Full description

Saved in:
Bibliographic Details
Main Author: Lim, Ze Xuan
Other Authors: Yap Kim Hui
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/177144
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-177144
record_format dspace
spelling sg-ntu-dr.10356-1771442024-05-31T15:43:31Z Hand pose/gesture analysis using artificial intelligence Lim, Ze Xuan Yap Kim Hui School of Electrical and Electronic Engineering EKHYap@ntu.edu.sg Computer and Information Science Gesture Hand gesture recognition is a crucial research topic in human-computer interaction, enabling efficient, intuitive, and natural communication between humans and computers. In short, hand gesture recognition is the task of classifying hand gestures. Despite tremendous progress, hand gesture recognition poses significant challenges due to its distinctive characteristics and difficulties, including self-occlusion or object occlusion, computational complexity in processing high-dimensional hand gesture data, and ambiguity in gesture interpretation due to similar hand configurations for different hand gestures. Over the last decade, the implementation of deep learning methods in the computer vision area has achieved state-of-the-art (SoTA) results. In this project, a skeleton-based dynamic hand gesture recognition model, namely Temporal Decoupling Graph Convolutional Network (TD-GCN), is selected to be the recognition model. TD-GCN is trained on two well-known hand gesture datasets: the SHREC’ 17 Track dataset and the DHG-14/28 dataset. Through retraining, modifications, and experiments, the trained model achieved a 96.79% accuracy on SHREC’ 17 Track 14-gesture dataset and a 92.74% accuracy on the 28-gesture dataset respectively. While on the DHG-14/28 dataset, the trained model achieved 92.14% and 89.29% accuracy respectively. Bachelor's degree 2024-05-27T06:17:19Z 2024-05-27T06:17:19Z 2024 Final Year Project (FYP) Lim, Z. X. (2024). Hand pose/gesture analysis using artificial intelligence. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177144 https://hdl.handle.net/10356/177144 en A3251-231 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Gesture
spellingShingle Computer and Information Science
Gesture
Lim, Ze Xuan
Hand pose/gesture analysis using artificial intelligence
description Hand gesture recognition is a crucial research topic in human-computer interaction, enabling efficient, intuitive, and natural communication between humans and computers. In short, hand gesture recognition is the task of classifying hand gestures. Despite tremendous progress, hand gesture recognition poses significant challenges due to its distinctive characteristics and difficulties, including self-occlusion or object occlusion, computational complexity in processing high-dimensional hand gesture data, and ambiguity in gesture interpretation due to similar hand configurations for different hand gestures. Over the last decade, the implementation of deep learning methods in the computer vision area has achieved state-of-the-art (SoTA) results. In this project, a skeleton-based dynamic hand gesture recognition model, namely Temporal Decoupling Graph Convolutional Network (TD-GCN), is selected to be the recognition model. TD-GCN is trained on two well-known hand gesture datasets: the SHREC’ 17 Track dataset and the DHG-14/28 dataset. Through retraining, modifications, and experiments, the trained model achieved a 96.79% accuracy on SHREC’ 17 Track 14-gesture dataset and a 92.74% accuracy on the 28-gesture dataset respectively. While on the DHG-14/28 dataset, the trained model achieved 92.14% and 89.29% accuracy respectively.
author2 Yap Kim Hui
author_facet Yap Kim Hui
Lim, Ze Xuan
format Final Year Project
author Lim, Ze Xuan
author_sort Lim, Ze Xuan
title Hand pose/gesture analysis using artificial intelligence
title_short Hand pose/gesture analysis using artificial intelligence
title_full Hand pose/gesture analysis using artificial intelligence
title_fullStr Hand pose/gesture analysis using artificial intelligence
title_full_unstemmed Hand pose/gesture analysis using artificial intelligence
title_sort hand pose/gesture analysis using artificial intelligence
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/177144
_version_ 1806059845723357184