Multi-view CNNS for hand gesture recognition
Hand gesture recognition plays a significant role in human-computer interaction and has broad applications in augmented/virtual reality. It is a practical project that it can be used to help those disabled people with special needs and requirements. Despite the recent progress on recognizing the...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Theses and Dissertations |
Language: | English |
Published: |
2017
|
Subjects: | |
Online Access: | http://hdl.handle.net/10356/72613 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-72613 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-726132023-07-04T15:05:36Z Multi-view CNNS for hand gesture recognition Dou, Yudi Yuan Junsong School of Electrical and Electronic Engineering DRNTU::Engineering::Electrical and electronic engineering Hand gesture recognition plays a significant role in human-computer interaction and has broad applications in augmented/virtual reality. It is a practical project that it can be used to help those disabled people with special needs and requirements. Despite the recent progress on recognizing the hand pose of Arabic numbers from 0 to 9, the accuracy of sign language’s hand gesture recognition is still far from satisfactory since the finger articulations are more complex. Di↵erent from the traditional model-driven method, we do not need colored markers to extract features or skin color models to represent hand region. We focus on data-driven approaches which do not need complex calibration. In this dissertation, we create our own dataset of 30 Chinese sign languages and propose to use multiview CNN-based methods to recognize them. We project hand depth image onto three orthogonal planes and feed every plane’s projected image into a convolutional neural network to generate three probabilities. It is a task of classification and then fuse three views’ output together to recognize the final hand gesture. In order to put this project into application, we also work on producing the real-time demo to output a string like Chinese spelling. Experiments show that the proposed approach could recognize 30 Chinese hand gestures accurately and produce demo in real-time. Master of Science (Signal Processing) 2017-08-30T07:25:31Z 2017-08-30T07:25:31Z 2017 Thesis http://hdl.handle.net/10356/72613 en 63 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Electrical and electronic engineering |
spellingShingle |
DRNTU::Engineering::Electrical and electronic engineering Dou, Yudi Multi-view CNNS for hand gesture recognition |
description |
Hand gesture recognition plays a significant role in human-computer interaction and
has broad applications in augmented/virtual reality. It is a practical project that it
can be used to help those disabled people with special needs and requirements.
Despite the recent progress on recognizing the hand pose of Arabic numbers from
0 to 9, the accuracy of sign language’s hand gesture recognition is still far from
satisfactory since the finger articulations are more complex.
Di↵erent from the traditional model-driven method, we do not need colored markers
to extract features or skin color models to represent hand region. We focus on
data-driven approaches which do not need complex calibration. In this dissertation,
we create our own dataset of 30 Chinese sign languages and propose to use multiview
CNN-based methods to recognize them. We project hand depth image onto
three orthogonal planes and feed every plane’s projected image into a convolutional
neural network to generate three probabilities. It is a task of classification and then
fuse three views’ output together to recognize the final hand gesture. In order to
put this project into application, we also work on producing the real-time demo to
output a string like Chinese spelling.
Experiments show that the proposed approach could recognize 30 Chinese hand
gestures accurately and produce demo in real-time. |
author2 |
Yuan Junsong |
author_facet |
Yuan Junsong Dou, Yudi |
format |
Theses and Dissertations |
author |
Dou, Yudi |
author_sort |
Dou, Yudi |
title |
Multi-view CNNS for hand gesture recognition |
title_short |
Multi-view CNNS for hand gesture recognition |
title_full |
Multi-view CNNS for hand gesture recognition |
title_fullStr |
Multi-view CNNS for hand gesture recognition |
title_full_unstemmed |
Multi-view CNNS for hand gesture recognition |
title_sort |
multi-view cnns for hand gesture recognition |
publishDate |
2017 |
url |
http://hdl.handle.net/10356/72613 |
_version_ |
1772828800829095936 |