Development of a vision-based gesture human-computer interface
A human computer interface (HCI) involves information passing between the human and the computer. Information is imparted from the computer to the human through visual and audio means. However, a physical interface is required for intention to be conveyed to the computer in a traditional HC...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
2010
|
Subjects: | |
Online Access: | http://hdl.handle.net/10356/40327 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-40327 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-403272023-03-04T18:34:38Z Development of a vision-based gesture human-computer interface Lim, Chun Ping. Seet Gim Lee, Gerald Sim Siang Kok School of Mechanical and Aerospace Engineering Robotics Research Centre DRNTU::Engineering::Mechanical engineering::Mechatronics A human computer interface (HCI) involves information passing between the human and the computer. Information is imparted from the computer to the human through visual and audio means. However, a physical interface is required for intention to be conveyed to the computer in a traditional HCI scheme. The keyboard and mouse, in particular, are pervasive in the world today. For a more intuitive HCI, the information passing from the human to the computer ought to be in the form of visual or audio. Hence, VG HCI emerged to fill this discontinuity. Hand gesture combined with vision, being the most fundamental and universal form of communicative means, is ideal for being developed as an intuitive HCI. However, detection of the hand from the image seen through the lenses of a video camera is not as intuitive to the computer as it is to people. As images are represented as structures of numbers, extracting accurate information about the hand from the data captured is a daunting task. To make matters worse, attaching meaning to the hand sign and hand motion detected involves high level intelligence which is still elusive to computers. Nonetheless, with reasonable assumptions and restrictions, algorithms capable of detecting the hand and understanding gestures do exist. Building on the foundation of such algorithms, a system of VG HCI was constructed. The VG HCI developed consists of two stages, namely, hand detection and gesture recognition. Hand detection involves localization of the hand and extraction of its geometrical information. Gesture recognition involves attaching meaning to the hand sign and hand motion detected. The first stage of the VG HCI was built based on 2-D appearance-based approach with an explicit shape model of the hand. With a few reasonable assumptions, this stage was achieved with 3 image processing techniques, namely background subtraction by codebook method, skin segmentation and contour convexity finding. To enhance the robustness of the system, a variant codebook method had been created to be used in skin segmentation. The gesture recognition stage comprises of static communicative gesture recognition and dynamic manipulative gesture extraction. A hand shape model, characterizing the hand sign, was defined as a means to identify static communicative gesture. Dynamic manipulative gesture was interpreted as the hand’s motion. The hand was tracked using a Kalman tracker to endow the system with noise rejection capability. The VG HCI system was subjected to a series of performance evaluation experiments and the system’s parameters were fine tuned with insights gained from the experiments. From the evaluations, it was concluded that the two key modules, which are the skin segmentation by codebook method and the Kalman tracker, are robust. This in turn translates into the robustness of the VG HCI. Bachelor of Engineering (Mechanical Engineering) 2010-06-14T08:52:00Z 2010-06-14T08:52:00Z 2010 2010 Final Year Project (FYP) http://hdl.handle.net/10356/40327 en Nanyang Technological University 121 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Mechanical engineering::Mechatronics |
spellingShingle |
DRNTU::Engineering::Mechanical engineering::Mechatronics Lim, Chun Ping. Development of a vision-based gesture human-computer interface |
description |
A human computer interface (HCI) involves information passing between the
human and the computer. Information is imparted from the computer to the human
through visual and audio means. However, a physical interface is required for intention
to be conveyed to the computer in a traditional HCI scheme. The keyboard and mouse, in particular, are pervasive in the world today. For a more intuitive HCI, the information passing from the human to the computer ought to be in the form of visual or audio. Hence, VG HCI emerged to fill this discontinuity.
Hand gesture combined with vision, being the most fundamental and universal
form of communicative means, is ideal for being developed as an intuitive HCI. However, detection of the hand from the image seen through the lenses of a video camera is not
as intuitive to the computer as it is to people. As images are represented as structures of numbers, extracting accurate information about the hand from the data captured is a
daunting task. To make matters worse, attaching meaning to the hand sign and hand
motion detected involves high level intelligence which is still elusive to computers. Nonetheless, with reasonable assumptions and restrictions, algorithms capable of detecting the hand and understanding gestures do exist. Building on the foundation of such algorithms, a system of VG HCI was constructed.
The VG HCI developed consists of two stages, namely, hand detection and
gesture recognition. Hand detection involves localization of the hand and extraction of
its geometrical information. Gesture recognition involves attaching meaning to the hand
sign and hand motion detected. The first stage of the VG HCI was built based on 2-D
appearance-based approach with an explicit shape model of the hand. With a few
reasonable assumptions, this stage was achieved with 3 image processing techniques,
namely background subtraction by codebook method, skin segmentation and contour
convexity finding. To enhance the robustness of the system, a variant codebook method
had been created to be used in skin segmentation. The gesture recognition stage
comprises of static communicative gesture recognition and dynamic manipulative
gesture extraction. A hand shape model, characterizing the hand sign, was defined as a
means to identify static communicative gesture. Dynamic manipulative gesture was
interpreted as the hand’s motion. The hand was tracked using a Kalman tracker to
endow the system with noise rejection capability.
The VG HCI system was subjected to a series of performance evaluation
experiments and the system’s parameters were fine tuned with insights gained from the
experiments. From the evaluations, it was concluded that the two key modules, which
are the skin segmentation by codebook method and the Kalman tracker, are robust. This
in turn translates into the robustness of the VG HCI. |
author2 |
Seet Gim Lee, Gerald |
author_facet |
Seet Gim Lee, Gerald Lim, Chun Ping. |
format |
Final Year Project |
author |
Lim, Chun Ping. |
author_sort |
Lim, Chun Ping. |
title |
Development of a vision-based gesture human-computer interface |
title_short |
Development of a vision-based gesture human-computer interface |
title_full |
Development of a vision-based gesture human-computer interface |
title_fullStr |
Development of a vision-based gesture human-computer interface |
title_full_unstemmed |
Development of a vision-based gesture human-computer interface |
title_sort |
development of a vision-based gesture human-computer interface |
publishDate |
2010 |
url |
http://hdl.handle.net/10356/40327 |
_version_ |
1759858300779757568 |