Natural gesture based control for quadcopter
Human-Computer Interaction has progressed rapidly throughout the years and the opportunities that it provides are massive. Through Natural User Interface, natural gestures can be used to control a quadcopter. This means that the user himself becomes the controller and is able to send commands to the...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
2015
|
Subjects: | |
Online Access: | http://hdl.handle.net/10356/65606 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-65606 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-656062023-03-03T20:43:27Z Natural gesture based control for quadcopter Chang, Yi Wei Suresh Sundaram School of Computer Engineering DRNTU::Engineering::Computer science and engineering Human-Computer Interaction has progressed rapidly throughout the years and the opportunities that it provides are massive. Through Natural User Interface, natural gestures can be used to control a quadcopter. This means that the user himself becomes the controller and is able to send commands to the quadcopter just by changing the posture of his body. Flying a quadcopter is quite different from controlling a remote controlled car. This is because unlike a car, a quadcopter can travel in three different axes. This is why the Microsoft Kinect is being used, as it is able to obtain depth information from the surroundings. The Microsoft Kinect is used to track the skeleton data of the user in order to determine the current posture of the user. Each posture is then classified as a specific command which is then sent to the quadcopter, which in this case is the Parrot AR.Drone 2.0. Certain thresholds were set for each posture so that there is a reasonable amount of leeway provided. This ensures that the user does not accidentally issue a command that he did not intend to do. The commands were sent to the quadcopter using Wi-Fi over the User Datagram Protocol to reduce bandwidth overhead and latency. This enables the quadcopter to respond instantly to the commands given by the user. Bachelor of Engineering (Computer Engineering) 2015-11-19T05:13:21Z 2015-11-19T05:13:21Z 2015 2015 Final Year Project (FYP) http://hdl.handle.net/10356/65606 en Nanyang Technological University 41 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Computer science and engineering |
spellingShingle |
DRNTU::Engineering::Computer science and engineering Chang, Yi Wei Natural gesture based control for quadcopter |
description |
Human-Computer Interaction has progressed rapidly throughout the years and the opportunities that it provides are massive. Through Natural User Interface, natural gestures can be used to control a quadcopter. This means that the user himself becomes the controller and is able to send commands to the quadcopter just by changing the posture of his body. Flying a quadcopter is quite different from controlling a remote controlled car. This is because unlike a car, a quadcopter can travel in three different axes. This is why the Microsoft Kinect is being used, as it is able to obtain depth information from the surroundings. The Microsoft Kinect is used to track the skeleton data of the user in order to determine the current posture of the user. Each posture is then classified as a specific command which is then sent to the quadcopter, which in this case is the Parrot AR.Drone 2.0. Certain thresholds were set for each posture so that there is a reasonable amount of leeway provided. This ensures that the user does not accidentally issue a command that he did not intend to do. The commands were sent to the quadcopter using Wi-Fi over the User Datagram Protocol to reduce bandwidth overhead and latency. This enables the quadcopter to respond instantly to the commands given by the user. |
author2 |
Suresh Sundaram |
author_facet |
Suresh Sundaram Chang, Yi Wei |
format |
Final Year Project |
author |
Chang, Yi Wei |
author_sort |
Chang, Yi Wei |
title |
Natural gesture based control for quadcopter |
title_short |
Natural gesture based control for quadcopter |
title_full |
Natural gesture based control for quadcopter |
title_fullStr |
Natural gesture based control for quadcopter |
title_full_unstemmed |
Natural gesture based control for quadcopter |
title_sort |
natural gesture based control for quadcopter |
publishDate |
2015 |
url |
http://hdl.handle.net/10356/65606 |
_version_ |
1759856252264906752 |