Hand gesture recognition using RF-sensing

In recent years, the focus on a device’s Human-Computer Interaction (HCI) has been placed into emphasis as the improvement of devices’ raw performance has started to slow down. This has led to research being done on hands-free devices and controllers. Hand gesture recognition is one of the methods...

Full description

Saved in:
Bibliographic Details
Main Author: Tan, Sheng Rong
Other Authors: LUO Jun
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/148119
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:In recent years, the focus on a device’s Human-Computer Interaction (HCI) has been placed into emphasis as the improvement of devices’ raw performance has started to slow down. This has led to research being done on hands-free devices and controllers. Hand gesture recognition is one of the methods that can be used to improve the overall HCI experience between a user and his/her device. With the rise of deep-learning techniques for analysing data, new possibilities in the field of smart-sensing can be made. Optical or acoustic sensors already exist within popular smartphone devices which many own today. Some of the features that utilise these sensors are face recognition in biometrics security and Voice Assistants. The radar, an alternate sensor that is seemingly less invasive, has been gaining significant research interest due to its non-invasive nature as compared to its counterparts. Radars have not been widely used in commercial products yet. In this project, a framework of machine learning using values from an ultra-wideband (UWB) radar sensor to recognise hand sign gestures shall be presented. Signals data shall be collected using the radar with each signature being a 1-dimensional tensor. These signatures will be pre- processing and then feed into a Convolutional Neural Network (CNN) to extract unique features before being passed to a classifier. Two different CNN architecture shall be compared in terms of correctness in the hand gesture prediction: (i) simple deep Convolutional Network (CN) (ii) extremely deep CN. The shape of the radar tensor and the parameters of the classifiers are optimized to maximize classification accuracy. The classification results of the proposed architecture (i) simple deep CN showed a moderate level of accuracy around 60%. The classification results of the proposed architecture (ii) Residual Network (ResNet) show a high level of accuracy above 90 % and a very low confusion probability even between similar gestures.