Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors
Gesture recognition using machine learning methods is valuable in the development of advanced cybernetics, robotics, and healthcare systems, and typically relies on images or videos. To improve recognition accuracy, such visual data can be fused with data from other sensors, but this approach is lim...
Saved in:
Main Authors: | , , , , , , , , , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/148021 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-148021 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1480212023-07-14T16:01:47Z Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors Wang, Ming Yan, Zheng Wang, Ting Cai, Pingqiang Gao, Siyu Zeng, Yi Wan, Changjin Wang, Hong Pan, Liang Yu, Jiancan Pan, Shaowu He, Ke Lu, Jie Chen, Xiaodong School of Materials Science and Engineering Innovative Centre for Flexible Devices Engineering::Materials Convolutional Neural Networks Network architecture Gesture recognition using machine learning methods is valuable in the development of advanced cybernetics, robotics, and healthcare systems, and typically relies on images or videos. To improve recognition accuracy, such visual data can be fused with data from other sensors, but this approach is limited by the quality of the sensor data and the incompatibility of the datasets. Here, we report a bioinspired data fusion architecture that can perform human gesture recognition by integrating visual data with somatosensory data from skin-like stretchable strain sensors. The learning architecture uses a convolutional neural network for visual processing, and then implements a sparse neural network for sensor data fusion and recognition. Our approach can achieve a recognition accuracy of 100%, and maintain recognition accuracy with noisy, under- or over-exposed images. We also show that our architecture can be implemented for robot navigation using hand gestures with a small error, even in the dark. Ministry of Education (MOE) National Research Foundation (NRF) Accepted version The project was partially supported by the National Research Foundation (NRF), Prime Minister’s office, Singapore, under its NRF Investigatorship (NRF2016NRF-NRF1001-21), Singapore Ministry of Education (MOE2015-T2-2-60), Advanced Manufacturing and Engineering (AME) Programmatic Grant (No. A19A1b0045), and the Australian Research Council (ARC) under Discovery Grant DP200100700. The authors thank all the volunteers for collecting data and also thank Dr. Ai Lin Chun for critically reading and editing the manuscript. 2021-04-26T02:31:10Z 2021-04-26T02:31:10Z 2020 Journal Article Wang, M., Yan, Z., Wang, T., Cai, P., Gao, S., Zeng, Y., Wan, C., Wang, H., Pan, L., Yu, J., Pan, S., He, K., Lu, J. & Chen, X. (2020). Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors. Nature Electronics, 3, 563-570. https://dx.doi.org/10.1038/s41928-020-0422-z 2520-1131 https://hdl.handle.net/10356/148021 10.1038/s41928-020-0422-z 3 563 570 en Nature Electronics © 2020 Macmillan Publishers Limited, part of Springer Nature. All rights reserved. This paper was published in Nature Electronics and is made available with permission of Macmillan Publishers Limited, part of Springer Nature. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Materials Convolutional Neural Networks Network architecture |
spellingShingle |
Engineering::Materials Convolutional Neural Networks Network architecture Wang, Ming Yan, Zheng Wang, Ting Cai, Pingqiang Gao, Siyu Zeng, Yi Wan, Changjin Wang, Hong Pan, Liang Yu, Jiancan Pan, Shaowu He, Ke Lu, Jie Chen, Xiaodong Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors |
description |
Gesture recognition using machine learning methods is valuable in the development of advanced cybernetics, robotics, and healthcare systems, and typically relies on images or videos. To improve recognition accuracy, such visual data can be fused with data from other sensors, but this approach is limited by the quality of the sensor data and the incompatibility of the datasets. Here, we report a bioinspired data fusion architecture that can perform human gesture recognition by integrating visual data with somatosensory data from skin-like stretchable strain sensors. The learning architecture uses a convolutional neural network for visual processing, and then implements a sparse neural network for sensor data fusion and recognition. Our approach can achieve a recognition accuracy of 100%, and maintain recognition accuracy with noisy, under- or over-exposed images. We also show that our architecture can be implemented for robot navigation using hand gestures with a small error, even in the dark. |
author2 |
School of Materials Science and Engineering |
author_facet |
School of Materials Science and Engineering Wang, Ming Yan, Zheng Wang, Ting Cai, Pingqiang Gao, Siyu Zeng, Yi Wan, Changjin Wang, Hong Pan, Liang Yu, Jiancan Pan, Shaowu He, Ke Lu, Jie Chen, Xiaodong |
format |
Article |
author |
Wang, Ming Yan, Zheng Wang, Ting Cai, Pingqiang Gao, Siyu Zeng, Yi Wan, Changjin Wang, Hong Pan, Liang Yu, Jiancan Pan, Shaowu He, Ke Lu, Jie Chen, Xiaodong |
author_sort |
Wang, Ming |
title |
Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors |
title_short |
Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors |
title_full |
Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors |
title_fullStr |
Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors |
title_full_unstemmed |
Gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors |
title_sort |
gesture recognition using a bioinspired learning architecture that integrates visual data with somatosensory data from stretchable sensors |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/148021 |
_version_ |
1773551318378479616 |