Arbitrary-view human action recognition for human robot interaction
This report presents the development and evaluation of an innovative Arbitrary-view Human Action Recognition (AVHAR) system, utilising the Robomaster Tello Talent for aerial camera feeds. The aim is to enhance capabilities in human-robot interaction by addressing the challenge of recognising human a...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/177242 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-177242 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1772422024-06-01T16:51:56Z Arbitrary-view human action recognition for human robot interaction Loh, Zachariah Jin Jun Li King Ho, Holden School of Mechanical and Aerospace Engineering HoldenLi@ntu.edu.sg Engineering View This report presents the development and evaluation of an innovative Arbitrary-view Human Action Recognition (AVHAR) system, utilising the Robomaster Tello Talent for aerial camera feeds. The aim is to enhance capabilities in human-robot interaction by addressing the challenge of recognising human actions from any viewpoint, the project leverages artificial intelligence, with a focus on Long Short-Term Memory (LSTM) networks.The objective of this project is to develop a system capable of processing skeletal data with high accuracy, achieving real-time action recognition. The evolution of AVHAR has been marked by several challenges, notably the variability of human actions when observed from different viewpoints. Occlusions also present a persistent challenge, where parts of the skeletal structure may be obscured by objects or other individuals, complicating accurate recognition. The project will focus on integrating TensorFlow, OpenCV, MediaPipe, and other libraries to build a comprehensive software infrastructure capable of extracting key points, preprocessing data, and applying sophisticated machine learning algorithms for action detection. This foundation facilitated the exploration of LSTM models, enhanced by three dense layers for refined data analysis and prediction accuracy. Bachelor's degree 2024-05-27T01:39:59Z 2024-05-27T01:39:59Z 2024 Final Year Project (FYP) Loh, Z. J. J. (2024). Arbitrary-view human action recognition for human robot interaction. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177242 https://hdl.handle.net/10356/177242 en C145 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering View |
spellingShingle |
Engineering View Loh, Zachariah Jin Jun Arbitrary-view human action recognition for human robot interaction |
description |
This report presents the development and evaluation of an innovative Arbitrary-view Human Action Recognition (AVHAR) system, utilising the Robomaster Tello Talent for aerial camera feeds. The aim is to enhance capabilities in human-robot interaction by addressing the challenge of recognising human actions from any viewpoint, the project leverages artificial intelligence, with a focus on Long Short-Term Memory (LSTM) networks.The objective of this project is to develop a system capable of processing skeletal data with high accuracy, achieving real-time action recognition.
The evolution of AVHAR has been marked by several challenges, notably the variability of human actions when observed from different viewpoints. Occlusions also present a persistent challenge, where parts of the skeletal structure may be obscured by objects or other individuals, complicating accurate recognition.
The project will focus on integrating TensorFlow, OpenCV, MediaPipe, and other libraries to build a comprehensive software infrastructure capable of extracting key points, preprocessing data, and applying sophisticated machine learning algorithms for action detection. This foundation facilitated the exploration of LSTM models, enhanced by three dense layers for refined data analysis and prediction accuracy. |
author2 |
Li King Ho, Holden |
author_facet |
Li King Ho, Holden Loh, Zachariah Jin Jun |
format |
Final Year Project |
author |
Loh, Zachariah Jin Jun |
author_sort |
Loh, Zachariah Jin Jun |
title |
Arbitrary-view human action recognition for human robot interaction |
title_short |
Arbitrary-view human action recognition for human robot interaction |
title_full |
Arbitrary-view human action recognition for human robot interaction |
title_fullStr |
Arbitrary-view human action recognition for human robot interaction |
title_full_unstemmed |
Arbitrary-view human action recognition for human robot interaction |
title_sort |
arbitrary-view human action recognition for human robot interaction |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/177242 |
_version_ |
1800916368109862912 |