Arbitrary-view human action recognition for human robot interaction

This report presents the development and evaluation of an innovative Arbitrary-view Human Action Recognition (AVHAR) system, utilising the Robomaster Tello Talent for aerial camera feeds. The aim is to enhance capabilities in human-robot interaction by addressing the challenge of recognising human a...

Full description

Saved in:
Bibliographic Details
Main Author: Loh, Zachariah Jin Jun
Other Authors: Li King Ho, Holden
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/177242
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-177242
record_format dspace
spelling sg-ntu-dr.10356-1772422024-06-01T16:51:56Z Arbitrary-view human action recognition for human robot interaction Loh, Zachariah Jin Jun Li King Ho, Holden School of Mechanical and Aerospace Engineering HoldenLi@ntu.edu.sg Engineering View This report presents the development and evaluation of an innovative Arbitrary-view Human Action Recognition (AVHAR) system, utilising the Robomaster Tello Talent for aerial camera feeds. The aim is to enhance capabilities in human-robot interaction by addressing the challenge of recognising human actions from any viewpoint, the project leverages artificial intelligence, with a focus on Long Short-Term Memory (LSTM) networks.The objective of this project is to develop a system capable of processing skeletal data with high accuracy, achieving real-time action recognition. The evolution of AVHAR has been marked by several challenges, notably the variability of human actions when observed from different viewpoints. Occlusions also present a persistent challenge, where parts of the skeletal structure may be obscured by objects or other individuals, complicating accurate recognition. The project will focus on integrating TensorFlow, OpenCV, MediaPipe, and other libraries to build a comprehensive software infrastructure capable of extracting key points, preprocessing data, and applying sophisticated machine learning algorithms for action detection. This foundation facilitated the exploration of LSTM models, enhanced by three dense layers for refined data analysis and prediction accuracy. Bachelor's degree 2024-05-27T01:39:59Z 2024-05-27T01:39:59Z 2024 Final Year Project (FYP) Loh, Z. J. J. (2024). Arbitrary-view human action recognition for human robot interaction. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177242 https://hdl.handle.net/10356/177242 en C145 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
View
spellingShingle Engineering
View
Loh, Zachariah Jin Jun
Arbitrary-view human action recognition for human robot interaction
description This report presents the development and evaluation of an innovative Arbitrary-view Human Action Recognition (AVHAR) system, utilising the Robomaster Tello Talent for aerial camera feeds. The aim is to enhance capabilities in human-robot interaction by addressing the challenge of recognising human actions from any viewpoint, the project leverages artificial intelligence, with a focus on Long Short-Term Memory (LSTM) networks.The objective of this project is to develop a system capable of processing skeletal data with high accuracy, achieving real-time action recognition. The evolution of AVHAR has been marked by several challenges, notably the variability of human actions when observed from different viewpoints. Occlusions also present a persistent challenge, where parts of the skeletal structure may be obscured by objects or other individuals, complicating accurate recognition. The project will focus on integrating TensorFlow, OpenCV, MediaPipe, and other libraries to build a comprehensive software infrastructure capable of extracting key points, preprocessing data, and applying sophisticated machine learning algorithms for action detection. This foundation facilitated the exploration of LSTM models, enhanced by three dense layers for refined data analysis and prediction accuracy.
author2 Li King Ho, Holden
author_facet Li King Ho, Holden
Loh, Zachariah Jin Jun
format Final Year Project
author Loh, Zachariah Jin Jun
author_sort Loh, Zachariah Jin Jun
title Arbitrary-view human action recognition for human robot interaction
title_short Arbitrary-view human action recognition for human robot interaction
title_full Arbitrary-view human action recognition for human robot interaction
title_fullStr Arbitrary-view human action recognition for human robot interaction
title_full_unstemmed Arbitrary-view human action recognition for human robot interaction
title_sort arbitrary-view human action recognition for human robot interaction
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/177242
_version_ 1800916368109862912