SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition

Skeleton-based action recognition can effectively solve the problem of reduced classification accuracy caused by background clutter. However, prevalent skeleton datasets predominantly rely on cameras to capture RGB frames and annotate skeletal keypoints, leading to susceptibility to ambient lighting...

Full description

Saved in:
Bibliographic Details
Main Author: Huang, Shunyu
Other Authors: Xie Lihua
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2024
Subjects:
GCN
Online Access:https://hdl.handle.net/10356/174832
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-174832
record_format dspace
spelling sg-ntu-dr.10356-1748322024-04-19T15:58:07Z SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition Huang, Shunyu Xie Lihua School of Electrical and Electronic Engineering ELHXIE@ntu.edu.sg Engineering Skeleton-based action recognition GCN Wireless human sensing Transfer learning Skeleton-based action recognition can effectively solve the problem of reduced classification accuracy caused by background clutter. However, prevalent skeleton datasets predominantly rely on cameras to capture RGB frames and annotate skeletal keypoints, leading to susceptibility to ambient lighting fluctuations and potential privacy violations. To mitigate these challenges, leveraging non-invasive sensors such as LiDAR and mmWave for wireless human sensing emerges as a viable alternative. Nevertheless, the diminutive data size associated with these non-invasive sensing methodologies renders the direct application of RGB-based skeleton action classification models less than optimal. Moreover, extracting keypoints from non-invasive data lacks the precision found in RGB modalities, culminating in skeleton data marred by noise and information loss. To address these issues, our work involves cross-model knowledge transfer acquired from the data-rich RGB modality to our classification task, naming this framework SkeFi. For specific instances of information loss, we integrate the enhanced Temporal Correlation Adaptive Graph Convolution (TC-AGC). Additionally, our research underscores the capability of augmenting the intrinsic potency of multiscale temporal modeling through the integration of ESPNet modules. By combining TC-AGC with this improved temporal modeling and implementing transfer learning, our framework realizes superior performance across three non-invasive modalities (RGB, mmWave, and LiDAR). Master's degree 2024-04-15T03:05:38Z 2024-04-15T03:05:38Z 2024 Thesis-Master by Coursework Huang, S. (2024). SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/174832 https://hdl.handle.net/10356/174832 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
Skeleton-based action recognition
GCN
Wireless human sensing
Transfer learning
spellingShingle Engineering
Skeleton-based action recognition
GCN
Wireless human sensing
Transfer learning
Huang, Shunyu
SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition
description Skeleton-based action recognition can effectively solve the problem of reduced classification accuracy caused by background clutter. However, prevalent skeleton datasets predominantly rely on cameras to capture RGB frames and annotate skeletal keypoints, leading to susceptibility to ambient lighting fluctuations and potential privacy violations. To mitigate these challenges, leveraging non-invasive sensors such as LiDAR and mmWave for wireless human sensing emerges as a viable alternative. Nevertheless, the diminutive data size associated with these non-invasive sensing methodologies renders the direct application of RGB-based skeleton action classification models less than optimal. Moreover, extracting keypoints from non-invasive data lacks the precision found in RGB modalities, culminating in skeleton data marred by noise and information loss. To address these issues, our work involves cross-model knowledge transfer acquired from the data-rich RGB modality to our classification task, naming this framework SkeFi. For specific instances of information loss, we integrate the enhanced Temporal Correlation Adaptive Graph Convolution (TC-AGC). Additionally, our research underscores the capability of augmenting the intrinsic potency of multiscale temporal modeling through the integration of ESPNet modules. By combining TC-AGC with this improved temporal modeling and implementing transfer learning, our framework realizes superior performance across three non-invasive modalities (RGB, mmWave, and LiDAR).
author2 Xie Lihua
author_facet Xie Lihua
Huang, Shunyu
format Thesis-Master by Coursework
author Huang, Shunyu
author_sort Huang, Shunyu
title SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition
title_short SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition
title_full SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition
title_fullStr SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition
title_full_unstemmed SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition
title_sort skefi: cross-modal knowledge transfer for wireless skeleton-based action recognition
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/174832
_version_ 1800916200170979328