SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition
Skeleton-based action recognition can effectively solve the problem of reduced classification accuracy caused by background clutter. However, prevalent skeleton datasets predominantly rely on cameras to capture RGB frames and annotate skeletal keypoints, leading to susceptibility to ambient lighting...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Coursework |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/174832 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-174832 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1748322024-04-19T15:58:07Z SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition Huang, Shunyu Xie Lihua School of Electrical and Electronic Engineering ELHXIE@ntu.edu.sg Engineering Skeleton-based action recognition GCN Wireless human sensing Transfer learning Skeleton-based action recognition can effectively solve the problem of reduced classification accuracy caused by background clutter. However, prevalent skeleton datasets predominantly rely on cameras to capture RGB frames and annotate skeletal keypoints, leading to susceptibility to ambient lighting fluctuations and potential privacy violations. To mitigate these challenges, leveraging non-invasive sensors such as LiDAR and mmWave for wireless human sensing emerges as a viable alternative. Nevertheless, the diminutive data size associated with these non-invasive sensing methodologies renders the direct application of RGB-based skeleton action classification models less than optimal. Moreover, extracting keypoints from non-invasive data lacks the precision found in RGB modalities, culminating in skeleton data marred by noise and information loss. To address these issues, our work involves cross-model knowledge transfer acquired from the data-rich RGB modality to our classification task, naming this framework SkeFi. For specific instances of information loss, we integrate the enhanced Temporal Correlation Adaptive Graph Convolution (TC-AGC). Additionally, our research underscores the capability of augmenting the intrinsic potency of multiscale temporal modeling through the integration of ESPNet modules. By combining TC-AGC with this improved temporal modeling and implementing transfer learning, our framework realizes superior performance across three non-invasive modalities (RGB, mmWave, and LiDAR). Master's degree 2024-04-15T03:05:38Z 2024-04-15T03:05:38Z 2024 Thesis-Master by Coursework Huang, S. (2024). SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/174832 https://hdl.handle.net/10356/174832 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering Skeleton-based action recognition GCN Wireless human sensing Transfer learning |
spellingShingle |
Engineering Skeleton-based action recognition GCN Wireless human sensing Transfer learning Huang, Shunyu SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition |
description |
Skeleton-based action recognition can effectively solve the problem of reduced classification accuracy caused by background clutter. However, prevalent skeleton datasets predominantly rely on cameras to capture RGB frames and annotate skeletal keypoints, leading to susceptibility to ambient lighting fluctuations and potential privacy violations. To mitigate these challenges, leveraging non-invasive sensors such as LiDAR and mmWave for wireless human sensing emerges as a viable alternative. Nevertheless, the diminutive data size associated with these non-invasive sensing methodologies renders the direct application of RGB-based skeleton action classification models less than optimal. Moreover, extracting keypoints from non-invasive data lacks the precision found in RGB modalities, culminating in skeleton data marred by noise and information loss. To address these issues, our work involves cross-model knowledge transfer acquired from the data-rich RGB modality to our classification task, naming this framework SkeFi. For specific instances of information loss, we integrate the enhanced Temporal Correlation Adaptive Graph Convolution (TC-AGC). Additionally, our research underscores the capability of augmenting the intrinsic potency of multiscale temporal modeling through the integration of ESPNet modules. By combining TC-AGC with this improved temporal modeling and implementing transfer learning, our framework realizes superior performance across three non-invasive modalities (RGB, mmWave, and LiDAR). |
author2 |
Xie Lihua |
author_facet |
Xie Lihua Huang, Shunyu |
format |
Thesis-Master by Coursework |
author |
Huang, Shunyu |
author_sort |
Huang, Shunyu |
title |
SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition |
title_short |
SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition |
title_full |
SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition |
title_fullStr |
SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition |
title_full_unstemmed |
SkeFi: cross-modal knowledge transfer for wireless skeleton-based action recognition |
title_sort |
skefi: cross-modal knowledge transfer for wireless skeleton-based action recognition |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/174832 |
_version_ |
1800916200170979328 |