Exploiting spatial-temporal relationships for 3D pose estimation via graph convolutional networks
Despite great progress in 3D pose estimation from single-view images or videos, it remains a challenging task due to the substantial depth ambiguity and severe selfocclusions. Motivated by the effectiveness of incorporating spatial dependencies and temporal consistencies to alleviate these issu...
Saved in:
Main Authors: | , , , , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2019
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/86102 http://hdl.handle.net/10220/49902 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Summary: | Despite great progress in 3D pose estimation from
single-view images or videos, it remains a challenging task
due to the substantial depth ambiguity and severe selfocclusions.
Motivated by the effectiveness of incorporating
spatial dependencies and temporal consistencies to alleviate
these issues, we propose a novel graph-based method
to tackle the problem of 3D human body and 3D hand
pose estimation from a short sequence of 2D joint detections.
Particularly, domain knowledge about the human
hand (body) configurations is explicitly incorporated into
the graph convolutional operations to meet the specific demand
of the 3D pose estimation. Furthermore, we introduce
a local-to-global network architecture, which is capable of
learning multi-scale features for the graph-based representations.
We evaluate the proposed method on challenging
benchmark datasets for both 3D hand pose estimation and
3D body pose estimation. Experimental results show that
our method achieves state-of-the-art performance on both
tasks. |
---|