Real-time 3D face-eye performance capture of a person wearing VR headset
Teleconference or telepresence based on virtual reality (VR) head-mount display (HMD) device is a very interesting and promising application since HMD can provide immersive feelings for users. However, in order to facilitate face-to-face communications for HMD users, real-time 3D facial performance...
Saved in:
Main Authors: | , , , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2020
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/138274 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-138274 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1382742020-04-30T02:36:32Z Real-time 3D face-eye performance capture of a person wearing VR headset Song, Guoxian Cai, Jianfei Cham, Tat-Jen Zheng, Jianmin Zhang, Juyong Fuchs, Henry School of Computer Science and Engineering 26th ACM international conference on Multimedia (MM '18) Institute for Media Innovation (IMI) Engineering::Computer science and engineering 3D Facial Reconstruction Gaze Estimation Teleconference or telepresence based on virtual reality (VR) head-mount display (HMD) device is a very interesting and promising application since HMD can provide immersive feelings for users. However, in order to facilitate face-to-face communications for HMD users, real-time 3D facial performance capture of a person wearing HMD is needed, which is a very challenging task due to the large occlusion caused by HMD. The existing limited solutions are very complex either in setting or in approach as well as lacking the performance capture of 3D eye gaze movement. In this paper, we propose a convolutional neural network (CNN) based solution for real-time 3D face-eye performance capture of HMD users without complex modification to devices. To address the issue of lacking training data, we generate massive pairs of HMD face-label dataset by data synthesis as well as collecting VR-IR eye dataset from multiple subjects. Then, we train a dense-fitting network for facial region and an eye gaze network to regress 3D eye model parameters. Extensive experimental results demonstrate that our system can efficiently and effectively produce in real time a vivid personalized 3D avatar with the correct identity, pose, expression and eye motion corresponding to the HMD user. NRF (Natl Research Foundation, S’pore) 2020-04-30T02:36:32Z 2020-04-30T02:36:32Z 2018 Conference Paper Song, G., Cai, J., Cham, T.-J., Zheng, J., Zhang, J., & Fuchs, H. (2018). Real-time 3D face-eye performance capture of a person wearing VR headset. Proceedings of the 26th ACM international conference on Multimedia (MM '18), 923-931. doi:10.1145/3240508.3240570 9781450356657 https://hdl.handle.net/10356/138274 10.1145/3240508.3240570 2-s2.0-85058230557 923 931 en © 2018 Association for Computing Machinery. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
country |
Singapore |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering 3D Facial Reconstruction Gaze Estimation |
spellingShingle |
Engineering::Computer science and engineering 3D Facial Reconstruction Gaze Estimation Song, Guoxian Cai, Jianfei Cham, Tat-Jen Zheng, Jianmin Zhang, Juyong Fuchs, Henry Real-time 3D face-eye performance capture of a person wearing VR headset |
description |
Teleconference or telepresence based on virtual reality (VR) head-mount display (HMD) device is a very interesting and promising application since HMD can provide immersive feelings for users. However, in order to facilitate face-to-face communications for HMD users, real-time 3D facial performance capture of a person wearing HMD is needed, which is a very challenging task due to the large occlusion caused by HMD. The existing limited solutions are very complex either in setting or in approach as well as lacking the performance capture of 3D eye gaze movement. In this paper, we propose a convolutional neural network (CNN) based solution for real-time 3D face-eye performance capture of HMD users without complex modification to devices. To address the issue of lacking training data, we generate massive pairs of HMD face-label dataset by data synthesis as well as collecting VR-IR eye dataset from multiple subjects. Then, we train a dense-fitting network for facial region and an eye gaze network to regress 3D eye model parameters. Extensive experimental results demonstrate that our system can efficiently and effectively produce in real time a vivid personalized 3D avatar with the correct identity, pose, expression and eye motion corresponding to the HMD user. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Song, Guoxian Cai, Jianfei Cham, Tat-Jen Zheng, Jianmin Zhang, Juyong Fuchs, Henry |
format |
Conference or Workshop Item |
author |
Song, Guoxian Cai, Jianfei Cham, Tat-Jen Zheng, Jianmin Zhang, Juyong Fuchs, Henry |
author_sort |
Song, Guoxian |
title |
Real-time 3D face-eye performance capture of a person wearing VR headset |
title_short |
Real-time 3D face-eye performance capture of a person wearing VR headset |
title_full |
Real-time 3D face-eye performance capture of a person wearing VR headset |
title_fullStr |
Real-time 3D face-eye performance capture of a person wearing VR headset |
title_full_unstemmed |
Real-time 3D face-eye performance capture of a person wearing VR headset |
title_sort |
real-time 3d face-eye performance capture of a person wearing vr headset |
publishDate |
2020 |
url |
https://hdl.handle.net/10356/138274 |
_version_ |
1681057586996051968 |