Real-time 3D face-eye performance capture of a person wearing VR headset

Teleconference or telepresence based on virtual reality (VR) head-mount display (HMD) device is a very interesting and promising application since HMD can provide immersive feelings for users. However, in order to facilitate face-to-face communications for HMD users, real-time 3D facial performance...

Full description

Saved in:
Bibliographic Details
Main Authors: Song, Guoxian, Cai, Jianfei, Cham, Tat-Jen, Zheng, Jianmin, Zhang, Juyong, Fuchs, Henry
Other Authors: School of Computer Science and Engineering
Format: Conference or Workshop Item
Language:English
Published: 2020
Subjects:
Online Access:https://hdl.handle.net/10356/138274
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-138274
record_format dspace
spelling sg-ntu-dr.10356-1382742020-04-30T02:36:32Z Real-time 3D face-eye performance capture of a person wearing VR headset Song, Guoxian Cai, Jianfei Cham, Tat-Jen Zheng, Jianmin Zhang, Juyong Fuchs, Henry School of Computer Science and Engineering 26th ACM international conference on Multimedia (MM '18) Institute for Media Innovation (IMI) Engineering::Computer science and engineering 3D Facial Reconstruction Gaze Estimation Teleconference or telepresence based on virtual reality (VR) head-mount display (HMD) device is a very interesting and promising application since HMD can provide immersive feelings for users. However, in order to facilitate face-to-face communications for HMD users, real-time 3D facial performance capture of a person wearing HMD is needed, which is a very challenging task due to the large occlusion caused by HMD. The existing limited solutions are very complex either in setting or in approach as well as lacking the performance capture of 3D eye gaze movement. In this paper, we propose a convolutional neural network (CNN) based solution for real-time 3D face-eye performance capture of HMD users without complex modification to devices. To address the issue of lacking training data, we generate massive pairs of HMD face-label dataset by data synthesis as well as collecting VR-IR eye dataset from multiple subjects. Then, we train a dense-fitting network for facial region and an eye gaze network to regress 3D eye model parameters. Extensive experimental results demonstrate that our system can efficiently and effectively produce in real time a vivid personalized 3D avatar with the correct identity, pose, expression and eye motion corresponding to the HMD user. NRF (Natl Research Foundation, S’pore) 2020-04-30T02:36:32Z 2020-04-30T02:36:32Z 2018 Conference Paper Song, G., Cai, J., Cham, T.-J., Zheng, J., Zhang, J., & Fuchs, H. (2018). Real-time 3D face-eye performance capture of a person wearing VR headset. Proceedings of the 26th ACM international conference on Multimedia (MM '18), 923-931. doi:10.1145/3240508.3240570 9781450356657 https://hdl.handle.net/10356/138274 10.1145/3240508.3240570 2-s2.0-85058230557 923 931 en © 2018 Association for Computing Machinery. All rights reserved.
institution Nanyang Technological University
building NTU Library
country Singapore
collection DR-NTU
language English
topic Engineering::Computer science and engineering
3D Facial Reconstruction
Gaze Estimation
spellingShingle Engineering::Computer science and engineering
3D Facial Reconstruction
Gaze Estimation
Song, Guoxian
Cai, Jianfei
Cham, Tat-Jen
Zheng, Jianmin
Zhang, Juyong
Fuchs, Henry
Real-time 3D face-eye performance capture of a person wearing VR headset
description Teleconference or telepresence based on virtual reality (VR) head-mount display (HMD) device is a very interesting and promising application since HMD can provide immersive feelings for users. However, in order to facilitate face-to-face communications for HMD users, real-time 3D facial performance capture of a person wearing HMD is needed, which is a very challenging task due to the large occlusion caused by HMD. The existing limited solutions are very complex either in setting or in approach as well as lacking the performance capture of 3D eye gaze movement. In this paper, we propose a convolutional neural network (CNN) based solution for real-time 3D face-eye performance capture of HMD users without complex modification to devices. To address the issue of lacking training data, we generate massive pairs of HMD face-label dataset by data synthesis as well as collecting VR-IR eye dataset from multiple subjects. Then, we train a dense-fitting network for facial region and an eye gaze network to regress 3D eye model parameters. Extensive experimental results demonstrate that our system can efficiently and effectively produce in real time a vivid personalized 3D avatar with the correct identity, pose, expression and eye motion corresponding to the HMD user.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Song, Guoxian
Cai, Jianfei
Cham, Tat-Jen
Zheng, Jianmin
Zhang, Juyong
Fuchs, Henry
format Conference or Workshop Item
author Song, Guoxian
Cai, Jianfei
Cham, Tat-Jen
Zheng, Jianmin
Zhang, Juyong
Fuchs, Henry
author_sort Song, Guoxian
title Real-time 3D face-eye performance capture of a person wearing VR headset
title_short Real-time 3D face-eye performance capture of a person wearing VR headset
title_full Real-time 3D face-eye performance capture of a person wearing VR headset
title_fullStr Real-time 3D face-eye performance capture of a person wearing VR headset
title_full_unstemmed Real-time 3D face-eye performance capture of a person wearing VR headset
title_sort real-time 3d face-eye performance capture of a person wearing vr headset
publishDate 2020
url https://hdl.handle.net/10356/138274
_version_ 1681057586996051968