Everybody's talkin': let me talk as you want
We present a method to edit a target portrait footage by taking a sequence of audio as input to synthesize a photo-realistic video. This method is unique because it is highly dynamic. It does not assume a person-specific rendering network yet capable of translating one source audio into one random c...
Saved in:
Main Authors: | Song, Linsen, Wu, Wayne, Qian, Chen, He, Ran, Loy, Chen Change |
---|---|
Other Authors: | School of Computer Science and Engineering |
Format: | Article |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/162986 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Similar Items
-
Silence and face-work in two chinese TV talk shows
by: Lee, C.L., et al.
Published: (2016) -
(IM)POLITENESS IN MANDARIN TV TALK SHOW —— A CASE STUDY OF THE JINXING SHOW
by: ZHU DAONING
Published: (2023) -
Text2Human: text-driven controllable human image generation
by: Jiang, Yuming, et al.
Published: (2022) -
ID preserving face super-resolution generative adversarial networks
by: Li, J., et al.
Published: (2021) -
Epistemic modality in TED talks on education
by: Ton Nu, My Nhat, et al.
Published: (2019)