Emotional facial expression transfer using motion capture data
An input-output temporal restricted Boltzmann machine is an artificial neural network that learns the probability distribution between input sequences and output sequences, and then uses the model to predict output sequences. Unlike other static models, IOTRBM can catch the details of non-linear fac...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
2015
|
Subjects: | |
Online Access: | http://hdl.handle.net/10356/63093 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-63093 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-630932023-03-03T20:34:21Z Emotional facial expression transfer using motion capture data Chandra, Ellensi Rey Huang Dongyan Lin Weisi School of Computer Engineering A*STAR Institute for Infocomm Research (I2R) DRNTU::Engineering::Computer science and engineering::Computing methodologies::Simulation and modeling DRNTU::Engineering::Computer science and engineering::Computing methodologies::Pattern recognition An input-output temporal restricted Boltzmann machine is an artificial neural network that learns the probability distribution between input sequences and output sequences, and then uses the model to predict output sequences. Unlike other static models, IOTRBM can catch the details of non-linear facial movements and eliminate irrelevant temporal noise, resulting in realistic predicted sequences. This project covers pre-processing raw motion capture data of neutral face expression and happy face expression and pass them as training data and testing data to the IOTRBM. The pre-processing tasks include recovering missing data points, removing silent parts in the motion capture data, and exercising canonical time warping (CTW) to temporally align the data frames based on visemes or phonemes. After passing the data sequences to IOTRBM and training the model, natural-looking happy expression sequences have been predicted based on the neutral expression sequences. Bachelor of Engineering (Computer Science) 2015-05-06T02:40:03Z 2015-05-06T02:40:03Z 2015 2015 Final Year Project (FYP) http://hdl.handle.net/10356/63093 en Nanyang Technological University 34 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Computer science and engineering::Computing methodologies::Simulation and modeling DRNTU::Engineering::Computer science and engineering::Computing methodologies::Pattern recognition |
spellingShingle |
DRNTU::Engineering::Computer science and engineering::Computing methodologies::Simulation and modeling DRNTU::Engineering::Computer science and engineering::Computing methodologies::Pattern recognition Chandra, Ellensi Rey Emotional facial expression transfer using motion capture data |
description |
An input-output temporal restricted Boltzmann machine is an artificial neural network that learns the probability distribution between input sequences and output sequences, and then uses the model to predict output sequences. Unlike other static models, IOTRBM can catch the details of non-linear facial movements and eliminate irrelevant temporal noise, resulting in realistic predicted sequences. This project covers pre-processing raw motion capture data of neutral face expression and happy face expression and pass them as training data and testing data to the IOTRBM. The pre-processing tasks include recovering missing data points, removing silent parts in the motion capture data, and exercising canonical time warping (CTW) to temporally align the data frames based on visemes or phonemes. After passing the data sequences to IOTRBM and training the model, natural-looking happy expression sequences have been predicted based on the neutral expression sequences. |
author2 |
Huang Dongyan |
author_facet |
Huang Dongyan Chandra, Ellensi Rey |
format |
Final Year Project |
author |
Chandra, Ellensi Rey |
author_sort |
Chandra, Ellensi Rey |
title |
Emotional facial expression transfer using motion capture data |
title_short |
Emotional facial expression transfer using motion capture data |
title_full |
Emotional facial expression transfer using motion capture data |
title_fullStr |
Emotional facial expression transfer using motion capture data |
title_full_unstemmed |
Emotional facial expression transfer using motion capture data |
title_sort |
emotional facial expression transfer using motion capture data |
publishDate |
2015 |
url |
http://hdl.handle.net/10356/63093 |
_version_ |
1759858393150914560 |