Deep learning for human motion generation
Human motion generation, a cornerstone technique in animation and video production, has widespread applications in various tasks like text-to-motion and music-to-dance. Previous works focus on developing specialist models tailored for each task without scalability. This project proposes Large Motion...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/176471 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-176471 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1764712024-05-24T15:49:19Z Deep learning for human motion generation Gu, Chenyang Liu Ziwei Wen Bihan School of Electrical and Electronic Engineering bihan.wen@ntu.edu.sg, ziwei.liu@ntu.edu.sg Computer and Information Science Computer vision Human motion generation, a cornerstone technique in animation and video production, has widespread applications in various tasks like text-to-motion and music-to-dance. Previous works focus on developing specialist models tailored for each task without scalability. This project proposes Large Motion Model (LMM), a motion-centric, multi-modal framework that unifies mainstream motion generation tasks into a generalist model. A unified motion model is appealing since it can leverage a wide range of motion data to achieve broad generalization beyond a single task. However, it is also challenging due to the heterogeneous nature of substantially different motion data and tasks. LMM tackles these challenges from three principled aspects: 1) Data: Datasets with different modalities, formats and tasks are consolidated into a comprehensive yet unified motion generation dataset, MotionVerse, comprising 10 tasks, 16 datasets, a total of 320k sequences, and 100 million frames. 2) Architecture: An articulated attention mechanism ArtAttention that incorporates body part-aware modeling into Diffusion Transformer backbone is designed. 3) Pre-Training: A novel pre-training strategy is proposed for LMM, which employs variable frame rates and masking forms, to better exploit knowledge from diverse training data. Extensive experiments demonstrate that our generalist LMM achieves competitive performance across various standard motion generation tasks over state-of-the-art specialist models. Notably, LMM exhibits strong generalization capabilities and emerging properties across many unseen tasks. Additionally, this project includes ablation studies that reveal valuable insights about training and scaling up large motion models for future research. Bachelor's degree 2024-05-21T01:46:06Z 2024-05-21T01:46:06Z 2024 Final Year Project (FYP) Gu, C. (2024). Deep learning for human motion generation. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/176471 https://hdl.handle.net/10356/176471 en A3234-231 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Computer vision |
spellingShingle |
Computer and Information Science Computer vision Gu, Chenyang Deep learning for human motion generation |
description |
Human motion generation, a cornerstone technique in animation and video production, has widespread applications in various tasks like text-to-motion and music-to-dance. Previous works focus on developing specialist models tailored for each task without scalability. This project proposes Large Motion Model (LMM), a motion-centric, multi-modal framework that unifies mainstream motion generation tasks into a generalist model. A unified motion model is appealing since it can leverage a wide range of motion data to achieve broad generalization beyond a single task. However, it is also challenging due to the heterogeneous nature of substantially different motion data and tasks. LMM tackles these challenges from three principled aspects: 1) Data: Datasets with different modalities, formats and tasks are consolidated into a comprehensive yet unified motion generation dataset, MotionVerse, comprising 10 tasks, 16 datasets, a total of 320k sequences, and 100 million frames. 2) Architecture: An articulated attention mechanism ArtAttention that incorporates body part-aware modeling into Diffusion Transformer backbone is designed. 3) Pre-Training: A novel pre-training strategy is proposed for LMM, which employs variable frame rates and masking forms, to better exploit knowledge from diverse training data. Extensive experiments demonstrate that our generalist LMM achieves competitive performance across various standard motion generation tasks over state-of-the-art specialist models. Notably, LMM exhibits strong generalization capabilities and emerging properties across many unseen tasks. Additionally, this project includes ablation studies that reveal valuable insights about training and scaling up large motion models for future research. |
author2 |
Liu Ziwei |
author_facet |
Liu Ziwei Gu, Chenyang |
format |
Final Year Project |
author |
Gu, Chenyang |
author_sort |
Gu, Chenyang |
title |
Deep learning for human motion generation |
title_short |
Deep learning for human motion generation |
title_full |
Deep learning for human motion generation |
title_fullStr |
Deep learning for human motion generation |
title_full_unstemmed |
Deep learning for human motion generation |
title_sort |
deep learning for human motion generation |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/176471 |
_version_ |
1814047310481457152 |