DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs
Recent advancements in 3D generative models have enabled the creation of high-quality 3D content, but intuitive user manipulation of these generated models remains a significant challenge. This project introduces DragGAN-3D, an innovative framework that extends the point-based manipulation capabilit...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/181164 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-181164 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1811642024-11-18T06:30:25Z DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs Wang, Haoxuan Xingang Pan College of Computing and Data Science xingang.pan@ntu.edu.sg Computer and Information Science Recent advancements in 3D generative models have enabled the creation of high-quality 3D content, but intuitive user manipulation of these generated models remains a significant challenge. This project introduces DragGAN-3D, an innovative framework that extends the point-based manipulation capabilities of DragGAN to the 3D domain, allowing users to intuitively edit 3D models by dragging control points in space. Building upon the tri-plane architecture of EG3D, our method enables precise, geometry-consistent modifications while maintaining multi-view consistency and high-fidelity 3D representations. The framework employs an iterative optimization process consisting of motion supervision and point tracking steps, with a dynamic discrete masking technique to control the scope of edits. We demonstrate the effectiveness of our approach through experiments on both randomly generated models and real-world images inverted into the latent space. Results show that DragGAN-3D successfully enables various edits, from subtle facial feature adjustments to geometric modifications, while preserving the overall quality and consistency of the 3D model. Our method is compatible with existing GAN inversion techniques, allowing the manipulation of real-world images, and proves robust across different datasets and pre-trained models. By bridging the gap between user intent and 3D model manipulation, DragGAN-3D represents a step forward in making 3D generative models more accessible and user-friendly for creative applications. Bachelor's degree 2024-11-18T06:30:25Z 2024-11-18T06:30:25Z 2024 Final Year Project (FYP) Wang, H. (2024). DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181164 https://hdl.handle.net/10356/181164 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science |
spellingShingle |
Computer and Information Science Wang, Haoxuan DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs |
description |
Recent advancements in 3D generative models have enabled the creation of high-quality 3D content, but intuitive user manipulation of these generated models remains a significant challenge. This project introduces DragGAN-3D, an innovative framework that extends the point-based manipulation capabilities of DragGAN to the 3D domain, allowing users to intuitively edit 3D models by dragging control points in space. Building upon the tri-plane architecture of EG3D, our method enables precise, geometry-consistent modifications while maintaining multi-view consistency and high-fidelity 3D representations.
The framework employs an iterative optimization process consisting of motion supervision and point tracking steps, with a dynamic discrete masking technique to control the scope of edits. We demonstrate the effectiveness of our approach through experiments on both randomly generated models and real-world images inverted into the latent space. Results show that DragGAN-3D successfully enables various edits, from subtle facial feature adjustments to geometric modifications, while preserving the overall quality and consistency of the 3D model.
Our method is compatible with existing GAN inversion techniques, allowing the manipulation of real-world images, and proves robust across different datasets and pre-trained models. By bridging the gap between user intent and 3D model manipulation, DragGAN-3D represents a step forward in making 3D generative models more accessible and user-friendly for creative applications. |
author2 |
Xingang Pan |
author_facet |
Xingang Pan Wang, Haoxuan |
format |
Final Year Project |
author |
Wang, Haoxuan |
author_sort |
Wang, Haoxuan |
title |
DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs |
title_short |
DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs |
title_full |
DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs |
title_fullStr |
DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs |
title_full_unstemmed |
DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs |
title_sort |
draggan-3d: interactive point-dragging manipulation of 3d gans |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/181164 |
_version_ |
1816859057101733888 |