DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs

Recent advancements in 3D generative models have enabled the creation of high-quality 3D content, but intuitive user manipulation of these generated models remains a significant challenge. This project introduces DragGAN-3D, an innovative framework that extends the point-based manipulation capabilit...

Full description

Saved in:
Bibliographic Details
Main Author: Wang, Haoxuan
Other Authors: Xingang Pan
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/181164
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-181164
record_format dspace
spelling sg-ntu-dr.10356-1811642024-11-18T06:30:25Z DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs Wang, Haoxuan Xingang Pan College of Computing and Data Science xingang.pan@ntu.edu.sg Computer and Information Science Recent advancements in 3D generative models have enabled the creation of high-quality 3D content, but intuitive user manipulation of these generated models remains a significant challenge. This project introduces DragGAN-3D, an innovative framework that extends the point-based manipulation capabilities of DragGAN to the 3D domain, allowing users to intuitively edit 3D models by dragging control points in space. Building upon the tri-plane architecture of EG3D, our method enables precise, geometry-consistent modifications while maintaining multi-view consistency and high-fidelity 3D representations. The framework employs an iterative optimization process consisting of motion supervision and point tracking steps, with a dynamic discrete masking technique to control the scope of edits. We demonstrate the effectiveness of our approach through experiments on both randomly generated models and real-world images inverted into the latent space. Results show that DragGAN-3D successfully enables various edits, from subtle facial feature adjustments to geometric modifications, while preserving the overall quality and consistency of the 3D model. Our method is compatible with existing GAN inversion techniques, allowing the manipulation of real-world images, and proves robust across different datasets and pre-trained models. By bridging the gap between user intent and 3D model manipulation, DragGAN-3D represents a step forward in making 3D generative models more accessible and user-friendly for creative applications. Bachelor's degree 2024-11-18T06:30:25Z 2024-11-18T06:30:25Z 2024 Final Year Project (FYP) Wang, H. (2024). DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181164 https://hdl.handle.net/10356/181164 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
spellingShingle Computer and Information Science
Wang, Haoxuan
DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs
description Recent advancements in 3D generative models have enabled the creation of high-quality 3D content, but intuitive user manipulation of these generated models remains a significant challenge. This project introduces DragGAN-3D, an innovative framework that extends the point-based manipulation capabilities of DragGAN to the 3D domain, allowing users to intuitively edit 3D models by dragging control points in space. Building upon the tri-plane architecture of EG3D, our method enables precise, geometry-consistent modifications while maintaining multi-view consistency and high-fidelity 3D representations. The framework employs an iterative optimization process consisting of motion supervision and point tracking steps, with a dynamic discrete masking technique to control the scope of edits. We demonstrate the effectiveness of our approach through experiments on both randomly generated models and real-world images inverted into the latent space. Results show that DragGAN-3D successfully enables various edits, from subtle facial feature adjustments to geometric modifications, while preserving the overall quality and consistency of the 3D model. Our method is compatible with existing GAN inversion techniques, allowing the manipulation of real-world images, and proves robust across different datasets and pre-trained models. By bridging the gap between user intent and 3D model manipulation, DragGAN-3D represents a step forward in making 3D generative models more accessible and user-friendly for creative applications.
author2 Xingang Pan
author_facet Xingang Pan
Wang, Haoxuan
format Final Year Project
author Wang, Haoxuan
author_sort Wang, Haoxuan
title DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs
title_short DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs
title_full DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs
title_fullStr DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs
title_full_unstemmed DRAGGAN-3D: Interactive point-dragging manipulation of 3D GANs
title_sort draggan-3d: interactive point-dragging manipulation of 3d gans
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/181164
_version_ 1816859057101733888