Exploring pre-trained diffusion models in a tuning-free manner

Diffusion models, which utilize a multi-step denoising sampling procedure and leverage extensive image-text pair datasets for training, have emerged as an innovative option among deep generative models. These models exhibit superior performance across various applications, including image synthes...

Full description

Saved in:
Bibliographic Details
Main Author: Wang, Jinghao
Other Authors: Liu Ziwei
Format: Thesis-Master by Research
Language:English
Published: Nanyang Technological University 2025
Subjects:
Online Access:https://hdl.handle.net/10356/181937
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-181937
record_format dspace
spelling sg-ntu-dr.10356-1819372025-02-05T01:58:53Z Exploring pre-trained diffusion models in a tuning-free manner Wang, Jinghao Liu Ziwei College of Computing and Data Science ziwei.liu@ntu.edu.sg Agricultural Sciences Deep learning Diffusion models, which utilize a multi-step denoising sampling procedure and leverage extensive image-text pair datasets for training, have emerged as an innovative option among deep generative models. These models exhibit superior performance across various applications, including image synthesis and video generation. In this thesis, we further explore applications of pre-trained diffusion models other than text-to-image generation applications in a tuning-free manner. In Chapter 1, we discuss image morphing between two real images via diffusion models. Our approach, FreeMorph, is based on key insights regarding attention interpolation and layout similarity in latent noise, which are critical for enhancing morphing quality. In Chapter 2, we discuss attention interpolation in diffusion models. This work introduces a novel training-free technique named Attention Interpolation via Diffusion (AID). AID has two key contributions: 1) a fused inner/outer interpolated attention layer to boost image consistency and fidelity; and 2) selection of interpolation coefficients via a beta distribution to increase smoothness. Master's degree 2025-01-03T02:33:11Z 2025-01-03T02:33:11Z 2024 Thesis-Master by Research Wang, J. (2024). Exploring pre-trained diffusion models in a tuning-free manner. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181937 https://hdl.handle.net/10356/181937 10.32657/10356/181937 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Agricultural Sciences
Deep learning
spellingShingle Agricultural Sciences
Deep learning
Wang, Jinghao
Exploring pre-trained diffusion models in a tuning-free manner
description Diffusion models, which utilize a multi-step denoising sampling procedure and leverage extensive image-text pair datasets for training, have emerged as an innovative option among deep generative models. These models exhibit superior performance across various applications, including image synthesis and video generation. In this thesis, we further explore applications of pre-trained diffusion models other than text-to-image generation applications in a tuning-free manner. In Chapter 1, we discuss image morphing between two real images via diffusion models. Our approach, FreeMorph, is based on key insights regarding attention interpolation and layout similarity in latent noise, which are critical for enhancing morphing quality. In Chapter 2, we discuss attention interpolation in diffusion models. This work introduces a novel training-free technique named Attention Interpolation via Diffusion (AID). AID has two key contributions: 1) a fused inner/outer interpolated attention layer to boost image consistency and fidelity; and 2) selection of interpolation coefficients via a beta distribution to increase smoothness.
author2 Liu Ziwei
author_facet Liu Ziwei
Wang, Jinghao
format Thesis-Master by Research
author Wang, Jinghao
author_sort Wang, Jinghao
title Exploring pre-trained diffusion models in a tuning-free manner
title_short Exploring pre-trained diffusion models in a tuning-free manner
title_full Exploring pre-trained diffusion models in a tuning-free manner
title_fullStr Exploring pre-trained diffusion models in a tuning-free manner
title_full_unstemmed Exploring pre-trained diffusion models in a tuning-free manner
title_sort exploring pre-trained diffusion models in a tuning-free manner
publisher Nanyang Technological University
publishDate 2025
url https://hdl.handle.net/10356/181937
_version_ 1823807346067898368