Exploring pre-trained diffusion models in a tuning-free manner
Diffusion models, which utilize a multi-step denoising sampling procedure and leverage extensive image-text pair datasets for training, have emerged as an innovative option among deep generative models. These models exhibit superior performance across various applications, including image synthes...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Research |
Language: | English |
Published: |
Nanyang Technological University
2025
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/181937 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-181937 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1819372025-02-05T01:58:53Z Exploring pre-trained diffusion models in a tuning-free manner Wang, Jinghao Liu Ziwei College of Computing and Data Science ziwei.liu@ntu.edu.sg Agricultural Sciences Deep learning Diffusion models, which utilize a multi-step denoising sampling procedure and leverage extensive image-text pair datasets for training, have emerged as an innovative option among deep generative models. These models exhibit superior performance across various applications, including image synthesis and video generation. In this thesis, we further explore applications of pre-trained diffusion models other than text-to-image generation applications in a tuning-free manner. In Chapter 1, we discuss image morphing between two real images via diffusion models. Our approach, FreeMorph, is based on key insights regarding attention interpolation and layout similarity in latent noise, which are critical for enhancing morphing quality. In Chapter 2, we discuss attention interpolation in diffusion models. This work introduces a novel training-free technique named Attention Interpolation via Diffusion (AID). AID has two key contributions: 1) a fused inner/outer interpolated attention layer to boost image consistency and fidelity; and 2) selection of interpolation coefficients via a beta distribution to increase smoothness. Master's degree 2025-01-03T02:33:11Z 2025-01-03T02:33:11Z 2024 Thesis-Master by Research Wang, J. (2024). Exploring pre-trained diffusion models in a tuning-free manner. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181937 https://hdl.handle.net/10356/181937 10.32657/10356/181937 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Agricultural Sciences Deep learning |
spellingShingle |
Agricultural Sciences Deep learning Wang, Jinghao Exploring pre-trained diffusion models in a tuning-free manner |
description |
Diffusion models, which utilize a multi-step denoising sampling procedure and
leverage extensive image-text pair datasets for training, have emerged as an innovative
option among deep generative models. These models exhibit superior
performance across various applications, including image synthesis and video generation.
In this thesis, we further explore applications of pre-trained diffusion
models other than text-to-image generation applications in a tuning-free manner.
In Chapter 1, we discuss image morphing between two real images via diffusion
models. Our approach, FreeMorph, is based on key insights regarding attention
interpolation and layout similarity in latent noise, which are critical for enhancing
morphing quality.
In Chapter 2, we discuss attention interpolation in diffusion models. This work
introduces a novel training-free technique named Attention Interpolation via
Diffusion (AID). AID has two key contributions: 1) a fused inner/outer interpolated
attention layer to boost image consistency and fidelity; and 2) selection of
interpolation coefficients via a beta distribution to increase smoothness. |
author2 |
Liu Ziwei |
author_facet |
Liu Ziwei Wang, Jinghao |
format |
Thesis-Master by Research |
author |
Wang, Jinghao |
author_sort |
Wang, Jinghao |
title |
Exploring pre-trained diffusion models in a tuning-free manner |
title_short |
Exploring pre-trained diffusion models in a tuning-free manner |
title_full |
Exploring pre-trained diffusion models in a tuning-free manner |
title_fullStr |
Exploring pre-trained diffusion models in a tuning-free manner |
title_full_unstemmed |
Exploring pre-trained diffusion models in a tuning-free manner |
title_sort |
exploring pre-trained diffusion models in a tuning-free manner |
publisher |
Nanyang Technological University |
publishDate |
2025 |
url |
https://hdl.handle.net/10356/181937 |
_version_ |
1823807346067898368 |