In-the-wild image quality assessment with diffusion priors
Blind image quality assessment (IQA) in the wild, which assesses the quality of images with complex authentic distortions and no reference images, presents significant challenges. Given the difficulty in collecting large-scale training data, leveraging limited data to develop a model with strong gen...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Coursework |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/181882 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-181882 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1818822024-12-27T15:46:13Z In-the-wild image quality assessment with diffusion priors Fu, Honghao Wen Bihan School of Electrical and Electronic Engineering Rapid-Rich Object Search (ROSE) Lab bihan.wen@ntu.edu.sg Computer and Information Science Blind IQA Diffusion prior Text-to-image model Knowledge distillation Blind image quality assessment (IQA) in the wild, which assesses the quality of images with complex authentic distortions and no reference images, presents significant challenges. Given the difficulty in collecting large-scale training data, leveraging limited data to develop a model with strong generalization remains an open problem. Motivated by the robust image perception capabilities of pretrained text-to image (T2I) diffusion models, we propose a novel IQA method, diffusion priors-based IQA (DP-IQA), to utilize the T2I model’s prior for improved performance and generalization ability. Specifically, we utilize pre-trained Stable Diffusion as the backbone, extracting multi-level features from the denoising U-Net guided by prompt embeddings through a tunable text adapter. Simultaneously, an image adapter compensates for information loss introduced by the lossy pre-trained encoder. Unlike T2I models that require full image distribution modeling, our approach targets image quality assessment, which inherently requires fewer parameters. To improve applicability, we distill the knowledge into a lightweight CNN-based student model, significantly reducing parameters while maintaining or even enhancing generalization performance. Experimental results demonstrate that DP-IQA achieves state-of-the-art performance on various in-the-wild datasets, highlighting the superior generalization capability of T2I priors in blind IQA tasks. To our knowledge, DP-IQA is the first method to apply pre-trained diffusion priors in blind IQA. Master's degree 2024-12-27T13:31:16Z 2024-12-27T13:31:16Z 2024 Thesis-Master by Coursework Fu, H. (2024). In-the-wild image quality assessment with diffusion priors. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181882 https://hdl.handle.net/10356/181882 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Blind IQA Diffusion prior Text-to-image model Knowledge distillation |
spellingShingle |
Computer and Information Science Blind IQA Diffusion prior Text-to-image model Knowledge distillation Fu, Honghao In-the-wild image quality assessment with diffusion priors |
description |
Blind image quality assessment (IQA) in the wild, which assesses the quality of images with complex authentic distortions and no reference images, presents significant challenges. Given the difficulty in collecting large-scale training data, leveraging limited data to develop a model with strong generalization remains an open problem. Motivated by the robust image perception capabilities of pretrained text-to image (T2I) diffusion models, we propose a novel IQA method, diffusion priors-based IQA (DP-IQA), to utilize the T2I model’s prior for improved performance and generalization ability. Specifically, we utilize pre-trained Stable Diffusion as the backbone, extracting multi-level features from the denoising U-Net guided by prompt embeddings through a tunable text adapter. Simultaneously, an image adapter compensates for information loss introduced by the lossy pre-trained encoder. Unlike T2I models that require full image distribution modeling, our approach targets image quality assessment, which inherently requires fewer parameters. To improve applicability, we distill the knowledge into a lightweight CNN-based student model, significantly reducing parameters while maintaining or even enhancing generalization performance. Experimental results demonstrate that DP-IQA achieves state-of-the-art performance on various in-the-wild datasets, highlighting the superior generalization capability of T2I priors in blind IQA tasks. To our knowledge, DP-IQA is the first method to apply pre-trained diffusion priors in blind IQA. |
author2 |
Wen Bihan |
author_facet |
Wen Bihan Fu, Honghao |
format |
Thesis-Master by Coursework |
author |
Fu, Honghao |
author_sort |
Fu, Honghao |
title |
In-the-wild image quality assessment with diffusion priors |
title_short |
In-the-wild image quality assessment with diffusion priors |
title_full |
In-the-wild image quality assessment with diffusion priors |
title_fullStr |
In-the-wild image quality assessment with diffusion priors |
title_full_unstemmed |
In-the-wild image quality assessment with diffusion priors |
title_sort |
in-the-wild image quality assessment with diffusion priors |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/181882 |
_version_ |
1820027784748072960 |