Global context with discrete diffusion in vector quantised modelling for image generation

The integration of Vector Quantised Variational AutoEncoder (VQ-VAE) with autoregressive models as generation part has yielded high-quality results on image generation. However, the autoregressive models will strictly follow the progressive scanning order during the sampling phase. This leads the ex...

Full description

Saved in:
Bibliographic Details
Main Authors: Hu, Minghui, Wang, Yujie, Cham, Tat-Jen, Yang, Jianfei, Suganthan, Ponnuthurai Nagaratnam
Other Authors: School of Computer Science and Engineering
Format: Conference or Workshop Item
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/172658
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-172658
record_format dspace
spelling sg-ntu-dr.10356-1726582023-12-19T04:51:15Z Global context with discrete diffusion in vector quantised modelling for image generation Hu, Minghui Wang, Yujie Cham, Tat-Jen Yang, Jianfei Suganthan, Ponnuthurai Nagaratnam School of Computer Science and Engineering 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Visualization Image Resolution The integration of Vector Quantised Variational AutoEncoder (VQ-VAE) with autoregressive models as generation part has yielded high-quality results on image generation. However, the autoregressive models will strictly follow the progressive scanning order during the sampling phase. This leads the existing VQ series models to hardly escape the trap of lacking global information. Denoising Diffusion Probabilistic Models (DDPM) in the continuous domain have shown a capability to capture the global context, while generating high-quality images. In the discrete state space, some works have demonstrated the potential to perform text generation and low resolution image generation. We show that with the help of a content-rich discrete visual codebook from VQ-VAE, the discrete diffusion model can also generate high fidelity images with global context, which compensates for the deficiency of the classical autoregressive model along pixel space. Meanwhile, the integration of the discrete VAE with the diffusion model resolves the drawback of conventional autoregressive models being oversized, and the diffusion model which demands excessive time in the sampling process when generating images. It is found that the quality of the generated images is heavily dependent on the discrete visual codebook. Extensive experiments demonstrate that the proposed Vector Quantised Discrete Diffusion Model (VQ-DDM) is able to achieve comparable performance to top-tier methods with low complexity. It also demonstrates outstanding advantages over other vectors quantised with autoregressive models in terms of image inpainting tasks without additional training. 2023-12-19T04:51:14Z 2023-12-19T04:51:14Z 2022 Conference Paper Hu, M., Wang, Y., Cham, T., Yang, J. & Suganthan, P. N. (2022). Global context with discrete diffusion in vector quantised modelling for image generation. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 11492-11501. https://dx.doi.org/10.1109/CVPR52688.2022.01121 9781665469463 https://hdl.handle.net/10356/172658 10.1109/CVPR52688.2022.01121 2-s2.0-85138499465 11492 11501 en © 2022 IEEE. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
Visualization
Image Resolution
spellingShingle Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
Visualization
Image Resolution
Hu, Minghui
Wang, Yujie
Cham, Tat-Jen
Yang, Jianfei
Suganthan, Ponnuthurai Nagaratnam
Global context with discrete diffusion in vector quantised modelling for image generation
description The integration of Vector Quantised Variational AutoEncoder (VQ-VAE) with autoregressive models as generation part has yielded high-quality results on image generation. However, the autoregressive models will strictly follow the progressive scanning order during the sampling phase. This leads the existing VQ series models to hardly escape the trap of lacking global information. Denoising Diffusion Probabilistic Models (DDPM) in the continuous domain have shown a capability to capture the global context, while generating high-quality images. In the discrete state space, some works have demonstrated the potential to perform text generation and low resolution image generation. We show that with the help of a content-rich discrete visual codebook from VQ-VAE, the discrete diffusion model can also generate high fidelity images with global context, which compensates for the deficiency of the classical autoregressive model along pixel space. Meanwhile, the integration of the discrete VAE with the diffusion model resolves the drawback of conventional autoregressive models being oversized, and the diffusion model which demands excessive time in the sampling process when generating images. It is found that the quality of the generated images is heavily dependent on the discrete visual codebook. Extensive experiments demonstrate that the proposed Vector Quantised Discrete Diffusion Model (VQ-DDM) is able to achieve comparable performance to top-tier methods with low complexity. It also demonstrates outstanding advantages over other vectors quantised with autoregressive models in terms of image inpainting tasks without additional training.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Hu, Minghui
Wang, Yujie
Cham, Tat-Jen
Yang, Jianfei
Suganthan, Ponnuthurai Nagaratnam
format Conference or Workshop Item
author Hu, Minghui
Wang, Yujie
Cham, Tat-Jen
Yang, Jianfei
Suganthan, Ponnuthurai Nagaratnam
author_sort Hu, Minghui
title Global context with discrete diffusion in vector quantised modelling for image generation
title_short Global context with discrete diffusion in vector quantised modelling for image generation
title_full Global context with discrete diffusion in vector quantised modelling for image generation
title_fullStr Global context with discrete diffusion in vector quantised modelling for image generation
title_full_unstemmed Global context with discrete diffusion in vector quantised modelling for image generation
title_sort global context with discrete diffusion in vector quantised modelling for image generation
publishDate 2023
url https://hdl.handle.net/10356/172658
_version_ 1787136421372362752