Exploring semantic information for deep image aesthetics assessment
Image Aesthetics Assessment (IAA) aims to evaluate the aesthetic pleasingness of an image. Constructing features that are discriminative in image aesthetics (i.e., aesthetic features) is essential for building a robust IAA model. Nevertheless, different from recognition-based tasks such as object cl...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Doctor of Philosophy |
Language: | English |
Published: |
Nanyang Technological University
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/169223 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-169223 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1692232023-08-01T07:08:34Z Exploring semantic information for deep image aesthetics assessment Hou, Jingwen Lin Weisi School of Computer Science and Engineering jingwen003@e.ntu.edu.sg Engineering::Computer science and engineering Image Aesthetics Assessment (IAA) aims to evaluate the aesthetic pleasingness of an image. Constructing features that are discriminative in image aesthetics (i.e., aesthetic features) is essential for building a robust IAA model. Nevertheless, different from recognition-based tasks such as object classification, the mapping between visual patterns on images and aesthetic labels is much more different, since a limited number of labels are related to much more diverse image contents. However, existing works on deep IAA have not sufficiently discussed how to improve deep IAA models by dealing with the connections between diverse image contents and limited aesthetic labels. Thus, in this thesis, we first propose to comprehend how a deep IAA model works by matching. That is, we assume a deep IAA model predicts the aesthetic label for a given image by implicitly 'matching' it to a seen training image with similar contents. Then the test image is assigned the aesthetic label of the training image with similar contents. Under this point of view, it is essential for a deep IAA model to have sufficient capability to represent the semantic contents of images in order to accurately match images with diverse contents. Therefore, this thesis explores how we can improve the discriminative power of aesthetic features constructed by deep IAA models by fully exploiting semantic information given by images. Specifically, we propose that aesthetic features should be constructed from sufficiently diverse semantic patterns. In the first work, we propose to distill knowledge on semantic patterns from a mixture of pre-trained object classification (POC) models to improve an end-to-end IAA model. We show that combining different POC models to help with feature extraction can more sufficiently exploit semantic information contained in the image, and can indeed lead to better IAA performance. However, what kinds of 'semantic information' have been captured by POC models are still unknown. Therefore, in the second work, we map visual semantic patterns captured by POC models to pre-defined textual tags to provide language-based descriptions of semantic information provided by an image, and construct aesthetic features from such language-based descriptions. Results show that we can indeed achieve good IAA performance with language-based descriptions of image semantic contents. In the third work, we further discuss how we can exploit personalized content preferences from features describing general semantic information of images when personal rating profiles are available. Finally, we come to a conclusion based on what we have investigated. Doctor of Philosophy 2023-07-13T01:01:23Z 2023-07-13T01:01:23Z 2023 Thesis-Doctor of Philosophy Hou, J. (2023). Exploring semantic information for deep image aesthetics assessment. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/169223 https://hdl.handle.net/10356/169223 10.32657/10356/169223 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering |
spellingShingle |
Engineering::Computer science and engineering Hou, Jingwen Exploring semantic information for deep image aesthetics assessment |
description |
Image Aesthetics Assessment (IAA) aims to evaluate the aesthetic pleasingness of an image. Constructing features that are discriminative in image aesthetics (i.e., aesthetic features) is essential for building a robust IAA model. Nevertheless, different from recognition-based tasks such as object classification, the mapping between visual patterns on images and aesthetic labels is much more different, since a limited number of labels are related to much more diverse image contents.
However, existing works on deep IAA have not sufficiently discussed how to improve deep IAA models by dealing with the connections between diverse image contents and limited aesthetic labels. Thus, in this thesis, we first propose to comprehend how a deep IAA model works by matching. That is, we assume a deep IAA model predicts the aesthetic label for a given image by implicitly 'matching' it to a seen training image with similar contents. Then the test image is assigned the aesthetic label of the training image with similar contents. Under this point of view, it is essential for a deep IAA model to have sufficient capability to represent the semantic contents of images in order to accurately match images with diverse contents. Therefore, this thesis explores how we can improve the discriminative power of aesthetic features constructed by deep IAA models by fully exploiting semantic information given by images. Specifically, we propose that aesthetic features should be constructed from sufficiently diverse semantic patterns.
In the first work, we propose to distill knowledge on semantic patterns from a mixture of pre-trained object classification (POC) models to improve an end-to-end IAA model. We show that combining different POC models to help with feature extraction can more sufficiently exploit semantic information contained in the image, and can indeed lead to better IAA performance. However, what kinds of 'semantic information' have been captured by POC models are still unknown. Therefore, in the second work, we map visual semantic patterns captured by POC models to pre-defined textual tags to provide language-based descriptions of semantic information provided by an image, and construct aesthetic features from such language-based descriptions. Results show that we can indeed achieve good IAA performance with language-based descriptions of image semantic contents. In the third work, we further discuss how we can exploit personalized content preferences from features describing general semantic information of images when personal rating profiles are available. Finally, we come to a conclusion based on what we have investigated. |
author2 |
Lin Weisi |
author_facet |
Lin Weisi Hou, Jingwen |
format |
Thesis-Doctor of Philosophy |
author |
Hou, Jingwen |
author_sort |
Hou, Jingwen |
title |
Exploring semantic information for deep image aesthetics assessment |
title_short |
Exploring semantic information for deep image aesthetics assessment |
title_full |
Exploring semantic information for deep image aesthetics assessment |
title_fullStr |
Exploring semantic information for deep image aesthetics assessment |
title_full_unstemmed |
Exploring semantic information for deep image aesthetics assessment |
title_sort |
exploring semantic information for deep image aesthetics assessment |
publisher |
Nanyang Technological University |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/169223 |
_version_ |
1773551421199745024 |