Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation
In computer vision, image classification has progressed rapidly with deep learning over the ten years. However, in the real world, we still face challenges to apply them when the datasets are highly imbalanced, or in some situations to deploy large networks. From the data perspective, in this th...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Research |
Language: | English |
Published: |
Nanyang Technological University
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/155131 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-155131 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1551312022-03-06T05:18:17Z Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation Wang, Yiming Lin Guosheng School of Computer Science and Engineering gslin@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision In computer vision, image classification has progressed rapidly with deep learning over the ten years. However, in the real world, we still face challenges to apply them when the datasets are highly imbalanced, or in some situations to deploy large networks. From the data perspective, in this thesis, we aim to improve data augmentations for long-tailed image classification, where only a few semantic classes possess many samples while most other classes have only a few samples. We propose a novel Hybrid Mixup strategy to increase the sample amount and diversity, where we uncover the efficacy of mixup in the latent space of StyleGAN2. Compared with the traditional mixup method on real images, the mixup images generated from the interpolated latent codes have better quality. Experiments on CIFAR-10-LT, CIFAR-100-LT demonstrate that our proposed Hybrid Mixup consistently boosts the head-, medium- and tail-class classification accuracy compared with the traditional mixup method on real images only. Moreover, our results are on par with the state of the arts or even surpass them in some settings. From the model viewpoint, we particularly research the knowledge distillation, which leverages large models to distill enriched knowledge into smaller ones. Here we focus on the scenario where teachers output one-hot predictions only. We find it still possible for students to boost classification accuracy by directly learning from these one-hot predictions. We further propose Patched One-hot Distillation that models empirical probability for teachers to capture the inter-class relationship. Experiments on CIFAR-100 and ImageNet datasets demonstrate that our proposed method helps students learn better than the baseline that directly learns from both the ground-truth labels and the predictions from teachers. Master of Engineering 2022-02-08T02:13:49Z 2022-02-08T02:13:49Z 2021 Thesis-Master by Research Wang, Y. (2021). Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/155131 https://hdl.handle.net/10356/155131 10.32657/10356/155131 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision Wang, Yiming Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation |
description |
In computer vision, image classification has progressed rapidly with deep learning over the ten years. However, in the real world, we still face challenges to apply them when the datasets are highly imbalanced, or in some situations to deploy large networks.
From the data perspective, in this thesis, we aim to improve data augmentations for long-tailed image classification, where only a few semantic classes possess many samples while most other classes have only a few samples. We propose a novel Hybrid Mixup strategy to increase the sample amount and diversity, where we uncover the efficacy of mixup in the latent space of StyleGAN2. Compared with the traditional mixup method on real images, the mixup images generated from the interpolated latent codes have better quality. Experiments on CIFAR-10-LT, CIFAR-100-LT demonstrate that our proposed Hybrid Mixup consistently boosts the head-, medium- and tail-class classification accuracy compared with the traditional mixup method on real images only. Moreover, our results are on par with the state of the arts or even surpass them in some settings.
From the model viewpoint, we particularly research the knowledge distillation, which leverages large models to distill enriched knowledge into smaller ones. Here we focus on the scenario where teachers output one-hot predictions only. We find it still possible for students to boost classification accuracy by directly learning from these one-hot predictions. We further propose Patched One-hot Distillation that models empirical probability for teachers to capture the inter-class relationship. Experiments on CIFAR-100 and ImageNet datasets demonstrate that our proposed method helps students learn better than the baseline that directly learns from both the ground-truth labels and the predictions from teachers. |
author2 |
Lin Guosheng |
author_facet |
Lin Guosheng Wang, Yiming |
format |
Thesis-Master by Research |
author |
Wang, Yiming |
author_sort |
Wang, Yiming |
title |
Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation |
title_short |
Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation |
title_full |
Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation |
title_fullStr |
Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation |
title_full_unstemmed |
Addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation |
title_sort |
addressing challenges in real-world image classification : long-tailed distribution and knowledge distillation |
publisher |
Nanyang Technological University |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/155131 |
_version_ |
1726885513131982848 |