Knowledge Distillation with Relative Representations for Image Representation Learning
Relative representations allow the alignment of latent spaces which embed data in extrinsically different manners but with similar relative distances between data points. This ability to compare different latent spaces for the same input lends itself to knowledge distillation techniques. We explore...
Saved in:
Main Authors: | , , |
---|---|
Format: | text |
Published: |
Archīum Ateneo
2023
|
Subjects: | |
Online Access: | https://archium.ateneo.edu/discs-faculty-pubs/387 https://doi.org/10.1007/978-3-031-41630-9_14 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Ateneo De Manila University |
id |
ph-ateneo-arc.discs-faculty-pubs-1387 |
---|---|
record_format |
eprints |
spelling |
ph-ateneo-arc.discs-faculty-pubs-13872024-02-21T02:51:52Z Knowledge Distillation with Relative Representations for Image Representation Learning Ramos, Patrick Alampay, Raphael Abu, Patricia Angela R Relative representations allow the alignment of latent spaces which embed data in extrinsically different manners but with similar relative distances between data points. This ability to compare different latent spaces for the same input lends itself to knowledge distillation techniques. We explore the applicability of relative representations to knowledge distillation by training a student model such that the relative representations of its outputs match the relative representations of the outputs of a teacher model. We test our Relative Representation Knowledge Distillation (RRKD) scheme on supervised and self-supervised image representation learning with MNIST and show that an encoder can be compressed to 47.71% of its original size while maintaining 91.92% of its full performance. We demonstrate that RRKD is competitive with or outperforms other relation-based distillation schemes in traditional distillation setups (CIFAR-10, CIFAR-100, SVHN) and in a transfer learning setting (Stanford Cars, Oxford-IIIT Pets, Oxford Flowers-102). Our results indicate that relative representations are an effective signal for knowledge distillation. Code is made available at https://github.com/Ramos-Ramos/rrkd. 2023-01-01T08:00:00Z text https://archium.ateneo.edu/discs-faculty-pubs/387 https://doi.org/10.1007/978-3-031-41630-9_14 Department of Information Systems & Computer Science Faculty Publications Archīum Ateneo Knowledge distillation Latent space Relative representations Computer Engineering Electrical and Computer Engineering Engineering |
institution |
Ateneo De Manila University |
building |
Ateneo De Manila University Library |
continent |
Asia |
country |
Philippines Philippines |
content_provider |
Ateneo De Manila University Library |
collection |
archium.Ateneo Institutional Repository |
topic |
Knowledge distillation Latent space Relative representations Computer Engineering Electrical and Computer Engineering Engineering |
spellingShingle |
Knowledge distillation Latent space Relative representations Computer Engineering Electrical and Computer Engineering Engineering Ramos, Patrick Alampay, Raphael Abu, Patricia Angela R Knowledge Distillation with Relative Representations for Image Representation Learning |
description |
Relative representations allow the alignment of latent spaces which embed data in extrinsically different manners but with similar relative distances between data points. This ability to compare different latent spaces for the same input lends itself to knowledge distillation techniques. We explore the applicability of relative representations to knowledge distillation by training a student model such that the relative representations of its outputs match the relative representations of the outputs of a teacher model. We test our Relative Representation Knowledge Distillation (RRKD) scheme on supervised and self-supervised image representation learning with MNIST and show that an encoder can be compressed to 47.71% of its original size while maintaining 91.92% of its full performance. We demonstrate that RRKD is competitive with or outperforms other relation-based distillation schemes in traditional distillation setups (CIFAR-10, CIFAR-100, SVHN) and in a transfer learning setting (Stanford Cars, Oxford-IIIT Pets, Oxford Flowers-102). Our results indicate that relative representations are an effective signal for knowledge distillation. Code is made available at https://github.com/Ramos-Ramos/rrkd. |
format |
text |
author |
Ramos, Patrick Alampay, Raphael Abu, Patricia Angela R |
author_facet |
Ramos, Patrick Alampay, Raphael Abu, Patricia Angela R |
author_sort |
Ramos, Patrick |
title |
Knowledge Distillation with Relative Representations for Image Representation Learning |
title_short |
Knowledge Distillation with Relative Representations for Image Representation Learning |
title_full |
Knowledge Distillation with Relative Representations for Image Representation Learning |
title_fullStr |
Knowledge Distillation with Relative Representations for Image Representation Learning |
title_full_unstemmed |
Knowledge Distillation with Relative Representations for Image Representation Learning |
title_sort |
knowledge distillation with relative representations for image representation learning |
publisher |
Archīum Ateneo |
publishDate |
2023 |
url |
https://archium.ateneo.edu/discs-faculty-pubs/387 https://doi.org/10.1007/978-3-031-41630-9_14 |
_version_ |
1792202616041635840 |