Knowledge Distillation with Relative Representations for Image Representation Learning

Relative representations allow the alignment of latent spaces which embed data in extrinsically different manners but with similar relative distances between data points. This ability to compare different latent spaces for the same input lends itself to knowledge distillation techniques. We explore...

Full description

Saved in:
Bibliographic Details
Main Authors: Ramos, Patrick, Alampay, Raphael, Abu, Patricia Angela R
Format: text
Published: Archīum Ateneo 2023
Subjects:
Online Access:https://archium.ateneo.edu/discs-faculty-pubs/387
https://doi.org/10.1007/978-3-031-41630-9_14
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Ateneo De Manila University
id ph-ateneo-arc.discs-faculty-pubs-1387
record_format eprints
spelling ph-ateneo-arc.discs-faculty-pubs-13872024-02-21T02:51:52Z Knowledge Distillation with Relative Representations for Image Representation Learning Ramos, Patrick Alampay, Raphael Abu, Patricia Angela R Relative representations allow the alignment of latent spaces which embed data in extrinsically different manners but with similar relative distances between data points. This ability to compare different latent spaces for the same input lends itself to knowledge distillation techniques. We explore the applicability of relative representations to knowledge distillation by training a student model such that the relative representations of its outputs match the relative representations of the outputs of a teacher model. We test our Relative Representation Knowledge Distillation (RRKD) scheme on supervised and self-supervised image representation learning with MNIST and show that an encoder can be compressed to 47.71% of its original size while maintaining 91.92% of its full performance. We demonstrate that RRKD is competitive with or outperforms other relation-based distillation schemes in traditional distillation setups (CIFAR-10, CIFAR-100, SVHN) and in a transfer learning setting (Stanford Cars, Oxford-IIIT Pets, Oxford Flowers-102). Our results indicate that relative representations are an effective signal for knowledge distillation. Code is made available at https://github.com/Ramos-Ramos/rrkd. 2023-01-01T08:00:00Z text https://archium.ateneo.edu/discs-faculty-pubs/387 https://doi.org/10.1007/978-3-031-41630-9_14 Department of Information Systems & Computer Science Faculty Publications Archīum Ateneo Knowledge distillation Latent space Relative representations Computer Engineering Electrical and Computer Engineering Engineering
institution Ateneo De Manila University
building Ateneo De Manila University Library
continent Asia
country Philippines
Philippines
content_provider Ateneo De Manila University Library
collection archium.Ateneo Institutional Repository
topic Knowledge distillation
Latent space
Relative representations
Computer Engineering
Electrical and Computer Engineering
Engineering
spellingShingle Knowledge distillation
Latent space
Relative representations
Computer Engineering
Electrical and Computer Engineering
Engineering
Ramos, Patrick
Alampay, Raphael
Abu, Patricia Angela R
Knowledge Distillation with Relative Representations for Image Representation Learning
description Relative representations allow the alignment of latent spaces which embed data in extrinsically different manners but with similar relative distances between data points. This ability to compare different latent spaces for the same input lends itself to knowledge distillation techniques. We explore the applicability of relative representations to knowledge distillation by training a student model such that the relative representations of its outputs match the relative representations of the outputs of a teacher model. We test our Relative Representation Knowledge Distillation (RRKD) scheme on supervised and self-supervised image representation learning with MNIST and show that an encoder can be compressed to 47.71% of its original size while maintaining 91.92% of its full performance. We demonstrate that RRKD is competitive with or outperforms other relation-based distillation schemes in traditional distillation setups (CIFAR-10, CIFAR-100, SVHN) and in a transfer learning setting (Stanford Cars, Oxford-IIIT Pets, Oxford Flowers-102). Our results indicate that relative representations are an effective signal for knowledge distillation. Code is made available at https://github.com/Ramos-Ramos/rrkd.
format text
author Ramos, Patrick
Alampay, Raphael
Abu, Patricia Angela R
author_facet Ramos, Patrick
Alampay, Raphael
Abu, Patricia Angela R
author_sort Ramos, Patrick
title Knowledge Distillation with Relative Representations for Image Representation Learning
title_short Knowledge Distillation with Relative Representations for Image Representation Learning
title_full Knowledge Distillation with Relative Representations for Image Representation Learning
title_fullStr Knowledge Distillation with Relative Representations for Image Representation Learning
title_full_unstemmed Knowledge Distillation with Relative Representations for Image Representation Learning
title_sort knowledge distillation with relative representations for image representation learning
publisher Archīum Ateneo
publishDate 2023
url https://archium.ateneo.edu/discs-faculty-pubs/387
https://doi.org/10.1007/978-3-031-41630-9_14
_version_ 1792202616041635840