Knowledge distillation in computer vision models
Knowledge distillation has gained significant popularity in the Vision Transformer (ViT) space as a powerful approach to enhance the efficiency of a small lightweight model. Knowledge distillation enables a larger and complex “teacher” model to relay its knowledge to a smaller “student” model. Th...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/181128 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-181128 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1811282024-11-15T11:46:47Z Knowledge distillation in computer vision models Yeoh, Yu Shyan Lin Guosheng College of Computing and Data Science gslin@ntu.edu.sg Computer and Information Science Classification Knowledge distillation has gained significant popularity in the Vision Transformer (ViT) space as a powerful approach to enhance the efficiency of a small lightweight model. Knowledge distillation enables a larger and complex “teacher” model to relay its knowledge to a smaller “student” model. This enables the student model to improve its own accuracy and retain its computational efficiency. Recent works, however, lack comprehensive exploration for Hybrid distillation techniques. This includes combining various distillation strategies to boost the efficiency of the student model. This project aims to research Hybrid distillation in the context of ViT models for image classification tasks. A series of experiments were conducted to compare the result of fine-tuned teacher and student models with distilled student models, including both traditional and Hybrid distillation approaches. The experiments on Hybrid distillation have shown to improve the accuracy of smaller student models with minimal impact on inference time, providing a possible solution for real-world applications. Bachelor's degree 2024-11-15T11:46:47Z 2024-11-15T11:46:47Z 2024 Final Year Project (FYP) Yeoh, Y. S. (2024). Knowledge distillation in computer vision models. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181128 https://hdl.handle.net/10356/181128 en SCSE23-1002 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science Classification |
spellingShingle |
Computer and Information Science Classification Yeoh, Yu Shyan Knowledge distillation in computer vision models |
description |
Knowledge distillation has gained significant popularity in the Vision Transformer
(ViT) space as a powerful approach to enhance the efficiency of a small lightweight
model. Knowledge distillation enables a larger and complex “teacher” model to relay
its knowledge to a smaller “student” model. This enables the student model to improve
its own accuracy and retain its computational efficiency. Recent works, however, lack
comprehensive exploration for Hybrid distillation techniques. This includes combining
various distillation strategies to boost the efficiency of the student model. This project
aims to research Hybrid distillation in the context of ViT models for image classification
tasks. A series of experiments were conducted to compare the result of fine-tuned
teacher and student models with distilled student models, including both traditional and
Hybrid distillation approaches. The experiments on Hybrid distillation have shown to
improve the accuracy of smaller student models with minimal impact on inference time,
providing a possible solution for real-world applications. |
author2 |
Lin Guosheng |
author_facet |
Lin Guosheng Yeoh, Yu Shyan |
format |
Final Year Project |
author |
Yeoh, Yu Shyan |
author_sort |
Yeoh, Yu Shyan |
title |
Knowledge distillation in computer vision models |
title_short |
Knowledge distillation in computer vision models |
title_full |
Knowledge distillation in computer vision models |
title_fullStr |
Knowledge distillation in computer vision models |
title_full_unstemmed |
Knowledge distillation in computer vision models |
title_sort |
knowledge distillation in computer vision models |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/181128 |
_version_ |
1816858992023961600 |