Knowledge distillation in computer vision models

Knowledge distillation has gained significant popularity in the Vision Transformer (ViT) space as a powerful approach to enhance the efficiency of a small lightweight model. Knowledge distillation enables a larger and complex “teacher” model to relay its knowledge to a smaller “student” model. Th...

Full description

Saved in:
Bibliographic Details
Main Author: Yeoh, Yu Shyan
Other Authors: Lin Guosheng
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/181128
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-181128
record_format dspace
spelling sg-ntu-dr.10356-1811282024-11-15T11:46:47Z Knowledge distillation in computer vision models Yeoh, Yu Shyan Lin Guosheng College of Computing and Data Science gslin@ntu.edu.sg Computer and Information Science Classification Knowledge distillation has gained significant popularity in the Vision Transformer (ViT) space as a powerful approach to enhance the efficiency of a small lightweight model. Knowledge distillation enables a larger and complex “teacher” model to relay its knowledge to a smaller “student” model. This enables the student model to improve its own accuracy and retain its computational efficiency. Recent works, however, lack comprehensive exploration for Hybrid distillation techniques. This includes combining various distillation strategies to boost the efficiency of the student model. This project aims to research Hybrid distillation in the context of ViT models for image classification tasks. A series of experiments were conducted to compare the result of fine-tuned teacher and student models with distilled student models, including both traditional and Hybrid distillation approaches. The experiments on Hybrid distillation have shown to improve the accuracy of smaller student models with minimal impact on inference time, providing a possible solution for real-world applications. Bachelor's degree 2024-11-15T11:46:47Z 2024-11-15T11:46:47Z 2024 Final Year Project (FYP) Yeoh, Y. S. (2024). Knowledge distillation in computer vision models. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/181128 https://hdl.handle.net/10356/181128 en SCSE23-1002 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Classification
spellingShingle Computer and Information Science
Classification
Yeoh, Yu Shyan
Knowledge distillation in computer vision models
description Knowledge distillation has gained significant popularity in the Vision Transformer (ViT) space as a powerful approach to enhance the efficiency of a small lightweight model. Knowledge distillation enables a larger and complex “teacher” model to relay its knowledge to a smaller “student” model. This enables the student model to improve its own accuracy and retain its computational efficiency. Recent works, however, lack comprehensive exploration for Hybrid distillation techniques. This includes combining various distillation strategies to boost the efficiency of the student model. This project aims to research Hybrid distillation in the context of ViT models for image classification tasks. A series of experiments were conducted to compare the result of fine-tuned teacher and student models with distilled student models, including both traditional and Hybrid distillation approaches. The experiments on Hybrid distillation have shown to improve the accuracy of smaller student models with minimal impact on inference time, providing a possible solution for real-world applications.
author2 Lin Guosheng
author_facet Lin Guosheng
Yeoh, Yu Shyan
format Final Year Project
author Yeoh, Yu Shyan
author_sort Yeoh, Yu Shyan
title Knowledge distillation in computer vision models
title_short Knowledge distillation in computer vision models
title_full Knowledge distillation in computer vision models
title_fullStr Knowledge distillation in computer vision models
title_full_unstemmed Knowledge distillation in computer vision models
title_sort knowledge distillation in computer vision models
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/181128
_version_ 1816858992023961600