On-demand deep model compression for mobile devices: A usage-driven model selection framework

Recent research has demonstrated the potential of deploying deep neural networks (DNNs) on resource-constrained mobile platforms by trimming down the network complexity using different compression techniques. The current practice only investigate stand-alone compression schemes even though each comp...

Full description

Saved in:
Bibliographic Details
Main Authors: LIU, Sicong, LIN, Yingyan, ZHOU, Zimu, NAN, Kaiming, LIU, Hui, DU, Junzhao
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2018
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/4734
https://ink.library.smu.edu.sg/context/sis_research/article/5737/viewcontent/mobisys18_liu.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-5737
record_format dspace
spelling sg-smu-ink.sis_research-57372020-01-16T10:44:23Z On-demand deep model compression for mobile devices: A usage-driven model selection framework LIU, Sicong LIN, Yingyan ZHOU, Zimu NAN, Kaiming LIU, Hui DU, Junzhao Recent research has demonstrated the potential of deploying deep neural networks (DNNs) on resource-constrained mobile platforms by trimming down the network complexity using different compression techniques. The current practice only investigate stand-alone compression schemes even though each compression technique may be well suited only for certain types of DNN layers. Also, these compression techniques are optimized merely for the inference accuracy of DNNs, without explicitly considering other application-driven system performance (e.g. latency and energy cost) and the varying resource availabilities across platforms (e.g. storage and processing capability). In this paper, we explore the desirable tradeoff between performance and resource constraints by user-specified needs, from a holistic system-level viewpoint. Specifically, we develop a usage-driven selection framework, referred to as AdaDeep, to automatically select a combination of compression techniques for a given DNN, that will lead to an optimal balance between user-specified performance goals and resource constraints. With an extensive evaluation on five public datasets and across twelve mobile devices, experimental results show that AdaDeep enables up to 9.8x latency reduction, 4.3x energy efficiency improvement, and 38x storage reduction in DNNs while incurring negligible accuracy loss. AdaDeep also uncovers multiple effective combinations of compression techniques unexplored in existing literature. 2018-06-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4734 info:doi/10.1145/3210240.3210337 https://ink.library.smu.edu.sg/context/sis_research/article/5737/viewcontent/mobisys18_liu.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Software Engineering
spellingShingle Software Engineering
LIU, Sicong
LIN, Yingyan
ZHOU, Zimu
NAN, Kaiming
LIU, Hui
DU, Junzhao
On-demand deep model compression for mobile devices: A usage-driven model selection framework
description Recent research has demonstrated the potential of deploying deep neural networks (DNNs) on resource-constrained mobile platforms by trimming down the network complexity using different compression techniques. The current practice only investigate stand-alone compression schemes even though each compression technique may be well suited only for certain types of DNN layers. Also, these compression techniques are optimized merely for the inference accuracy of DNNs, without explicitly considering other application-driven system performance (e.g. latency and energy cost) and the varying resource availabilities across platforms (e.g. storage and processing capability). In this paper, we explore the desirable tradeoff between performance and resource constraints by user-specified needs, from a holistic system-level viewpoint. Specifically, we develop a usage-driven selection framework, referred to as AdaDeep, to automatically select a combination of compression techniques for a given DNN, that will lead to an optimal balance between user-specified performance goals and resource constraints. With an extensive evaluation on five public datasets and across twelve mobile devices, experimental results show that AdaDeep enables up to 9.8x latency reduction, 4.3x energy efficiency improvement, and 38x storage reduction in DNNs while incurring negligible accuracy loss. AdaDeep also uncovers multiple effective combinations of compression techniques unexplored in existing literature.
format text
author LIU, Sicong
LIN, Yingyan
ZHOU, Zimu
NAN, Kaiming
LIU, Hui
DU, Junzhao
author_facet LIU, Sicong
LIN, Yingyan
ZHOU, Zimu
NAN, Kaiming
LIU, Hui
DU, Junzhao
author_sort LIU, Sicong
title On-demand deep model compression for mobile devices: A usage-driven model selection framework
title_short On-demand deep model compression for mobile devices: A usage-driven model selection framework
title_full On-demand deep model compression for mobile devices: A usage-driven model selection framework
title_fullStr On-demand deep model compression for mobile devices: A usage-driven model selection framework
title_full_unstemmed On-demand deep model compression for mobile devices: A usage-driven model selection framework
title_sort on-demand deep model compression for mobile devices: a usage-driven model selection framework
publisher Institutional Knowledge at Singapore Management University
publishDate 2018
url https://ink.library.smu.edu.sg/sis_research/4734
https://ink.library.smu.edu.sg/context/sis_research/article/5737/viewcontent/mobisys18_liu.pdf
_version_ 1770575014964232192