Multi-task zipping via layer-wise neuron sharing

Future mobile devices are anticipated to perceive, understand and react to the world on their own by running multiple correlated deep neural networks on-device. Yet the complexity of these neural networks needs to be trimmed down both within-model and cross-model to fit in mobile storage and memory....

Full description

Saved in:
Bibliographic Details
Main Authors: HE, Xiaoxi, ZHOU, Zimu, THIELE, Lothar
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2018
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/4554
https://ink.library.smu.edu.sg/context/sis_research/article/5557/viewcontent/neurips18_he.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-5557
record_format dspace
spelling sg-smu-ink.sis_research-55572019-12-26T08:34:37Z Multi-task zipping via layer-wise neuron sharing HE, Xiaoxi ZHOU, Zimu THIELE, Lothar Future mobile devices are anticipated to perceive, understand and react to the world on their own by running multiple correlated deep neural networks on-device. Yet the complexity of these neural networks needs to be trimmed down both within-model and cross-model to fit in mobile storage and memory. Previous studies squeeze the redundancy within a single model. In this work, we aim to reduce the redundancy across multiple models. We propose Multi-Task Zipping (MTZ), a framework to automatically merge correlated, pre-trained deep neural networks for cross-model compression. Central in MTZ is a layer-wise neuron sharing and incoming weight updating scheme that induces a minimal change in the error function. MTZ inherits information from each model and demands light retraining to re-boost the accuracy of individual tasks. Evaluations show that MTZ is able to fully merge the hidden layers of two VGG-16 networks with a 3.18% increase in the test error averaged on ImageNet and CelebA, or share 39.61% parameters between the two networks with .5% increase in the test errors for both tasks. The number of iterations to retrain the combined network is at least 17.8 × lower than that of training a single VGG-16 network. Moreover, experiments show that MTZ is also able to effectively merge multiple residual networks. 2018-12-08T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4554 https://ink.library.smu.edu.sg/context/sis_research/article/5557/viewcontent/neurips18_he.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Software Engineering
spellingShingle Software Engineering
HE, Xiaoxi
ZHOU, Zimu
THIELE, Lothar
Multi-task zipping via layer-wise neuron sharing
description Future mobile devices are anticipated to perceive, understand and react to the world on their own by running multiple correlated deep neural networks on-device. Yet the complexity of these neural networks needs to be trimmed down both within-model and cross-model to fit in mobile storage and memory. Previous studies squeeze the redundancy within a single model. In this work, we aim to reduce the redundancy across multiple models. We propose Multi-Task Zipping (MTZ), a framework to automatically merge correlated, pre-trained deep neural networks for cross-model compression. Central in MTZ is a layer-wise neuron sharing and incoming weight updating scheme that induces a minimal change in the error function. MTZ inherits information from each model and demands light retraining to re-boost the accuracy of individual tasks. Evaluations show that MTZ is able to fully merge the hidden layers of two VGG-16 networks with a 3.18% increase in the test error averaged on ImageNet and CelebA, or share 39.61% parameters between the two networks with .5% increase in the test errors for both tasks. The number of iterations to retrain the combined network is at least 17.8 × lower than that of training a single VGG-16 network. Moreover, experiments show that MTZ is also able to effectively merge multiple residual networks.
format text
author HE, Xiaoxi
ZHOU, Zimu
THIELE, Lothar
author_facet HE, Xiaoxi
ZHOU, Zimu
THIELE, Lothar
author_sort HE, Xiaoxi
title Multi-task zipping via layer-wise neuron sharing
title_short Multi-task zipping via layer-wise neuron sharing
title_full Multi-task zipping via layer-wise neuron sharing
title_fullStr Multi-task zipping via layer-wise neuron sharing
title_full_unstemmed Multi-task zipping via layer-wise neuron sharing
title_sort multi-task zipping via layer-wise neuron sharing
publisher Institutional Knowledge at Singapore Management University
publishDate 2018
url https://ink.library.smu.edu.sg/sis_research/4554
https://ink.library.smu.edu.sg/context/sis_research/article/5557/viewcontent/neurips18_he.pdf
_version_ 1770574886285082624