Stitching weight-shared deep neural networks for efficient multitask inference on GPU
Intelligent personal and home applications demand multiple deep neural networks (DNNs) running on resourceconstrained platforms for compound inference tasks, known as multitask inference. To fit multiple DNNs into low-resource devices, emerging techniques resort to weight sharing among DNNs to reduc...
Saved in:
Main Authors: | WANG, Zeyu, HE, Xiaoxi, ZHOU, Zimu, WANG, Xu, MA, Qiang, MIAO, Xin, LIU, Zhuo, THIELE, Lothar, YANG, Zheng. |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2022
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/7486 https://ink.library.smu.edu.sg/context/sis_research/article/8489/viewcontent/secon22_wang.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Pruning-aware merging for efficient multitask inference
by: GAO, Dawei, et al.
Published: (2021) -
On-device deep multi-task inference via multi-task zipping
by: HE, Xiaoxi, et al.
Published: (2023) -
Pruning meta-trained networks for on-device adaptation
by: GAO, Dawei, et al.
Published: (2021) -
Win: Weight-decay-integrated Nesterov acceleration for faster network training
by: ZHOU, Pan, et al.
Published: (2024) -
Evolutionary multitasking : a computer science view of cognitive multitasking
by: Ong, Yew-Soon, et al.
Published: (2021)