Adaptive loss-aware quantization for multi-bit networks
We investigate the compression of deep neural networks by quantizing their weights and activations into multiple binary bases, known as multi-bit networks (MBNs), which accelerate the inference and reduce the storage for the deployment on low-resource mobile and embedded platforms. We propose Adapti...
Saved in:
Main Authors: | QU, Zhongnan, ZHOU, Zimu, CHENG, Yun, THIELE, Lothar |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2020
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/5251 https://ink.library.smu.edu.sg/context/sis_research/article/6254/viewcontent/cvpr20_qu.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
p-Meta: Towards on-device deep model adaptation
by: QU, Zhongnan, et al.
Published: (2022) -
Bounds on the optimal quantization performance of dynamically quantized linear systems with bounded noise
by: Ling, Q., et al.
Published: (2014) -
Numerical studies on quantized vortex dynamics in superfludity and superconductivity
by: TANG QINGLIN
Published: (2013) -
Adaptive vertex quantization for mesh compression
by: Qiu, Z.M., et al.
Published: (2014) -
Certified quantization strategy synthesis for neural networks
by: ZHANG, Yedi, et al.
Published: (2024)