Certified quantization strategy synthesis for neural networks

Quantization plays an important role in deploying neural networks on embedded, real-time systems with limited computing and storage resources (e.g., edge devices). It significantly reduces the model storage cost and improves inference efficiency by using fewer bits to represent the parameters. Howev...

Full description

Saved in:
Bibliographic Details
Main Authors: ZHANG, Yedi, CHEN, Guangke, SUN, Jun
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2024
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/9325
https://ink.library.smu.edu.sg/context/sis_research/article/10325/viewcontent/978_3_031_71162_6_18_pvoa_cc_by.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-10325
record_format dspace
spelling sg-smu-ink.sis_research-103252024-09-26T07:44:53Z Certified quantization strategy synthesis for neural networks ZHANG, Yedi CHEN, Guangke SUN, Jun SUN, Jun Quantization plays an important role in deploying neural networks on embedded, real-time systems with limited computing and storage resources (e.g., edge devices). It significantly reduces the model storage cost and improves inference efficiency by using fewer bits to represent the parameters. However, it was recently shown that critical properties may be broken after quantization, such as robustness and backdoor-freeness. In this work, we introduce the first method for synthesizing quantization strategies that verifiably maintain desired properties after quantization, leveraging a key insight that quantization leads to a data distribution shift in each layer. We propose to compute the preimage for each layer based on which the preceding layer is quantized, ensuring that the quantized reachable region of the preceding layer remains within the preimage. To tackle the challenge of computing the exact preimage, we propose an MILP-based method to compute its under-approximation. We implement our method into a tool Quadapter and demonstrate its effectiveness and efficiency by providing certified quantization that successfully preserves model robustness and backdoor-freeness. 2024-09-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/9325 info:doi/10.1007/978-3-031-71162-6_18 https://ink.library.smu.edu.sg/context/sis_research/article/10325/viewcontent/978_3_031_71162_6_18_pvoa_cc_by.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Numerical Analysis and Scientific Computing
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Databases and Information Systems
Numerical Analysis and Scientific Computing
spellingShingle Databases and Information Systems
Numerical Analysis and Scientific Computing
ZHANG, Yedi
CHEN, Guangke
SUN, Jun
SUN, Jun
Certified quantization strategy synthesis for neural networks
description Quantization plays an important role in deploying neural networks on embedded, real-time systems with limited computing and storage resources (e.g., edge devices). It significantly reduces the model storage cost and improves inference efficiency by using fewer bits to represent the parameters. However, it was recently shown that critical properties may be broken after quantization, such as robustness and backdoor-freeness. In this work, we introduce the first method for synthesizing quantization strategies that verifiably maintain desired properties after quantization, leveraging a key insight that quantization leads to a data distribution shift in each layer. We propose to compute the preimage for each layer based on which the preceding layer is quantized, ensuring that the quantized reachable region of the preceding layer remains within the preimage. To tackle the challenge of computing the exact preimage, we propose an MILP-based method to compute its under-approximation. We implement our method into a tool Quadapter and demonstrate its effectiveness and efficiency by providing certified quantization that successfully preserves model robustness and backdoor-freeness.
format text
author ZHANG, Yedi
CHEN, Guangke
SUN, Jun
SUN, Jun
author_facet ZHANG, Yedi
CHEN, Guangke
SUN, Jun
SUN, Jun
author_sort ZHANG, Yedi
title Certified quantization strategy synthesis for neural networks
title_short Certified quantization strategy synthesis for neural networks
title_full Certified quantization strategy synthesis for neural networks
title_fullStr Certified quantization strategy synthesis for neural networks
title_full_unstemmed Certified quantization strategy synthesis for neural networks
title_sort certified quantization strategy synthesis for neural networks
publisher Institutional Knowledge at Singapore Management University
publishDate 2024
url https://ink.library.smu.edu.sg/sis_research/9325
https://ink.library.smu.edu.sg/context/sis_research/article/10325/viewcontent/978_3_031_71162_6_18_pvoa_cc_by.pdf
_version_ 1814047910043582464