HACScale : hardware-aware compound scaling for resource-efficient DNNs
Model scaling is an effective way to improve the accuracy of deep neural networks (DNNs) by increasing the model capacity. However, existing approaches seldom consider the underlying hardware, causing inefficient utilization of hardware resources and consequently high inference latency. In this pape...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/155808 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-155808 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1558082023-12-15T03:06:01Z HACScale : hardware-aware compound scaling for resource-efficient DNNs Kong, Hao Liu, Di Luo, Xiangzhong Liu, Weichen Subramaniam, Ravi School of Computer Science and Engineering 2022 27th Asia and South Pacific Design Automation Conference (ASP-DAC) HP-NTU Digital Manufacturing Corporate Lab Engineering::Computer science and engineering Deep Learning Design Automation Model scaling is an effective way to improve the accuracy of deep neural networks (DNNs) by increasing the model capacity. However, existing approaches seldom consider the underlying hardware, causing inefficient utilization of hardware resources and consequently high inference latency. In this paper, we propose HACScale, a hardware-aware model scaling strategy to fully exploit hardware resources for higher accuracy. In HACScale, different dimensions of DNNs are jointly scaled with consideration of their contributions to hardware utilization and accuracy. To improve the efficiency of width scaling, we introduce importance-aware width scaling in HACScale, which computes the importance of each layer to the accuracy and scales each layer accordingly to optimize the trade-off between accuracy and model parameters. Experiments show that HACScale improves the hardware utilization by 1.92× on ImageNet, as a result, it achieves 2.41% accuracy improvement with a negligible latency increase of 0.6%. On CIFAR-10, HACScale improves the accuracy by 2.23% with only 6.5% latency growth. Nanyang Technological University National Research Foundation (NRF) Submitted/Accepted version This study is supported under the RIE2020 Industry Alignment Fund – Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry partner, HP Inc., through the HP-NTU Digital Manufacturing Corporate Lab. This work is also partially supported by NTU NAP M4082282 and SUG M4082087, Singapore. 2022-03-22T02:32:35Z 2022-03-22T02:32:35Z 2022 Conference Paper Kong, H., Liu, D., Luo, X., Liu, W. & Subramaniam, R. (2022). HACScale : hardware-aware compound scaling for resource-efficient DNNs. 2022 27th Asia and South Pacific Design Automation Conference (ASP-DAC), 708-713. https://dx.doi.org/10.1109/ASP-DAC52403.2022.9712593 9781665421355 https://hdl.handle.net/10356/155808 10.1109/ASP-DAC52403.2022.9712593 2-s2.0-85122940747 708 713 en M4082282 M4082087 10.21979/N9/KSXK4T © 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The published version is available at: https://doi.org/10.1109/ASP-DAC52403.2022.9712593. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Deep Learning Design Automation |
spellingShingle |
Engineering::Computer science and engineering Deep Learning Design Automation Kong, Hao Liu, Di Luo, Xiangzhong Liu, Weichen Subramaniam, Ravi HACScale : hardware-aware compound scaling for resource-efficient DNNs |
description |
Model scaling is an effective way to improve the accuracy of deep neural networks (DNNs) by increasing the model capacity. However, existing approaches seldom consider the underlying hardware, causing inefficient utilization of hardware resources and consequently high inference latency. In this paper, we propose HACScale, a hardware-aware model scaling strategy to fully exploit hardware resources for higher accuracy. In HACScale, different dimensions of DNNs are jointly scaled with consideration of their contributions to hardware utilization and accuracy. To improve the efficiency of width scaling, we introduce importance-aware width scaling in HACScale, which computes the importance of each layer to the accuracy and scales each layer accordingly to optimize the trade-off between accuracy and model parameters. Experiments show that HACScale improves the hardware utilization by 1.92× on ImageNet, as a result, it achieves 2.41% accuracy improvement with a negligible latency increase of 0.6%. On CIFAR-10, HACScale improves the accuracy by 2.23% with only 6.5% latency growth. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Kong, Hao Liu, Di Luo, Xiangzhong Liu, Weichen Subramaniam, Ravi |
format |
Conference or Workshop Item |
author |
Kong, Hao Liu, Di Luo, Xiangzhong Liu, Weichen Subramaniam, Ravi |
author_sort |
Kong, Hao |
title |
HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_short |
HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_full |
HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_fullStr |
HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_full_unstemmed |
HACScale : hardware-aware compound scaling for resource-efficient DNNs |
title_sort |
hacscale : hardware-aware compound scaling for resource-efficient dnns |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/155808 |
_version_ |
1787136612860166144 |