FAT: an in-memory accelerator with fast addition for ternary weight neural networks
Convolutional Neural Networks (CNNs) demonstrate excellent performance in various applications but have high computational complexity. Quantization is applied to reduce the latency and storage cost of CNNs. Among the quantization methods, Binary and Ternary Weight Networks (BWNs and TWNs) have a uni...
Saved in:
Main Authors: | Zhu, Shien, Duong, Luan H. K., Chen, Hui, Liu, Di, Liu, Weichen |
---|---|
Other Authors: | School of Computer Science and Engineering |
Format: | Article |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/162483 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Similar Items
-
TAB : unified and optimized ternary, binary and mixed-precision neural network inference on the edge
by: Zhu, Shien, et al.
Published: (2022) -
iMAT: energy-efficient in-memory acceleration for ternary neural networks with sparse dot product
by: Zhu, Shien, et al.
Published: (2023) -
iMAD: an in-memory accelerator for AdderNet with efficient 8-bit addition and subtraction operations
by: Zhu, Shien, et al.
Published: (2022) -
Crossbar-aligned & integer-only neural network compression for efficient in-memory acceleration
by: Huai, Shuo, et al.
Published: (2023) -
Smart scissor: coupling spatial redundancy reduction and CNN compression for embedded hardware
by: Kong, Hao, et al.
Published: (2023)