Efficient implementation of activation functions for LSTM accelerators

Activation functions such as hyperbolic tangent (tanh) and logistic sigmoid (sigmoid) are critical computing elements in a long short term memory (LSTM) cell and network. These activation functions are non-linear, leading to challenges in their hardware implementations. Area-efficient and high perf...

Full description

Saved in:
Bibliographic Details
Main Authors: Chong, Yi Sheng, Goh, Wang Ling, Ong, Yew-Soon, Nambiar, Vishnu P., Do, Anh Tuan
Other Authors: School of Electrical and Electronic Engineering
Format: Conference or Workshop Item
Language:English
Published: 2021
Subjects:
Online Access:https://hdl.handle.net/10356/153121
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-153121
record_format dspace
spelling sg-ntu-dr.10356-1531212021-12-18T20:11:40Z Efficient implementation of activation functions for LSTM accelerators Chong, Yi Sheng Goh, Wang Ling Ong, Yew-Soon Nambiar, Vishnu P. Do, Anh Tuan School of Electrical and Electronic Engineering Interdisciplinary Graduate School (IGS) School of Computer Science and Engineering 2021 29th IFIP International Conference on Very Large Scale Integration (VLSI-SoC) Energy Research Institute @ NTU (ERI@N) Engineering::Electrical and electronic engineering::Integrated circuits Activation function Long Short Term Memory Accelerator Activation functions such as hyperbolic tangent (tanh) and logistic sigmoid (sigmoid) are critical computing elements in a long short term memory (LSTM) cell and network. These activation functions are non-linear, leading to challenges in their hardware implementations. Area-efficient and high performance hardware implementation of these activation functions thus becomes crucial to allow high throughput in a LSTM accelerator. In this work, we propose an approximation scheme which is suitable for both tanh and sigmoid functions. The proposed hardware for sigmoid function is 8.3 times smaller than the state-of-the-art, while for tanh function, it is the second smallest design. When applying the approximated tanh and sigmoid of 2% error in a LSTM cell computation, its final hidden state and cell state record errors of 3.1% and 5.8% respectively. When the same approximated functions are applied to a single layer LSTM network of 64 hidden nodes, the accuracy drops by 2.8% only. This proposed small yet accurate activation function hardware is promising to be used in Internet of Things (IoT) applications where accuracy can be traded off for ultra-low power consumption. Accepted version We thank the Programmatic grant no. A1687b0033, Singapore RIE 2020, AME domain. 2021-12-14T02:10:19Z 2021-12-14T02:10:19Z 2021 Conference Paper Chong, Y. S., Goh, W. L., Ong, Y., Nambiar, V. P. & Do, A. T. (2021). Efficient implementation of activation functions for LSTM accelerators. 2021 29th IFIP International Conference on Very Large Scale Integration (VLSI-SoC). https://dx.doi.org/10.1109/VLSI-SoC53125.2021.9606971 978-1-6654-2614-5 2324-8440 https://hdl.handle.net/10356/153121 10.1109/VLSI-SoC53125.2021.9606971 en A1687b0033 © 2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The published version is available at: https://doi.org/10.1109/VLSI-SoC53125.2021.9606971. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering::Integrated circuits
Activation function
Long Short Term Memory Accelerator
spellingShingle Engineering::Electrical and electronic engineering::Integrated circuits
Activation function
Long Short Term Memory Accelerator
Chong, Yi Sheng
Goh, Wang Ling
Ong, Yew-Soon
Nambiar, Vishnu P.
Do, Anh Tuan
Efficient implementation of activation functions for LSTM accelerators
description Activation functions such as hyperbolic tangent (tanh) and logistic sigmoid (sigmoid) are critical computing elements in a long short term memory (LSTM) cell and network. These activation functions are non-linear, leading to challenges in their hardware implementations. Area-efficient and high performance hardware implementation of these activation functions thus becomes crucial to allow high throughput in a LSTM accelerator. In this work, we propose an approximation scheme which is suitable for both tanh and sigmoid functions. The proposed hardware for sigmoid function is 8.3 times smaller than the state-of-the-art, while for tanh function, it is the second smallest design. When applying the approximated tanh and sigmoid of 2% error in a LSTM cell computation, its final hidden state and cell state record errors of 3.1% and 5.8% respectively. When the same approximated functions are applied to a single layer LSTM network of 64 hidden nodes, the accuracy drops by 2.8% only. This proposed small yet accurate activation function hardware is promising to be used in Internet of Things (IoT) applications where accuracy can be traded off for ultra-low power consumption.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Chong, Yi Sheng
Goh, Wang Ling
Ong, Yew-Soon
Nambiar, Vishnu P.
Do, Anh Tuan
format Conference or Workshop Item
author Chong, Yi Sheng
Goh, Wang Ling
Ong, Yew-Soon
Nambiar, Vishnu P.
Do, Anh Tuan
author_sort Chong, Yi Sheng
title Efficient implementation of activation functions for LSTM accelerators
title_short Efficient implementation of activation functions for LSTM accelerators
title_full Efficient implementation of activation functions for LSTM accelerators
title_fullStr Efficient implementation of activation functions for LSTM accelerators
title_full_unstemmed Efficient implementation of activation functions for LSTM accelerators
title_sort efficient implementation of activation functions for lstm accelerators
publishDate 2021
url https://hdl.handle.net/10356/153121
_version_ 1720447110896156672