A low-complexity complex-valued activation function for fast and accurate spectral domain convolutional neural network

Conventional Convolutional Neural Networks (CNNs), which are realized in spatial domain, exhibit high computational complexity. This results in high resource utilization and memory usage and makes them unsuitable for implementation in resource and energy-constrained embedded systems. A promising app...

Full description

Saved in:
Bibliographic Details
Main Authors: Rizvi, S. M., Ab. Rahman, A. A. H., Hani, M. K., Ayat, S. O.
Format: Article
Language:English
Published: Institute of Advanced Engineering and Science 2021
Subjects:
Online Access:http://eprints.utm.my/id/eprint/95359/1/ShahriyarMasudRizvi2021_ALowComplexityComplex.pdf
http://eprints.utm.my/id/eprint/95359/
http://dx.doi.org/10.11591/ijeei.v9i1.2737
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Universiti Teknologi Malaysia
Language: English
Description
Summary:Conventional Convolutional Neural Networks (CNNs), which are realized in spatial domain, exhibit high computational complexity. This results in high resource utilization and memory usage and makes them unsuitable for implementation in resource and energy-constrained embedded systems. A promising approach for low-complexity and high-speed solution is to apply CNN modeled in the spectral domain. One of the main challenges in this approach is the design of activation functions. Some of the proposed solutions perform activation functions in spatial domain, necessitating multiple and computationally expensive spatial-spectral domain switching. On the other hand, recent work on spectral activation functions resulted in very computationally intensive solutions. This paper proposes a complex-valued activation function for spectral domain CNNs that only transmits input values that have positive-valued real or imaginary component. This activation function is computationally inexpensive in both forward and backward propagation and provides sufficient nonlinearity that ensures high classification accuracy. We apply this complex-valued activation function in a LeNet-5 architecture and achieve an accuracy gain of up to 7% for MNIST and 6% for Fashion MNIST dataset, while providing up to 79% and 85% faster inference times, respectively, over state-of-the-art activation functions for spectral domain.