Optogenetics inspired transition metal dichalcogenide neuristors for in-memory deep recurrent neural networks

Shallow feed-forward networks are incapable of addressing complex tasks such as natural language processing that require learning of temporal signals. To address these requirements, we need deep neuromorphic architectures with recurrent connections such as deep recurrent neural networks. However, th...

Full description

Saved in:
Bibliographic Details
Main Authors: John, Rohit Abraham, Acharya, Jyotibdha, Zhu, Chao, Surendran, Abhijith, Bose, Sumon Kumar, Chaturvedi, Apoorva, Tiwari, Nidhi, Gao, Yang, He, Yongmin, Zhang, Keke K., Xu, Manzhang, Leong, Wei Lin, Liu, Zheng, Basu, Arindam, Mathews, Nripan
Other Authors: School of Materials Science and Engineering
Format: Article
Language:English
Published: 2021
Subjects:
Online Access:https://hdl.handle.net/10356/152915
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:Shallow feed-forward networks are incapable of addressing complex tasks such as natural language processing that require learning of temporal signals. To address these requirements, we need deep neuromorphic architectures with recurrent connections such as deep recurrent neural networks. However, the training of such networks demand very high precision of weights, excellent conductance linearity and low write-noise- not satisfied by current memristive implementations. Inspired from optogenetics, here we report a neuromorphic computing platform comprised of photo-excitable neuristors capable of in-memory computations across 980 addressable states with a high signal-to-noise ratio of 77. The large linear dynamic range, low write noise and selective excitability allows high fidelity opto-electronic transfer of weights with a two-shot write scheme, while electrical in-memory inference provides energy efficiency. This method enables implementing a memristive deep recurrent neural network with twelve trainable layers with more than a million parameters to recognize spoken commands with >90% accuracy.