Lightweight and efficient neural natural language processing with quaternion networks

Many state-of-the-art neural models for NLP are heavily parameterized and thus memory inefficient. This paper proposes a series of lightweight and memory efficient neural architectures for a potpourri of natural language processing (NLP) tasks. To this end, our models exploit computation using Quate...

Full description

Saved in:
Bibliographic Details
Main Authors: TAY, Yi, ZHANG, Aston, LUU, Anh Tuan, RAO, Jinfeng, ZHANG, Shuai, WANG, Shuohang, FU, Jie, HUI, Siu Cheung
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2019
Subjects:
Online Access:https://ink.library.smu.edu.sg/scis_studentpub/2
https://ink.library.smu.edu.sg/context/scis_studentpub/article/1002/viewcontent/P19_1145.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.scis_studentpub-1002
record_format dspace
spelling sg-smu-ink.scis_studentpub-10022025-03-18T03:08:39Z Lightweight and efficient neural natural language processing with quaternion networks TAY, Yi ZHANG, Aston LUU, Anh Tuan RAO, Jinfeng ZHANG, Shuai WANG, Shuohang FU, Jie HUI, Siu Cheung Many state-of-the-art neural models for NLP are heavily parameterized and thus memory inefficient. This paper proposes a series of lightweight and memory efficient neural architectures for a potpourri of natural language processing (NLP) tasks. To this end, our models exploit computation using Quaternion algebra and hypercomplex spaces, enabling not only expressive inter-component interactions but also significantly (75%) reduced parameter size due to lesser degrees of freedom in the Hamilton product. We propose Quaternion variants of models, giving rise to new architectures such as the Quaternion attention Model and Quaternion Transformer. Extensive experiments on a battery of NLP tasks demonstrates the utility of proposed Quaternion-inspired models, enabling up to 75% reduction in parameter size without significant loss in performance. 2019-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/scis_studentpub/2 info:doi/10.18653/v1/P19-1145 https://ink.library.smu.edu.sg/context/scis_studentpub/article/1002/viewcontent/P19_1145.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ SCIS Student Publications eng Institutional Knowledge at Singapore Management University OS and Networks Programming Languages and Compilers
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic OS and Networks
Programming Languages and Compilers
spellingShingle OS and Networks
Programming Languages and Compilers
TAY, Yi
ZHANG, Aston
LUU, Anh Tuan
RAO, Jinfeng
ZHANG, Shuai
WANG, Shuohang
FU, Jie
HUI, Siu Cheung
Lightweight and efficient neural natural language processing with quaternion networks
description Many state-of-the-art neural models for NLP are heavily parameterized and thus memory inefficient. This paper proposes a series of lightweight and memory efficient neural architectures for a potpourri of natural language processing (NLP) tasks. To this end, our models exploit computation using Quaternion algebra and hypercomplex spaces, enabling not only expressive inter-component interactions but also significantly (75%) reduced parameter size due to lesser degrees of freedom in the Hamilton product. We propose Quaternion variants of models, giving rise to new architectures such as the Quaternion attention Model and Quaternion Transformer. Extensive experiments on a battery of NLP tasks demonstrates the utility of proposed Quaternion-inspired models, enabling up to 75% reduction in parameter size without significant loss in performance.
format text
author TAY, Yi
ZHANG, Aston
LUU, Anh Tuan
RAO, Jinfeng
ZHANG, Shuai
WANG, Shuohang
FU, Jie
HUI, Siu Cheung
author_facet TAY, Yi
ZHANG, Aston
LUU, Anh Tuan
RAO, Jinfeng
ZHANG, Shuai
WANG, Shuohang
FU, Jie
HUI, Siu Cheung
author_sort TAY, Yi
title Lightweight and efficient neural natural language processing with quaternion networks
title_short Lightweight and efficient neural natural language processing with quaternion networks
title_full Lightweight and efficient neural natural language processing with quaternion networks
title_fullStr Lightweight and efficient neural natural language processing with quaternion networks
title_full_unstemmed Lightweight and efficient neural natural language processing with quaternion networks
title_sort lightweight and efficient neural natural language processing with quaternion networks
publisher Institutional Knowledge at Singapore Management University
publishDate 2019
url https://ink.library.smu.edu.sg/scis_studentpub/2
https://ink.library.smu.edu.sg/context/scis_studentpub/article/1002/viewcontent/P19_1145.pdf
_version_ 1827070818428911616