Lightweight and efficient neural natural language processing with quaternion networks
Many state-of-the-art neural models for NLP are heavily parameterized and thus memory inefficient. This paper proposes a series of lightweight and memory efficient neural architectures for a potpourri of natural language processing (NLP) tasks. To this end, our models exploit computation using Quate...
Saved in:
Main Authors: | , , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2019
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/scis_studentpub/2 https://ink.library.smu.edu.sg/context/scis_studentpub/article/1002/viewcontent/P19_1145.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.scis_studentpub-1002 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.scis_studentpub-10022025-03-18T03:08:39Z Lightweight and efficient neural natural language processing with quaternion networks TAY, Yi ZHANG, Aston LUU, Anh Tuan RAO, Jinfeng ZHANG, Shuai WANG, Shuohang FU, Jie HUI, Siu Cheung Many state-of-the-art neural models for NLP are heavily parameterized and thus memory inefficient. This paper proposes a series of lightweight and memory efficient neural architectures for a potpourri of natural language processing (NLP) tasks. To this end, our models exploit computation using Quaternion algebra and hypercomplex spaces, enabling not only expressive inter-component interactions but also significantly (75%) reduced parameter size due to lesser degrees of freedom in the Hamilton product. We propose Quaternion variants of models, giving rise to new architectures such as the Quaternion attention Model and Quaternion Transformer. Extensive experiments on a battery of NLP tasks demonstrates the utility of proposed Quaternion-inspired models, enabling up to 75% reduction in parameter size without significant loss in performance. 2019-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/scis_studentpub/2 info:doi/10.18653/v1/P19-1145 https://ink.library.smu.edu.sg/context/scis_studentpub/article/1002/viewcontent/P19_1145.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ SCIS Student Publications eng Institutional Knowledge at Singapore Management University OS and Networks Programming Languages and Compilers |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
OS and Networks Programming Languages and Compilers |
spellingShingle |
OS and Networks Programming Languages and Compilers TAY, Yi ZHANG, Aston LUU, Anh Tuan RAO, Jinfeng ZHANG, Shuai WANG, Shuohang FU, Jie HUI, Siu Cheung Lightweight and efficient neural natural language processing with quaternion networks |
description |
Many state-of-the-art neural models for NLP are heavily parameterized and thus memory inefficient. This paper proposes a series of lightweight and memory efficient neural architectures for a potpourri of natural language processing (NLP) tasks. To this end, our models exploit computation using Quaternion algebra and hypercomplex spaces, enabling not only expressive inter-component interactions but also significantly (75%) reduced parameter size due to lesser degrees of freedom in the Hamilton product. We propose Quaternion variants of models, giving rise to new architectures such as the Quaternion attention Model and Quaternion Transformer. Extensive experiments on a battery of NLP tasks demonstrates the utility of proposed Quaternion-inspired models, enabling up to 75% reduction in parameter size without significant loss in performance. |
format |
text |
author |
TAY, Yi ZHANG, Aston LUU, Anh Tuan RAO, Jinfeng ZHANG, Shuai WANG, Shuohang FU, Jie HUI, Siu Cheung |
author_facet |
TAY, Yi ZHANG, Aston LUU, Anh Tuan RAO, Jinfeng ZHANG, Shuai WANG, Shuohang FU, Jie HUI, Siu Cheung |
author_sort |
TAY, Yi |
title |
Lightweight and efficient neural natural language processing with quaternion networks |
title_short |
Lightweight and efficient neural natural language processing with quaternion networks |
title_full |
Lightweight and efficient neural natural language processing with quaternion networks |
title_fullStr |
Lightweight and efficient neural natural language processing with quaternion networks |
title_full_unstemmed |
Lightweight and efficient neural natural language processing with quaternion networks |
title_sort |
lightweight and efficient neural natural language processing with quaternion networks |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2019 |
url |
https://ink.library.smu.edu.sg/scis_studentpub/2 https://ink.library.smu.edu.sg/context/scis_studentpub/article/1002/viewcontent/P19_1145.pdf |
_version_ |
1827070818428911616 |