Modeling intra-relation in math word problems with different functional multi-head attentions
Several deep learning models have been proposed for solving math word problems (MWPs) automatically. Although these models have the ability to capture features without manual efforts, their approaches to capturing features are not specifically designed for MWPs. To utilize the merits of deep learnin...
Saved in:
Main Authors: | , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2019
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/4867 https://ink.library.smu.edu.sg/context/sis_research/article/5870/viewcontent/P19_1619_pvoa.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-5870 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-58702024-04-18T06:04:43Z Modeling intra-relation in math word problems with different functional multi-head attentions LI, Jierui WANG, Lei ZHANG, Jipeng WANG, Yan DAI, Bing Tian ZHANG, Dongxiang Several deep learning models have been proposed for solving math word problems (MWPs) automatically. Although these models have the ability to capture features without manual efforts, their approaches to capturing features are not specifically designed for MWPs. To utilize the merits of deep learning models with simultaneous consideration of MWPs’ specific features, we propose a group attention mechanism to extract global features, quantity-related features, quantity-pair features and question-related features in MWPs respectively. The experimental results show that the proposed approach performs significantly better than previous state-of-the-art methods, and boost performance from 66.9% to 69.5% on Math23K with training-test split, from 65.8% to 66.9% on Math23K with 5-fold cross-validation and from 69.2% to 76.1% on MAWPS. 2019-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4867 info:doi/10.18653/v1/P19-1619 https://ink.library.smu.edu.sg/context/sis_research/article/5870/viewcontent/P19_1619_pvoa.pdf http://creativecommons.org/licenses/by-nc-sa/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Attention mechanisms Cross validation Global feature Learning models State-of-the-art methods Word problem Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Attention mechanisms Cross validation Global feature Learning models State-of-the-art methods Word problem Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing |
spellingShingle |
Attention mechanisms Cross validation Global feature Learning models State-of-the-art methods Word problem Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing LI, Jierui WANG, Lei ZHANG, Jipeng WANG, Yan DAI, Bing Tian ZHANG, Dongxiang Modeling intra-relation in math word problems with different functional multi-head attentions |
description |
Several deep learning models have been proposed for solving math word problems (MWPs) automatically. Although these models have the ability to capture features without manual efforts, their approaches to capturing features are not specifically designed for MWPs. To utilize the merits of deep learning models with simultaneous consideration of MWPs’ specific features, we propose a group attention mechanism to extract global features, quantity-related features, quantity-pair features and question-related features in MWPs respectively. The experimental results show that the proposed approach performs significantly better than previous state-of-the-art methods, and boost performance from 66.9% to 69.5% on Math23K with training-test split, from 65.8% to 66.9% on Math23K with 5-fold cross-validation and from 69.2% to 76.1% on MAWPS. |
format |
text |
author |
LI, Jierui WANG, Lei ZHANG, Jipeng WANG, Yan DAI, Bing Tian ZHANG, Dongxiang |
author_facet |
LI, Jierui WANG, Lei ZHANG, Jipeng WANG, Yan DAI, Bing Tian ZHANG, Dongxiang |
author_sort |
LI, Jierui |
title |
Modeling intra-relation in math word problems with different functional multi-head attentions |
title_short |
Modeling intra-relation in math word problems with different functional multi-head attentions |
title_full |
Modeling intra-relation in math word problems with different functional multi-head attentions |
title_fullStr |
Modeling intra-relation in math word problems with different functional multi-head attentions |
title_full_unstemmed |
Modeling intra-relation in math word problems with different functional multi-head attentions |
title_sort |
modeling intra-relation in math word problems with different functional multi-head attentions |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2019 |
url |
https://ink.library.smu.edu.sg/sis_research/4867 https://ink.library.smu.edu.sg/context/sis_research/article/5870/viewcontent/P19_1619_pvoa.pdf |
_version_ |
1814047490435973120 |