Modeling intra-relation in math word problems with different functional multi-head attentions

Several deep learning models have been proposed for solving math word problems (MWPs) automatically. Although these models have the ability to capture features without manual efforts, their approaches to capturing features are not specifically designed for MWPs. To utilize the merits of deep learnin...

Full description

Saved in:
Bibliographic Details
Main Authors: LI, Jierui, WANG, Lei, ZHANG, Jipeng, WANG, Yan, DAI, Bing Tian, ZHANG, Dongxiang
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2019
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/4867
https://ink.library.smu.edu.sg/context/sis_research/article/5870/viewcontent/P19_1619_pvoa.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-5870
record_format dspace
spelling sg-smu-ink.sis_research-58702024-04-18T06:04:43Z Modeling intra-relation in math word problems with different functional multi-head attentions LI, Jierui WANG, Lei ZHANG, Jipeng WANG, Yan DAI, Bing Tian ZHANG, Dongxiang Several deep learning models have been proposed for solving math word problems (MWPs) automatically. Although these models have the ability to capture features without manual efforts, their approaches to capturing features are not specifically designed for MWPs. To utilize the merits of deep learning models with simultaneous consideration of MWPs’ specific features, we propose a group attention mechanism to extract global features, quantity-related features, quantity-pair features and question-related features in MWPs respectively. The experimental results show that the proposed approach performs significantly better than previous state-of-the-art methods, and boost performance from 66.9% to 69.5% on Math23K with training-test split, from 65.8% to 66.9% on Math23K with 5-fold cross-validation and from 69.2% to 76.1% on MAWPS. 2019-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4867 info:doi/10.18653/v1/P19-1619 https://ink.library.smu.edu.sg/context/sis_research/article/5870/viewcontent/P19_1619_pvoa.pdf http://creativecommons.org/licenses/by-nc-sa/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Attention mechanisms Cross validation Global feature Learning models State-of-the-art methods Word problem Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Attention mechanisms
Cross validation
Global feature
Learning models
State-of-the-art methods
Word problem
Artificial Intelligence and Robotics
Numerical Analysis and Scientific Computing
spellingShingle Attention mechanisms
Cross validation
Global feature
Learning models
State-of-the-art methods
Word problem
Artificial Intelligence and Robotics
Numerical Analysis and Scientific Computing
LI, Jierui
WANG, Lei
ZHANG, Jipeng
WANG, Yan
DAI, Bing Tian
ZHANG, Dongxiang
Modeling intra-relation in math word problems with different functional multi-head attentions
description Several deep learning models have been proposed for solving math word problems (MWPs) automatically. Although these models have the ability to capture features without manual efforts, their approaches to capturing features are not specifically designed for MWPs. To utilize the merits of deep learning models with simultaneous consideration of MWPs’ specific features, we propose a group attention mechanism to extract global features, quantity-related features, quantity-pair features and question-related features in MWPs respectively. The experimental results show that the proposed approach performs significantly better than previous state-of-the-art methods, and boost performance from 66.9% to 69.5% on Math23K with training-test split, from 65.8% to 66.9% on Math23K with 5-fold cross-validation and from 69.2% to 76.1% on MAWPS.
format text
author LI, Jierui
WANG, Lei
ZHANG, Jipeng
WANG, Yan
DAI, Bing Tian
ZHANG, Dongxiang
author_facet LI, Jierui
WANG, Lei
ZHANG, Jipeng
WANG, Yan
DAI, Bing Tian
ZHANG, Dongxiang
author_sort LI, Jierui
title Modeling intra-relation in math word problems with different functional multi-head attentions
title_short Modeling intra-relation in math word problems with different functional multi-head attentions
title_full Modeling intra-relation in math word problems with different functional multi-head attentions
title_fullStr Modeling intra-relation in math word problems with different functional multi-head attentions
title_full_unstemmed Modeling intra-relation in math word problems with different functional multi-head attentions
title_sort modeling intra-relation in math word problems with different functional multi-head attentions
publisher Institutional Knowledge at Singapore Management University
publishDate 2019
url https://ink.library.smu.edu.sg/sis_research/4867
https://ink.library.smu.edu.sg/context/sis_research/article/5870/viewcontent/P19_1619_pvoa.pdf
_version_ 1814047490435973120