Template-based math word problem solvers with recursive neural networks

The design of automatic solvers to arithmetic math word problems has attracted considerable attention in recent years and a large number of datasets and methods have been published. Among them, Math23K is the largest data corpus that is very helpful to evaluate the generality and robustness of a pro...

Full description

Saved in:
Bibliographic Details
Main Authors: WANG, Lei, ZHANG, Dongxiang, ZHANG, Jipeng, XU, Xing, GAO, Lianli, DAI, Bing Tian, SHEN, Heng Tao
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2019
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/4866
https://ink.library.smu.edu.sg/context/sis_research/article/5869/viewcontent/4697_Article_Text_7736_1_10_20190707.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-5869
record_format dspace
spelling sg-smu-ink.sis_research-58692024-04-18T05:53:53Z Template-based math word problem solvers with recursive neural networks WANG, Lei ZHANG, Dongxiang ZHANG, Jipeng XU, Xing GAO, Lianli DAI, Bing Tian SHEN, Heng Tao The design of automatic solvers to arithmetic math word problems has attracted considerable attention in recent years and a large number of datasets and methods have been published. Among them, Math23K is the largest data corpus that is very helpful to evaluate the generality and robustness of a proposed solution. The best performer in Math23K is a seq2seq model based on LSTM to generate the math expression. However, the model suffers from performance degradation in large space of target expressions. In this paper, we propose a template-based solution based on recursive neural network for math expression construction. More specifically, we first apply a seq2seq model to predict a tree-structure template, with inferred numbers as leaf nodes and unknown operators as inner nodes. Then, we design a recursive neural network to encode the quantity with Bi-LSTM and self attention, and infer the unknown operator nodes in a bottom-up manner. The experimental results clearly establish the superiority of our new framework as we improve the accuracy by a wide margin in two of the largest datasets, i.e., from 58.1% to 66.9% in Math23K and from 62.8% to 66.8% in MAWPS. 2019-01-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4866 info:doi/10.1609/aaai.v33i01.33017144 https://ink.library.smu.edu.sg/context/sis_research/article/5869/viewcontent/4697_Article_Text_7736_1_10_20190707.pdf http://creativecommons.org/licenses/by-nc-sa/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial Intelligence and Robotics Numerical Analysis and Scientific Computing
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Artificial Intelligence and Robotics
Numerical Analysis and Scientific Computing
spellingShingle Artificial Intelligence and Robotics
Numerical Analysis and Scientific Computing
WANG, Lei
ZHANG, Dongxiang
ZHANG, Jipeng
XU, Xing
GAO, Lianli
DAI, Bing Tian
SHEN, Heng Tao
Template-based math word problem solvers with recursive neural networks
description The design of automatic solvers to arithmetic math word problems has attracted considerable attention in recent years and a large number of datasets and methods have been published. Among them, Math23K is the largest data corpus that is very helpful to evaluate the generality and robustness of a proposed solution. The best performer in Math23K is a seq2seq model based on LSTM to generate the math expression. However, the model suffers from performance degradation in large space of target expressions. In this paper, we propose a template-based solution based on recursive neural network for math expression construction. More specifically, we first apply a seq2seq model to predict a tree-structure template, with inferred numbers as leaf nodes and unknown operators as inner nodes. Then, we design a recursive neural network to encode the quantity with Bi-LSTM and self attention, and infer the unknown operator nodes in a bottom-up manner. The experimental results clearly establish the superiority of our new framework as we improve the accuracy by a wide margin in two of the largest datasets, i.e., from 58.1% to 66.9% in Math23K and from 62.8% to 66.8% in MAWPS.
format text
author WANG, Lei
ZHANG, Dongxiang
ZHANG, Jipeng
XU, Xing
GAO, Lianli
DAI, Bing Tian
SHEN, Heng Tao
author_facet WANG, Lei
ZHANG, Dongxiang
ZHANG, Jipeng
XU, Xing
GAO, Lianli
DAI, Bing Tian
SHEN, Heng Tao
author_sort WANG, Lei
title Template-based math word problem solvers with recursive neural networks
title_short Template-based math word problem solvers with recursive neural networks
title_full Template-based math word problem solvers with recursive neural networks
title_fullStr Template-based math word problem solvers with recursive neural networks
title_full_unstemmed Template-based math word problem solvers with recursive neural networks
title_sort template-based math word problem solvers with recursive neural networks
publisher Institutional Knowledge at Singapore Management University
publishDate 2019
url https://ink.library.smu.edu.sg/sis_research/4866
https://ink.library.smu.edu.sg/context/sis_research/article/5869/viewcontent/4697_Article_Text_7736_1_10_20190707.pdf
_version_ 1814047490175926272