Template-Based Math Word Problem Solvers with Recursive Neural Networks

  • Lei Wang University of Electronic Science and Technology of China
  • Dongxiang Zhang University of Electronic Science and Technology of China
  • Jipeng Zhang University of Electronic Science and Technology of China
  • Xing Xu University of Electronic Science and Technology of China
  • Lianli Gao University of Electronic Science and Technology of China
  • Bing Tian Dai Singapore Management University
  • Heng Tao Shen University of Electronic Science and Technology of China

Abstract

The design of automatic solvers to arithmetic math word problems has attracted considerable attention in recent years and a large number of datasets and methods have been published. Among them, Math23K is the largest data corpus that is very helpful to evaluate the generality and robustness of a proposed solution. The best performer in Math23K is a seq2seq model based on LSTM to generate the math expression. However, the model suffers from performance degradation in large space of target expressions. In this paper, we propose a template-based solution based on recursive neural network for math expression construction. More specifically, we first apply a seq2seq model to predict a tree-structure template, with inferred numbers as leaf nodes and unknown operators as inner nodes. Then, we design a recursive neural network to encode the quantity with Bi-LSTM and self attention, and infer the unknown operator nodes in a bottom-up manner. The experimental results clearly establish the superiority of our new framework as we improve the accuracy by a wide margin in two of the largest datasets, i.e., from 58.1% to 66.9% in Math23K and from 62.8% to 66.8% in MAWPS.

Published
2019-07-17
Section
AAAI Technical Track: Natural Language Processing