학술논문

Combining Transformers and Tree-based Decoders for Solving Math Word Problems
Document Type
Conference
Source
2023 IEEE International Conference on Big Data (BigData) Big Data (BigData), 2023 IEEE International Conference on. :5940-5945 Dec, 2023
Subject
Bioengineering
Computing and Processing
Geoscience
Robotics and Control Systems
Signal Processing and Analysis
Recurrent neural networks
Semantics
Big Data
Transformers
Mathematical models
Natural language processing
Graph neural networks
Math Word Problem
Self-Attention Mechanism
Transformer
Sequence-to-Sequence Model
Tree-based Decoder
Language
Abstract
Solving math word problems is a popular topic in natural language processing. We not only need to classify the grammatical structures in the questions, but also understand the mathematical logic expressed between words. Errors in semantic understanding may lead to the failure to generate correct solution equations. Thus, the correct answer cannot be calculated. Previous studies mainly used sequence-to-sequence recurrent neural networks (RNNs) to obtain meaning in words, or combined graph neural networks to capture more information in questions to achieve better results. In addition, recent studies also showed that, a tree-based decoder leads to better results than a decoder of RNNs. In this paper, we propose to combine transformers and tree-based decoders for solving math word problems. Firstly, we use a transformer encoder to read math word problems, whose outputs are given to two different decoders, including Transformer decoder, and a treebased decoder. Secondly, from the answer equations generated from the two decoders, the better solution is selected. The experimental results on the two commonly used math word problem datasets, MAWPS and ASDiv-A, show that our model achieves 89.3% and 81.9% accuracy, which are 2.2% and 4.2% higher than the vanilla transformer model, respectively. For the MAWPS dataset the performance is comparable to state-of-the-art model Graph2Tree. This shows the effectiveness of our proposed method.