透過您的圖書館登入
IP:3.144.248.24
  • 學位論文

以簡化法處理數學文字題

A Reduction Based Approach for Math Word Problem

指導教授 : 項傑

摘要


數學文字題包含自然語言理解與邏輯推斷,所以一直以來都被視為人工智慧領域重要的應用,其中小學數學問題並沒有複雜的數學計算與文字敘述,研究人員能讓其設計出來的系統專注於語言理解與邏輯推斷的部分,而不需要涵蓋廣大的文法以及數學相關的Domain Knowledge,也因此非常適合用來當作衡量的數據集。 本篇論文提出一個新的小學數學解題系統,與現今其他以神經網路建構的系統的不同之處在於此系統著重於邏輯推斷過程。我們的系統除了可以算出答案以外還能產生生出此答案的解釋,並且透過簡化題目來減少所需要設計的規則。此外針對此系統目前處理不了的問題我們也做了錯誤分析,並設計了新的句子簡化方法,更進一步的將其與系統進行結合,最終此系統在我們提出的小學數學數據集上得到顯著的成果。

並列摘要


Math Word Problem involves natural language understanding and logical inference, thus it has been regarded as a major application of AI. Because there is no complicated mathematical calculation in elementary math word problem, researchers could concentrate on the task of understanding and inference. In this thesis, we propose a novel elementary math solver system, named “elementary mathematics problem solver”. The difference between prevailing neural network based systems and ours is that we focus more on the process of logical inference and domain common sense. Our solver not only answers the question, but also gives corresponding explanations. We adopt label sequence as “pattern” to identify similar sentences and problems. Solution strategy is described in natural language script so that teachers can make change at will. We greatly reduce the patterns needed by filtering out non-essential words in problem sentences. Since the explanations are written in natural language, error analysis is very intuitive. Our problem solving philosophy is based on mimicking how humans learn to solve a problem. Ultimately, our system achieves significant performance on our elementary mathematics dataset.

參考文獻


Sutskever and I.Vinyals and Le, Q., “Sequence to sequence learning with neuralnetworks,”In Advances in Neural Information Processing Systems,NIPS 2014, 2014.ix, 5
A. Vaswani an d N. Shazeer and N. Parmar and J. Uszkoreit, L. Jones and A. N.Gomez. Kaiser and I. Polosukhin, “Attention is all you need,”in Advances in neuralinformation processing systems, pp. 5998–6008, 2017. ix, 6, 7, 27
J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deepbidirectional transformers for language understanding,”Proceedings of the 2019Conference of the North American Chapter of the Association for ComputationalLinguistics: Human Language Technologies, 2019. ix, 9, 10, 27
J. Elman, “Finding structure in time,”Cognitive Science, vol. 14, no. 2, pp. 179–211,1990. 4
S. Hochreiter and J. Schmidhuber, “Long short-term memory,”Neural Computation,vol. 9, no. 8, pp. 1735–1780, 1997. 5

延伸閱讀