Triple-attention computation model for question answering

In order to assess the degree of intelligence the machine, the machine's understanding of the language is an indispensable and important aspect. The question answering system is an important task for the machine to understand human language. Thesis proposes a question and answer system model...

Full description

Saved in:
Bibliographic Details
Main Author: Yu, Sicheng
Other Authors: Mao Kezhi
Format: Theses and Dissertations
Language:English
Published: 2018
Subjects:
Online Access:http://hdl.handle.net/10356/75953
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-75953
record_format dspace
spelling sg-ntu-dr.10356-759532023-07-04T15:56:21Z Triple-attention computation model for question answering Yu, Sicheng Mao Kezhi School of Electrical and Electronic Engineering DRNTU::Engineering::Electrical and electronic engineering In order to assess the degree of intelligence the machine, the machine's understanding of the language is an indispensable and important aspect. The question answering system is an important task for the machine to understand human language. Thesis proposes a question and answer system model based on three kinds of attention calculations. Comparing with other existing models, the calculation of the three attention fully extracted the information between the context and the question in different aspects, making the neural network better learn the context-based representation of the question. The model consists of three layers, embedding layer, attention layer, and predict layer. The role of embedding layer is to vectorize the words in the context and question. The attention layer first calculates the mutual attention between the context and the question, and then calculates the Self-Attention. Finally, the predictive layer is used to predict the start and end of the answer. Through experiments on the SQuAD dataset, the performance of the model using the different RNN architectures is better than that of the main reference model in both EM and F1 values. In addition, the performance of this model has performed well in many question answering models proposed in recent years, surpassing many classical models, and has strong competitiveness. Keywords: Natural language processing, Recurrent neural network, Question answering, Attention mechanism Master of Science (Signal Processing) 2018-09-10T07:05:50Z 2018-09-10T07:05:50Z 2018 Thesis http://hdl.handle.net/10356/75953 en 69 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Electrical and electronic engineering
spellingShingle DRNTU::Engineering::Electrical and electronic engineering
Yu, Sicheng
Triple-attention computation model for question answering
description In order to assess the degree of intelligence the machine, the machine's understanding of the language is an indispensable and important aspect. The question answering system is an important task for the machine to understand human language. Thesis proposes a question and answer system model based on three kinds of attention calculations. Comparing with other existing models, the calculation of the three attention fully extracted the information between the context and the question in different aspects, making the neural network better learn the context-based representation of the question. The model consists of three layers, embedding layer, attention layer, and predict layer. The role of embedding layer is to vectorize the words in the context and question. The attention layer first calculates the mutual attention between the context and the question, and then calculates the Self-Attention. Finally, the predictive layer is used to predict the start and end of the answer. Through experiments on the SQuAD dataset, the performance of the model using the different RNN architectures is better than that of the main reference model in both EM and F1 values. In addition, the performance of this model has performed well in many question answering models proposed in recent years, surpassing many classical models, and has strong competitiveness. Keywords: Natural language processing, Recurrent neural network, Question answering, Attention mechanism
author2 Mao Kezhi
author_facet Mao Kezhi
Yu, Sicheng
format Theses and Dissertations
author Yu, Sicheng
author_sort Yu, Sicheng
title Triple-attention computation model for question answering
title_short Triple-attention computation model for question answering
title_full Triple-attention computation model for question answering
title_fullStr Triple-attention computation model for question answering
title_full_unstemmed Triple-attention computation model for question answering
title_sort triple-attention computation model for question answering
publishDate 2018
url http://hdl.handle.net/10356/75953
_version_ 1772828368301981696