Encoder-decoder based neural machine translation

With economic globalization and the rapid development of the Internet, the connections between different countries and languages have become closer and closer, which sharply increase people's demand for cross-language communication. Traditional human translation which has many weaknesses cannot...

Full description

Saved in:
Bibliographic Details
Main Author: Luo, Wenhao
Other Authors: Goh Wang Ling
Format: Theses and Dissertations
Language:English
Published: 2019
Subjects:
Online Access:http://hdl.handle.net/10356/78408
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:With economic globalization and the rapid development of the Internet, the connections between different countries and languages have become closer and closer, which sharply increase people's demand for cross-language communication. Traditional human translation which has many weaknesses cannot meet the needs of a wide range of translations. Artificial intelligence technology has been applied, Machine translation technology is an effective way to realize automatic translation and solve the increasingly common cross-language communication. The statistic machine translation can previously satisfy the minimum requirement of translation, but it requires many improvements. The work of this dissertation is to explore possible application of deep neural network and combine the current popular Recurrent Neural Network (RNN) system to achieve high performance machine translation. After comparing the advantages and disadvantages of different improvement models for RNN, Long Short-Term Memory (LSTM) [27] which is a more complete experimental algorithm model for the encoding and decoding process is engaged in this dissertation. Two kinds of Neural Machine Translation (NMT) models are available, the classical NMT model with greedy decoding, and the NMT model with attention mechanism [32], both were reviewed and explored in this study. Following which the BLEU i evaluation method is used to index the performance of two models, and the results obtained verify that the NMT model with attention mechanism has 1.9 BLEU value higher than the greedy decoding NMT model in training, and 2.3 BLEU value higher than the greedy decoding NMT model in testing, which directly proves the NMT model of the attention mechanism is improved for the performance of neural machine translation. The translation results have also been compared in three different size of sentences, from which conclusions can be got that normal NMT does well in short sentences only, but lose its power in middle and long sentences, while NMT model with attention mechanism act nicely in all three types of sentences. Whereas, some problems occur like excessive translation, which needs future exploration.