Unsupervised word translation with adversarial autoencoder

Crosslingual word embeddings learned from monolingual embeddings have a crucial role in many downstream tasks, ranging from machine translation to transfer learning. Adversarial training has shown impressive success in learning crosslingual embeddings and the associated word translation task without...

Full description

Saved in:
Bibliographic Details
Main Authors: Mohiuddin, Tasnim, Joty, Shafiq
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2021
Subjects:
Online Access:https://hdl.handle.net/10356/148677
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-148677
record_format dspace
spelling sg-ntu-dr.10356-1486772021-05-12T08:20:06Z Unsupervised word translation with adversarial autoencoder Mohiuddin, Tasnim Joty, Shafiq School of Computer Science and Engineering Engineering::Computer science and engineering Embeddings Mapping Crosslingual word embeddings learned from monolingual embeddings have a crucial role in many downstream tasks, ranging from machine translation to transfer learning. Adversarial training has shown impressive success in learning crosslingual embeddings and the associated word translation task without any parallel data by mapping monolingual embeddings to a shared space. However, recent work has shown superior performance for non-adversarial methods in more challenging language pairs. In this article, we investigate adversarial autoencoder for unsupervised word translation and propose two novel extensions to it that yield more stable training and improved results. Our method includes regularization terms to enforce cycle consistency and input reconstruction, and puts the target encoders as an adversary against the corresponding discriminator. We use two types of refinement procedures sequentially after obtaining the trained encoders and mappings from the adversarial training, namely, refinement with Procrustes solution and refinement with symmetric re-weighting. Extensive experimentations with high- and low-resource languages from two different data sets show that our method achieves better performance than existing adversarial and non-adversarial approaches and is also competitive with the supervised system. Along with performing comprehensive ablation studies to understand the contribution of different components of our adversarial model, we also conduct a thorough analysis of the refinement procedures to understand their effects. Published version 2021-05-12T08:20:06Z 2021-05-12T08:20:06Z 2020 Journal Article Mohiuddin, T. & Joty, S. (2020). Unsupervised word translation with adversarial autoencoder. Computational Linguistics, 46(2), 257-288. https://dx.doi.org/10.1162/COLI_a_00374 0891-2017 https://hdl.handle.net/10356/148677 10.1162/COLI_a_00374 2-s2.0-85087454517 2 46 257 288 en Computational Linguistics © 2020 Association for Computational Linguistics Published under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0) license. This is an open-access article distributed under the terms of the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits you to copy and redistribute in any medium or format, for non-commercial use only, provided that the original work is not remixed, transformed, or built upon, and that appropriate credit to the original source is given. For a full description of the license, please visit https://creativecommons.org/licenses/by-nc-nd/4.0/legalcode. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Embeddings
Mapping
spellingShingle Engineering::Computer science and engineering
Embeddings
Mapping
Mohiuddin, Tasnim
Joty, Shafiq
Unsupervised word translation with adversarial autoencoder
description Crosslingual word embeddings learned from monolingual embeddings have a crucial role in many downstream tasks, ranging from machine translation to transfer learning. Adversarial training has shown impressive success in learning crosslingual embeddings and the associated word translation task without any parallel data by mapping monolingual embeddings to a shared space. However, recent work has shown superior performance for non-adversarial methods in more challenging language pairs. In this article, we investigate adversarial autoencoder for unsupervised word translation and propose two novel extensions to it that yield more stable training and improved results. Our method includes regularization terms to enforce cycle consistency and input reconstruction, and puts the target encoders as an adversary against the corresponding discriminator. We use two types of refinement procedures sequentially after obtaining the trained encoders and mappings from the adversarial training, namely, refinement with Procrustes solution and refinement with symmetric re-weighting. Extensive experimentations with high- and low-resource languages from two different data sets show that our method achieves better performance than existing adversarial and non-adversarial approaches and is also competitive with the supervised system. Along with performing comprehensive ablation studies to understand the contribution of different components of our adversarial model, we also conduct a thorough analysis of the refinement procedures to understand their effects.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Mohiuddin, Tasnim
Joty, Shafiq
format Article
author Mohiuddin, Tasnim
Joty, Shafiq
author_sort Mohiuddin, Tasnim
title Unsupervised word translation with adversarial autoencoder
title_short Unsupervised word translation with adversarial autoencoder
title_full Unsupervised word translation with adversarial autoencoder
title_fullStr Unsupervised word translation with adversarial autoencoder
title_full_unstemmed Unsupervised word translation with adversarial autoencoder
title_sort unsupervised word translation with adversarial autoencoder
publishDate 2021
url https://hdl.handle.net/10356/148677
_version_ 1701270543702949888