dc.contributor.author | Artetxe Zurutuza, Mikel | |
dc.contributor.author | Labaka Intxauspe, Gorka | |
dc.contributor.author | Agirre Bengoa, Eneko | |
dc.contributor.author | Cho, Kyunghyun | |
dc.date.accessioned | 2024-10-15T17:37:13Z | |
dc.date.available | 2024-10-15T17:37:13Z | |
dc.date.issued | 2018 | |
dc.identifier.citation | ICLR 2018 Conference | es_ES |
dc.identifier.uri | http://hdl.handle.net/10810/69964 | |
dc.description.abstract | In spite of the recent success of neural machine translation (NMT) in standard benchmarks, the lack of large parallel corpora poses a major practical problem for many language pairs. There have been several proposals to alleviate this issue with, for instance, triangulation and semi-supervised learning techniques, but they still require a strong cross-lingual signal. In this work, we completely remove the need of parallel data and propose a novel method to train an NMT system in a completely unsupervised manner, relying on nothing but monolingual corpora. Our model builds upon the recent work on unsupervised embedding mappings, and consists of a slightly modified attentional encoder-decoder model that can be trained on monolingual corpora alone using a combination of denoising and backtranslation. Despite the simplicity of the approach, our system obtains 15.56 and 10.21 BLEU points in WMT 2014 French-to-English and German-to-English translation. The model can also profit from small parallel corpora, and attains 21.81 and 15.24 points when combined with 100,000 parallel sentences, respectively. Our implementation is released as an open source project. | es_ES |
dc.description.sponsorship | This research was partially supported by a Google Faculty Award, the Spanish MINECO (TUNER TIN2015-65308-C5-1-R, MUSTER PCIN-2015-226 and TADEEP TIN2015-70214-P, cofunded by EU FEDER), the Basque Government (MODELA KK-2016/00082), the UPV/EHU (excellence research group), and the NVIDIA GPU grant program. Mikel Artetxe enjoys a doctoral grant from the Spanish MECD. Kyunghyun Cho thanks support by eBay, TenCent, Facebook, Google, NVIDIA and CIFAR, and was partly supported by Samsung Advanced Institute of Technology (Next Generation Deep Learning: from pattern recognition to AI). | es_ES |
dc.language.iso | eng | es_ES |
dc.publisher | OpenReview | es_ES |
dc.rights | info:eu-repo/semantics/openAccess | es_ES |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/ | * |
dc.title | Unsupervised Neural Machine Translation | es_ES |
dc.type | info:eu-repo/semantics/conferenceObject | es_ES |
dc.rights.holder | (c) 2018 The authors under the Creative Commons Attribution 4.0 International (CC BY 4.0) | es_ES |
dc.relation.publisherversion | https://openreview.net/forum?id=Sy2ogebAW | es_ES |
dc.departamentoes | Lenguajes y sistemas informáticos | es_ES |
dc.departamentoeu | Hizkuntza eta sistema informatikoak | es_ES |