Investigating Backtranslation in Neural Machine Translation

Por favor, use este identificador para citar o enlazar este ítem: http://hdl.handle.net/10045/76085
Registro completo de metadatos
Registro completo de metadatos
Campo DCValorIdioma
dc.contributor.authorPoncelas, Alberto-
dc.contributor.authorShterionov, Dimitar-
dc.contributor.authorWay, Andy-
dc.contributor.authorMaillette de Buy Wenniger, Gideon-
dc.contributor.authorPassban, Peyman-
dc.date.accessioned2018-05-31T10:09:17Z-
dc.date.available2018-05-31T10:09:17Z-
dc.date.issued2018-
dc.identifier.citationPoncelas, Alberto, et al. “Investigating Backtranslation in Neural Machine Translation”. In: Pérez-Ortiz, Juan Antonio, et al. (Eds.). Proceedings of the 21st Annual Conference of the European Association for Machine Translation: 28-30 May 2018, Universitat d'Alacant, Alacant, Spain, pp. 249-258es_ES
dc.identifier.isbn978-84-09-01901-4-
dc.identifier.urihttp://hdl.handle.net/10045/76085-
dc.description.abstractA prerequisite for training corpus-based machine translation (MT) systems – either Statistical MT (SMT) or Neural MT (NMT) – is the availability of high-quality parallel data. This is arguably more important today than ever before, as NMT has been shown in many studies to outperform SMT, but mostly when large parallel corpora are available; in cases where data is limited, SMT can still outperform NMT. Recently researchers have shown that back-translating monolingual data can be used to create synthetic parallel corpora, which in turn can be used in combination with authentic parallel data to train a high-quality NMT system. Given that large collections of new parallel text become available only quite rarely, back-translation has become the norm when building state-of-the-art NMT systems, especially in resource-poor scenarios. However, we assert that there are many unknown factors regarding the actual effects of back-translated data on the translation capabilities of an NMT model. Accordingly, in this work we investigate how using back-translated data as a training corpus – both as a separate standalone dataset as well as combined with human-generated parallel data – affects the performance of an NMT model. We use incrementally larger amounts of back-translated data to train a range of NMT systems for German-to-English, and analyse the resulting translation performance.es_ES
dc.description.sponsorshipThis research has been supported by the ADAPT Centre for Digital Content Technology which is funded under the SFI Research Centres Programme (Grant 13/RC/2106) and is co-funded under the European Regional Development Fund. This work has also received funding from the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No 713567.es_ES
dc.languageenges_ES
dc.publisherEuropean Association for Machine Translationes_ES
dc.rights© 2018 The authors. This article is licensed under a Creative Commons 3.0 licence, no derivative works, attribution, CC-BY-ND.es_ES
dc.subjectMachine Translationes_ES
dc.subject.otherLenguajes y Sistemas Informáticoses_ES
dc.titleInvestigating Backtranslation in Neural Machine Translationes_ES
dc.typeinfo:eu-repo/semantics/conferenceObjectes_ES
dc.peerreviewedsies_ES
dc.relation.publisherversionhttp://eamt2018.dlsi.ua.es/proceedings-eamt2018.pdfes_ES
dc.rights.accessRightsinfo:eu-repo/semantics/openAccesses_ES
dc.relation.projectIDinfo:eu-repo/grantAgreement/EC/H2020/713567es_ES
Aparece en las colecciones:EAMT2018 - Proceedings
Investigaciones financiadas por la UE

Archivos en este ítem:
Archivos en este ítem:
Archivo Descripción TamañoFormato 
ThumbnailEAMT2018-Proceedings_27.pdf1,7 MBAdobe PDFAbrir Vista previa


Este ítem está licenciado bajo Licencia Creative Commons Creative Commons