Show simple item record

dc.contributor.authorOrmazabal Oregi, Aitor
dc.contributor.authorArtetxe Zurutuza, Mikel
dc.contributor.authorSoroa Echave, Aitor ORCID
dc.contributor.authorLabaka Intxauspe, Gorka ORCID
dc.contributor.authorAgirre Bengoa, Eneko ORCID
dc.date.accessioned2024-10-15T18:00:03Z
dc.date.available2024-10-15T18:00:03Z
dc.date.issued2022
dc.identifier.citationProceedings of the 60th Annual Meeting of the Association for Computational Linguistics 1 : 1621-1638 (2022)es_ES
dc.identifier.urihttp://hdl.handle.net/10810/69967
dc.description.abstractRound-trip Machine Translation (MT) is a popular choice for paraphrase generation, which leverages readily available parallel corpora for supervision. In this paper, we formalize the implicit similarity function induced by this approach, and show that it is susceptible to non-paraphrase pairs sharing a single ambiguous translation. Based on these insights, we design an alternative similarity metric that mitigates this issue by requiring the entire translation distribution to match, and implement a relaxation of it through the Information Bottleneck method. Our approach incorporates an adversarial term into MT training in order to learn representations that encode as much information about the reference translation as possible, while keeping as little information about the input as possible. Paraphrases can be generated by decoding back to the source from this representation, without having to generate pivot translations. In addition to being more principled and efficient than round-trip MT, our approach offers an adjustable parameter to control the fidelity-diversity trade-off, and obtains better results in our experiments.es_ES
dc.description.sponsorshipAitor Ormazabal, Gorka Labaka, Aitor Soroa and Eneko Agirre were supported by the Basque Government (excellence research group IT1343-19 and DeepText project KK-2020/00088) and the Spanish MINECO (project DOMINO PGC2018-102041-B-I00 MCIU/AEI/FEDER, UE). Aitor Ormazabal was supported by a doctoral grant from the Spanish MECD. Computing infrastructure funded by UPV/EHU and Gipuzkoako Foru Aldundia.es_ES
dc.language.isoenges_ES
dc.publisherACLes_ES
dc.rightsinfo:eu-repo/semantics/openAccesses_ES
dc.rights.urihttp://creativecommons.org/licenses/by/3.0/es/*
dc.titlePrincipled Paraphrase Generation with Parallel Corporaes_ES
dc.typeinfo:eu-repo/semantics/conferenceObjectes_ES
dc.rights.holder(c)2022 The Association for Computational Linguistics, licensed on a Creative Commons Attribution 4.0 International License.es_ES
dc.relation.publisherversionhttps://doi.org/10.18653/v1/2022.acl-long.114es_ES
dc.identifier.doi10.18653/v1/2022.acl-long.114
dc.departamentoesLenguajes y sistemas informáticoses_ES
dc.departamentoeuHizkuntza eta sistema informatikoakes_ES


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record

(c)2022 The Association for Computational Linguistics, licensed on a Creative Commons Attribution 4.0 International License.
Except where otherwise noted, this item's license is described as (c)2022 The Association for Computational Linguistics, licensed on a Creative Commons Attribution 4.0 International License.