Mostra el registre d'ítem simple
Semantic and syntactic information for neural machine translation: Injecting features to the transformer
dc.contributor.author | Armengol Estapé, Jordi |
dc.contributor.author | Ruiz Costa-Jussà, Marta |
dc.contributor.other | Universitat Politècnica de Catalunya. Departament de Ciències de la Computació |
dc.contributor.other | Barcelona Supercomputing Center |
dc.date.accessioned | 2021-06-17T08:42:28Z |
dc.date.available | 2021-06-17T08:42:28Z |
dc.date.issued | 2021-05-18 |
dc.identifier.citation | Armengol, J.; Costa-jussà, M.R. Semantic and syntactic information for neural machine translation: Injecting features to the transformer. "Machine translation", 18 Maig 2021, vol. 35, p. 3-17. |
dc.identifier.issn | 0922-6567 |
dc.identifier.uri | http://hdl.handle.net/2117/347441 |
dc.description.abstract | Introducing factors such as linguistic features has long been proposed in machine translation to improve the quality of translations. More recently, factored machine translation has proven to still be useful in the case of sequence-to-sequence systems. In this work, we investigate whether this gains hold in the case of the state-of-the-art architecture in neural machine translation, the Transformer, instead of recurrent architectures. We propose a new model, the Factored Transformer, to introduce an arbitrary number of word features in the source sequence in an attentional system. Specifically, we suggest two variants depending on the level at which the features are injected. Moreover, we suggest two combination mechanisms for the word features and words themselves. We experiment both with classical linguistic features and semantic features extracted from a linked data database, and with two low-resource datasets. With the best-found configuration, we show improvements of 0.8 BLEU over the baseline Transformer in the IWSLT German-to-English task. Moreover, we experiment with the more challenging FLoRes English-to-Nepali benchmark, which includes both low-resource and very distant languages, and obtain an improvement of 1.2 BLEU. These improvements are achieved with linguistic and not with semantic information. |
dc.description.sponsorship | This work is supported by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (Grant Agreement No. 947657). |
dc.format.extent | 15 p. |
dc.language.iso | eng |
dc.rights | Attribution 4.0 International |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ |
dc.subject | Àrees temàtiques de la UPC::Informàtica::Intel·ligència artificial::Llenguatge natural |
dc.subject.lcsh | Machine translating |
dc.subject.lcsh | Computational linguistics |
dc.subject.other | Transformer |
dc.subject.other | Factored neural machine translation |
dc.subject.other | Linguistic features |
dc.subject.other | Semantic features |
dc.title | Semantic and syntactic information for neural machine translation: Injecting features to the transformer |
dc.type | Article |
dc.subject.lemac | Traducció automàtica |
dc.subject.lemac | Lingüística computacional |
dc.contributor.group | Universitat Politècnica de Catalunya. VEU - Grup de Tractament de la Parla |
dc.identifier.doi | 10.1007/s10590-021-09264-2 |
dc.description.peerreviewed | Peer Reviewed |
dc.relation.publisherversion | https://link.springer.com/article/10.1007/s10590-021-09264-2 |
dc.rights.access | Open Access |
local.identifier.drac | 31831082 |
dc.description.version | Postprint (published version) |
dc.relation.projectid | info:eu-repo/grantAgreement/EC/H2020/947657/EU/Lifelong UNiversal lAnguage Representation/LUNAR |
local.citation.author | Armengol, J.; Costa-jussà, Marta R. |
local.citation.publicationName | Machine translation |
local.citation.volume | 35 |
local.citation.startingPage | 3 |
local.citation.endingPage | 17 |
Fitxers d'aquest items
Aquest ítem apareix a les col·leccions següents
-
Articles de revista [1.046]
-
Articles de revista [172]
-
Articles de revista [370]