TY - GEN
T1 - Monotonic Simultaneous Translation with Chunk-wise Reordering and Refinement
AU - Han, Hyojung
AU - Ahn, Seokchan
AU - Choi, Yoonjung
AU - Chung, Insoo
AU - Kim, Sangha
AU - Cho, Kyunghyun
N1 - Publisher Copyright:
© 2021 Association for Computational Linguistics
PY - 2021
Y1 - 2021
N2 - Recent work in simultaneous machine translation is often trained with conventional full sentence translation corpora, leading to either excessive latency or necessity to anticipate as-yet-unarrived words, when dealing with a language pair whose word orders significantly differ. This is unlike human simultaneous interpreters who produce largely monotonic translations at the expense of the grammaticality of a sentence being translated. In this paper, we thus propose an algorithm to reorder and refine the target side of a full sentence translation corpus, so that the words/phrases between the source and target sentences are aligned largely monotonically, using word alignment and non-autoregressive neural machine translation. We then train a widely used wait-k simultaneous translation model on this reordered- and-refined corpus. The proposed approach improves BLEU scores and resulting translations exhibit enhanced monotonicity with source sentences.
AB - Recent work in simultaneous machine translation is often trained with conventional full sentence translation corpora, leading to either excessive latency or necessity to anticipate as-yet-unarrived words, when dealing with a language pair whose word orders significantly differ. This is unlike human simultaneous interpreters who produce largely monotonic translations at the expense of the grammaticality of a sentence being translated. In this paper, we thus propose an algorithm to reorder and refine the target side of a full sentence translation corpus, so that the words/phrases between the source and target sentences are aligned largely monotonically, using word alignment and non-autoregressive neural machine translation. We then train a widely used wait-k simultaneous translation model on this reordered- and-refined corpus. The proposed approach improves BLEU scores and resulting translations exhibit enhanced monotonicity with source sentences.
UR - http://www.scopus.com/inward/record.url?scp=85127091798&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85127091798&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85127091798
T3 - WMT 2021 - 6th Conference on Machine Translation, Proceedings
SP - 1110
EP - 1123
BT - WMT 2021 - 6th Conference on Machine Translation, Proceedings
PB - Association for Computational Linguistics (ACL)
T2 - 6th Conference on Machine Translation, WMT 2021
Y2 - 10 November 2021 through 11 November 2021
ER -