TY - GEN
T1 - HSSA tree structures for BTG-based preordering in machine translation
AU - Zhang, Yujia
AU - Wang, Hao
AU - Lepage, Yves
N1 - Funding Information:
The second author is supported in part by China Scholarship Council (CSC) under CSC Grant No. 201406890026. We would like to thank Tetsuji Nakagawa for his most helpful comments on the experiment setting details.
PY - 2016
Y1 - 2016
N2 - The Hierarchical Sub-Sentential Alignment (HSSA) method is a method to obtain aligned binary tree structures for two aligned sentences in translation correspondence. We propose to use the binary aligned tree structures delivered by this method as training data for preordering prior to machine translation. For that, we learn a Bracketing Transduction Grammar (BTG) from these binary aligned tree structures. In two oracle experiments in English to Japanese and Japanese to English translation, we show that it is theoretically possible to outperform a baseline system with a default distortion limit of 6, by about 2.5 and 5 BLEU points and, 7 and 10 RIBES points respectively, when preordering the source sentences using the learnt preordering model and using a distortion limit of 0. An attempt at learning a preordering model and its results are also reported.
AB - The Hierarchical Sub-Sentential Alignment (HSSA) method is a method to obtain aligned binary tree structures for two aligned sentences in translation correspondence. We propose to use the binary aligned tree structures delivered by this method as training data for preordering prior to machine translation. For that, we learn a Bracketing Transduction Grammar (BTG) from these binary aligned tree structures. In two oracle experiments in English to Japanese and Japanese to English translation, we show that it is theoretically possible to outperform a baseline system with a default distortion limit of 6, by about 2.5 and 5 BLEU points and, 7 and 10 RIBES points respectively, when preordering the source sentences using the learnt preordering model and using a distortion limit of 0. An attempt at learning a preordering model and its results are also reported.
UR - http://www.scopus.com/inward/record.url?scp=85015885084&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85015885084&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85015885084
T3 - Proceedings of the 30th Pacific Asia Conference on Language, Information and Computation, PACLIC 2016
SP - 123
EP - 132
BT - Proceedings of the 30th Pacific Asia Conference on Language, Information and Computation, PACLIC 2016
A2 - Park, Jong C.
A2 - Chung, Jin-Woo
PB - Institute for the Study of Language and Information
T2 - 30th Pacific Asia Conference on Language, Information and Computation, PACLIC 2016
Y2 - 28 October 2016 through 30 October 2016
ER -