Abstract
Machine translation (MT) converts one language to another automatically. One of the major challenges of MT is evaluating the performance of the system. There are many automatic evaluation metrics available these days. But the results of automatic evaluation metrics are sometimes not reliable. In this paper, we have attempted to address this issue by considering another type of evaluation strategy, i.e., syntactic evaluation in Bangla-to-English translation. We have attempted to address the problems of automatic evaluation metric BLEU and, thereby, how syntactic evaluation could be helpful in achieving higher accuracy is discussed. In our syntactic evaluation, we have exploited the use of parts of speech (POS) during computing evaluation scores. A comparative analysis is done on different types of evaluations such as syntactic, human, and automatic on a low-resourced English–Bangla language pair. A correlation indicates syntactic evaluation score correlates more with the human evaluation score compared to the normal BLEU score.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Hutchins J, Lovtskii E (2000) Petr Petrovich Troyanskii (1894–1950): a forgotten pioneer of mechanical translation. Mach Transl 15:187–221
Brown PF, Della Pietra SA, Della Pietra VJ, Mercer RL (1991) Statistical approach to sense disambiguation in machine translation 146–151. https://doi.org/10.3115/112405.112427
Xiong D, Meng F, Liu Q (2016) Topic-based term translation models for statistical machine translation. Artif Intell 232:54–75
Koehn P et al (2007) Moses: open source toolkit for statistical machine translation. In: Proceedings of the 45th annual meeting ACL interaction poster demonstration session—ACL ’07 177. https://doi.org/10.3115/1557769.1557821
Vaswani A et al (2017) Attention is all you need. Adv Neural Inf Process Syst 2017-Decem, 5999–6009
Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. Adv Neural Inf Process Syst 4:3104–3112
Stahlberg F (2020) Neural machine translation: a review. J Artif Intell Res 69:343–418
Vathsala MK, Holi G (2020) RNN based machine translation and transliteration for Twitter data. Int J Speech Technol 23:499–504
Duh K (2008) Ranking vs. regression in machine translation evaluation. In: Third workshop on statistical machine translation WMT 2008 annual meeting association on computer linguist ACL 2008 191–194. https://doi.org/10.3115/1626394.1626425
Liu D, Gildea D (2005) Syntactic features for evaluation of machine translation. In: Proceedings of the ACL workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization ACL 2005 25–32
Banerjee S, Lavie A (2005) METEOR: an automatic metric for mt evaluation with improved correlation with human judgments. In: Proceedings of the acl workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization 2005 65–72
Doddington G (2002) Automatic evaluation of machine translation quality using n-gram co-occurrence statistics 138. https://doi.org/10.3115/1289189.1289273
Park Y, Patwardhan S, Visweswariah K, Gates SC (2008) An empirical analysis of word error rate and keyword error rate. In: Proceedings of annual conference on international speech communication association INTERSPEECH 2070–2073. https://doi.org/10.21437/interspeech.2008-537
Guzmán F, Joty S, Mà rquez L, Nakov P (2017) Machine translation evaluation with neural networks. Comput Speech Lang 45:180–200
PopovÃc M, Ney H (2009) Syntax-oriented evaluation measures for machine translation output. In: EACL 2009—Proceedings of the Fourth Workshop on Statistical Machine Translation 29–32. https://doi.org/10.3115/1626431.1626435
Duma M, Vertan C, Park VM, Menzel W (2013) A new syntactic metric for evaluation of machine translation. ACL Student Res Work 130–135
Haque R, Hasanuzzaman M, Way A (2020) Analysing terminology translation errors in statistical and neural machine translation. Mach Transl 34:149–195
Papineni K, Roukos S, Ward T, Zhu WJ (2002) {B}leu: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting of the association for computational linguistics 311–318 (Association for Computational Linguistics, 2002). https://doi.org/10.3115/1073083.1073135
Agnihotri S (2019) Hyperparameter optimization on neural machine translation. Creat Components 124
Lim R, Heafield K, Hoang H, Briers M, Malony A (2018) Exploring hyper-parameter optimization for neural machine translation on GPU architectures 1–8
Tran N, Schneider J-G, Weber I, Qin AK (2020) Hyper-parameter optimization in classification: to-do or not-to-do. Pattern Recognit 103:107245
Lankford S, Afli H, Way A (2022) Human evaluation of English–Irish transformer-Based NMT 1–19
Newman B, Ang KS, Gong J, Hewitt J (2021) Refining targeted syntactic evaluation of language models 3710–3723. https://doi.org/10.18653/v1/2021.naacl-main.290
Manning C et al (2015) The Stanford CoreNLP natural language processing toolkit 55–60. https://doi.org/10.3115/v1/p14-5010
Marcus MP, Santorini B, Marcinkiewicz MA (1993) Building a large annotated corpus of English: the Penn Treebank. Comput Linguist 19:313–330
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Datta, G., Joshi, N., Gupta, K. (2023). Exploiting Parts of Speech in Bangla-To-English Machine Translation Evaluation. In: Singh, Y., Verma, C., Zoltán, I., Chhabra, J.K., Singh, P.K. (eds) Proceedings of International Conference on Recent Innovations in Computing. ICRIC 2022. Lecture Notes in Electrical Engineering, vol 1011. Springer, Singapore. https://doi.org/10.1007/978-981-99-0601-7_5
Download citation
DOI: https://doi.org/10.1007/978-981-99-0601-7_5
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-0600-0
Online ISBN: 978-981-99-0601-7
eBook Packages: Computer ScienceComputer Science (R0)