Abstract
Over the past decade, there has been noteworthy progress in the field of Automatic Grammatical Error Correction (GEC). Despite this growth, current GEC models possess limitations as they primarily concentrate on single sentences, neglecting the significance of contextual understanding in error correction. While a few models have begun to factor in context alongside target sentences, they frequently depend on inflexible boundaries, which leads to the omission of vital information necessary for rectifying certain errors. To address this issue, we introduce the Dynamic Context Learner (DCL) model, which identifies optimal breakpoints within paragraphs or documents to retain maximum context. Our method surpasses those employing fixed sequence lengths or assuming a limited number of preceding sentences as context. Through extensive evaluation on the CoNLL-2014, BEA-Dev, and FCE-Test datasets, we substantiate the efficacy of our approach, achieving substantial F\(_{0.5}\) score enhancements: 77% increase, 19.61% boost, and 10.49% rise respectively, compared to state-of-the-art models. Furthermore, we contrast our model’s performance with LLaMA’s GEC capabilities. We extend our investigation to scientific writing encompassing various context lengths and validate our technique on the GEC S2ORC dataset, yielding cutting-edge results in scholarly publications.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Awasthi, A., Sarawagi, S., Goyal, R., Ghosh, S., Piratla, V.: Parallel iterative edit models for local sequence transduction. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 4260–4270 (2019)
Beltagy, I., Lo, K., Cohan, A.: SciBERT: a pretrained language model for scientific text. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3615–3620. Association for Computational Linguistics, Hong Kong, China, November 2019. https://doi.org/10.18653/v1/D19-1371. https://aclanthology.org/D19-1371
Bryant, C., Felice, M., Briscoe, T.: Automatic annotation and evaluation of error types for grammatical error correction. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 793–805. Association for Computational Linguistics, Vancouver, Canada, July 2017. https://doi.org/10.18653/v1/P17-1074. https://aclanthology.org/P17-1074
Bryant, C., Felice, M., Briscoe, T.: The BEA 2019 shared task on grammatical error correction. In: Proceedings of the 14th Workshop on Innovative Use of NLP for Building Educational Applications, pp. 54–75. Association for Computational Linguistics (2019)
Chollampatt, S., Wang, W., Ng, H.T.: Cross-sentence grammatical error correction. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 435–445 (2019)
Dahlmeier, D., Ng, H.T.: Better evaluation for grammatical error correction. In: Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 568–572. Association for Computational Linguistics, Montréal, Canada, June 2012. https://aclanthology.org/N12-1067
Dahlmeier, D., Ng, H.T., Wu, S.M.: Building a large annotated corpus of learner English: the NUS corpus of learner English. In: Proceedings of the Eighth Workshop on Innovative Use of NLP for Building Educational Applications, pp. 22–31 (2013)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186. Association for Computational Linguistics, Minneapolis, Minnesota, June 2019. https://doi.org/10.18653/v1/N19-1423. https://aclanthology.org/N19-1423
Liu, Y., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)
Lo, K., Wang, L.L., Neumann, M., Kinney, R., Weld, D.: S2ORC: the semantic scholar open research corpus. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 4969–4983. Association for Computational Linguistics, July 2020. https://doi.org/10.18653/v1/2020.acl-main.447. https://aclanthology.org/2020.acl-main.447
Malmi, E., Krause, S., Rothe, S., Mirylenka, D., Severyn, A.: Encode, tag, realize: high-precision text editing. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 5054–5065 (2019)
Mizumoto, T., Hayashibe, Y., Komachi, M., Nagata, M., Matsumoto, Y.: The effect of learner corpus size in grammatical error correction of ESL writings (2012)
Nallapati, R., Zhou, B., Gulcehre, C., Xiang, B., et al.: Abstractive text summarization using sequence-to-sequence RNNs and beyond. arXiv preprint arXiv:1602.06023 (2016)
Narayan, S., Cohen, S.B., Lapata, M.: Don’t give me the details, just the summary! Topic-aware convolutional neural networks for extreme summarization. arXiv preprint arXiv:1808.08745 (2018)
Ng, H.T., Wu, S.M., Briscoe, T., Hadiwinoto, C., Susanto, R.H., Bryant, C.: The CoNLL-2014 shared task on grammatical error correction. In: Proceedings of the Eighteenth Conference on Computational Natural Language Learning: Shared Task, pp. 1–14. Association for Computational Linguistics (2014). https://doi.org/10.3115/v1/W14-1701. http://www.aclweb.org/anthology/W14-1701
Omelianchuk, K., Atrasevych, V., Chernodub, A., Skurzhanskyi, O.: GECToR-grammatical error correction: tag, not rewrite. In: Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications, pp. 163–170 (2020)
Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training. OpenAI Blog 1(8) (2018)
Sennrich, R., Haddow, B., Birch, A.: Edinburgh neural machine translation systems for WMT 16. arXiv preprint arXiv:1606.02891 (2016)
Tajiri, T., Komachi, M., Matsumoto, Y.: Tense and aspect error correction for ESL learners using global context. In: Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 198–202. Association for Computational Linguistics, Jeju Island, Korea, July 2012. https://aclanthology.org/P12-2039
Touvron, H., et al.: LLaMA: open and efficient foundation language models (2023). https://doi.org/10.48550/arXiv.2302.13971
Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)
Wang, L., Tu, Z., Way, A., Liu, Q.: Exploiting cross-sentence context for neural machine translation. arXiv preprint arXiv:1704.04347 (2017)
Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding. In: Advances in Neural Information Processing Systems, vol. 32 (2019)
Yannakoudakis, H., Briscoe, T., Medlock, B.: A new dataset and method for automatically grading ESOL texts. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 180–189. Association for Computational Linguistics, Portland, Oregon, USA, June 2011. https://aclanthology.org/P11-1019
Yuan, Z., Bryant, C.: Document-level grammatical error correction. In: Proceedings of the 16th Workshop on Innovative Use of NLP for Building Educational Applications, pp. 75–84 (2021)
Zhang, Y., Cui, L., Cai, D., Huang, X., Fang, T., Bi, W.: Multi-task instruction tuning of llama for specific scenarios: a preliminary study on writing assistance. ArXiv abs/2305.13225 (2023)
Acknowledgements
Rajiv Ratn Shah is partly supported by the Infosys Center for AI, the Center for Design and New Media, and the Center of Excellence in Healthcare at IIIT Delhi.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
A Appendix
A Appendix
Table 9 deduces that corrections dependent on the more extended context, like tense and verb choice, have seen significant improvement with our DCL model and show a mean of 24\(\%\) of lift across correction made in missing (M), replacement (R) and unnecessary (U) error in FCE-Test. The formula for calculating the \(Diff F_{0.5}\) can be taken from Eq. 1. In the Eq. 1, A stands for the \(F_{0.5}\) score of the baseline model, and B stands for \(F_{0.5}\) GEC-DCL, the proposed model.
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Anand, A. et al. (2023). GEC-DCL: Grammatical Error Correction Model with Dynamic Context Learning for Paragraphs and Scholarly Papers. In: Goyal, V., Kumar, N., Bhowmick, S.S., Goyal, P., Goyal, N., Kumar, D. (eds) Big Data and Artificial Intelligence. BDA 2023. Lecture Notes in Computer Science, vol 14418. Springer, Cham. https://doi.org/10.1007/978-3-031-49601-1_7
Download citation
DOI: https://doi.org/10.1007/978-3-031-49601-1_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-49600-4
Online ISBN: 978-3-031-49601-1
eBook Packages: Computer ScienceComputer Science (R0)