Skip to main content

GEC-DCL: Grammatical Error Correction Model with Dynamic Context Learning for Paragraphs and Scholarly Papers

  • Conference paper
  • First Online:
Big Data and Artificial Intelligence (BDA 2023)

Abstract

Over the past decade, there has been noteworthy progress in the field of Automatic Grammatical Error Correction (GEC). Despite this growth, current GEC models possess limitations as they primarily concentrate on single sentences, neglecting the significance of contextual understanding in error correction. While a few models have begun to factor in context alongside target sentences, they frequently depend on inflexible boundaries, which leads to the omission of vital information necessary for rectifying certain errors. To address this issue, we introduce the Dynamic Context Learner (DCL) model, which identifies optimal breakpoints within paragraphs or documents to retain maximum context. Our method surpasses those employing fixed sequence lengths or assuming a limited number of preceding sentences as context. Through extensive evaluation on the CoNLL-2014, BEA-Dev, and FCE-Test datasets, we substantiate the efficacy of our approach, achieving substantial F\(_{0.5}\) score enhancements: 77% increase, 19.61% boost, and 10.49% rise respectively, compared to state-of-the-art models. Furthermore, we contrast our model’s performance with LLaMA’s GEC capabilities. We extend our investigation to scientific writing encompassing various context lengths and validate our technique on the GEC S2ORC dataset, yielding cutting-edge results in scholarly publications.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 49.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 64.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Awasthi, A., Sarawagi, S., Goyal, R., Ghosh, S., Piratla, V.: Parallel iterative edit models for local sequence transduction. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 4260–4270 (2019)

    Google Scholar 

  2. Beltagy, I., Lo, K., Cohan, A.: SciBERT: a pretrained language model for scientific text. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3615–3620. Association for Computational Linguistics, Hong Kong, China, November 2019. https://doi.org/10.18653/v1/D19-1371. https://aclanthology.org/D19-1371

  3. Bryant, C., Felice, M., Briscoe, T.: Automatic annotation and evaluation of error types for grammatical error correction. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 793–805. Association for Computational Linguistics, Vancouver, Canada, July 2017. https://doi.org/10.18653/v1/P17-1074. https://aclanthology.org/P17-1074

  4. Bryant, C., Felice, M., Briscoe, T.: The BEA 2019 shared task on grammatical error correction. In: Proceedings of the 14th Workshop on Innovative Use of NLP for Building Educational Applications, pp. 54–75. Association for Computational Linguistics (2019)

    Google Scholar 

  5. Chollampatt, S., Wang, W., Ng, H.T.: Cross-sentence grammatical error correction. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 435–445 (2019)

    Google Scholar 

  6. Dahlmeier, D., Ng, H.T.: Better evaluation for grammatical error correction. In: Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 568–572. Association for Computational Linguistics, Montréal, Canada, June 2012. https://aclanthology.org/N12-1067

  7. Dahlmeier, D., Ng, H.T., Wu, S.M.: Building a large annotated corpus of learner English: the NUS corpus of learner English. In: Proceedings of the Eighth Workshop on Innovative Use of NLP for Building Educational Applications, pp. 22–31 (2013)

    Google Scholar 

  8. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186. Association for Computational Linguistics, Minneapolis, Minnesota, June 2019. https://doi.org/10.18653/v1/N19-1423. https://aclanthology.org/N19-1423

  9. Liu, Y., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)

  10. Lo, K., Wang, L.L., Neumann, M., Kinney, R., Weld, D.: S2ORC: the semantic scholar open research corpus. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 4969–4983. Association for Computational Linguistics, July 2020. https://doi.org/10.18653/v1/2020.acl-main.447. https://aclanthology.org/2020.acl-main.447

  11. Malmi, E., Krause, S., Rothe, S., Mirylenka, D., Severyn, A.: Encode, tag, realize: high-precision text editing. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 5054–5065 (2019)

    Google Scholar 

  12. Mizumoto, T., Hayashibe, Y., Komachi, M., Nagata, M., Matsumoto, Y.: The effect of learner corpus size in grammatical error correction of ESL writings (2012)

    Google Scholar 

  13. Nallapati, R., Zhou, B., Gulcehre, C., Xiang, B., et al.: Abstractive text summarization using sequence-to-sequence RNNs and beyond. arXiv preprint arXiv:1602.06023 (2016)

  14. Narayan, S., Cohen, S.B., Lapata, M.: Don’t give me the details, just the summary! Topic-aware convolutional neural networks for extreme summarization. arXiv preprint arXiv:1808.08745 (2018)

  15. Ng, H.T., Wu, S.M., Briscoe, T., Hadiwinoto, C., Susanto, R.H., Bryant, C.: The CoNLL-2014 shared task on grammatical error correction. In: Proceedings of the Eighteenth Conference on Computational Natural Language Learning: Shared Task, pp. 1–14. Association for Computational Linguistics (2014). https://doi.org/10.3115/v1/W14-1701. http://www.aclweb.org/anthology/W14-1701

  16. Omelianchuk, K., Atrasevych, V., Chernodub, A., Skurzhanskyi, O.: GECToR-grammatical error correction: tag, not rewrite. In: Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications, pp. 163–170 (2020)

    Google Scholar 

  17. Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training. OpenAI Blog 1(8) (2018)

    Google Scholar 

  18. Sennrich, R., Haddow, B., Birch, A.: Edinburgh neural machine translation systems for WMT 16. arXiv preprint arXiv:1606.02891 (2016)

  19. Tajiri, T., Komachi, M., Matsumoto, Y.: Tense and aspect error correction for ESL learners using global context. In: Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 198–202. Association for Computational Linguistics, Jeju Island, Korea, July 2012. https://aclanthology.org/P12-2039

  20. Touvron, H., et al.: LLaMA: open and efficient foundation language models (2023). https://doi.org/10.48550/arXiv.2302.13971

  21. Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)

    Google Scholar 

  22. Wang, L., Tu, Z., Way, A., Liu, Q.: Exploiting cross-sentence context for neural machine translation. arXiv preprint arXiv:1704.04347 (2017)

  23. Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding. In: Advances in Neural Information Processing Systems, vol. 32 (2019)

    Google Scholar 

  24. Yannakoudakis, H., Briscoe, T., Medlock, B.: A new dataset and method for automatically grading ESOL texts. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 180–189. Association for Computational Linguistics, Portland, Oregon, USA, June 2011. https://aclanthology.org/P11-1019

  25. Yuan, Z., Bryant, C.: Document-level grammatical error correction. In: Proceedings of the 16th Workshop on Innovative Use of NLP for Building Educational Applications, pp. 75–84 (2021)

    Google Scholar 

  26. Zhang, Y., Cui, L., Cai, D., Huang, X., Fang, T., Bi, W.: Multi-task instruction tuning of llama for specific scenarios: a preliminary study on writing assistance. ArXiv abs/2305.13225 (2023)

    Google Scholar 

Download references

Acknowledgements

Rajiv Ratn Shah is partly supported by the Infosys Center for AI, the Center for Design and New Media, and the Center of Excellence in Healthcare at IIIT Delhi.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Avinash Anand .

Editor information

Editors and Affiliations

A Appendix

A Appendix

Table 9 deduces that corrections dependent on the more extended context, like tense and verb choice, have seen significant improvement with our DCL model and show a mean of 24\(\%\) of lift across correction made in missing (M), replacement (R) and unnecessary (U) error in FCE-Test. The formula for calculating the \(Diff F_{0.5}\) can be taken from Eq. 1. In the Eq. 1, A stands for the \(F_{0.5}\) score of the baseline model, and B stands for \(F_{0.5}\) GEC-DCL, the proposed model.

$$\begin{aligned} Diff F_{0.5} = 100 * \frac{2 * |A - B|}{(A + B)} \end{aligned}$$
(1)
Table 8. Types of error and their distribution used in the errorify script.
Table 9. Error type-specific performance comparison of GECToR and DCL-GEC on FCE-Test in F\(_{0.5}\), difference between our proposed model and GECToR in last column. 24\(\%\) increase shown by our DCL-GEC model.

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Anand, A. et al. (2023). GEC-DCL: Grammatical Error Correction Model with Dynamic Context Learning for Paragraphs and Scholarly Papers. In: Goyal, V., Kumar, N., Bhowmick, S.S., Goyal, P., Goyal, N., Kumar, D. (eds) Big Data and Artificial Intelligence. BDA 2023. Lecture Notes in Computer Science, vol 14418. Springer, Cham. https://doi.org/10.1007/978-3-031-49601-1_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-49601-1_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-49600-4

  • Online ISBN: 978-3-031-49601-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics