Skip to main content

Identification and Generation of Actions Using Pre-trained Language Models

  • Conference paper
  • First Online:
Web Information Systems Engineering – WISE 2023 (WISE 2023)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 14306))

Included in the following conference series:

  • 686 Accesses

Abstract

Email is important in day-to-day communication and its volume is increasing rapidly in the workplace. Many tasks or actions that need to be tracked are contained in emails, but workers find this tracking difficult. An automatic way to find these actions would improve workplace productivity. We focused on the following problems in this paper: 1) identifying actions such as requests and commitments in emails, and 2) generating actionable text from the context of the emails. Recently, pre-trained language models trained on large unlabelled corpus have achieved state-of-the-art results in NLP tasks. In the present study, a combination of two pre-trained models, Bidirectional Encoder Representations from Transformers (BERT) and Text-to-Text Transfer Transformer (T5), is used to identify and generate actions from emails. In our method, the first step is to extract actions from emails using BERT sequence classification. The second step is to generate meaningful actionable text using T5 summarization. The Enron People Assignment (EPA) dataset is used for the evaluation of these methods on both large and small datasets. The BERT sequence classification model is evaluated against other language models and machine learning models. The results show that the BERT model outperforms other machine learning models for the action identification task, and the generated text from the summarization model shows significant improvement over the action sentence. Thus, the contribution of this paper is a state-of-the-art model for identifying actions and generating actionable text by leveraging pre-trained models.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bellotti, V., Ducheneaut, N., Howard, M., Smith, I.: Taking email to task: the design and evaluation of a task management centered email tool. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 345–352. ACM (2003)

    Google Scholar 

  2. Cohen, W.W., Carvalho, V.R., Mitchell, T.M.: Learning to classify email into “speech acts”. In: Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing (2004)

    Google Scholar 

  3. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint: arXiv:1810.04805 (2018)

  4. Garg, A., et al.: NEWS article summarization with pretrained transformer. In: Garg, D., Wong, K., Sarangapani, J., Gupta, S.K. (eds.) IACC 2020. CCIS, vol. 1367, pp. 203–211. Springer, Singapore (2021). https://doi.org/10.1007/978-981-16-0401-0_15

    Chapter  Google Scholar 

  5. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  6. Lampert, A., Dale, R., Paris, C.: Detecting emails containing requests for action. In: Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pp. 984–992. Association for Computational Linguistics (2010)

    Google Scholar 

  7. Lampert, A., Dale, R., Paris, C., et al.: The nature of requests and commitments in email messages. In: Proceedings of the AAAI Workshop on Enhanced Messaging, pp. 42–47 (2008)

    Google Scholar 

  8. Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004)

    Google Scholar 

  9. Mukherjee, S., Mukherjee, S., Hasegawa, M., Awadallah, A.H., White, R.: Smart to-do: automatic generation of to-do items from emails. arXiv preprint: arXiv:2005.06282 (2020)

  10. Nezhad, H.R.M., Gunaratna, K., Cappi, J.: eAssistant: cognitive assistance for identification and auto-triage of actionable conversations. In: Proceedings of the 26th International Conference on World Wide Web Companion, pp. 89–98 (2017)

    Google Scholar 

  11. Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1), 5485–5551 (2020)

    MathSciNet  Google Scholar 

  12. Rameshkumar, R., Bailey, P., Jha, A., Quirk, C.: Assigning people to tasks identified in email: the EPA dataset for addressee tagging for detected task intent. In: Proceedings of the 2018 EMNLP Workshop W-NUT: The 4th Workshop on Noisy User-Generated Text, pp. 28–32 (2018)

    Google Scholar 

  13. Sanh, V., Debut, L., Chaumond, J., Wolf, T.: DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint: arXiv:1910.01108 (2019)

  14. Shu, K., Mukherjee, S., Zheng, G., Awadallah, A.H., Shokouhi, M., Dumais, S.: Learning with weak supervision for email intent detection. In: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1051–1060 (2020)

    Google Scholar 

  15. Sun, C., Qiu, X., Xu, Y., Huang, X.: How to fine-tune BERT for text classification? In: Sun, M., Huang, X., Ji, H., Liu, Z., Liu, Y. (eds.) CCL 2019. LNCS (LNAI), vol. 11856, pp. 194–206. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-32381-3_16

    Chapter  Google Scholar 

  16. Wang, W., Hosseini, S., Awadallah, A.H., Bennett, P.N., Quirk, C.: Context-aware intent identification in email conversations. In: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 585–594 (2019)

    Google Scholar 

  17. Whittaker, S., Sidner, C.: Email overload: exploring personal information management of email. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 276–283 (1996)

    Google Scholar 

  18. Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding. In: Advances in Neural Information Processing Systems, vol. 32 (2019)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Senthil Ganesan Yuvaraj .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yuvaraj, S.G., Benatallah, B., Motahari-Nezhad, H.R., Rabhi, F. (2023). Identification and Generation of Actions Using Pre-trained Language Models. In: Zhang, F., Wang, H., Barhamgi, M., Chen, L., Zhou, R. (eds) Web Information Systems Engineering – WISE 2023. WISE 2023. Lecture Notes in Computer Science, vol 14306. Springer, Singapore. https://doi.org/10.1007/978-981-99-7254-8_52

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-7254-8_52

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-7253-1

  • Online ISBN: 978-981-99-7254-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics