Abstract
Email is important in day-to-day communication and its volume is increasing rapidly in the workplace. Many tasks or actions that need to be tracked are contained in emails, but workers find this tracking difficult. An automatic way to find these actions would improve workplace productivity. We focused on the following problems in this paper: 1) identifying actions such as requests and commitments in emails, and 2) generating actionable text from the context of the emails. Recently, pre-trained language models trained on large unlabelled corpus have achieved state-of-the-art results in NLP tasks. In the present study, a combination of two pre-trained models, Bidirectional Encoder Representations from Transformers (BERT) and Text-to-Text Transfer Transformer (T5), is used to identify and generate actions from emails. In our method, the first step is to extract actions from emails using BERT sequence classification. The second step is to generate meaningful actionable text using T5 summarization. The Enron People Assignment (EPA) dataset is used for the evaluation of these methods on both large and small datasets. The BERT sequence classification model is evaluated against other language models and machine learning models. The results show that the BERT model outperforms other machine learning models for the action identification task, and the generated text from the summarization model shows significant improvement over the action sentence. Thus, the contribution of this paper is a state-of-the-art model for identifying actions and generating actionable text by leveraging pre-trained models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bellotti, V., Ducheneaut, N., Howard, M., Smith, I.: Taking email to task: the design and evaluation of a task management centered email tool. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 345–352. ACM (2003)
Cohen, W.W., Carvalho, V.R., Mitchell, T.M.: Learning to classify email into “speech acts”. In: Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing (2004)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv preprint: arXiv:1810.04805 (2018)
Garg, A., et al.: NEWS article summarization with pretrained transformer. In: Garg, D., Wong, K., Sarangapani, J., Gupta, S.K. (eds.) IACC 2020. CCIS, vol. 1367, pp. 203–211. Springer, Singapore (2021). https://doi.org/10.1007/978-981-16-0401-0_15
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Lampert, A., Dale, R., Paris, C.: Detecting emails containing requests for action. In: Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pp. 984–992. Association for Computational Linguistics (2010)
Lampert, A., Dale, R., Paris, C., et al.: The nature of requests and commitments in email messages. In: Proceedings of the AAAI Workshop on Enhanced Messaging, pp. 42–47 (2008)
Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004)
Mukherjee, S., Mukherjee, S., Hasegawa, M., Awadallah, A.H., White, R.: Smart to-do: automatic generation of to-do items from emails. arXiv preprint: arXiv:2005.06282 (2020)
Nezhad, H.R.M., Gunaratna, K., Cappi, J.: eAssistant: cognitive assistance for identification and auto-triage of actionable conversations. In: Proceedings of the 26th International Conference on World Wide Web Companion, pp. 89–98 (2017)
Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(1), 5485–5551 (2020)
Rameshkumar, R., Bailey, P., Jha, A., Quirk, C.: Assigning people to tasks identified in email: the EPA dataset for addressee tagging for detected task intent. In: Proceedings of the 2018 EMNLP Workshop W-NUT: The 4th Workshop on Noisy User-Generated Text, pp. 28–32 (2018)
Sanh, V., Debut, L., Chaumond, J., Wolf, T.: DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint: arXiv:1910.01108 (2019)
Shu, K., Mukherjee, S., Zheng, G., Awadallah, A.H., Shokouhi, M., Dumais, S.: Learning with weak supervision for email intent detection. In: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1051–1060 (2020)
Sun, C., Qiu, X., Xu, Y., Huang, X.: How to fine-tune BERT for text classification? In: Sun, M., Huang, X., Ji, H., Liu, Z., Liu, Y. (eds.) CCL 2019. LNCS (LNAI), vol. 11856, pp. 194–206. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-32381-3_16
Wang, W., Hosseini, S., Awadallah, A.H., Bennett, P.N., Quirk, C.: Context-aware intent identification in email conversations. In: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 585–594 (2019)
Whittaker, S., Sidner, C.: Email overload: exploring personal information management of email. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 276–283 (1996)
Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R.R., Le, Q.V.: XLNet: generalized autoregressive pretraining for language understanding. In: Advances in Neural Information Processing Systems, vol. 32 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Yuvaraj, S.G., Benatallah, B., Motahari-Nezhad, H.R., Rabhi, F. (2023). Identification and Generation of Actions Using Pre-trained Language Models. In: Zhang, F., Wang, H., Barhamgi, M., Chen, L., Zhou, R. (eds) Web Information Systems Engineering – WISE 2023. WISE 2023. Lecture Notes in Computer Science, vol 14306. Springer, Singapore. https://doi.org/10.1007/978-981-99-7254-8_52
Download citation
DOI: https://doi.org/10.1007/978-981-99-7254-8_52
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-7253-1
Online ISBN: 978-981-99-7254-8
eBook Packages: Computer ScienceComputer Science (R0)