Towards Equipping Transformer with the Ability of Systematic Compositionality

Authors

  • Chen Huang Sichuan University Engineering Research Center of Machine Learning and Industry Intelligence
  • Peixin Qin Sichuan University Engineering Research Center of Machine Learning and Industry Intelligence
  • Wenqiang Lei Sichuan University Engineering Research Center of Machine Learning and Industry Intelligence
  • Jiancheng Lv Sichuan University Engineering Research Center of Machine Learning and Industry Intelligence

DOI:

https://doi.org/10.1609/aaai.v38i16.29788

Keywords:

NLP: (Large) Language Models

Abstract

One of the key factors in language productivity and human cognition is the ability of Systematic Compositionality, which refers to understanding composed, unseen examples of seen primitives. However, recent evidence reveals that the Transformers have difficulty in generalizing the composed context based on the seen primitives. To this end, we take the first step to propose a compositionality-aware Transformer called CAT and two novel pre-training tasks to facilitate the systematic compositionality. We tentatively provide a successful implementation of a multi-layer CAT on the basis of the especially popular BERT. The experimental results demonstrate that CAT outperforms baselines on compositionality-aware tasks with minimal impact on effectiveness on standardized language understanding tasks.

Published

2024-03-24

How to Cite

Huang, C., Qin, P., Lei, W., & Lv, J. (2024). Towards Equipping Transformer with the Ability of Systematic Compositionality. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16), 18289-18297. https://doi.org/10.1609/aaai.v38i16.29788

Issue

Section

AAAI Technical Track on Natural Language Processing I