Graph Component Contrastive Learning for Concept Relatedness Estimation

Authors

  • Yueen Ma The Chinese University of Hong Kong
  • Zixing Song The Chinese University of Hong Kong
  • Xuming Hu Tsinghua University
  • Jingjing Li The Chinese University of Hong Kong
  • Yifei Zhang The Chinese University of Hong Kong
  • Irwin King The Chinese University of Hong Kong

DOI:

https://doi.org/10.1609/aaai.v37i11.26568

Keywords:

SNLP: Text Classification, SNLP: Sentence-Level Semantics and Textual Inference

Abstract

Concept relatedness estimation (CRE) aims to determine whether two given concepts are related. Existing methods only consider the pairwise relationship between concepts, while overlooking the higher-order relationship that could be encoded in a concept-level graph structure. We discover that this underlying graph satisfies a set of intrinsic properties of CRE, including reflexivity, commutativity, and transitivity. In this paper, we formalize the CRE properties and introduce a graph structure named ConcreteGraph. To address the data scarcity issue in CRE, we introduce a novel data augmentation approach to sample new concept pairs from the graph. As it is intractable for data augmentation to fully capture the structural information of the ConcreteGraph due to a large amount of potential concept pairs, we further introduce a novel Graph Component Contrastive Learning framework to implicitly learn the complete structure of the ConcreteGraph. Empirical results on three datasets show significant improvement over the state-of-the-art model. Detailed ablation studies demonstrate that our proposed approach can effectively capture the high-order relationship among concepts.

Downloads

Published

2023-06-26

How to Cite

Ma, Y., Song, Z., Hu, X., Li, J., Zhang, Y., & King, I. (2023). Graph Component Contrastive Learning for Concept Relatedness Estimation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 13362-13370. https://doi.org/10.1609/aaai.v37i11.26568

Issue

Section

AAAI Technical Track on Speech & Natural Language Processing