Next Article in Journal
Time–Frequency Analysis of Experimental Measurements for the Determination of EMI Noise Generators in Power Converters
Next Article in Special Issue
RefinePose: Towards More Refined Human Pose Estimation
Previous Article in Journal
Detent Force Reduction in Linear Interior Permanent Magnet Generator for Direct-Drive Wave Power Conversion
Previous Article in Special Issue
Railway Obstacle Intrusion Detection Based on Convolution Neural Network Multitask Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Knowledge Base Question Answering Method Based on Graph Convolutional Network and Optimized Search Space

Computer School, Beijing Information Science & Technology University, Beijing 100101, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(23), 3897; https://doi.org/10.3390/electronics11233897
Submission received: 25 October 2022 / Revised: 19 November 2022 / Accepted: 23 November 2022 / Published: 25 November 2022

Abstract

:
Knowledge base question answering (KBQA) aims to provide answers to natural language questions from information in the knowledge base. Although many methods perform well when dealing with simple questions, there are still two challenges for complex questions: huge search space and information missing from the query graphs’ structure. To solve these problems, we propose a novel KBQA method based on a graph convolutional network and optimized search space. When generating the query graph, we rank the query graphs by both their semantic and structural similarities with the question. Then, we just use the top k for the next step. In this process, we specifically extract the structure information of the query graphs by a graph convolutional network while extracting semantic information by a pre-trained model. Thus, we can enhance the method’s ability to understand complex questions. We also introduce a constraint function to optimize the search space. Furthermore, we use the beam search algorithm to reduce the search space further. Experiments on the WebQuestionsSP dataset demonstrate that our method outperforms some baseline methods, showing that the structural information of the query graph has a significant impact on the KBQA task.

1. Introduction

A knowledge graph is a heterogeneous multi-digraph, which means it is directed, and multiple edges can exist between two nodes. An agent generates knowledge by relating elements of a graph to real-world objects and actions. A knowledge graph (KG), also known as a knowledge base (KB), is a structured representation of facts that describes a collection of interlinked descriptions of entities, relationships, and semantic descriptions of entities [1]. Knowledge bases store a large amount of factual knowledge from the real world. Many large KBs, such as DBPedia [2], Freebase, YAGO [3] and NELL [4], have been built to serve downstream tasks. Knowledge base question answering (KBQA), which aims to answer natural language questions by knowledge bases, has received a lot of attention as an important research direction [5,6,7,8]. Figure 1 shows the process of finding the answer to a question by the knowledge in KB.
Semantic parsing-based methods (SP-based methods) are one of the mainstream approaches for KBQA [9,10]. The SP-based methods first convert natural language questions into symbolic logical forms; after that, the answers are obtained by executing them in a knowledge base [11]. Such methods can visualize the reasoning process, which makes the results have high interpretability. However, they rely heavily on the design of logical forms and parsing algorithms.
Some works combine graphical structures with SP-based methods to solve the problem [12,13]. These methods transform question answering into a query graph generation process and show powerful expressive power in the complex KBQA task. However, such approaches still face two problems. (1) The number of query graphs grows exponentially with the growth of the knowledge base size and the emergence of complex questions [14]. (2) Most works only consider the semantic information of the query graphs, while they ignore the natural graph structure features. However, the latter information is also useful for selecting the correct query graphs [15]. Therefore, how to reduce the number of candidate query graphs and how to precisely select the correct query graphs are still the key challenges of the current KBQA work.
In this paper, we focus on how to address the two challenges. For challenge 1, we observe that usually, the correct answer to a complex question cannot be found just once in the large search space. Therefore, we can use staged queries to decompose complex questions into multiple simple questions. In addition, a complex question has more than one constraint, which can be used to further reduce the search overhead. We note that some approaches use the graph structure information to improve the effect in some other Natural Language Processing (NLP) tasks but not KBQA [16,17]. In fact, the structure of the query graph is also useful for KBQA. Therefore, for challenge 2, we extract the structure information of the query graphs to enhance the ability of our method to select the correct answer in KBQA.
Based on the motivation above, we propose a novel KBQA method based on a graph convolutional network by optimized search space. We transform the process of answering complex questions into a hierarchical process of generating query graphs. We extract the constraint function from the complex question and use it to reduce the number of candidate query graphs. After that, we design a novel ranker that scores the candidate query graphs using two components: semantic similarity matching and graph structural similarity. Finally, it uses the beam search algorithm to select the Top K highest-rated query graphs from the candidate query graphs. Due to the addition of the graph structure similarity matching module, our method can select query graphs more accurately. Our main contributions are as follows:
  • To reduce the huge search space for KBQA, we use a constraint function as well as the beam search algorithm to limit the number of candidate query graphs and reduce the computational overhead.
  • To update the correctness of query graphs, we add structural information to the semantic information of the query graphs and score the query graphs from multiple perspectives, which enhances the model’s ability to understand complex questions.
  • Experimental results on the publicly available KBQA dataset WebQuestionsSP show that our method achieves good experimental results compared to the baseline methods.

2. Related Work

2.1. Semantic Parsing-Based Methods for KBQA

Semantic parsing-based methods are the most dominant class of KBQA methods, which aim to parse natural language discourse into logical forms [18,19]. Specifically, this category of methods first encodes the question through semantic and syntactic analysis. Afterward, the encoded questions are converted into logical forms of statements (e.g., SPARQL Protocol and RDF Query Language (SPARQL) and Structured Query Language (SQL)) by using a logical parsing module. Finally, the obtained logical form statements are executed on the knowledge base to query the answers [20,21].
The earlier methods [22,23] can handle simple questions well. However, in the subsequent large-scale knowledge bases, these traditional methods are no longer applicable in the face of complex questions with complex semantic syntax involving multiple entities.

2.2. Query Graph-Based Methods for KBQA

The concept of query graph was first proposed by Yih et al., 2015 [12], which is a new idea to simplify the traditional semantic parsing-based methods [13,14]. The query graph-based method introduces the semantic information formed by entities and relations in the knowledge base during the parsing of a question. It transforms the semantic understanding process of a question into a query graph generation process, which shows the semantic matching process more intuitively and thus has very good interpretability.
However, the query graph generation process usually relies on predefined manual rules, which are not well suited for a large number of complex questions in a large-scale knowledge base. To alleviate this, Ding et al., 2019 [24] used the substructure of frequently occurring queries to assist query graph generation. Abujabal et al., 2017 [25] automatically generated templates based on question–answer pairs to reduce manual operations. Hu et al., 2018 [26] applied aggregation operations and coreference resolution techniques to accommodate complex questions.
In addition, earlier methods only consider the degree of predicate matching in the natural language question and the query graph. They use the core query path in the query graph to measure the similarity to the question [12,27]. These methods omit much useful information and lead to less accurate filtering of the query graph. Based on this, Lan et al., 2020 [28] more comprehensively utilized the information from nodes, relations, and constraints in the query graph generation process. They transformed the query graph into a serialized form containing nodes, relations, and constraints before performing the semantic similarity measure, which enhances the matching ability of their method to the correct query graph. However, the serialization process causes two nodes that are originally adjacent to each other to be split in the sequence, distorting part of the semantic information and destroying the graph structure information that the query graph naturally has.

3. Method

3.1. Overview of the Method

Task Description: A KB collects knowledge data in the form of triples K = { h , r , t } , where r R (the set of relations) and h , t E (the set of entities). For a given natural language question q, the KBQA task is to find the answer a, where a E .
Method overview: We propose a novel KBQA method based on a graph convolutional network by an optimized search space. We formalize the KBQA task as maximizing the probability distribution p ( a | K , q ) . Instead of reasoning directly about K, we retrieve a query graph g K and infer a on g. Since g is unknown, we treat it as a latent variable and rewrite p ( a | K , q ) as:
p ( a | K , q ) = g p ( a | q , g ) p ( g | q )
To obtain the query graph g, our method starts from the topic entity in question q and generates the query graph hierarchically using the e x t e n d or c o n s t r a i n operations, which are described in Section 3.2.
We assume that the correct query graph has a high degree of similarity to the question q. We can use this to select the correct path from the generated candidate query graphs. To measure this similarity, we design a R a n k e r (described in Section 3.3) that selects the candidate query graphs based on semantic matching and structural similarity of the graphs.
Specifically, we use the pre-trained language model RoBERTa to measure the semantic similarity between the question q and the candidate query graphs. At the same time, we use a graph convolutional network to encode the semantic and structural information of the candidate query graphs together, after which we can measure the similarity of these candidate query graphs. Finally, we combine a constrain function and a beam search algorithm to select the query graphs with high similarity for the next step. The beam search algorithm improves the greedy search algorithm by selecting b e a m s i z e candidates from the set of candidates generated by each search as the starting point for subsequent searches. Therefore, we can select the b e a m s i z e query graphs with high similarity scores from all candidate query graphs, which largely reduces the number of query graphs to optimize the search space.
We repeat the above generation-ranking operation until we find the correct answer or reach the maximum hop count limit. An example in Figure 2 shows the process of our method to find the correct answer to a question.

3.2. Query Graph Generation

This module uses two actions: e x t e n d and c o n s t r a i n to generate query graphs.
The e x t e n d action extends the core relational path by adding relations (selected by R a n k e r ) to the query graph. Specifically, we connect the relation r chosen by R a n k e r to the lambda variable X (or the topic entity e t ). After the connection, the original lambda variable X becomes an intermediate variable y (the topic entity e t remains unchanged), while the other end of r becomes the new lambda variable X.
Referring to Luo et al., 2018 [29], we generate a constraint function by matching the keywords (e.g., first, last, biggest, etc.) in the question. The c o n s t r a i n action attaches the detected constraint function to the lambda variable X or an intermediate variable connected to X. In the example in Figure 2, when our method detects the keyword ‘first’, it generates a constraint function a r g m i n , which limits the search to the nodes around which it is connected. Such a constraint helps the model limit the search to a certain range, which reduces the search space.
This module starts with the topic entity (topic entity linking results are from the paper [28]) and uses the e x t e n d action or c o n s t r a i n action to generate the query graph step by step. Some previous methods [12,27] place the process of adding constraints after the core path is fully generated. However, such methods are too simple and have a limited reduction in the number of candidate query graphs. Therefore, our method performs the c o n s t r a i n action before the e x t e n d action, which reduces the number of candidate query graphs.

3.3. Query Graph Ranker

For the method that uses enumeration to search [12], the number of candidate query graphs approaches k n , where k is the core path length and n is the average number of single-hop candidate paths. For complex questions, k n varies from thousands to millions. Such an order of magnitude cannot be handled with current methods.
Therefore, to prevent the number of candidate query graphs from growing exponentially with the number of query steps, we use a beam search algorithm to limit the number of query graphs obtained at each step. Further, in order to select query graphs associated with the correct answers, we design a scoring function to rank the query graphs by both semantic and graph structure perspectives of the query graphs and some simple features. Figure 3 is an example of the query graph R a n k e r .

3.3.1. Semantic Similarity Measure

This module aims to measure the semantic similarity of the natural language question q and the query graph g. This module starts from the topic entity in the question and transforms the query graph into a sequence form g containing entities and relations according to the query graph generation process.
Specifically, we compose the question q and the query graph sequence g into a statement pair as the input to RoBERTa (robustly optimized BERT approach) [30]. Then, their semantic similarity s c o r e ( q , g ) is obtained. The formulas are as follows:
H q g = R o B E R T a C L S ( [ q ; g ] )
s c o r e ( q , g ) = L I N E A R ( H q g )
where RoBERTaCLS denotes the (CLS) representation of the concatenated input (Figure 4), and L I N E A R is a projection layer reducing the representation to a scalar similarity score.

3.3.2. Graph Structure Similarity Measure

The semantic similarity metric module lacks the structural information of the query graph. Furthermore, the sequence transformation process leads to the segmentation of adjacent nodes in the query graph. Therefore, in addition to the semantic information mentioned in Section 3.3.1), this module also parses the query graph from the view of its structure.
First, the module vectorizes a node and its type as N e (using the Global Vectors for Word Representation (GloVe)). After that, N e is fed into the Bi-directional Long Short-Term Memory (Bi-LSTM), and the hidden state h e of the last time step of the Bi-LSTM is selected as the final encoding of the node, i.e.,
h e = B i L S T M ( N e )
At this point, the initial description of each node in the query graph is obtained, but each node in the current graph contains only its own information and lacks the description of its neighboring nodes. Therefore, this module uses the Graph Convolutional Network (GCN) to represent the query graph g. The GCN hierarchically aggregates nodes and their neighbor representations. After several aggregations, the nodes contain more information about their neighborhoods. Then, h g , the final representation of the graph g, is obtained by averaging over all nodes’ representations. The formulas are as follows:
h i ( l + 1 ) = R e L U ( j N ( i ) h j ( l ) D j i W ( l ) + b ( l ) )
D j i = | N ( j ) | | N ( i ) |
h g ( l ) = 1 | V | i V h i ( l )
where N ( i ) is the set of neighbor nodes of node h i ; h j ( l ) is the representation of node h j in the l-th iteration; W ( l ) is the parameter matrix of each layer of linear transformation; b ( l ) is the bias value of each aggregation; and V denotes the set of nodes in graph g.
Finally, the graph structure similarity s c o r e ( q , g ) is measured by using the cosine similarity:
s c o r e ( q , g ) = c o s ( h q , h g )
where h q is the vector representation of the question q (obtained from RoBERTa).

3.3.3. Candidate Query Graph Selection

We design a scoring function that uses the previously obtained semantic similarity and structure similarity as well as some simple features as evaluation criteria to rank the candidate query graphs, the formulas are as follows:
F e a t u r e s = s c o r e ( q , s ) s c o r e ( q , g ) F a n s w e r F t o p i c F c o n s
S C O R E = s i g m o i d ( W [ F e a t u r e s ] + b )
where F a n s w e r is the number of candidate answers; F t o p i c is the topic entity score; F c o n s is the number of constraints; and W and b parameters are to be learned during model training.
Finally, we use the beam search algorithm to select the top K candidate query graphs for the next iteration.

4. Experimentals

4.1. Datasets

We conduct experiments using the WebQuestionsSP (WebQSP [31]) dataset to evaluate the effectiveness of our method. WebQuestionsSP is a widely used publicly available dataset containing 4737 questions based on Freebase KB. Following Sun et al., 2018 [32], we partitioned the dataset into the training/validation/testing sets with the number of 2848/250/1639 questions.

4.2. Methods for Comparison

We have selected several methods in related fields within the last few years as baseline methods. First, we compare the method proposed by Lan et al., 2019 [33], which considers the complexity of multi-hop relational paths but does not use set searches or constraints to reduce the search space. After that, we compare the method of Chen et al., 2019 [34], who transforms the extraction of multi-hop relationships into multiple single-pick extractions, thus reducing the search space. We also compare the method that uses additional information: Han et al., 2020 [35] take textual information as hyper-edges and update entity states using GCN. Next, we compare the method of Yan et al., 2021 [36] that uses auxiliary tasks to enhance the pre-trained model. Then, we compare the method of Qin et al., 2021 [37], who use the relational graph to reduce the search space of the query graph. Finally, we compared some of the latest methods [7,8,14,38]. Among them, Zhang et al., 2022 [7] composed subgraphs from multiple entities. Chen et al., 2022 [14] used abstract query graphs to enhance query graph accuracy. Ye et al., 2022 [8] and Hu et al., 2022 [38] used generative methods to find answers.

4.3. Results

The results of our method compared with the baseline methods on WebQuestionsSP are shown in Table 1.
The method of Qin et al., 2021 [37] reduces the number of candidate query graphs but does not extract the graph structure information of query graphs. Although Han et al., 2020 [35] use GCN to extract graph structure information, they ignore the matching of semantic information. Yan et al. 2021 [36] reformulate the retrieval-based KBQA task to make it a question-context matching form and propose three auxiliary tasks for relation learning, namely relation extraction, relation matching, and relation reasoning, which gives the best results (Hit@1-score) among all baseline methods. Due to the clear supervised signal, these supervised models show excellent performance. In particular, the method of Ye et al., 2022 [8] achieved a surprising F1-score of 76.6.
In contrast, our method not only extracts semantic information by using a pre-trained model but also uses GCN to extract graph structure information. Furthermore, we also combine the beam search algorithm and constraint function to enhance the performance of our method. Thus, our method achieves competitive performance on the WebQSP dataset compared to other baseline methods.

4.4. Ablation Study

In order to verify the validity of each component in the model, we performed an ablation study. Table 2 shows the experimental results.
Variant 1 (w/o RoBERTa): We use Gate Recurrent Unit (GRU) to replace RoBERTa in the model. The performance of the model decreased by 6.0% due to the prevalence of missing links in the knowledge base. For example, 71% of the person entities in Freebase are missing birthplace information [39]. This leads to the fact that two logically related nodes are not linked in the knowledge base, which reduces the likelihood of finding the correct answer. However, the pre-trained model contains knowledge of many open domains and can make predictions about the missing links in the KB.
Variant 2 (w/o GCN): We remove part of the graph structure similarity measure. The performance of the model decreases by 2.2%, which confirms that for query graph-based KBQA methods, extracting the graph structure of the query graph is important. The query graph cannot be filtered well by semantic information matching alone.
Variant 3 (w/o Other features): We remove the simple features of the candidate query graph selection module. This variant has the lowest performance degradation of 0.6%. This proves that these simple features are much less capable of filtering query graphs than semantic matching as well as graph structure matching.
Furthermore, in order to evaluate the impact of the model components more extensively, we continued our experiments based on Variant 1. The results are shown in Table 3.
Variant 1-a (w/o GCN): We removed the graph structure extraction and matching module from variant 1. In this case, the model uses only semantic similarity to select candidate query graphs. The results of the model decreased by 2.4%. This demonstrates that graph structure matching can improve the performance of the query graph-based KBQA model very well.
Variant 1-b (w/o Other features): We removed the simple features from variant 1. The model effect is reduced by 0.4%.
The results of these two variants demonstrate that both graph structure and simple features can have some improvement on the KBQA task under different settings.
We also compared the change in the F1-score score during training for each variant (excluding the variant with simple features removed since the difference in their effectiveness was not significant). As can be seen from Figure 5, although the graph structure metric makes the model fluctuate more sharply in the early stages, which makes the model less effective than the variants without considering the graph structure at some point, it also gives the model a higher upper limit.
The ablation study results prove that each module in our model improves the effectiveness of the model. Moreover, the above variants still outperform some of the baseline models, which proves that the effectiveness of our method comes not only from the individual modules but also depends on the overall process design of the model.

5. Conclusions

In this paper, we propose a novel KBQA method based on graph convolutional networks and optimized search spaces. By constraining the search process, the model is able to handle complex problems with multiple hops. It solves the problem of graph structure information missing in previous query graph-based KBQA methods, and the results show that the addition of the graph structure matching module improves the model performance by 2.2% (F1-score). Experiments on the WebQSP dataset show that our method has excellent performance.
Limitations: In the process of using keywords to detect constraint functions, there may be ambiguity issues. In addition, while using graph structures to improve model performance, our approach leads to an increase in model training time. Further, the large-scale pre-training of the model implies a large resource overhead.
Future work: We plan to optimize the model, reduce the resource overhead, and resolve the ambiguity in the constraint function. We also intend to study the effect of different dataset partitions on the experiment.

Author Contributions

Conceptualization, X.H. and J.L. (Junzhe Li ); Data curation, J.L. (Jintao Luo), J.L. (Junzhe Li) and H.Y.; Formal analysis, J.L. (Jintao Luo); Investigation, J.L. (Junzhe Li); Methodology, X.H., J.L. (Jintao Luo) and J.L. (Junzhe Li); Project administration, X.H.; Resources, X.H., H.Y. and L.W.; Software, J.L. (JintaoLuo) and J.L. (Junzhe Li); Supervision, X.H., H.Y. and L.W.; Visualization, X.H. and J.L. (Jintao Luo); Writing—original draft, X.H. and J.L (Jintao Luo); Writing—review and editing, X.H., H.Y. and L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by: Undergraduate Teaching Reform and Innovation Project of Beijing Higher Education, China, Grant Number: 5112210807, and by the project of Excellent teaching management personnel in Beijing universities, Grant Number: 5112210823.

Data Availability Statement

The WebQuestionsSP dataset can be accessed via the following link (http://aka.ms/WebQSP).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zamini, M.; Reza, H.; Rabiei, M. A Review of Knowledge Graph Completion. Information 2022, 13, 396. [Google Scholar] [CrossRef]
  2. Lehmann, J.; Isele, R.; Jakob, M.; Jentzsch, A.; Kontokostas, D.; Mendes, P.N.; Hellmann, S.; Morsey, M.; van Kleef, P.; Auer, S.; et al. DBpedia—A large-scale, multilingual knowledge base extracted from Wikipedia. Semant. Web 2015, 6, 167–195. [Google Scholar] [CrossRef] [Green Version]
  3. Suchanek, F.M.; Kasneci, G.; Weikum, G. Yago: A core of semantic knowledge. In Proceedings of the WWW’07, 16th International Conference on World Wide Web, Banff, AB, Canada, 8–12 May 2007; pp. 697–706. [Google Scholar]
  4. Mitchell, T.; Cohen, W.; Hruschka, E.; Talukdar, P.; Yang, B.; Betteridge, J.; Carlson, A.; Dalvi, B.; Gardner, M.; Kisiel, B.; et al. Never-ending learning. Commun. ACM 2018, 61, 103–115. [Google Scholar] [CrossRef] [Green Version]
  5. Liang, C.; Berant, J.; Le, Q.V.; Forbus, K.D.; Lao, N. Neural Symbolic Machines: Learning Semantic Parsers on Freebase with Weak Supervision. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, July 30–4 August 2017; Volume 1, pp. 23–33. [Google Scholar]
  6. He, G.; Lan, Y.; Jiang, J.; Zhao, W.X.; Wen, J.R. Improving multi-hop knowledge base question answering by learning intermediate supervision signals. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, Online, 8–12 March 2021; pp. 553–561. [Google Scholar]
  7. Zhang, J.; Zhang, X.; Yu, J.; Tang, J.; Tang, J.; Li, C.; Chen, H. Subgraph Retrieval Enhanced Model for Multi-hop Knowledge Base Question Answering. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 5773–5784. [Google Scholar]
  8. Ye, X.; Yavuz, S.; Hashimoto, K.; Zhou, Y.; Xiong, C. RNG-KBQA: Generation Augmented Iterative Ranking for Knowledge Base Question Answering. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 6032–6043. [Google Scholar]
  9. Abujabal, A.; Roy, R.S.; Yahya, M.; Weikum, G. Never-Ending Learning for Open-Domain Question Answering over Knowledge Bases. In Proceedings of the 2018 World Wide Web Conference on World Wide Web, Lyon, France, 23–27 April 2018; pp. 1053–1062. [Google Scholar]
  10. Zhu, S.; Cheng, X.; Su, S. Knowledge-based question answering by tree-to-sequence learning. Neurocomputing 2020, 372, 64–72. [Google Scholar] [CrossRef]
  11. Lan, Y.; He, G.; Jiang, J.; Jiang, J.; Zhao, W.X.; Wen, J.R. A survey on complex knowledge base question answering: Methods, challenges and solutions. In Proceedings of the 13th International Joint Conference on Artificial Intelligence, Montreal, QC, Canada, 19–27 August 2021; pp. 4483–4491. [Google Scholar]
  12. Yih, W.; Chang, M.; He, X.; Gao, J. Semantic Parsing via Staged Query Graph Generation: Question Answering with Knowledge Base. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing, Beijing, China, 26–31 July 2015; Volume 1, pp. 1321–1331. [Google Scholar]
  13. Qiu, Y.; Zhang, K.; Wang, Y.; Jin, X.; Bai, L.; Guan, S.; Cheng, X. Hierarchical Query Graph Generation for Complex Question Answering over Knowledge Graph. In Proceedings of the CIKM’20: The 29th ACM International Conference on Information and Knowledge Management, Online, 19–23 October 2020; pp. 1285–1294. [Google Scholar]
  14. Chen, Y.; Li, H.; Qi, G.; Wu, T.; Wang, T. Outlining and Filling: Hierarchical Query Graph Generation for Answering Complex Questions Over Knowledge Graphs. IEEE Trans. Knowl. Data Eng. 2022, 1–14. [Google Scholar] [CrossRef]
  15. Sorokin, D.; Gurevych, I. Modeling Semantics with Gated Graph Neural Networks for Knowledge Base Question Answering. In Proceedings of the 27th International Conference on Computational Linguistics, Santa Fe, NM, USA, 20–26 August 2018; pp. 3306–3317. [Google Scholar]
  16. Yu, T.; Yasunaga, M.; Yang, K.; Zhang, R.; Wang, D.; Li, Z.; Radev, D.R. SyntaxSQLNet: Syntax Tree Networks for Complex and Cross-Domain Text-to-SQL Task. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 1653–1663. [Google Scholar]
  17. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the 5th International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  18. Berant, J.; Liang, P. Semantic parsing via paraphrasing. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Baltimore, MD, USA, 22–27 June 2014; pp. 1415–1425. [Google Scholar]
  19. Reddy, S.; Lapata, M.; Steedman, M. Large-scale semantic parsing without question-answer pairs. Trans. Assoc. Comput. Linguist. 2014, 2, 377–392. [Google Scholar] [CrossRef]
  20. Sun, Y.; Zhang, L.; Cheng, G.; Qu, Y. SPARQA: Skeleton-based semantic parsing for complex questions over knowledge bases. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 8952–8959. [Google Scholar]
  21. Chen, Y.; Li, H.; Hua, Y.; Qi, G. Formal Query Building with Query Structure Prediction for Complex Question Answering over Knowledge Base. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, Yokohama, Japan, 11–17 July 2020; pp. 3751–3758. [Google Scholar]
  22. Cai, Q.; Yates, A. Large-scale semantic parsing via schema matching and lexicon extension. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Sofia, Bulgaria, 4–9 August 2013; pp. 423–433. [Google Scholar]
  23. Kwiatkowski, T.; Choi, E.; Artzi, Y.; Zettlemoyer, L. Scaling semantic parsers with on-the-fly ontology matching. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, Seattle, WA, USA, 18–21 October 2013; pp. 1545–1556. [Google Scholar]
  24. Ding, J.; Hu, W.; Xu, Q.; Qu, Y. Leveraging frequent query substructures to generate formal queries for complex question answering. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China, 3–7 November 2019; pp. 2614–2622. [Google Scholar]
  25. Abujabal, A.; Yahya, M.; Riedewald, M.; Weikum, G. Automated template generation for question answering over knowledge graphs. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 1191–1200. [Google Scholar]
  26. Hu, S.; Zou, L.; Zhang, X. A state-transition framework to answer complex questions over knowledge base. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 2098–2108. [Google Scholar]
  27. Xu, K.; Reddy, S.; Feng, Y.; Huang, S.; Zhao, D. Question Answering on Freebase via Relation Extraction and Textual Evidence. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016. [Google Scholar]
  28. Lan, Y.; Jiang, J. Query graph generation for answering multi-hop complex questions from knowledge bases. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 969–974. [Google Scholar]
  29. Luo, K.; Lin, F.; Luo, X.; Zhu, K. Knowledge base question answering via encoding of complex query graphs. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 2185–2194. [Google Scholar]
  30. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A robustly optimized bert pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  31. Yih, W.t.; Richardson, M.; Meek, C.; Chang, M.W.; Suh, J. The value of semantic parse labeling for knowledge base question answering. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Berlin, Germany, 7–12 August 2016; pp. 201–206. [Google Scholar]
  32. Sun, H.; Dhingra, B.; Zaheer, M.; Mazaitis, K.; Salakhutdinov, R.; Cohen, W.W. Open Domain Question Answering Using Early Fusion of Knowledge Bases and Text. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 4231–4242. [Google Scholar]
  33. Lan, Y.; Wang, S.; Jiang, J. Knowledge base question answering with topic units. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 5046–5052. [Google Scholar]
  34. Chen, Z.Y.; Chang, C.H.; Chen, Y.P.; Nayak, J.; Ku, L.W. UHop: An Unrestricted-Hop Relation Extraction Framework for Knowledge-Based Question Answering. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seoul, Republic of Korea, 21–24 June 2019; pp. 345–356. [Google Scholar]
  35. Han, J.; Cheng, B.; Wang, X. Open domain question answering based on text enhanced knowledge graph with hyperedge infusion. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP, Online, 16–20 November 2020; pp. 1475–1481. [Google Scholar]
  36. Yan, Y.; Li, R.; Wang, S.; Zhang, H.; Daoguang, Z.; Zhang, F.; Wu, W.; Xu, W. Large-Scale Relation Learning for Question Answering over Knowledge Bases with Pre-trained Language Models. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online, 7–11 November 2021; pp. 3653–3660. [Google Scholar]
  37. Qin, K.; Li, C.; Pavlu, V.; Aslam, J. Improving query graph generation for complex question answering over knowledge base. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online, 7–11 November 2021; pp. 4201–4207. [Google Scholar]
  38. Hu, X.; Wu, X.; Shu, Y.; Qu, Y. Logical Form Generation via Multi-task Learning for Complex Question Answering over Knowledge Bases. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 1687–1696. [Google Scholar]
  39. Krompaß, D.; Baier, S.; Tresp, V. Type-constrained representation learning in knowledge graphs. In Proceedings of the International Semantic Web Conference, Bethlehem, PA, USA, 11–15 October 2015; Springer: New York, NY, USA, 2015; pp. 640–655. [Google Scholar]
Figure 1. An example of a KBQA task. For the question “In which stadium did Player A’s team win the 1998 World Championship?”, the orange circle and the orange line represent the inference process from Player A (the topic entity) to Stadium B (the answer).
Figure 1. An example of a KBQA task. For the question “In which stadium did Player A’s team win the 1998 World Championship?”, the orange circle and the orange line represent the inference process from Player A (the topic entity) to Stadium B (the answer).
Electronics 11 03897 g001
Figure 2. An example of the query process. Starting from a topic entity in the question, the e x t e n d and c o n s t r a i n operations are applied to generate the query graphs and eventually find the answer. The orange circles represent the constraint function to reduce the search space. R a n k e r is used to select the path with a higher score after ranking the candidate paths, such as the path made up of the orange arrow and the lambda variable X.
Figure 2. An example of the query process. Starting from a topic entity in the question, the e x t e n d and c o n s t r a i n operations are applied to generate the query graphs and eventually find the answer. The orange circles represent the constraint function to reduce the search space. R a n k e r is used to select the path with a higher score after ranking the candidate paths, such as the path made up of the orange arrow and the lambda variable X.
Electronics 11 03897 g002
Figure 3. The structure of the query graph R a n k e r .
Figure 3. The structure of the query graph R a n k e r .
Electronics 11 03897 g003
Figure 4. The input and output of RoBERTa for measuring the semantic similarity.
Figure 4. The input and output of RoBERTa for measuring the semantic similarity.
Electronics 11 03897 g004
Figure 5. Comparison of the F1-score for each variant. (a) Comparison of our method and its variants. (b) Comparison of variant 1 and its variant.
Figure 5. Comparison of the F1-score for each variant. (a) Comparison of our method and its variants. (b) Comparison of variant 1 and its variant.
Electronics 11 03897 g005
Table 1. Experimental results for comparison with baselines.
Table 1. Experimental results for comparison with baselines.
MethodF1Hits@1
Lan el al. (2019) [33]67.968.2
Chen et al. (2019) [34]68.5-
Han et al. (2020) [35]60.668.4
Yan et al. (2021) [36]64.572.9
Qin et al. (2021) [37]66-
Zhang et al. (2022) [7]64.169.5
Chen et al. (2022) * [14]70.370.6
Ye et al. (2022) * [8]75.6-
Hu et al. (2022) * [38]76.6-
our method68.968.5
* denotes supervised methods that use gold SPARQL (or ground truth logical form) as a supervised signal. Our method uses only question–answer pairs, which is a weakly supervised method. The bolded scores represent the highest scores.
Table 2. Experimental results of the ablation study.
Table 2. Experimental results of the ablation study.
MothodF1 Δ F1
our method68.90.0
w/o RoBERTa62.9−6.0
w/o GCN66.7−2.2
w/o Other features68.3−0.6
Table 3. Ablation study for Variant 1.
Table 3. Ablation study for Variant 1.
MethodF1 Δ F1
Variant 162.90.0
w/o GCN60.5−2.4
w/o Other features62.4−0.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hou, X.; Luo, J.; Li, J.; Wang, L.; Yang, H. A Novel Knowledge Base Question Answering Method Based on Graph Convolutional Network and Optimized Search Space. Electronics 2022, 11, 3897. https://doi.org/10.3390/electronics11233897

AMA Style

Hou X, Luo J, Li J, Wang L, Yang H. A Novel Knowledge Base Question Answering Method Based on Graph Convolutional Network and Optimized Search Space. Electronics. 2022; 11(23):3897. https://doi.org/10.3390/electronics11233897

Chicago/Turabian Style

Hou, Xia, Jintao Luo, Junzhe Li, Liangguo Wang, and Hongbo Yang. 2022. "A Novel Knowledge Base Question Answering Method Based on Graph Convolutional Network and Optimized Search Space" Electronics 11, no. 23: 3897. https://doi.org/10.3390/electronics11233897

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop