Recommendation method for fusion of knowledge graph convolutional network

In the application of internet of vehicles system, it is particularly important to obtain real-time and effective vehicle information and provide personalized functional services for vehicle operation. This algorithm combines knowledge graph technology with convolutional network and presents a new algorithm model, that is, when calculating the representation of a given entity in the knowledge graph, the information of the neighboring entity is combined with the deviation. Through the integration of neighbor entity information, the local neighborhood structure can be better captured and stored in each entity, and the weight of different neighbor entities depends on the relationship between them and the specific user, which can better reflect the user's personalized interests, in order to fully demonstrate the characteristics of the entity. Compared with the traditional coordinated filtering technology SVD model, this model has improved accuracy and F1 value.

may be many in extreme conditions since the neighbor nodes of each entity are largely different. For this reason, drawing lessons from the idea of graph convolutional network, a concept of receptive field is defined in knowledge graph convolutional network [7]. The receptive field refers to a set of neighborhoods of a fixed size. In this way, the calculation amount of the knowledge graph convolutional network has a certain controllability, and this greatly improves the scalability of the algorithm. So as to achieve effective access to vehicle information, real-time information of the vehicle.

KGCN layer
The predictive equation for judging whether the users are interested in vehicle information in the IOV system.
where G is the knowledge graph composed of entity-relation-entity triples, Y represents the sparse matrix of a given user's vehicle, u represents user, v represents Vehicles, θ represents the set of learning parameters of the whole model.

Measure users' preference for different vehicle information relationships
where g is aggregator, r represents relationship, π represents weight. Define function g: a certain function that can make R d × R d → R, here is an aggregator, used to calculate the score between the user and a certain. Where d is the dimension of the feature representation of the knowledge graph. The real meaning here can be understood as (calculating the user's preference for different relationships). For example, if a user pays much attention to the vehicle service information of the corresponding online ride-hailing, he will pay more attention to the number of vehicle service times, user evaluation and other information, and will be more willing to use high-quality vehicles based on these judgments. Here, g function is used to measure the user's preference for different vehicle information relationships.

Use the linear combination of neighbor information to describe the neighborhood information of a node
where N(v) is the set of all entities directly connected to entity v (that is, within one hop), Note that the weight π here is not only related to the relationship between the two nodes v and e, but also related to the characteristics of the corresponding user u at this time. The weight here is the g function scores of all entities e and relations r in N(v) corresponding to v and then normalized to formula 4.
(1) where N(v) represents the set of all entities directly connected to entity v (that is, within one hop), where e is the characteristics of entity. When calculating the representation of a given entity neighborhood, the vector scores of users and relationships are similar to the role of personalized filters. This processing is because the neighbor feature vectors are focused on the basis of these user-specific vectors.

Control the number of neighbor users
where N(v) represents the set of all entities directly connected to entity v (that is, within one hop), S(v) is entity, (single-layer) accepting field, e is the characteristics of entity, K represents Hyperparameter, select K neighbor users.
In a real knowledge graph [7], there will be too many neighbor users in a certain node v, which will bring huge pressure to the calculation of the overall model. It does not use all its neighbor users, but randomly and uniformly sample a fixed-size set from its neighbor users. Instead, a hyperparameter K is defined. For each node v, only K neighbor users are selected for calculation. In other words, the neighborhood representation of v at this time is written as. In KGCN, because the calculation of the final feature of v is sensitive to these regions v u S(v) .

Define aggregator
The key point in the KGCN model is the fusion of the feature v of the entity and the neighbor feature v u S(v) . Three types of aggregators are defined in KGCN. In this experiment, an additive aggregator is used for comparison experiments.
Sum Aggregator. The sum aggregator is to perform arithmetic addition of two eigenvectors, and then add a nonlinear transformation.
In the formula, w is conversion weight, b represents bias, S(v) represents physical (singlelayer) acceptance field.
Splicing aggregator The splicing aggregator performs splicing operations on two feature vectors, and then adds a nonlinear transformation.
Neighbor aggregator The neighbor aggregator only uses the neighbor features of the entity v, and directly uses the neighborhood representation to replace the representation of the v node.

Algorithm
For a given entity, its final vector feature indicates that it must have a significant relationship with its neighbor users directly connected, which is called a first-order entity feature.
In order to further study the relationship between the two entities in the knowledge graph and explore other potential interests of users, the user's neighborhood set is expanded from one layer to multiple layers, that is, the selected single entity feature is transferred to a layer of adjacent neighborhoods, Then the first-order feature vector can be obtained, as shown in Fig. 1. In other words, the h-level features of a single entity emphatically aggregate the neighbor feature representations within h hops near itself. For this design, different aggregators are used to collect neighborhood information.
H represents the maximum number of receptive fields. For the obtained entity pairs of users and Vehicles, first obtain the receptive field size of the item through iteration, and then repeat the aggregation for H times. In the h time, calculate the neighborhood e ∈ M[h] features of each entity. Then aggregate it with its own features e u [h − 1] to obtain the feature representation of v in the next iteration, as shown in Fig. 2. V u and user feature u are passed into the function together: R d × R d → R d to get the final predicted probability.
where V u represents H-order entity feature, Since the algorithm traverses all possible user-item pairs, in order to make the algorithm more efficient, the negative sampling technique is selected in the training of this article. Therefore, the specific loss function of the KGCN model is as follows: Ev i ∼P(v i ) ℓ y uv ,ŷ uv + �F � 2 In the formul, ℓ represents a-cross-direction loss function, P represents the distribution of negative samples, T u is the number of negative samples of user u, p represents uniform distribution.
The code of the KGCN algorithm is shown in Table 1.

Results
This experiment uses a medical data database as a training experiment. The specific data is shown in Table 2.
The experimental data and various parameters are shown in Table 3.
In this experiment, a classic hidden factor model based on collaborative filtering, singular value decomposition (SVD) and the fusion knowledge graph convolutional network KGCN [8,9], are used in the recommended technology to compare experiments, and the results can be compared to Table 4.
The F1 and accurate value line graphs of the training set, validation set, and test set of the data are shown in Figs. 3 and 4.

Discussion
The traditional collaborative filtering algorithm SVD does not have the auxiliary effect of the knowledge map, and the experimental effect is worse than that of KGCN. In the real relationship, using the attributes of users and vehicles, the various attributes do not exist alone, but are related to each other, forming a knowledge graph. Therefore, KGCN can better represent the relationship and perform better than SVD [9]. Because the multihop neighborhood structure is used, it shows that the knowledge graph captures neighborhood information is helpful for recommendation.
First, the influence of the number of samples in the neighborhood, the performance of the algorithm is further analyzed by changing the number of samples, as shown in Table 5. It can be concluded from the table that the best performance is achieved when the sampling (12) T u = v : y uv = 1       format is 8. This is because a too small K does not have enough capacity to contain neighbor information, and a large K is easily affected by noise [10]. Second, adjust the influence of the number of receiving field layers on the experimental results. After experimenting with the number of layers H of the receiving field from 1 to 4, it is found that it has an effect on the performance of KGCN. The results are shown in Table 6. By comparing the data, it can be concluded that H is more sensitive to the experimental effect of the model than the number of samples K in the field. When H = 4, the accuracy rate drops significantly. This is because too many layers of the receiving field have a negative impact on the experimental effect. This result is consistent with common sense judgment, because if the connected relationship in the knowledge graph is longer, these influencing factors at the end of the relationship chain are meaningless when recommending the similarity of Vehicles. In other words, H being 1 or 2 is sufficient for real life scenarios.
In the IOV system, for physical vehicle information extraction, the data required for training data is too high, and high-quality data has a large labor cost. Therefore, unsupervised learning is considered in subsequent studies to reduce the manual pressure as much as possible. For recommendation technology, and the algorithm is based on knowledge graph construction, the workload is relatively large.

Conclusions
The main content of this chapter is to in-depth study of the recommendation technology based on the knowledge graph-KGCN, this method is to aggregate the feature vector representation of the adjacent entities of a given entity [11,12], and calculate the neighborhood features of the entity by weighting, and by setting a fixed-size neighborhood collection as the acceptance field, the calculation amount of the algorithm is controlled. By comparing with the traditional collaborative filtering algorithm, the result proves the superiority of the algorithm, and it is more suitable for the recommendation of knowledge graph. In the application of IOV system. The new algorithm is proposed to better obtain real-time and effective vehicle information, and provide personalized functions for vehicle operation to provide better services.