Skip to main content
Log in

Adaptive self-propagation graph convolutional network for recommendation

  • Published:
World Wide Web Aims and scope Submit manuscript

Abstract

Graph Convolutional Networks (GCNs) have received a lot of attention in recommender systems due to their powerful representation learning ability on graph data. Depending on whether to combine ego embeddings after aggregating neighbor embeddings, GCNs can be divided into two categories: without self-propagation and with self-propagation. (1) The GCNs without self-propagation bring the loss of inherent information (e.g., income, age) because of discarding ego embeddings. (2) The existing GCNs with self-propagation treat all nodes (i.e., users and items) indistinguishably, so that the distinctive and diverse characteristics of users and items are overlooked. In light of these problems existed in two types of GCNs, we propose a novel GCN model, Adaptive Self-propagation Graph Convolutional Network (ASP-GCN), to retain inherent information and distinctive characteristics of users and items simultaneously. We first conduct pilot experiments to prove that existing GCNs actually suffer from the aforementioned problems. Then, we use Gumbel-Softmax trick to generate categorical distributions for each node in each layer between two types of embeddings: neighborhood embeddings and hybrid embeddings. Neighborhood embeddings are the aggregation of neighbor embeddings and hybrid embeddings consist of ego and neighborhood embeddings. Next, user and item embeddings are updated by aggregating these two types of embeddings proportionally according to corresponding categorical distributions. After obtaining node embeddings in the last convolution layer, the Bayesian Personalized Ranking loss optimized with a similarity term is used to refine the model parameters. Comprehensive experiments are conducted on three benchmark datasets to demonstrate the effectiveness of ASP-GCN over present state-of-the-art approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Availability of data and materials

All datasets used in this paper are open datasets.

References

  1. Berg, R.v.d., Kipf, T.N., Welling, M.: Graph convolutional matrix completion. arXiv preprint arXiv:1706.02263 (2017)

  2. Chen, L., Wu, L., Hong, R., Zhang, K., Wang, M.: Revisiting graph based collaborative filtering: A linear residual graph convolutional network approach. In: Proceedings of the AAAI conference on artificial intelligence, vol. 34, pp. 27–34. (2020)

  3. He, X., Deng, K., Wang, X., Li, Y., Zhang, Y., Wang, M.: Lightgcn: Simplifying and powering graph convolution network for recommendation. In: Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, pp. 639–648. (2020)

  4. Ying, R., He, R., Chen, K., Eksombatchai, P., Hamilton, W.L., Leskovec, J.: Graph convolutional neural networks for web-scale recommender systems. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, pp. 974–983. (2018)

  5. Liu, M., Li, J., Li, G., Pan, P.: Cross domain recommendation via bi-directional transfer graph collaborative filtering networks. In: Proceedings of the 29th ACM International Conference on Information & Knowledge Management, pp. 885–894. (2020)

  6. Wang, H., Lian, D., Zhang, Y., Qin, L., He, X., Lin, Y., Lin, X.: Binarized graph neural network. World Wide Web 24(3), 825–848 (2021)

    Article  Google Scholar 

  7. Wang, L., Hu, F., Wu, S., Wang, L.: Fully hyperbolic graph convolution network for recommendation. In: Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pp. 3483–3487. (2021)

  8. Mao, K., Zhu, J., Xiao, X., Lu, B., Wang, Z., He, X.: Ultragcn: Ultra simplification of graph convolutional networks for recommendation. In: Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pp. 1253–1262. (2021)

  9. Wang, X., He, X., Wang, M., Feng, F., Chua, T.S.: Neural graph collaborative filtering. In: Proceedings of the 42nd international ACM SIGIR conference on Research and development in Information Retrieval, pp. 165–174. (2019)

  10. Liu, M., Gao, H., Ji, S.: Towards deeper graph neural networks. In: Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pp. 338–348. (2020)

  11. Liu, X., Ding, J., Jin, W., Xu, H., Ma, Y., Liu, Z., Tang, J.: Graph neural networks with adaptive residual. Adv. Neural Inf. Process. Syst. 34, 9720–9733 (2021)

    Google Scholar 

  12. Rendle, S., Freudenthaler, C., Gantner, Z., Schmidt-Thieme, L.: Bpr: Bayesian personalized ranking from implicit feedback. arXiv preprint arXiv:1205.2618 (2012)

  13. Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)

  14. Hamilton, W.L., Ying, R., Leskovec, J.: Inductive representation learning on large graphs. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 1025–1035. (2017)

  15. Zhuang, C.: Ma, Q.: Dual graph convolutional networks for graph-based semi-supervised classification. In: Proceedings of the 2018 World Wide Web Conference, pp. 499–508. (2018)

  16. Qu, L., Zhu, H., Duan, Q., Shi, Y.: Continuous-time link prediction via temporal dependent graph neural network. In: Proceedings of The Web Conference 2020, pp. 3026–3032. (2020)

  17. Chen, H., Yin, H., Sun, X., Chen, T., Gabrys, B., Musial, K.: Multi-level graph convolutional networks for cross-platform anchor link prediction. In: Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pp. 1503–1511. (2020)

  18. Yadati, N., Nitin, V., Nimishakavi, M., Yadav, P., Louis, A., Talukdar, P.: Nhp: Neural hypergraph link prediction. In: Proceedings of the 29th ACM International Conference on Information & Knowledge Management, pp. 1705–1714. (2020)

  19. Ranganathan, V., Barbosa, D.: Hoplop: multi-hop link prediction over knowledge graph embeddings. World Wide Web 25(2), 1037–1065 (2022)

    Article  Google Scholar 

  20. Guo, S., Lin, Y., Feng, N., Song, C., Wan, H.: Attention based spatial-temporal graph convolutional networks for traffic flow forecasting. In: Proceedings of the AAAI conference on artificial intelligence, vol. 33, pp. 922–929. (2019)

  21. Lu, B., Gan, X., Jin, H., Fu, L., Zhang, H.: Spatiotemporal adaptive gated graph convolution network for urban traffic flow forecasting. In: Proceedings of the 29th ACM International Conference on Information & Knowledge Management, pp. 1025–1034. (2020)

  22. Lu, B., Gan, X., Jin, H., Fu, L., Wang, X., Zhang, H.: Make more connections: Urban traffic flow forecasting with spatiotemporal adaptive gated graph convolution network. ACM Trans. Intell. Syst. Technol. (TIST) 13(2), 1–25 (2022)

    Google Scholar 

  23. Xu, M., Li, X., Wang, F., Shang, J.S., Chong, T., Cheng, W., Xu, J.: Learning to effectively model spatial-temporal heterogeneity for traffic flow forecasting. World Wide Web 26(3), 849–865 (2023)

    Article  Google Scholar 

  24. Defferrard, M., Bresson, X., Vandergheynst, P.: Convolutional neural networks on graphs with fast localized spectral filtering. In: Proceedings of the 30th International Conference on Neural Information Processing Systems, pp. 3844–3852. (2016)

  25. Yu, W., Qin, Z.: Graph convolutional network for recommendation with low-pass collaborative filters. In: International Conference on Machine Learning, pp. 10936–10945. (2020)

  26. Liao, R., Zhao, Z., Urtasun, R., Zemel, R.S.: Lanczosnet: Multi-scale deep graph convo-lutional networks. In: 7th International Conference on Learning Representations, ICLR 2019. (2019)

  27. Estrach, J.B., Zaremba, W., Szlam, A., LeCun, Y.: Spectral networks and deep locally connected networks on graphs. In: 2nd international conference on learning representations, ICLR, vol. 2014. (2014)

  28. Zhu, H., Feng, F., He, X., Wang, X., Li, Y., Zheng, K., Zhang, Y.: Bilinear graph neural network with neighbor interactions. In: Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, pp. 1452–1458. (2021)

  29. Huang, T., Dong, Y., Ding, M., Yang, Z., Feng, W., Wang, X., Tang, J.: Mixgcf: An improved training method for graph neural network-based recommender systems. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 665–674. (2021)

  30. Zheng, Y., Gao, C., Chen, L., Jin, D., Li, Y.: Dgcn: Diversified recommendation with graph convolutional networks. In: Proceedings of the Web Conference 2021, pp. 401–412. (2021)

  31. Liu, F., Cheng, Z., Zhu, L., Gao, Z., Nie, L.: Interest-aware message-passing gcn for recommendation. In: Proceedings of the Web Conference 2021, pp. 1296–1305. (2021)

  32. Huijben, I.A., Kool, W., Paulus, M.B., Van Sloun, R.J.: A review of the gumbel-max trick and its extensions for discrete stochasticity in machine learning. IEEE Trans. Pattern. Anal. Mach. Intell. 45(2), 1353–1371 (2022)

    Article  Google Scholar 

  33. Maddison, C.J., Tarlow, D., Minka, T.: A* sampling. In: Proceedings of the 27th International Conference on Neural Information Processing Systems-Volume 2, pp. 3086–3094. (2014)

  34. Oberst, M., Sontag, D.: Counterfactual off-policy evaluation with gumbel-max structural causal models. In: International Conference on Machine Learning, pp. 4881–4890. (2019)

  35. Lorberbom, G., Johnson, D., Maddison, C.J., Tarlow, D., Hazan, T.: Learning generalized gumbel-max causal mechanisms. Adv. Neural Inf. Syst. 34 (2021)

  36. Jang, E., Gu, S., Poole, B.: Categorical reparameterization with gumbel-softmax. arXiv preprint arXiv:1611.01144 (2016)

  37. Potapczynski, A., Loaiza-Ganem, G., Cunningham, J.P.: Invertible gaussian reparameterization: Revisiting the gumbel-softmax. Adv. Neural Inf. Process Syst. 33, 12311–12321 (2020)

    Google Scholar 

  38. Hu, S., Xie, S., Zheng, H., Liu, C., Shi, J., Liu, X., Lin, D.: Dsnas: Direct neural architecture search without parameter retraining. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12084–12092. (2020)

  39. Wu, B., Dai, X., Zhang, P., Wang, Y., Sun, F., Wu, Y., Tian, Y., Vajda, P., Jia, Y., Keutzer, K.: Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10734–10742. (2019)

  40. Xiao, T., Chen, Z., Wang, D., Wang, S.: Learning how to propagate messages in graph neural networks. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 1894–1903. (2021)

  41. He, C., Ye, H., Shen, L., Zhang, T.: Milenas: Efficient neural architecture search via mixed-level reformulation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11993–12002. (2020)

  42. Li, Y., Dong, M., Wang, Y., Xu, C.: Neural architecture search in a proxy validation loss landscape. In: International Conference on Machine Learning, pp. 5853–5862. (2020)

  43. Li, D., Yang, T., Du, L., He, Z., Jiang, L.: Adaptivegcn: Efficient gcn through adaptively sparsifying graphs. In: Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pp. 3206–3210. (2021)

  44. Zheng, C., Zong, B., Cheng, W., Song, D., Ni, J., Yu, W., Chen, H., Wang, W.: Robust graph representation learning via neural sparsification. In: International Conference on Machine Learning, pp. 11458–11468. (2020)

  45. Kong, T., Kim, T., Jeon, J., Choi, J., Lee, Y.C., Park, N., Kim, S.W.: Linear, or non-linear, that is the question! In: Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, pp. 517–525. (2022)

  46. Klambauer, G., Unterthiner, T., Mayr, A., Hochreiter, S.: Self-normalizing neural networks. In: Proceedings of the 31st international conference on neural information processing systems, pp. 972–981. (2017)

  47. He, X., Liao, L., Zhang, H., Nie, L., Hu, X., Chua, T.S.: Neural collaborative filtering. In: Proceedings of the 26th international conference on world wide web, pp. 173–182. (2017)

  48. He, X., Zhang, H., Kan, M.Y., Chua, T.S.: Fast matrix factorization for online recommendation with implicit feedback. In: Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, pp. 549–558. (2016)

  49. Hu, Y., Koren, Y., Volinsky, C.: Collaborative filtering for implicit feedback datasets. In: 2008 Eighth IEEE international conference on data mining, pp. 263–272. IEEE (2008)

  50. Wang, X., Jin, H., Zhang, A., He, X., Xu, T., Chua, T.S.: Disentangled graph collaborative filtering. In: Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval, pp. 1001–1010. (2020)

  51. Wu, J., Wang, X., Feng, F., He, X., Chen, L., Lian, J., Xie, X.: Self-supervised graph learning for recommendation. In: Proceedings of the 44th international ACM SIGIR conference on research and development in information retrieval, pp. 726–735. (2021)

  52. Zhu, T., Sun, L., Chen, G.: Embedding disentanglement in graph convolutional networks for recommendation. IEEE Trans. Knowl. Data Eng. (2021)

  53. Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp. 249–256. (2010)

  54. Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. In: ICLR (Poster). (2015)

Download references

Acknowledgements

We acknowledge the editorial committee’s support and all anonymous reviewers for their insightful comments and suggestions, which can improve the content and presentation of this manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grants 71774159 and 62272066, China Postdoctoral Science Foundation under Grants 2021T140707, Jiangsu Postdoctoral Science Foundation under Grants 2021K565C, and State Key Laboratory of NBC Protection for Civilian under grant SKLNBC2020-23.

Author information

Authors and Affiliations

Authors

Contributions

Zhuo Cai wrote the main manuscript text and designed the experiments, Guan Yuan provided the main idea of this manuscript and organized the full text, Xiaobao Zhuang prepared Figures 1-4 and formated the references. Senzhang Wang checked the English grammar and smoothed the representation of main text, Shaojie Qiao gave many suggestions and instructions in experiments, and Mu Zhu sorted out experimental data. All authors have proofread and approved the final manuscript.

Corresponding author

Correspondence to Guan Yuan.

Ethics declarations

Competing interests

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Cai, Z., Yuan, G., Zhuang, X. et al. Adaptive self-propagation graph convolutional network for recommendation. World Wide Web 26, 3183–3206 (2023). https://doi.org/10.1007/s11280-023-01182-y

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11280-023-01182-y

Keywords

Navigation