Journal of Applied Sciences ›› 2021, Vol. 39 ›› Issue (4): 532-544.doi: 10.3969/j.issn.0255-8297.2021.04.002

• Special Issue on CCF NCCA 2020 • Previous Articles    

Heterogeneous Information Network Representation Learning Based on Generative Adversarial Network

LIU Xinghong1, WANG Ying1,2,3, WANG Xin3,4, LAN Shumei1,2,3   

  1. 1. College of Computer Science and Technology, Jilin University, Changchun 130012, Jilin, China;
    2. College of software, Jilin University, Changchun 130012, Jilin, China;
    3. Key Laboratory of Symbol Computation and Knowledge Engineering, Ministry of Education, Jilin University, Changchun 130012, Jilin, China;
    4. College of Computer Technology and Engineering, Changchun Institute of Technology, Changchun 130012, Jilin, China
  • Received:2020-08-26 Published:2021-08-04

Abstract: In view of the high-dimensional sparsity shortcomings of traditional heterogeneous information networks, we firstly proposed an unsupervised learning model-heterogeneous network representation learning based on generative adversarial network (HNRL-GAN) that embeds the high-dimensional vertices of heterogeneous information networks into low-dimensional vector spaces. Secondly, having analyzed the shortcomings of HNRL-GAN, we proposed an improved model, called as heterogeneous network representation learning based on generative adversarial network plus plus (HNRL-GAN++). Finally, we used HNRL-GAN and HNRL-GAN++ in three data sets, including DBLP, Yelp, and Aminer, to perform node classification and node clustering for testing the effectiveness of the two models. Experimental results show that: 1) Both HNRL-GAN and HNRL-GAN++ achieve the goal of representing high-dimensional sparse nodes in heterogeneous information networks as low-dimensional dense vectors; 2) Compared with HNRL-GAN, HNRL-GAN++ has better performance in retaining network structure information and semantic information in high-dimensional space.

Key words: heterogeneous information network, generative adversarial network, network representation learning

CLC Number: