Skip to main content
Log in

Extractive convolutional adversarial networks for network embedding

World Wide Web Aims and scope Submit manuscript

Abstract

Network embedding plays an important role in various real-world applications. Most traditional algorithms focus on the topological structure while ignore the information from node attributes. The attributed information is potentially valuable to network embedding. To solve this problem, we propose a deep learning model named Extractive Convolutional Adversarial Network (ECAN) for network embedding. This model aims to extract the latent representations from the topological structure, the attributed information, and labels via three components. In the first part, ECAN extracts features from the topological structure and the attributed information of nodes separately. The second part is a prediction model, which aims to exploit labels of vertices. The third part is a convolutional adversarial model. We train it to distinguish the extractive features which are generated by the hidden layers in the extractive network from either the attributed information or the topological structure. Experiments on six real-world datasets demonstrate the effectiveness of ECAN when compared with state-of-the-art embedding algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Figure 1
Figure 2
Figure 3
Figure 4

References

  1. Aggarwal, C.C. (ed.): Social Network Data Analytics. Springer (2011)

  2. Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, pp 214–223 (2017)

  3. Belkin, M., Niyogi, P.: Laplacian eigenmaps and spectral techniques for embedding and clustering. In: Advances in Neural Information Processing Systems 14 [Neural Information Processing Systems: Natural and Synthetic, NIPS 2001, December 3-8, 2001, Vancouver, British Columbia, Canada], pp 585–591 (2001)

  4. Carterette, B.: Precision and recall. In: Encyclopedia of Database Systems. 2nd edn. Springer (2018)

  5. Chen, R., Xie, H., Chen, J., Rao, Y., Zhao, Y., Wang, F.L.: Extractive adversarial networks for network embedding. In: Proceedings of the 5th International Conference on Behavioral, Economic, and Socio-Cultural Computing (BESC), pp 162–167 (2018)

  6. Dai, Q., Li, Q., Tang, J., Wang, D.: Adversarial network embedding. In: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, Louisiana, USA, February 2-7, 2018, pp 2167–2174 (2018)

  7. Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., Lempitsky, V.S.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17, 59:1–59:35 (2016)

    MathSciNet  MATH  Google Scholar 

  8. Gao, S., Denoyer, L., Gallinari, P.: Temporal link prediction by integrating content and structure information. In: Proceedings of the 20th ACM Conference on Information and Knowledge Management, CIKM 2011, Glasgow, United Kingdom, October 24-28, 2011, pp 1169–1174 (2011)

  9. Gao, H., Wang, Z., Ji, S.: Large-scale learnable graph convolutional networks. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2018, London, UK, August 19-23, 2018, pp 1416–1424 (2018)

  10. Gutiérrez-Gómez, L., Delvenne, J.: Unsupervised network embedding for graph visualization, clustering and classification. arXiv:1903.05980 (2019)

  11. Huang, X., Li, J., Hu, X.: Label informed attributed network embedding. In: Proceedings of the 10th ACM International Conference on Web Search and Data Mining, WSDM 2017, Cambridge, United Kingdom, February 6-10, 2017, pp 731–739 (2017)

  12. Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings (2017)

  13. Kirincic, V., Ceperic, E., Vlahinic, S., Lerga, J.: Support vector machine state estimation. Appl. Artif. Intell. 33(6), 517–530 (2019)

    Article  Google Scholar 

  14. Kynkäänniemi, T., Karras, T., Laine, S., Lehtinen, J., Aila, T.: Improved precision and recall metric for assessing generative models. arXiv:1904.06991 (2019)

  15. Le, T.M.V., Lauw, H.W.: Probabilistic latent document network embedding. In: 2014 IEEE International Conference on Data Mining, ICDM 2014, Shenzhen, China, December 14-17, 2014, pp. 270–279 (2014)

  16. Lu, Q., Getoor, L.: Link-based classification. In: Proceedings of the 20th International Conference on Machine Learning, (ICML 2003), August 21-24, 2003, Washington, DC, USA, pp. 496–503 (2003)

  17. Malliaros, F.D., Vazirgiannis, M.: Clustering and community detection in directed networks: A survey. arXiv:1308.0971 (2013)

  18. Perozzi, B., Al-Rfou, R., Skiena, S.: Deepwalk: Online learning of social representations. In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’14, New York, NY, USA - August 24 - 27, 2014, pp. 701–710 (2014)

  19. Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. In: Proceedings of the 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings (2016)

  20. Ran, Z., Wang, W., Hu, B.: On connections between rényi entropy principal component analysis, kernel learning and graph embedding. Pattern Recogn. Lett. 112, 125–130 (2018)

    Article  Google Scholar 

  21. Rao, H., Shi, X., Rodrigue, A.K., Feng, J., Xia, Y., Elhoseny, M., Yuan, X., Gu, L.: Feature selection based on artificial bee colony and gradient boosting decision tree. Appl. Soft Comput. 74, 634–642 (2019)

    Google Scholar 

  22. Ribeiro, B., Vieira, A., das Neves, J.C.: Supervised isomap with dissimilarity measures in embedding learning. In: Proceedings of the 13th Iberoamerican Congress on Pattern Recognition, Progress in Pattern Recognition, Image Analysis and Applications, CIARP 2008, Havana, Cuba, September 9-12, 2008. Proceedings, pp. 389–396 (2008)

  23. Tang, J., Qu, M.,Wang, M., Zhang, M., Yan, J., Mei, Q.: LINE: large-scale information network embedding. In: Proceedings of the 24th International Conference on World Wide Web, WWW 2015, Florence, Italy, May 18-22, 2015, pp. 1067–1077 (2015)

  24. Wakitani, S., Yamamoto, T., Ishimura, A.: Study on an adaptive GMDH-PID controller using adaptive moment estimation. In: 2017 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2017, Banff, AB, Canada, October 5-8, 2017, pp. 1587–1591 (2017)

  25. Wang, X., Cui, P., Wang, J., Pei, J., Zhu, W., Yang, S.: Community preserving network embedding. In: Proceedings of the 31st AAAI Conference on Artificial Intelligence, February 4-9, 2017, San Francisco, California, USA, pp. 203–209 (2017)

  26. Wang, H., Wang, J., Wang, J., Zhao, M., Zhang, W., Zhang, F., Xie, X., Guo, M.: Graphgan: Graph representation learning with generative adversarial nets. In: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, Louisiana, USA, February 2-7, 2018, pp. 2508–2515 (2018)

  27. Xue, L., Qian, T.: Speech analysis based on locally linear embedding(lle). In: Proceedings of the 6th International Conference on Natural Computation, ICNC 2010, Yantai, Shandong, China, 10-12 August 2010, pp. 2159–2162 (2010)

  28. Yang, T., Jin, R., Chi, Y., Zhu, S.: Combining link and content for community detection: a discriminative approach. In: Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, June 28 - July 1, 2009, pp. 927–936 (2009)

  29. Yin, H.: On multidimensional scaling and the embedding of self-organising maps. Neural Netw. 21(2-3), 160–169 (2008)

    Google Scholar 

  30. Zhang, D., Wang, J., Zhao, X.: Estimating the uncertainty of average F1 scores. In: Proceedings of the 2015 International Conference on The Theory of Information Retrieval, ICTIR 2015, Northampton, Massachusetts, USA, September 27-30, 2015, pp. 317–320 (2015)

  31. Zhang, S., Li, X., Zong, M., Zhu, X., Cheng, D.: Learning k for knn classification. ACM TIST 8(3), 43:1–43:19 (2017)

    Google Scholar 

  32. Zhang, Z., Yang, H., Bu, J., Zhou, S., Yu, P., Zhang, J., Ester, M., Wang, C.: ANRL: attributed network representation learning via deep neural networks. In: Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI 2018, July 13-19, 2018, Stockholm, Sweden, pp. 3155–3161 (2018)

  33. Zhao, B., Sen, P., Getoor, L.: Entity and relationship labeling in affiliation networks. In: ICML Workshop on Statistical Network Analysis (2006)

Download references

Acknowledgments

This work was supported by Top-Up Fund (TFG-04) and Seed Fund (SFG-10) for General Research Fund / Early Career Scheme and Interdisciplinary Research Scheme of the Dean’s Research Fund 2018-19 (FLASS/DRF/IDS-3), Departmental Collaborative Research Fund 2019 (MIT/DCRF-R2/18-19), Funding Support to General Research Fund Proposal (RG 39/2019-2020R) and the Internal Research Grant (RG 90/2018-2019R) of The Education University of Hong Kong, and LEO Dr David P. Chan Institute of Data Science, Lingnan University, Hong Kong. This work was also supported by the National Key R&D Program of China (2018YFB1004404), Key R&D Program of Guangdong Province (2018B010107005), and National Natural Science Foundation of China (U1711262, U1501252, U1611264, U1711261). This article is an extended journal version of a conference paper published at BESC 2018 [5]. Some contents from the conference version are re-used in this journal article as this article is a follow-up work of the conference paper. The new contents of this article are more than 70% according to the regulation of the published journal. The new contents can be summarized in the following aspects: (1) To overcome the drawback that the training process of our basic method may be unstable, we propose a new method which exploits the idea of convolution in the network structure; (2) We add three metrics (i.e., Precision, Recall, and F1 score) to evaluate the effectiveness of different models comprehensively; (3) To test the influence of the kernel size in the hidden layers of the extractive model, the prediction model, and the convolutional adversarial model, we set this parameter to 1 and 3 for comparison; (4) Some recently developed approaches are included as baselines; (5) We conduct in-depth experiments on the proposed model to validate the effectiveness of the employed dimension reduction strategy and classifier.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yanghui Rao.

Additional information

Publisher’snote

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

This article belongs to the Topical Collection: Special Issue on Computational Social Science as the Ultimate Web Intelligence

Guest Editors: Xiaohui Tao, Juan D. Velasquez, Jiming Liu, and Ning Zhong

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Qin, X., Rao, Y., Xie, H. et al. Extractive convolutional adversarial networks for network embedding. World Wide Web 23, 1925–1944 (2020). https://doi.org/10.1007/s11280-019-00740-7

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11280-019-00740-7

Keywords

Navigation