Back to Search Start Over

Improving Textual Network Learning with Variational Homophilic Embeddings

Authors :
Wang, Wenlin
Tao, Chenyang
Gan, Zhe
Wang, Guoyin
Chen, Liqun
Zhang, Xinyuan
Zhang, Ruiyi
Yang, Qian
Henao, Ricardo
Carin, Lawrence
Publication Year :
2019

Abstract

The performance of many network learning applications crucially hinges on the success of network embedding algorithms, which aim to encode rich network information into low-dimensional vertex-based vector representations. This paper considers a novel variational formulation of network embeddings, with special focus on textual networks. Different from most existing methods that optimize a discriminative objective, we introduce Variational Homophilic Embedding (VHE), a fully generative model that learns network embeddings by modeling the semantic (textual) information with a variational autoencoder, while accounting for the structural (topology) information through a novel homophilic prior design. Homophilic vertex embeddings encourage similar embedding vectors for related (connected) vertices. The proposed VHE promises better generalization for downstream tasks, robustness to incomplete observations, and the ability to generalize to unseen vertices. Extensive experiments on real-world networks, for multiple tasks, demonstrate that the proposed method consistently achieves superior performance relative to competing state-of-the-art approaches.<br />Comment: Accepted to NeurIPS 2019

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.1909.13456
Document Type :
Working Paper