Skip to main content

Exploit Label Embeddings for Enhancing Network Classification

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 10439))

Abstract

Learning representations for network has aroused great research interests in recent years. Existing approaches embed vertices into a low dimensional continuous space which encodes local or global network structures. While these methods show improvements over traditional representations, they do not utilize the label information until the learned embeddings are used for training classifier. That is, the process of representation learning is separated from the labels and thus is unsupervised. In this paper, we propose a novel method which learns the embeddings for vertices under the supervision of labels. The key idea is to regard the label as the context of a vertex. More specifically, we attach a true or virtual label node for each training or test sample, and update the embeddings for vertices and labels to maximize the probability of both the neighbors and their labels in the context. We conduct extensive experiments on three real datasets. Results demonstrate that our method outperforms the state-of-the-art approaches by a large margin.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Ahmed, A., Shervashidze, N., Narayanamurthy, S., Josifovski, V., Smola, A.J.: Distributed large-scale natural graph factorization. In: Proceedings of WWW, pp. 37–48 (2013)

    Google Scholar 

  2. Angelova, R., Weikum, G.: Graph-based text classification: learn from your neighbors. In: Proceeding of SIGIR, pp. 485–492 (2006)

    Google Scholar 

  3. Balasubramanian, M., Schwartz, E.L.: The isomap algorithm and topological stability. Science 295(5552), 7 (2002)

    Article  Google Scholar 

  4. Belkin, M., Niyogi, P.: Laplacian eigenmaps and spectral techniques for embedding and clustering. In: International Conference on Neural Information Processing Systems: Natural and Synthetic, pp. 585–591 (2001)

    Google Scholar 

  5. Bottou, L.: Stochastic gradient learning in neural networks. In: Neuro-Nîmes (1991)

    Google Scholar 

  6. Fan, R.E., Chang, K.W., Hsieh, C.J., Wang, X.R., Lin, C.J.: Liblinear: a library for large linear classification. J. Mach. Learn. Res. 9, 1871–1874 (2008)

    MATH  Google Scholar 

  7. Grover, A., Leskovec, J.: node2vec: Scalable feature learning for networks. In: Proceedings of SIGKDD, pp. 855–864 (2016)

    Google Scholar 

  8. Macskassy, S.A., Provost, F.: A simple relational classier. In: Proceedings of the Second Workshop on Multi-Relational Data Mining (MRDM) at KDD-2003, pp. 64–76 (2003)

    Google Scholar 

  9. Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. CoRR abs/1301.3781 (2013)

    Google Scholar 

  10. Mikolov, T., Sutskever, I., Chen, K., Corrado, G., Dean., J.: Distributed representations of words and phrases and their compositionality. In: Proceedings of NIPS, pp. 3111–3119 (2013)

    Google Scholar 

  11. Perozzi, B., Al-Rfou, R., Skiena., S.: Deepwalk: online learning of social representations. In: Proceedings of SIGKDD, pp. 701–710 (2014)

    Google Scholar 

  12. Roweis, S.T., Saul, L.K.: Nonlinear dimensionality reduction by locally linear embedding. Science 290(5500), 2323–2326 (2000)

    Article  Google Scholar 

  13. Tang, J., Qu, M., Wang, M., Zhang, M., Yan, J., Mei, Q.: Line: large-scale information network embedding. In: Proceedings of WWW, pp. 1067–1077 (2015)

    Google Scholar 

  14. Tang, L., Liu, H.: Relational learning via latent social dimensions. In: ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, pp. 817–826, 28 June–July 2009

    Google Scholar 

  15. Tang, L., Liu, H.: Scalable learning of collective behavior based on sparse social dimensions. In: Proceeding of ACM Conference on Information and Knowledge Management (CIKM), pp. 1107–1116 (2009)

    Google Scholar 

  16. Tang, L., Liu, H.: Leveraging social media networks for classification. Data Min. Knowl. Disc. 23(3), 447–478 (2011)

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgment

The work described in this paper has been supported in part by the NSFC project (61572376), and the 111 project(B07037).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tieyun Qian .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Chen, Y., Qian, T., Zhong, M., Li, X. (2017). Exploit Label Embeddings for Enhancing Network Classification. In: Benslimane, D., Damiani, E., Grosky, W., Hameurlain, A., Sheth, A., Wagner, R. (eds) Database and Expert Systems Applications. DEXA 2017. Lecture Notes in Computer Science(), vol 10439. Springer, Cham. https://doi.org/10.1007/978-3-319-64471-4_36

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-64471-4_36

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-64470-7

  • Online ISBN: 978-3-319-64471-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics