Cross-Media Semantic Matching via Sparse Neural Network Pre-trained by Deep Restricted Boltzmann Machines

  • Conference paper
  • First Online:
Internet Multimedia Computing and Service (ICIMCS 2017)

Abstract

Cross-media retrieval arouses considerable attentions and becomes a more and more worthwhile research direction in the domain of information retrieval. Different from many related works which perform retrieval by map** heterogeneous data into a common representation subspace using a couple of projection matrices, we input multi-modal media data into a model of neural network which utilize a deep sparse neural network pre-trained by restricted Boltzmann machines and output their semantic understanding for semantic matching (RSNN-SM). Consequently, the heterogeneous modality data are represented by their top-level semantic outputs, and cross-media retrieval is performed by measuring their semantic similarities. Experimental results on several real-world datasets show that, RSNN-SM obtains the best performance and outperforms the state-of-the-art approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 107.00
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Zhao, M., Zhang, H., Meng, L.: An angle structure descriptor for image retrieval. China Commun. 13(8), 222–230 (2016)

    Article  Google Scholar 

  2. Zhao, M., Zhang, H., Sun, J.: A novel image retrieval method based on multi-trend structure descriptor. J. Vis. Commun. Image Represent. 38(c), 73–81 (2016)

    Article  Google Scholar 

  3. Hong, R., Wang, M., Li, G., Nie, L., Zha, Z.J., Chua, T.S.: Multimedia question answering. IEEE Multimed. 19(4), 72–78 (2012)

    Article  Google Scholar 

  4. Mitchell, T.M., Carbonell, J.G., Michalski, R.S.: Machine Learning. China Machine Press, Bei**g (2003)

    Google Scholar 

  5. Zhang, H., Ji, H., Wang, X.: Transfer learning from unlabeled data via neural networks, pp. 173–187 (2012)

    Google Scholar 

  6. Lee, H., Ekanadham, C., Ng, A.Y.: Sparse deep belief net model for visual area V2. In: International Conference on Neural Information Processing Systems, pp. 873–880 (2007)

    Google Scholar 

  7. Hinton, G.E., Osindero, S., Teh, Y.W.: A fast learning algorithm for deep belief nets. Neural Comput. 18(7), 1527–1554 (2014)

    Article  MathSciNet  MATH  Google Scholar 

  8. Sun, J., Liu, X., Wan, W., Li, J., Zhao, D., Zhang, H.: Video hashing based on appearance and attention features fusion via DBN. Neurocomputing 213, 84–94 (2016)

    Article  Google Scholar 

  9. Rasiwasia, N., Pereira, J.C., Coviello, E., Doyle, G., Lanckriet, G.R.G., Levy, R., Vasconcelos, N.: A new approach to cross-modal multimedia retrieval. In: International Conference on Multimedia, pp. 251–260 (2010)

    Google Scholar 

  10. Tenenbaum, J.B., Freeman, W.T.: Separating style and content with bilinear models. Neural Comput. 12(6), 1247–1283 (2014)

    Article  Google Scholar 

  11. Rosipal, R., Krämer, N.: Overview and recent advances in partial least squares. In: Saunders, C., Grobelnik, M., Gunn, S., Shawe-Taylor, J. (eds.) SLSFS 2005. LNCS, vol. 3940, pp. 34–51. Springer, Heidelberg (2006). https://doi.org/10.1007/11752790_2

    Chapter  Google Scholar 

  12. Gong, Y., Ke, Q., Isard, M., Lazebnik, S.: A multi-view embedding space for modeling internet images, tags, and their semantics. Int. J. Comput. Vis. 106(2), 210–233 (2014)

    Article  Google Scholar 

  13. Sharma, A., Kumar, A., Daume, H., Jacobs, D.W.: Generalized multiview analysis: a discriminative latent space. In: Computer Vision and Pattern Recognition, pp. 2160–2167 (2012)

    Google Scholar 

  14. Wang, K., Yin, Q., Wang, W., Wu, S., Wang, L.: A comprehensive survey on cross-modal retrieval (2016)

    Google Scholar 

  15. Wei, Y., Zhao, Y., Zhu, Z., Wei, S., **ao, Y., Feng, J., Yan, S.: Modality-dependent cross-media retrieval. ACM Trans. Intell. Syst. Technol. 7(4), 57 (2016)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Huaxiang Zhang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhang, B., Zhang, H., Sun, J., Wang, Z., Wu, H., Dong, X. (2018). Cross-Media Semantic Matching via Sparse Neural Network Pre-trained by Deep Restricted Boltzmann Machines. In: Huet, B., Nie, L., Hong, R. (eds) Internet Multimedia Computing and Service. ICIMCS 2017. Communications in Computer and Information Science, vol 819. Springer, Singapore. https://doi.org/10.1007/978-981-10-8530-7_27

Download citation

  • DOI: https://doi.org/10.1007/978-981-10-8530-7_27

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-10-8529-1

  • Online ISBN: 978-981-10-8530-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics

Navigation