Skip to main content
Log in

Efficient multi-view clustering networks

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

In the last decade, deep learning has made remarkable progress on multi-view clustering (MvC), with existing literature adopting a broad target to guide the network learning process, such as minimizing the reconstruction loss. However, despite this strategy being effective, it lacks efficiency. Hence, in this paper, we proposed a novel framework, entitled Efficient Multi-view Clustering Networks (EMC-Nets), which guarantees the network’s learning efficiency and produces a common discriminative representation from multiple sources. Specifically, we developed an alternating process, involving an approximation and an instruction process, which effectively stimulate the process of multi-view feature fusion to force network to learn a discriminative common representation. The approximation process employs a standard clustering algorithm, i.e., k-means, to generate pseudo labels corresponding to the current common representation, and then it leverages the pseudo labels to force the network to approximate a reasonable cluster distribution. Considering the instruction process, it aims to provide a correct learning direction for the approximation process and prevent the network from obtaining trivial solutions. Experiment results on four real-world datasets demonstrate that the proposed method outperforms state-of-the-art methods. Our source code will be available soon at https://github.com/Guanzhou-Ke/EMC-Nets.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

References

  1. Lowe DG (2004) Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60(2):91–110

    Article  Google Scholar 

  2. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. CVPR 1:886–893

    Google Scholar 

  3. Ojala T, Pietikainen M, Maenpaa T (2002) Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE TPAMI 24(7):971–987

    Article  Google Scholar 

  4. Yang Y, Wang H (2018) Multi-view clustering: A survey. Big Data Mining and Analytics 1(2):83–107

    Article  Google Scholar 

  5. Liu J et al (2013) Multi-view clustering via joint nonnegative matrix factorization, In: Proceedings of the 2013 SIAM international conference on data mining, society for industrial and applied mathematics

  6. Zhang Z et al (2018) Binary multi-view clustering. IEEE Transactions on Pattern Analysis and Machine Intelligence 41(7):1774–1782

    Article  Google Scholar 

  7. Peng X, Huang Z, Lv J, Zhu H, Zhou JT (2019) COMIC: Multi-view clustering without parameter selection. International Conference on Machine Learning 97(ICML’19):5092–5101

  8. Chaudhuri K et al (2009) Multi-view clustering via canonical correlation analysis. In: Proceedings of the 26th annual international conference on machine learning, ACM, pp 129–136

  9. Melzer T, Reiter M, Bischof H (2001) Nonlinear feature extraction using generalized canonical correlation analysis. In: Proc. intern. conf. artificial neural networks (ICANN2001), pp 353–360

  10. Galen A, Raman A, Jeff B, Livescu K (2013) Deep canonical correlation analysis. In: International conference on machine learning, pp 1247–1255

  11. Zhao H, Zhengming D, Yun F (2017) Multi-view clustering via deep matrix factorization. In: Proceedings of the AAAI conference on artificial intelligence, vol 31, No 1

  12. Abavisani M, Patel VM (2018) Deep multimodal subspace clustering networks. IEEE Journal ofSelected Topics in Signal Processing 12(6):1601–1614

    Article  Google Scholar 

  13. Zhang C, Liu Y, Fu H (2019) AE\(^2\)-Nets: Autoencoder in autoencoder networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2577–2585

  14. Zhang Y et al (2019) A multitask multiview clustering algorithm in heterogeneous situations based on LLE and LE. Knowledge-Based Systems 163:776–786

    Article  Google Scholar 

  15. Zhaoyang L, Wang Q, Tao Z, Gao Q, Yang Z (2019) Deep Adversarial multi-view clustering network. In: IJCAI, pp 2952–2958

  16. Zhou R, Shen Y-D(2020) End-to-End adversarial-attention network for multi-modal clustering. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 14619–14628

  17. Goodfellow IJ et al (2014) Generative adversarial nets. In: Proceedings of the 27th international conference on neural information processing systems - volume 2 (NIPS’14). MIT Press, Cambridge, MA, USA, pp 2672–2680

  18. Zhan X, Xie J, Liu Z, Ong Y-S, Loy CC (2020) Online deep clustering for unsupervised representation learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 6688–6697

  19. Caron M, Piotr B, Joulin A, Douze M (2018) Deep clustering for unsupervised learning of visual features. In: Proceedings of the european conference on computer vision (ECCV), pp 132–149

  20. Vaswani A et al (2017) Attention is all you need. In: NIPS

  21. Jing L, Tian Y (2020) Self-supervised visual feature learning with deep neural networks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence

  22. Tu W et al (2021) Deep fusion clustering network. In: AAAI

  23. Xu LL, Neufeld J, Larson B, Schuurmans D (2005) Maximum margin clustering. In: Proc. 18th annu. conf. advances in neural information processing systems 17, Cambridge, MA, USA, pp 1537–1544

  24. Du L, Zhou P, Shi L, Wang HM, Fan MY, Wang WJ, Shen YD (2015) Robust multiple kernel k-means using ’2;1-norm. In: Proc. 24th int. conf. artificial intelligence, Buenos Aires, Argentina, pp 3476–3482

  25. Chaudhuri K, Kakade SM, Livescu K, Sridharan K (2009) Multi-view clustering via canonical correlation analysis. In: Proc. 26th annu. int. conf. machine learning, Montreal, Canada, pp 129–136

  26. Zhao XR, Evans N, Dugelay JL (2014) A subspace co-training framework for multi-view clustering. Pattern Recogn. Lett. 41:73–82

    Article  Google Scholar 

  27. Guo YH (2013) Convex subspace representation learning from multi-view data. In: Proc. 27th AAAI conf. artificial intelligence, Bellevue, WA, USA, pp 387–393

  28. Xue Z, Li GR, Wang SH, Zhang CJ, Zhang WG, Huang QM (2015) (2015) GOMES: A group-aware multi-view fusion approach towards real-world image clustering. In: Proc. IEEE int. conf. multimedia and expo, Turin, Italy, pp 1–6

  29. Wang W et al (2015) On deep multi-view representation learning. In: International conference on machine learning, PMLR, pp 1083–1092

  30. Bottou L (2012) Stochastic gradient descent tricks. In: Neural networks: Tricks of the trade, Springer, pp 421–436

  31. Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313(5786):504–507

    Article  MathSciNet  Google Scholar 

  32. Asuncion A, Newman D (2007) Uci machine learning repository

  33. Shang C, Palmer A, Sun J, Chen K-S, Lu J, Bi J (2017) VIGAN: missing view imputation with generative adversarial networks. arxiv:1708.06724

  34. Cai X, Wang H, Huang H, Ding C (2012) Joint stage recognition and anatomical annotation of drosophila gene expression patterns. Bioinformatics 28(12):i16–i24

    Article  Google Scholar 

  35. Jiang Y-G, Ye G, Chang S-F, Ellis D, Loui AC (2011) Consumer video understanding: A benchmark database and an evaluation of human and machine performance. In: ICMR, pp 29

  36. Kumar A, Rai P, Daume H (2011) Co-regularized multi-view spectral clustering. In: NIPS, pp 1413–1421

  37. Ng AY, Jordan MI, Weiss Y (2001) On spectral clustering: Analysis and an algorithm. In: Proceedings of the 14th international conference on neural information processing systems: natural and synthetic

  38. Cai X, Nie F, Huang H (2013) Multi-view k-means clustering on big data. In: Twenty-Third international Joint conference on artificial intelligence

  39. Xia R, Pan Y, Du L, Yin J (2014) Robust multi-view spectral clustering via low-rank and sparse decomposition. In: AAAI, pp 2149–2155

  40. Zhan K, Nie F, Wang J, Yang Y (2019) Multiview consensus graph clustering. In: A publication of the IEEE signal processing society, IEEE transactions on image processing

  41. Nie F, Li J, Li X et al (2016) Parameter-free auto-weighted multiple graph learning: A framework for multiview clustering and semi-supervised classification. In: IJCAI, pp 1881–1887

  42. Glorot X, Bordes A, Bengio Y (2011) Deep sparse rectifier neural networks. In: Proceedings of the fourteenth international conference on artificial intelligence and statistics, JMLR Workshop and Conference Proceedings, pp 315–323

  43. Glorot X, Bengio Y (2010) Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics, JMLR workshop and conference proceedings, pp 249–256

  44. Hinton GE, Nitish S, Krizhevsky A, Sutskever I, Salakhutdinov RR (2012) Improving neural networks by preventing co-adaptation of feature detectors. arXiv:1207.0580

  45. Ketchen DJ, Shook CL (1996) The application of cluster analysis in strategic management research: an analysis and critique. Strategic Management Journal 17(6):441–458

    Article  Google Scholar 

  46. van der Maaten L, Hinton G (2008) Visualizing data using t-SNE. Journal of Machine Learning Research 9:2579–2605

    MATH  Google Scholar 

  47. Dempster AP, Laird NM, Rubin DB (1977) Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society: Series B (Methodological) 39(1):1–22

    MathSciNet  MATH  Google Scholar 

  48. Kuhn HW (1955) The Hungarian method for the assignment problem. Naval research logistics quarterly 2(1–2):83–97

Download references

Acknowledgements

This work was supported by Natural Science Foundation of Guangdong Province, China (No. 2020A1515011468; 2016A0303010003); Guangdong University Scientific Research Project, China (No. 2019KTSCX189; 2018KTSCX235; 2018WZDXM014); Joint Research and Development Fund of Wuyi University and Hong Kong and Macau (2019WGALH21).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhiyong Hong.

Additional information

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix: Network architectures

Appendix: Network architectures

In this section, we report the details of the network architecture, includes network layer, input and output size, hyper-parameters of the layer, and the total number of parameters, used in the experiments. Note that all the layer of network are used Pytorch-style API. For the fully-connected (linear) layer, we use bias as default, and we report the head and the dimension of feedforward (dff) of the TransformerEncoder (Tables 7, 8, 9, 10).

Table 7 EMC-Nets in the BDGP experiments
Table 8 EMC-Nets in the HW experiments
Table 9 EMC-Nets in the CCV experiments
Table 10 EMC-Nets in the MNIST experiments

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ke, G., Hong, Z., Yu, W. et al. Efficient multi-view clustering networks. Appl Intell 52, 14918–14934 (2022). https://doi.org/10.1007/s10489-021-03129-0

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-021-03129-0

Keywords

Navigation