Skip to main content
Log in

Support matrix machine with pinball loss for classification

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Support vector machine (SVM) is one of the highly efficient classification algorithms. Unfortunately, it is designed only for input samples expressed as vectors. In real life, most input samples are naturally in matrix form and include structural information, such as electroencephalogram (EEG) signals and gray images. Support matrix machine (SMM), which can capture the latent structure within input matrices by regularizing the regression matrix to be low rank, is more suitable for matrix-form data than the SVM. However, the SMM adopts hinge loss, which is easily sensitive to noise and unstable to re-sampling. In this paper, to tackle this issue, we propose a new SMM with pinball loss (Pin−SMM), which can simultaneously consider the intrinsic structural information of input matrices and noise insensitivity. Our Pin−SMM is defined as a spectral elastic net with pinball loss, penalizing the rightly classified points. The optimization problem of Pin−SMM is also convex, which motivates us to construct the fast alternating direction method of multipliers (Fast ADMM) to solve it. Comprehensive experiments on two popular image datasets and an EEG dataset with different noises are conducted, and the experimental results confirm the effectiveness of our presented algorithm.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Notes

  1. For the detailed proof, please see Appendix A.

  2. http://pascal.inrialpes.fr/data/human/.

  3. http://www.vision.caltech.edu/Image_Datasets/Caltech101/.

  4. http://bbci.de/competition/iv/#dataset2a.

  5. https://sccn.ucsd.edu/eeglab/index.php.

References

  1. Vapnik V (1995) The nature of statistical learning theory, 267–290. Springer, New York

    Book  Google Scholar 

  2. Hua S, Sun Z (2001) Support vector machine approach for protein subcellular localization prediction. Bioinformatics 17:721–728

    Article  Google Scholar 

  3. Zhang W, Yoshida T, Tang X (2008) Text classification based on multi-word with support vector machine. Knowl-Based Syst 21(8):879–886

    Article  Google Scholar 

  4. Zhao Y, Huang G, Hu Q, Li B (2020) An improved weighted one class support vector machine for turboshaft engine fault detection. Eng Appl Artif Intell 94:103796

    Article  Google Scholar 

  5. Deng N, Tian Y, Zhang C (2012) Support vector machines: optimization based theory, algorithms, and extensions, 41–60. CRC Press, Philadelphia, PA

    Book  Google Scholar 

  6. Kotsia I, Patras I (2011) Support tucker machines. IEEE Conf Computer Vision Pattern Recogn (CVPR) 6:633–640

    Google Scholar 

  7. Wolf L, Jhuang H, Hazan T (2007) Modeling Appearances with Low-Rank SVM, in: IEEE Conference on Computer Vision and Pattern Recognition, pp 1−6

  8. Pirsiavash H, Ramanan D, Fowlkes C.C (2009) Bilinear classifiers for visual recognition, in: Proceedings of the Advances in Neural Information Processing Systems, pp 1482-1490

  9. You C, Palade V, Wu X (2019) Robust structure low-rank representation in latent space. Eng Appl Artif Intell 77:117–124

    Article  Google Scholar 

  10. Luo L, Xie Y, Zhang Z, Li W. J (2015) Support matrix machines, in: Proceedings of the 32nd International Conference on Machine Learning, pp 333-342

  11. Kobayashi T, Otsu N (2012) Efficient optimization for low-rank integrated bilinear classifiers, in: European Conference on Computer Vision, pp 474-487

  12. Zheng Q, Zhu F, Qin J, Chen B, Heng P (2018) Sparse support matrix machine. Pattern Recogn 76:715–726

    Article  Google Scholar 

  13. Zhu C, Wang Z (2017) Entropy-based matrix learning machine for imbalanced datasets. Pattern Recogn Lett 88:72–80

    Article  Google Scholar 

  14. Bi J, Zhang T (2005) Support vector classification with input data uncertainty, in: Proceedings of the Advances in Neural Information Processing Systems, pp 161-168

  15. Song Q, Hu W, Xie W (2002) Robust support vector machine with bullet hole image classification. IEEE Trans Syst, Man, Cybern Part C (Appl Rev) 32(4):440–448

    Article  Google Scholar 

  16. Shivaswamy PK, Bhattacharyya C, Smola AJ (2006) Second order cone programming approaches for handling missing and uncertain data. J Mach Learn Res 7(7):1283–1314

    MathSciNet  MATH  Google Scholar 

  17. Zhang J, Wang Y (2008) A rough margin based support vector machine. Inf Sci 178(9):2204–2214

    Article  MathSciNet  Google Scholar 

  18. Zhu W, Song Y, Xiao Y (2020) Support vector machine classifier with huberized pinball loss. Eng Appl Artif Intell 91:103635

    Article  Google Scholar 

  19. Huang X, Shi L, Suykens JAK (2014) Support vector machine classifier with pinball loss. IEEE Trans Pattern Anal Mach Intell 36(5):984–997

    Article  Google Scholar 

  20. Xu Y, Yang Z, Pan X (2017) A novel twin support vector machine with pinball loss. IEEE Trans Neural Netw Learn Syst 28(2):359–370

    Article  MathSciNet  Google Scholar 

  21. Shen X, Niu L, Qi Z, Tian Y (2017) Support vector machine classifier with truncated pinball loss. Pattern Recogn 68:199–210

    Article  Google Scholar 

  22. Tanveer M, Sharma A, Suganthan PN (2019) General twin support vector machine with pinball loss function. Inf Sci 494:311–327

    Article  MathSciNet  Google Scholar 

  23. Goldstein T, O’Donoghue B, Setzer S (2014) Fast alternating direction optimization methods. Siam J Imag Sci 7(3):1588–1623

    Article  MathSciNet  Google Scholar 

  24. Glowinski R, Marrocco A (1975) Sur lápproximation, par éléments finis dórdre un, et. la résolution, par pénalisation-dualité, dúne classe de problèmes de Dirichlet non linéaires, Journal of Equine Veterinary Science, 2 (R2), 41-76

  25. Cai JF, Candès EJ, Shen Z (2010) A singular value thresholding algorithm for matrix completion. SIAM J Optim 20(4):1956–1982

    Article  MathSciNet  Google Scholar 

  26. Watson GA (1991) Characterization of the subdifferential of some matrix norms. Linear Algebra Appl 170(6):33–45

    MathSciNet  MATH  Google Scholar 

  27. Koenker R, Hallock KF (2001) Quantile regression. J Econ Perspect 15(4):143–156

    Article  Google Scholar 

  28. Barat C, Ducottet C (2016) String representations and distances in deep convolutional neural networks for image classification. Pattern Recogn 54:104–115

    Article  Google Scholar 

  29. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection, in: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp 886-893

  30. Li F, Fergus R, Perona P (2006) One-shot learning of object categories. IEEE Trans Pattern Anal Mach Intell 28(4):561–594

    Google Scholar 

  31. Ang KK, Chin ZY, Wang C, Guan C, Zhang H (2012) Filter bank common spatial pattern algorithm on BCI competition IV datasets 2a and 2b. Front Neurosci 6:39

    Article  Google Scholar 

  32. Nam CS, Jeon Y, Kim YJ, Lee I, Park K (2011) Movement imagery-related lateralization of event-related (de) synchronization (ERD/ERS): motor-imagery duration effects. Clin Neurophysiol Off J Int Feder Clin Neurophysiol 122(3):567–577

    Article  Google Scholar 

  33. Basar E, Güntekin B (2013) Chapter 19-Review of delta, theta, alpha, beta, and gamma response oscillations in neuropsychiatric disorders, Supplements to Clinical Neurophysiology, Elsevier Health Sciences, 63, 303−341

  34. Lotte F (2014) A tutorial on eeg signal-processing techniques for mental-state recognition in brain-computer interfaces, in: Guide to Brain-Computer Music Interfacing, Springer, pp 133-161

  35. Pfurtscheller G, Neuper C, Flotzinger D, Pregenzer M (1997) EEG-based discrimination between imagination of right and left hand movement. Electroencephal Clin Neurophysiol 103(6):642

    Article  Google Scholar 

  36. Vidaurre C, Krämer N, Blankertz B, Schlögl A (2009) Time domain parameters as a feature for EEG-based brain-computer interfaces. Neural Netw 22(9):1313–1319

    Article  Google Scholar 

  37. Rieke F, Warland D, Steveninck RDRV, Bialek W (1999) Spikes: exploring the neural code, 15–25. MIT Press, Cambridge, Mass

    MATH  Google Scholar 

  38. Crone N.E, Miglioretti D.L, Gordon B, Lesser R.P (1998) Functional mapping of human sensorimotor cortex with electrocorticographic spectral analysis. II. Event-related synchronization in the gamma band, Brain 121 (12), 2301-2315

  39. Crone NE, Miglioretti DL, Gordon B, Sieracki JM, Wilson MT, Uematsu S, Lesser RP (1998) Functional mapping of human sensorimotor cortex with electrocorticographic spectral analysis, I Alpha and beta event-related desynchronization. Brain 121(12):2271–2299

    Article  Google Scholar 

  40. Miller K, Leuthardt E, Schalk G, Rao R, Anderson N, Moran D, Miller J, Ojemann J (2007) Spectral changes in cortical surface potentials during motor movement. J Neurosci 27(9):2424–2432

    Article  Google Scholar 

  41. Demsšar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res 7:1–30

    MathSciNet  Google Scholar 

  42. Salvador G, Alberto F, Julián L, Herrera F (2010) Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: experimental analysis of power. Inf Sci 180(10):2044–2064

    Article  Google Scholar 

  43. Fazel M, Hindi H, Boyd S. P (2001) A rank minimization heuristic with application to minimum order system approximation, Proceedings of the 2001 American Control Conference, 6, 4734−4739

Download references

Acknowledgements

The authors gratefully acknowledge the helpful comments and suggestions of the reviewers, which have improved the presentation. This work was supported in part by the National Natural Science Foundation of China (No. 12071475, 11671010) and Beijing Natural Science Foundation (No.4172035).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yitian Xu.

Ethics declarations

Conflict of interest

The authors declared that they have no conflicts of interest to this work.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix A. Convexity of Pin−SMM

Appendix A. Convexity of Pin−SMM

In order to prove that problem (9) is convex, we first define

$$\begin{aligned}&\varPhi _{1}({\mathbf {W}}) = \frac{1}{2}{\text {tr}}({\mathbf {W}}^{T}{\mathbf {W}}), \nonumber \\&\varPhi _{2}({\mathbf {W}}) = \Vert {\mathbf {W}}\Vert _{*}\nonumber , \\&\varPhi _{3}({\mathbf {W}}, b) = \sum _{i=1}^{n}L_{\tau } ({\mathbf {X}}_{i},y_{i},f({\mathbf {X}}_{i})). \nonumber \end{aligned}$$

It is obvious that the function \(\varPhi _{1}({\mathbf {W}})\) is convex. The nuclear norm of a matrix is a convex alternative of the matrix rank [10], which has been proved in Appendix A of the literature [43]. Therefore, \(\varPhi _{2}({\mathbf {W}})\) is also convex. Next, we demonstrate that the \(\varPhi _{3}({\mathbf {W}}, b)\) is convex. The sum of the convex functions is a convex function. We only need to prove that the function \(L_{\tau }({\mathbf {X}}_{i},y_{i},f({\mathbf {X}}_{i}))\) is convex. Let \(y_{i}f({\mathbf {X}}_{i}) - 1 = \mu\), and the pinball loss can be rewritten as:

$$\begin{aligned} L_{\tau }(\mu )=\left\{ \begin{aligned}&-\mu ,&\mu < 0, \\&\tau \mu ,&\mu \ge 0. \end{aligned}\right. \end{aligned}$$
(30)

Then we have

$$\begin{aligned} L_{\tau }(\mu )= \tau \max (0, \mu ) + \max (0, -\mu ), \end{aligned}$$
(31)

where \(\tau > 0\). Thus the function \(L_{\tau }(\mu )\) is convex. Because \(f({\mathbf {X}}_{i}) = {\text {tr}}({\mathbf {W}}^{T}{\mathbf {X}}_{i})+b\), we have \(L_{\tau }({\mathbf {X}}_{i},y_{i},f({\mathbf {X}}_{i}))\) is convex. So \(\varPhi _{3}({\mathbf {W}}, b)\) is convex with respect to \({\mathbf {W}}\) and b, respectively.

Moreover, parameters \(\lambda\) and C are positive; therefore, \(\varPhi _{1}({\mathbf {W}}) + \lambda \varPhi _{2}({\mathbf {W}}) + C\varPhi _{3}({\mathbf {W}}, b)\) is convex. That is, the Pin−SMM is convex.

This completes the proof. \(\Box\).

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Feng, R., Xu, Y. Support matrix machine with pinball loss for classification. Neural Comput & Applic 34, 18643–18661 (2022). https://doi.org/10.1007/s00521-022-07460-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-022-07460-6

Keywords

Navigation