Abstract
We consider the problem of inductive matrix completion, i.e., the reconstruction of a matrix using side features of its rows and columns. In numerous applications, however, side information of this kind includes redundant or uninformative features, so feature selection is required. An approach based on matrix factorization with group LASSO regularization on the coefficients of the side features is proposed, which combines feature selection with matrix completion. It is proved that the theoretical sample complexity for the proposed approach is lower than for methods without sparsifying. A computationally efficient iterative procedure for simultaneous matrix completion and feature selection is proposed. Experiments on synthetic and real-world data demonstrate that, due to the feature selection procedure, the proposed approach outperforms other methods.
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1134%2FS0965542521050079/MediaObjects/11470_2021_1499_Fig1_HTML.gif)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1134%2FS0965542521050079/MediaObjects/11470_2021_1499_Fig2_HTML.gif)
Similar content being viewed by others
REFERENCES
J. D. M. Rennie and N. Srebro, “Fast maximum margin matrix factorization for collaborative prediction,” Proceedings of the 22nd International Conference on Machine Learning (2005), pp. 713–719.
Y. Koren, R. Bell, and C. Volinsky, “Matrix factorization techniques for recommender systems,” Computer 42 (8), 30–37 (2009).
J. Yi, T. Yang, R. Jin, A. K. Jain, and M. Mahdavi, “Robust ensemble clustering by matrix completion,” 2012 IEEE 12th International Conference on Data Mining (ICDM) (2012), pp. 1176–1181.
A. Argyriou, T. Evgeniou, and M. Pontil, “Convex multi-task feature learning,” Mach. Learn. 73 (3), 243–272 (2008).
R. S. Cabral, F. Torre, J. P. Costeira, and A. Bernardino, “Matrix completion for multi-label image classification,” Advances in Neural Information Processing Systems (2011), pp. 190–198.
Z. Weng and X. Wang, “Low-rank matrix completion for array signal processing,” 2012 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP) (2012), pp. 2697–2700.
P. Chen and D. Suter, “Recovering the missing components in a large noisy low-rank matrix: Application to SFM,” IEEE Trans. Pattern Anal. Mach. Intell. 26 (8), 1051–1063 (2004).
E. J. Candès and B. Recht, “Exact matrix completion via convex optimization,” Found. Comput. Math. 9 (6), 717–772 (2009).
E. J. Candès and T. Tao, “The power of convex relaxation: Near-optimal matrix completion,” IEEE Trans. Inf. Theory 56 (5), 2053–2080 (2010).
O. Shamir and S. Shalev-Shwartz, “Matrix completion with the trace norm: Learning, bounding, and transducing,” J. Mach. Learn. Res. 15 (1), 3401–3423 (2014).
J. Hannon, M. Bennett, and B. Smyth, “Recommending twitter users to follow using content and collaborative filtering approaches,” Proceedings of the Fourth ACM Conference on Recommender Systems (2010), pp. 199–206.
M. Xu, R. Jin, and Z.-H. Zhou, “Speedup matrix completion with side information: Application to multi-label learning,” Advances in Neural Information Processing Systems (2013), pp. 2301–2309.
N. Natarajan and I. S. Dhillon, “Inductive matrix completion for predicting gene-disease associations,” Bioinformatics 30 (12), i60–i68 (2014).
K.-Y. Chiang, C.-J. Hsieh, and I. S. Dhillon, “Matrix completion with noisy side information,” Proceedings of the 28th International Conference on Neural Information Processing Systems (2015), Vol. 2, pp. 3447–3455.
S. Si, K.-Y. Chiang, C.-J. Hsieh, N. Rao, and I. S. Dhillon, “Goal-directed inductive matrix completion,” Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (2016), pp. 1165–1174.
J. Lu, G. Liang, J. Sun, and J. Bi, “A sparse interactive model for matrix completion with side information,” Advances in Neural Information Processing Systems 29 (2016), pp. 4071–4079.
Y. Guo, “Convex co-embedding for matrix completion with predictive side information,” Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17) (2017), pp. 1955–1961.
X. Zhang, S. Du, and Q. Gu, “Fast and sample efficient inductive matrix completion via multi-phase Procrustes flow,” Proceedings of the 35th International Conference on Machine Learning (2018), pp. 5756–5765.
A. Soni, T. Chevalier, and S. Jain, “Noisy inductive matrix completion under sparse factor models,” 2017 IEEE Intern. Symposium on Information Theory (ISIT) (2017), pp. 2990–2994.
P. Jain, P. Netrapalli, and S. Sanghavi, “Low-rank matrix completion using alternating minimization,” Proceedings of the 45 Annual ACM Symposium on Theory of Computing (2013), pp. 665–674.
Q. Berthet and N. Baldin, “Statistical and computational rates in graph logistic regression,” International Conference on Artificial Intelligence and Statistics (2020), pp. 2719–2730.
D. P. Bertsekas and J. N. Tsitsiklis, Parallel and Distributed Computation: Numerical Methods (Prentice Hall, Englewood Cliffs, 1989).
S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, “Distributed optimization and statistical learning via the alternating direction method of multipliers,” Found. Trends Mach. Learn. 3 (1), 1–122 (2011).
A. Maurer and M. Pontil, “Structured sparsity and generalization,” J. Mach. Learn. Res. 13, 671–690 (2012).
P. L. Bartlett and S. Mendelson, “Rademacher and Gaussian complexities: Risk bounds and structural results,” J. Mach. Learn. Res. 3, 463–482 (2002).
M. Mohri, A. Rostamizadeh, and A. Talwalkar, Foundations of Machine Learning (MIT Press, US, 2012).
R. Glowinski and A. Marroco, “Sur l’approximation, par éléments finis d’ordre un, et la résolution, par pénalisation: Dualité d’une classe de problémes de Dirichlet non linкires,” ESAIM: Math. Model. Numer. Anal. 9, 41–76 (1975).
D. Gabay and B. Mercier, “A dual algorithm for the solution of nonlinear variational problems via finite element approximation,” Comput. Math. Appl. 2 (1), 17–40 (1976).
D. Gabay, “Chapter 9: Applications of the method of multipliers to variational inequalities,” in Augmented Lagrangian Methods: Applications to the Numerical Solution of Boundary-Value Problems (North-Holland, Amsterdam, 1983), pp. 299–331.
J. Eckstein and D. P. Bertsekas, “On the Douglas–Rachford splitting method and the proximal point algorithm for maximal monotone operators,” Math. Progr. 55 (1), 293–318 (1992).
H.-F. Yu, P. Jain, P. Kar, and I. S. Dhillon, “Large-scale multi-label learning with missing labels,” Proceedings of the 31st International Conference on Machine Learning (2014), pp. 593–601.
C.-J. Lin, R. C. Weng, and S. S. Keerthi, “Trust region newton method for logistic regression,” J. Mach. Learn. Res. 9, 627–650 (2008).
N. Simon, J. Friedman, T. Hastie, and R. Tibshirani, “A sparse-group lasso,” J. Comput. Graphic. Stat. 22 (2), 231–245 (2013).
C.-C. Chang and C.-J. Lin, “LIBSVM: A library for support vector machines,” ACM Trans. Intell. Syst. Technol. 2 (3), 1–27 (2011).
M. R. Farhat, B. J. Shapiro, K. J. Kieser, et al., “Genomic analysis identifies targets of convergent positive selection in drug-resistant mycobacterium tuberculosis,” Nat. Genet. 45, 1183–1189 (2013).
T. M. Walker, T. A. Kohl, S. V. Omar, et al., “Whole-genome sequencing for prediction of mycobacterium tuberculosis drug susceptibility and resistance: A retrospective cohort study,” Lancet Infect. Dis. Appl. 15 (10), 1193–1202 (2015).
L. J. Pankhurst, C. Elias, A. A. Votintseva, et al., “Rapid, comprehensive, and affordable mycobacterial diagnosis with whole-genome sequencing: A prospective study,” Lancet Respir. Med. 4 (1), 49–58 (2016).
F. Coll, J. Phelan, G. A. Hill-Cawthorne, et al., “Genome-wide analysis of multi- and extensively drug-resistant mycobacterium tuberculosis,” Nat. Genet. 50 (2), 307–316 (2018).
Funding
This work was supported by the Russian Foundation for Basic Research, project no. 18-37-00489.
Author information
Authors and Affiliations
Corresponding authors
Additional information
Translated by I. Ruzanova
Rights and permissions
About this article
Cite this article
Burkina, M., Nazarov, I., Panov, M. et al. Inductive Matrix Completion with Feature Selection. Comput. Math. and Math. Phys. 61, 719–732 (2021). https://doi.org/10.1134/S0965542521050079
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1134/S0965542521050079