Improving empirical efficiency of CUR decomposition | The Journal of Supercomputing Skip to main content
Log in

Improving empirical efficiency of CUR decomposition

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

A technique that has recently become popular in data analysis and machine learning is the CUR decomposition of a large matrix. Existing approaches for CUR decomposition are usually slow, so that they are not applicable to large real-world matrices. In this paper, we propose a method to efficiently compute a CUR decomposition of a matrix. In our method, we use the fast CUR method as a basis and improve its runtime by estimating approximate Frobenius norms, rather than computing their exact values. We compare our method against several state of the art algorithms for CUR decomposition and show that it outperforms the other methods, in terms of running time.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Japan)

Instant access to the full article PDF.

Fig. 1

Similar content being viewed by others

Availability of data and materials

The used data are publicly available and their links are provided in Sect. 4.2.

Notes

  1. https://www.worldometers.info/world-population/population-by-country/.

  2. https://www.imageprocessingplace.com/root_files_V3/image_databases.htm.

  3. https://www.imageprocessingplace.com/root_files_V3/image_databases.htm.

  4. https://archive.ics.uci.edu/ml/datasets/Malware+static+and+dynamic+features+VxHeaven+and+Virus+Total.

  5. https://archive.ics.uci.edu/ml/datasets/cnae-9.

References

  1. Worldometers 09 June 2020. Countries in the world by population, 2020

  2. Akoglu L, Tong H, Koutra D (2015) Graph-based anomaly detection and description: a survey. Data Min Knowl Discov 29:626–688

    Article  MathSciNet  Google Scholar 

  3. Boutsidis C, Woodruff David P (2017) Optimal CUR matrix decompositions. SIAM J Comput 46(2):543–589

    Article  MathSciNet  MATH  Google Scholar 

  4. Cai H, Hamm K, Huang L, Li J, Wang T (2021) Rapid robust principal component analysis: CUR accelerated inexact low rank estimation. IEEE Signal Process Lett 28:116–120

    Article  Google Scholar 

  5. Cai H, Hamm K, Huang L, Needell D (2021) Robust CUR decomposition: theory and imaging applications. SIAM J Imag Sci 14(4):1472–1503

    Article  MathSciNet  MATH  Google Scholar 

  6. Candès E, Recht B (2012) Exact matrix completion via convex optimization. Commun ACM 55(6):111–119

  7. Chiu J, Demanet L (2013) Sublinear randomized algorithms for skeleton decompositions. SIAM J Matrix Anal Appl 34(3):1361–1383

    Article  MathSciNet  MATH  Google Scholar 

  8. Drineas P, Kannan R, Mahoney MW (2006) Fast Monte Carlo algorithms for matrices III: computing a compressed approximate matrix decomposition. SIAM J Comput 36(1):184–206

    Article  MathSciNet  MATH  Google Scholar 

  9. Dua D, Graff C (2017) UCI machine learning repository

  10. Goreinov SA, Tyrtyshnikov EE, Zamarashkin NL (1997) A theory of pseudoskeleton approximations. Linear Algebr Appl 261(1):1–21

    Article  MathSciNet  MATH  Google Scholar 

  11. Guennebaud G, Jacob B et al (2020) Eigen v 3.3.9. http://eigen.tuxfamily.org,

  12. Hamm Keaton, Huang Longxiu (2020) Perspectives on CUR decompositions. Appl Comput Harmonic Anal 48(3):1088–1099

    Article  MathSciNet  MATH  Google Scholar 

  13. Hamm K, Huang L (2020) Stability of sampling for CUR decompositions. Found Data Sci 2(2):83–99

    Article  Google Scholar 

  14. Keshavan RH, Montanari A, Sewoong Oh (2010) Matrix completion from a few entries. IEEE Trans Inf Theor 56(6):2980–2998

    Article  MathSciNet  MATH  Google Scholar 

  15. Khetan A, Oh S (2017) Matrix norm estimation from a few entries. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pages 6424–6433

  16. Klema V, Laub A (1980) The singular value decomposition: its computation and some applications. IEEE Trans Autom Control 25(2):164–176

    Article  MathSciNet  MATH  Google Scholar 

  17. Li Yi, Woodruff David P(2020) Input-sparsity low rank approximation in schatten norm. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, pages 6001–6009

  18. Li Yi, Woodruff David P (2022) Tight bounds for sketching the operator norm, schatten norms, and subspace embeddings. CoRR, abs/2202.09797

  19. Mahoney Michael W, Drineas Petros (2009) CUR matrix decompositions for improved data analysis. Proc Nat Acad Sci 106(3):697–702

    Article  MathSciNet  MATH  Google Scholar 

  20. Mitrovic N, Asif Muhammad T, Rasheed U, Dauwels J, Jaillet P (2013) CUR decomposition for compression and compressed sensing of large-scale traffic data. In 16th International IEEE Conference on Intelligent Transportation Systems, ITSC 2013, The Hague, The Netherlands, October 6-9, 2013, pages 1475–1480. IEEE

  21. Song Z, Woodruff David P, Zhong P (2017) Low rank approximation with entrywise l\({}_{\text{1}}\)-norm error. In STOC, pages 688–701. ACM

  22. Song Z, Woodruff DP, Zhong P (2019) Relative error tensor low rank approximation. In SODA, pages 2772–2789. SIAM

  23. Sun J, Xie Y, Zhang H, Faloutsos C (2008) Less is more: sparse graph mining with compact matrix decomposition. Stat Anal Data Min 1(1):6–22

    Article  MathSciNet  MATH  Google Scholar 

  24. Voronin S, Martinsson PG (2017) Efficient algorithms for CUR and interpolative matrix decompositions. Adv Comput Math 43(3):495–516

    Article  MathSciNet  MATH  Google Scholar 

  25. Wang S, Zhang Z (2013) Improving CUR matrix decomposition and the nyström approximation via adaptive sampling. J Mach Learn Res 14(1):2729–2769

    MathSciNet  MATH  Google Scholar 

  26. Williams Christopher KI, Seeger M (2001) Using the nyström method to speed up kernel machines. In T.K. Leen, T.G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13 (NIPS 2000), pages 682–688. MIT Press

Download references

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

MHC: developed ideas, analyzed results, wrote paper, supervised work. ZY: implemented algorithms, ran experiments.

Corresponding author

Correspondence to Mostafa Haghir Chehreghani.

Ethics declarations

Conflict of interest

The authors declare that they have no competing interests as defined by Springer, or other interests that might be perceived to influence the results and/or discussion reported in this paper.

Ethical approval

Not applicable.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Haghir Chehreghani, M., Yaghoobi, Z. Improving empirical efficiency of CUR decomposition. J Supercomput 79, 9350–9366 (2023). https://doi.org/10.1007/s11227-022-05039-5

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-022-05039-5

Keywords

Navigation