[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/1150402.1150440acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
Article

Tensor-CUR decompositions for tensor-based data

Published: 20 August 2006 Publication History

Abstract

Motivated by numerous applications in which the data may be modeled by a variable subscripted by three or more indices, we develop a tensor-based extension of the matrix CUR decomposition. The tensor-CUR decomposition is most relevant as a data analysis tool when the data consist of one mode that is qualitatively different than the others. In this case, the tensor-CUR decomposition approximately expresses the original data tensor in terms of a basis consisting of underlying subtensors that are actual data elements and thus that have natural interpretation in terms ofthe processes generating the data. In order to demonstrate the general applicability of this tensor decomposition, we apply it to problems in two diverse domains of data analysis: hyperspectral medical image analysis and consumer recommendation system analysis. In the hyperspectral data application, the tensor-CUR decomposition is used to compress the data, and we show that classification quality is not substantially reduced even after substantial data compression. In the recommendation system application, the tensor-CUR decomposition is used to reconstruct missing entries in a user-product-product preference tensor, and we show that high quality recommendations can be made on the basis of a small number of basis users and a small number of product-product comparisons from a new user.

References

[1]
G. Adomavicius and A. Tuzhilin. Toward the next generation of recommender systems: A survey of the state-of-the-art and possible extensions. IEEE Transactions on Knowledge and Data Engineering, 17:734--749, 2005.
[2]
Y. Azar, A. Fiat, A. R. Karlin, F. McSherry, and J. Saia. Spectral analysis of data. In Proceedings of the 33rd Annual ACM Symposium on Theory of Computing, pages 619--626, 2001.
[3]
M. W. Berry, S. A. Pulatova, and G. W. Stewart. Computing sparse reduced-rank approximations to sparse matrices. Technical Report UMIACS TR-2004-32 CMSC TR-4589, University of Maryland, College Park, MD, 2004.
[4]
D. Billsus and M. J. Pazzani. Learning collaborative information filters. In Proceedings of the 15th International Conference on Machine Learning, pages 46--54, 1998.
[5]
J. Breese, D. Heckerman, and C. Kadie. Empirical analysis of predictive algorithms for collaborative filtering. In Proceedings of the 14th Annual Conference on Uncertainty in Artificial Intelligence, pages 43--52, 1998.
[6]
J. D. Carroll and J. J. Chang. Analysis of individual differences in multidimensional scaling via an N-way generalization of "Eckart- Young" decomposition. Psychometrika, 35(3):283--319, 1970.
[7]
W. W. Cohen, R. E. Schapire, and Y. Singer. Learning to order things. In Annual Advances in Neural Information Processing Systems 10: Proceedings of the 1997 Conference, pages 451--457, 1998.
[8]
P. Drineas, R. Kannan, and M. W. Mahoney. Fast Monte Carlo algorithms for matrices I: Approximating matrix multiplication. To appear in: SIAM Journal on Computing.
[9]
P. Drineas, R. Kannan, and M. W. Mahoney. Fast Monte Carlo algorithms for matrices II: Computing a low-rank approximation to a matrix. To appear in: SIAM Journal on Computing.
[10]
P. Drineas, R. Kannan, and M. W. Mahoney. Fast Monte Carlo algorithms for matrices III: Computing a compressed approximate matrix decomposition. To appear in: SIAM Journal on Computing.
[11]
P. Drineas, I. Kerenidis, and P. Raghavan. Competitive recommendation systems. In Proceedings of the 34th Annual ACM Symposium on Theory of Computing, pages 82--90, 2002.
[12]
P. Drineas and M. W. Mahoney. A randomized algorithm for a tensor-based generalization of the Singular Value Decomposition. To appear in: Linear Algebra and its Applications.
[13]
P. Drineas and M. W. Mahoney. On the Nyström method for approximating a Gram matrix for improved kernel-based learning. Journal of Machine Learning Research, 6:2153--2175, 2005.
[14]
Y. Freund, R. Iyer, R. E. Schapire, and Y Singer. An efficient boosting algorithm for combining preferences. Journal of Machine Learning Research, 4:933--969, 2003.
[15]
K. Goldberg, T. Roeder, D. Gupta, and C. Perkins. Eigentaste: A constant time collaborative filtering algorithm. Information Retrieval, 4:133--151, 2001.
[16]
G. H. Golub and C. F. Van Loan. Matrix Computations. Johns Hopkins University Press, Baltimore, 1989.
[17]
S. A. Goreinov and E. E. Tyrtyshnikov. The maximum-volume concept in approximation by low-rank matrices. Contemporary Mathematics, 280:47--51, 2001.
[18]
S. A. Goreinov, E. E. Tyrtyshnikov, and N. L. Zamarashkin. A theory of pseudoskeleton approximations. Linear Algebra and Its Applications, 261:1--21, 1997.
[19]
R. A. Harshman and M. E. Lundy. The PARAFAC model for three-way factor analysis and multidimensional scaling. In H. G. Law, C. W. Snyder Jr., J. Hattie, and R. P. McDonald, editors, Research Methods for Multimode Data Analysis, pages 122--215. Praeger, 1984.
[20]
J. Håstad Tensor rank is NP-complete. Journal of Algorithms, 11(2):644--654, 1990.
[21]
R. Jin, L. Si, and C. X. Zhai. Preference-based graphic models for collaborative filtering. In Proceedings of the 19th Conference in Uncertainty in Artificial Intelligence, pages 329--336, 2003.
[22]
R. Jin, L. Si, C. X. Zhai, and J. Callan. Collaborative filtering with decoupled models for preferences and ratings. In Proceedings of the 12th International Conference on Information and Knowledge Management, pages 309--316, 2003.
[23]
J. Kleinberg and M. Sandler. Using mixture models for collaborative filtering. In Proceedings of the 36th Annual ACM Symposium on Theory of Computing, pages 569--578, 2004.
[24]
P. M. Kroonenberg and J. De Leeuw. Principal component analysis of three-mode data by means of alternating least squares algorithms. Psychometrika, 45(1):69--97, 1980.
[25]
J. B. Kruskal. Three-way arrays: Rank and uniqueness of trilinear decompositions, with application to arithmetic complexity and statistics. Linear Algebra and Its Applications, 18:95--138, 1977.
[26]
J. B. Kruskal. Rank, decomposition, and uniqueness for 3-way and N-way arrays. In R. Coppi and S. Bolasco, editors, Multiway Data Analysis, pages 7--18. Elesvier Science Publishers, 1989.
[27]
R. Kumar, P. Raghavan, S. Rajagopalan, and A. Tomkins. Recommendation systems: a probabilistic analysis. In Proceedings of the 39th Annual IEEE Symposium on Foundations of Computer Science, pages 664--673, 1998.
[28]
L. De Lathauwer, B. De Moor, and J. Vandewalle. A multilinear singular value decomposition. SIAM Journal on Matrix Analysis and Applications, 21(4):1253--1278, 2000.
[29]
L. De Lathauwer, B. De Moor, and J. Vandewalle. On the best rank-1 and rank-(R1, R2, .., RN)approximation of higher-order tensors. SIAM Journal on Matrix Analysis and Applications, 21(4):1324--1342, 2000.
[30]
D. Leibovici and R. Sabatier. A singular value decomposition of a k-way array for a principal component analysis of multiway data, PTA-k. Linear Algebra and Its Applications, 269:307--329, 1998.
[31]
L.-H. Lim and G. H. Golub. Tensors for numerical multilinear algebra: ranks and basic decompositions. Technical Report 05-02, Stanford University SCCM, Stanford, CA, April 2005.
[32]
G. Linden, B. Smith, and J. York. Amazon.com recommendations: Item-to-item collaborative filtering. IEEE Internet Computing, 7:76--80, 2003.
[33]
M. Maggioni, G. L. Davis, F. J. Warner, F. B. Geshwind, A. C. Coppi, R. A. DeVerse,and R. R. Coifman. Algorithms from signal and data processing applied to hyperspectral analysis: Discriminating normal and malignant microarray colon tissue sections using a novel digital mirror device system. Manuscript, 2006.
[34]
M. Maggioni, G. L. Davis, F. J. Warner, F. B. Geshwind, A. C. Coppi, R. A. DeVerse,and R. R. Coifman. Hyperspectral microscopic analysis of normal, benign and carcinoma microarray tissue sections. Manuscript, 2006.
[35]
P. Resnick and H. R. Varian. Recommender systems. Communications of the ACM, 40(3):56--58, 1997.
[36]
B. Sarwar, G. Karypis, J. Konstan, and J. Riedl. Application of dimensionality reduction in recommender system - a case study. In Proceedings of the WebKDD 2000 Workshop, pages 000--000, 2000.
[37]
G. W. Stewart. Four algorithms for the efficient computation of truncated QR approximations to a sparse matrix. Numerische Mathematik, 83:313--323, 1999.
[38]
G. W. Stewart. Error analysis of the quasi-Gram-Schmidt algorithm. Technical Report UMIACS TR-2004-17 CMSC TR-4572, University of Maryland, College Park, MD, 2004.
[39]
L. R. Tucker. Some mathematical notes on three-mode factor analysis. Psychometrika, 31(3):279--311, 1966.
[40]
C. K. I. Williams and M. Seeger. Using the Nyström method to speed up kernel machines. In Annual Advances in Neural Information Processing Systems 13: Proceedings of the 2000 Conference, pages 682--688, 2001.

Cited By

View all
  • (2024)Maximal volume matrix cross approximation for image compression and least squares solutionAdvances in Computational Mathematics10.1007/s10444-024-10196-750:5Online publication date: 16-Sep-2024
  • (2023)Tensor Robust CUR for Compression and Denoising of Hyperspectral DataIEEE Access10.1109/ACCESS.2023.329763011(77492-77505)Online publication date: 2023
  • (2023)One-Pass Additive-Error Subset Selection for $$\ell _{p}$$ Subspace Approximation and (k, p)-ClusteringAlgorithmica10.1007/s00453-023-01124-085:10(3144-3167)Online publication date: 11-May-2023
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '06: Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining
August 2006
986 pages
ISBN:1595933395
DOI:10.1145/1150402
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 20 August 2006

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. CUR decomposition
  2. hyperspectral image analysis
  3. recommendation system analysis
  4. tensor CUR

Qualifiers

  • Article

Conference

KDD06

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Upcoming Conference

KDD '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)50
  • Downloads (Last 6 weeks)8
Reflects downloads up to 06 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Maximal volume matrix cross approximation for image compression and least squares solutionAdvances in Computational Mathematics10.1007/s10444-024-10196-750:5Online publication date: 16-Sep-2024
  • (2023)Tensor Robust CUR for Compression and Denoising of Hyperspectral DataIEEE Access10.1109/ACCESS.2023.329763011(77492-77505)Online publication date: 2023
  • (2023)One-Pass Additive-Error Subset Selection for $$\ell _{p}$$ Subspace Approximation and (k, p)-ClusteringAlgorithmica10.1007/s00453-023-01124-085:10(3144-3167)Online publication date: 11-May-2023
  • (2022)Error analysis of tensor-train cross approximationProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3601305(14236-14249)Online publication date: 28-Nov-2022
  • (2020)Fast deterministic CUR matrix decomposition with accuracy assuranceProceedings of the 37th International Conference on Machine Learning10.5555/3524938.3525365(4594-4603)Online publication date: 13-Jul-2020
  • (2019)SingleshotProceedings of the 33rd International Conference on Neural Information Processing Systems10.5555/3454287.3454853(6304-6315)Online publication date: 8-Dec-2019
  • (2018)Semantic Representation of Documents Based on Matrix Decomposition2018 International Conference on Data Science and Engineering (ICDSE)10.1109/ICDSE.2018.8527824(1-6)Online publication date: Aug-2018
  • (2016)Decomposition-by-normalization (DBN)Data Mining and Knowledge Discovery10.1007/s10618-015-0401-630:1(1-46)Online publication date: 1-Jan-2016
  • (2016)Turbo-SMTStatistical Analysis and Data Mining10.1002/sam.113159:4(269-290)Online publication date: 1-Aug-2016
  • (2015)ParCubeACM Transactions on Knowledge Discovery from Data10.1145/272998010:1(1-25)Online publication date: 22-Jul-2015
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media