all AI news
Efficient Algorithms for Regularized Nonnegative Scale-invariant Low-rank Approximation Models
March 28, 2024, 4:41 a.m. | Jeremy E. Cohen, Valentin Leplat
cs.LG updates on arXiv.org arxiv.org
Abstract: Regularized nonnegative low-rank approximations such as sparse Nonnegative Matrix Factorization or sparse Nonnegative Tucker Decomposition are an important branch of dimensionality reduction models with enhanced interpretability. However, from a practical perspective, the choice of regularizers and regularization coefficients, as well as the design of efficient algorithms, is challenging because of the multifactor nature of these models and the lack of theory to back these choices. This paper aims at improving upon these issues. By studying …
abstract algorithms approximation arxiv cs.lg cs.na design dimensionality factorization however interpretability low math.na math.oc matrix perspective practical regularization scale tucker type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Principal Data Engineering Manager
@ Microsoft | Redmond, Washington, United States
Machine Learning Engineer
@ Apple | San Diego, California, United States