all AI news
Adapprox: Adaptive Approximation in Adam Optimization via Randomized Low-Rank Matrices
March 25, 2024, 4:41 a.m. | Pengxiang Zhao, Ping Li, Yingjie Gu, Yi Zheng, Stephan Ludger K\"olker, Zhefeng Wang, Xiaoming Yuan
cs.LG updates on arXiv.org arxiv.org
Abstract: As deep learning models exponentially increase in size, optimizers such as Adam encounter significant memory consumption challenges due to the storage of first and second moment data. Current memory-efficient methods like Adafactor and CAME often compromise accuracy with their matrix factorization techniques. Addressing this, we introduce Adapprox, a novel approach that employs randomized low-rank matrix approximation for a more effective and accurate approximation of Adam's second moment. Adapprox features an adaptive rank selection mechanism, finely …
abstract accuracy adam approximation arxiv challenges consumption cs.cl cs.lg current data deep learning factorization low math.oc matrix memory memory consumption optimization storage type via
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US