all AI news
Harmony: Overcoming the Hurdles of GPU Memory Capacity to Train Massive DNN Models on Commodity Servers. (arXiv:2202.01306v2 [cs.DC] UPDATED)
Aug. 2, 2022, 2:11 a.m. | Youjie Li, Amar Phanishayee, Derek Murray, Jakub Tarnawski, Nam Sung Kim
cs.LG updates on arXiv.org arxiv.org
Deep neural networks (DNNs) have grown exponentially in size over the past
decade, leaving only those who have massive datacenter-based resources with the
ability to develop and train such models. One of the main challenges for the
long tail of researchers who might have only limited resources (e.g., a single
multi-GPU server) is limited GPU memory capacity compared to model size. The
problem is so acute that the memory requirement of training massive DNN models
can often exceed the aggregate …
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote