Aug. 2, 2022, 2:11 a.m. | Youjie Li, Amar Phanishayee, Derek Murray, Jakub Tarnawski, Nam Sung Kim

cs.LG updates on arXiv.org arxiv.org

Deep neural networks (DNNs) have grown exponentially in size over the past
decade, leaving only those who have massive datacenter-based resources with the
ability to develop and train such models. One of the main challenges for the
long tail of researchers who might have only limited resources (e.g., a single
multi-GPU server) is limited GPU memory capacity compared to model size. The
problem is so acute that the memory requirement of training massive DNN models
can often exceed the aggregate …

arxiv capacity dnn gpu massive memory servers

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote