all AI news
OLLA: Optimizing the Lifetime and Location of Arrays to Reduce the Memory Usage of Neural Networks. (arXiv:2210.12924v2 [cs.LG] UPDATED)
Nov. 4, 2022, 1:13 a.m. | Benoit Steiner, Mostafa Elhoushi, Jacob Kahn, James Hegarty
cs.LG updates on arXiv.org arxiv.org
The size of deep neural networks has grown exponentially in recent years.
Unfortunately, hardware devices have not kept pace with the rapidly increasing
memory requirements. To cope with this, researchers have turned to techniques
such as spilling and recomputation, which increase training time, or reduced
precision and model pruning, which can affect model accuracy. We present OLLA,
an algorithm that optimizes the lifetime and memory location of the tensors
used to train neural networks. Our method reduces the memory usage …
arrays arxiv location memory networks neural networks reduce
More from arxiv.org / cs.LG updates on arXiv.org
A Single-Loop Algorithm for Decentralized Bilevel Optimization
1 day, 3 hours ago |
arxiv.org
CLEANing Cygnus A deep and fast with R2D2
1 day, 3 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Stagista Technical Data Engineer
@ Hager Group | BRESCIA, IT
Data Analytics - SAS, SQL - Associate
@ JPMorgan Chase & Co. | Mumbai, Maharashtra, India