Web: http://arxiv.org/abs/2206.07446

June 16, 2022, 1:10 a.m. | Rongjie Yi, Ting Cao, Ao Zhou, Xiao Ma, Shangguang Wang, Mengwei Xu

cs.LG updates on arXiv.org arxiv.org

DNNs are ubiquitous on edge devices nowadays. With its increasing importance
and use cases, it's not likely to pack all DNNs into device memory and expect
that each inference has been warmed up. Therefore, cold inference, the process
to read, initialize, and execute a DNN model, is becoming commonplace and its
performance is urgently demanded to be optimized. To this end, we present
NNV12, the first on-device inference engine that optimizes for cold inference
NNV12 is built atop 3 novel …

arxiv deep deep learning devices edge edge devices learning lg on

More from arxiv.org / cs.LG updates on arXiv.org

Machine Learning Researcher - Saalfeld Lab

@ Howard Hughes Medical Institute - Chevy Chase, MD | Ashburn, Virginia

Project Director, Machine Learning in US Health

@ ideas42.org | Remote, US

Data Science Intern

@ NannyML | Remote

Machine Learning Engineer NLP/Speech

@ Play.ht | Remote

Research Scientist, 3D Reconstruction

@ Yembo | Remote, US

Clinical Assistant or Associate Professor of Management Science and Systems

@ University at Buffalo | Buffalo, NY