Jan. 31, 2024, 3:46 p.m. | Kun Wang Jiani Cao Zimu Zhou Zhenjiang Li

cs.LG updates on arXiv.org arxiv.org

Executing deep neural networks (DNNs) on edge artificial intelligence (AI) devices enables various autonomous mobile computing applications. However, the memory budget of edge AI devices restricts the number and complexity of DNNs allowed in such applications. Existing solutions, such as model compression or cloud offloading, reduce the memory footprint of DNN inference at the cost of decreased model accuracy or autonomy. To avoid these drawbacks, we divide DNN into blocks and swap them in and out in order, such that …

applications artificial artificial intelligence autonomous beyond budget cloud complexity compression computing cs.ai cs.dc cs.lg devices dnn edge edge ai inference intelligence memory mobile mobile computing networks neural networks reduce solutions

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote