all AI news
MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning
April 4, 2024, 4:45 a.m. | Ji Lin, Wei-Ming Chen, Han Cai, Chuang Gan, Song Han
cs.CV updates on arXiv.org arxiv.org
Abstract: Tiny deep learning on microcontroller units (MCUs) is challenging due to the limited memory size. We find that the memory bottleneck is due to the imbalanced memory distribution in convolutional neural network (CNN) designs: the first several blocks have an order of magnitude larger memory usage than the rest of the network. To alleviate this issue, we propose a generic patch-by-patch inference scheduling, which operates only on a small spatial region of the feature map …
abstract arxiv cnn convolutional neural network cs.cv deep learning designs distribution inference mcus memory microcontroller network neural network type units usage
More from arxiv.org / cs.CV updates on arXiv.org
Compact 3D Scene Representation via Self-Organizing Gaussian Grids
1 day, 20 hours ago |
arxiv.org
Fingerprint Matching with Localized Deep Representation
1 day, 20 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne