all AI news
MOHAQ: Multi-Objective Hardware-Aware Quantization of Recurrent Neural Networks. (arXiv:2108.01192v3 [cs.LG] UPDATED)
Jan. 21, 2022, 2:11 a.m. | Nesma M. Rezk, Tomas Nordström, Dimitrios Stathis, Zain Ul-Abdin, Eren Erdal Aksoy, Ahmed Hemani
cs.LG updates on arXiv.org arxiv.org
The compression of deep learning models is of fundamental importance in
deploying such models to edge devices. The selection of compression parameters
can be automated to meet changes in the hardware platform and application using
optimization algorithms. This article introduces a Multi-Objective
Hardware-Aware Quantization (MOHAQ) method, which considers hardware efficiency
and inference error as objectives for mixed-precision quantization. The
proposed method feasibly evaluates candidate solutions in a large search space
by relying on two steps. First, post-training quantization is applied …
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Business Intelligence Analyst
@ Rappi | COL-Bogotá
Applied Scientist II
@ Microsoft | Redmond, Washington, United States