Jan. 21, 2022, 2:11 a.m. | Nesma M. Rezk, Tomas Nordström, Dimitrios Stathis, Zain Ul-Abdin, Eren Erdal Aksoy, Ahmed Hemani

cs.LG updates on arXiv.org arxiv.org

The compression of deep learning models is of fundamental importance in
deploying such models to edge devices. The selection of compression parameters
can be automated to meet changes in the hardware platform and application using
optimization algorithms. This article introduces a Multi-Objective
Hardware-Aware Quantization (MOHAQ) method, which considers hardware efficiency
and inference error as objectives for mixed-precision quantization. The
proposed method feasibly evaluates candidate solutions in a large search space
by relying on two steps. First, post-training quantization is applied …

arxiv hardware networks neural networks

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Business Intelligence Analyst

@ Rappi | COL-Bogotá

Applied Scientist II

@ Microsoft | Redmond, Washington, United States