Feb. 28, 2022, 2:11 a.m. | Amir Ardakani, Arash Ardakani, Brett Meyer, James J. Clark, Warren J. Gross

cs.LG updates on arXiv.org arxiv.org

Quantization of deep neural networks is a promising approach that reduces the
inference cost, making it feasible to run deep networks on resource-restricted
devices. Inspired by existing methods, we propose a new framework to learn the
quantization intervals (discrete values) using the knowledge of the network's
weight and activation distributions, i.e., standard deviation. Furthermore, we
propose a novel base-2 logarithmic quantization scheme to quantize weights to
power-of-two discrete values. Our proposed scheme allows us to replace
resource-hungry high-precision multipliers with …

arxiv deviation networks neural networks quantization standard

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote