Sept. 10, 2022, 6:24 p.m. | /u/markurtz

Machine Learning www.reddit.com

Utilizing the oBERT research we published at Neural Magic and some further iteration, we’ve enabled an increase in NLP performance of 175X on CPUs while retaining 99% accuracy on the question-answering task in MLPerf. A combination of distillation, layer dropping, quantization, and unstructured pruning with oBERT enabled these large performance gains through the [DeepSparse Engine](https://github.com/neuralmagic/deepsparse). All of our contributions and research are open-sourced or free to use. Read through the [oBERT paper on arxiv](https://arxiv.org/abs/2203.07259), try out the research in [SparseML](https://github.com/neuralmagic/sparseml). …

machinelearning mlperf nlp performance sparsity

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Sr. Software Development Manager, AWS Neuron Machine Learning Distributed Training

@ Amazon.com | Cupertino, California, USA