April 20, 2023, 3:08 p.m. | Michał Oleszak

Towards Data Science - Medium towardsdatascience.com

Reduce your models' latency, storage, and inference costs with quantization and pruning

artificial intelligence costs data data science inference latency machine learning mlops model optimization optimization programming quantization reading reduce science storage tensorflow

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Software Engineer, Generative AI (C++)

@ SoundHound Inc. | Toronto, Canada