Nov. 17, 2022, 6:05 p.m. | Heiko Hotz

Towards Data Science - Medium towardsdatascience.com

DeepSpeed Deep Dive — Model Implementations for Inference (MII)

A closer look at the latest open-source library from DeepSpeed

Image by author — created with Stable Diffusion

What is this about?

The DeepSpeed team has recently released a new open-source library called Model Implementation for Inference (MII), aimed towards making low-latency, low-cost inference of powerful models not only feasible but also easily accessible. You can read all about it in their blog post.

When I started exploring the MII …

ai deep dive deepspeed inference large language models machine learning nlp

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead Data Engineer

@ WorkMoney | New York City, United States - Remote