July 2, 2023, 4:48 a.m. | /u/Accretence

Machine Learning www.reddit.com

Theoretically we'll be serving a large number of users and I am worried about clogging entire GPUs for example on a Stable Diffusion request that would take Nvidia A100 about 20 seconds to fulfill.

Do I need as many GPUs as my concurrent users?

Is there any relevant guide on GPU virtualizations?

a100 api diffusion example gpu gpus machinelearning nvidia nvidia a100 serve server stable diffusion tts

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote