June 28, 2024, 2:33 p.m. | /u/mcharytoniuk

Machine Learning www.reddit.com

I have started this project recently. It allows us to self-host llama.cpp and use it with open-source models.

It started to gain some traction recently, and it is production-ready.

It allows scaling from zero instances, so if you are using cloud providers to prototype your ideas with open-source LLMs, you will only pay for what you actually use. If there is a period of inactivity, you can use it to shut down expensive GPU instances and only leave some cheap …

cloud cloud providers cpp custom ideas instances llama llms machinelearning open-source models production project prototype scaling will you

Junior Senior Reliability Engineer

@ NielsenIQ | Bogotá, Colombia

[Job - 15712] Vaga Afirmativa para Mulheres - QA (Automation), SR

@ CI&T | Brazil

Production Reliability Engineer, Trade Desk

@ Jump Trading | Sydney, Australia

Senior Process Engineer, Prenatal

@ BillionToOne | Union City and Menlo Park, CA

Senior Scientist, Sustainability Science and Innovation

@ Microsoft | Redmond, Washington, United States

Data Scientist

@ Ford Motor Company | Chennai, Tamil Nadu, India