March 11, 2024, 2:02 p.m. | Marie Stephen Leo

Towards AI - Medium pub.towardsai.net

Reduce LLM latency and cost by over 95% using OpenAI, LiteLLM, Qdrant, and Sentence Transformers!

ai chatbots artificial intelligence caching chatbots cost data science generative generative ai chatbots latency llm machine learning openai programming qdrant reading reduce semantic technology transformers

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

AIML - Sr Machine Learning Engineer, Data and ML Innovation

@ Apple | Seattle, WA, United States

Senior Data Engineer

@ Palta | Palta Cyprus, Palta Warsaw, Palta remote