April 10, 2024, 9:59 p.m. | MLOps.community

MLOps.community www.youtube.com

Huge thank you to Zilliz for sponsoring this episode. Zilliz RAG - https://zilliz.com/vector-database-use-cases/llm-retrieval-augmented-generation

MLOps podcast #217 with Sam Bean, Software Engineer (Applied AI) at Rewind.ai, The Real E2E RAG Stack.

// Abstract
What does a fully operational LLM + Search stack look like when you're running your own retrieval and inference infrastructure? What does the flywheel really mean for RAG applications? How do you maintain the quality of your responses? How do you prune/dedupe documents to maintain your document quality? …

abstract applications applied ai e2e engineer flywheel inference infrastructure llm look mean mlops mlops podcast podcast quality rag responses retrieval running sam search software software engineer stack

More from www.youtube.com / MLOps.community

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Software Engineer, Generative AI (C++)

@ SoundHound Inc. | Toronto, Canada