July 31, 2023, 5:54 p.m. | MLOps.community

MLOps.community www.youtube.com

// Abstract
Document Question-Answering is a popular LLM use case. LangChain makes it easy to assemble LLM components (e.g., models and retrievers) into chains that support question-answering. But, it is not always obvious to (1) evaluate the answer quality and (2) use this evaluation to guide improved QA chain settings (e.g., chunk size, retrieved docs count) or components (e.g., model or retriever choice). We recently released an open-source, hosted app to address these limitations (see blog post here). We have …

abstract auto benchmarking case components easy evaluation langchain llm llm performance llms part performance popular prod quality support

More from www.youtube.com / MLOps.community

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne