all AI news
Benchmarking LLM performance with LangChain Auto-Evaluator // Lance Martin //LLMs in Prod Con Part 2
July 31, 2023, 5:54 p.m. | MLOps.community
MLOps.community www.youtube.com
Document Question-Answering is a popular LLM use case. LangChain makes it easy to assemble LLM components (e.g., models and retrievers) into chains that support question-answering. But, it is not always obvious to (1) evaluate the answer quality and (2) use this evaluation to guide improved QA chain settings (e.g., chunk size, retrieved docs count) or components (e.g., model or retriever choice). We recently released an open-source, hosted app to address these limitations (see blog post here). We have …
abstract auto benchmarking case components easy evaluation langchain llm llm performance llms part performance popular prod quality support
More from www.youtube.com / MLOps.community
Jobs in AI, ML, Big Data
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US