all AI news
Benchmarking LLM performance with LangChain Auto-Evaluator // Lance Martin //LLMs in Prod Con Part 2
July 31, 2023, 5:54 p.m. | MLOps.community
MLOps.community www.youtube.com
Document Question-Answering is a popular LLM use case. LangChain makes it easy to assemble LLM components (e.g., models and retrievers) into chains that support question-answering. But, it is not always obvious to (1) evaluate the answer quality and (2) use this evaluation to guide improved QA chain settings (e.g., chunk size, retrieved docs count) or components (e.g., model or retriever choice). We recently released an open-source, hosted app to address these limitations (see blog post here). We have …
abstract auto benchmarking case components easy evaluation langchain llm llm performance llms part performance popular prod quality support
More from www.youtube.com / MLOps.community
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne