all AI news
Fireside Chat - The Future of LLMs // David Hershey & Daniel Jeffries // LLMs in Prod Con Part 2
Aug. 17, 2023, 1:09 p.m. | MLOps.community
MLOps.community www.youtube.com
Evaluating the performance of language models (LLMs) is a pressing issue for companies working with generative AI. Defining what makes a model "good" and measuring its performance are challenging due to the diverse range of LLM applications. Existing evaluation methods, including benchmarks and user preference comparisons, have limitations in scalability and objectivity. The future of LLM evaluation lies in scaling testing with machine learning systems, such as reward models that capture user preferences, and simulating user sessions to …
abstract applications chat companies daniel daniel jeffries david diverse future generative good issue language language models llm llm applications llms part performance prod
More from www.youtube.com / MLOps.community
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne