all AI news
Fireside Chat - The Future of LLMs // David Hershey & Daniel Jeffries // LLMs in Prod Con Part 2
Aug. 17, 2023, 1:09 p.m. | MLOps.community
MLOps.community www.youtube.com
Evaluating the performance of language models (LLMs) is a pressing issue for companies working with generative AI. Defining what makes a model "good" and measuring its performance are challenging due to the diverse range of LLM applications. Existing evaluation methods, including benchmarks and user preference comparisons, have limitations in scalability and objectivity. The future of LLM evaluation lies in scaling testing with machine learning systems, such as reward models that capture user preferences, and simulating user sessions to …
abstract applications chat companies daniel daniel jeffries david diverse future generative good issue language language models llm llm applications llms part performance prod
More from www.youtube.com / MLOps.community
Jobs in AI, ML, Big Data
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US