all AI news
Evaluation
March 14, 2023, 7:46 p.m. | LangChain
LangChain blog.langchain.dev
Evaluation of language models, and by extension applications built on top of language models, is hard. With recent model releases (OpenAI, Anthropic, Google) evaluation is becoming a bigger and bigger issue. People are starting to try to tackle this, with OpenAI releasing OpenAI/evals
- focused on evaluating OpenAI models.
anthropic applications bigger evals evaluation extension google issue language language models openai people releases
More from blog.langchain.dev / LangChain
[Week of 4/15] LangChain Release Notes
1 day, 6 hours ago |
blog.langchain.dev
LangSmith: Production Monitoring & Automations
2 weeks, 3 days ago |
blog.langchain.dev
LangFriend: a Journal with Long-Term Memory
3 weeks, 1 day ago |
blog.langchain.dev
Open Source Extraction Service
3 weeks, 3 days ago |
blog.langchain.dev
Using Feedback to Improve Your Application: Self Learning GPTs
4 weeks, 2 days ago |
blog.langchain.dev
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Lead Software Engineer - Artificial Intelligence, LLM
@ OpenText | Hyderabad, TG, IN
Lead Software Engineer- Python Data Engineer
@ JPMorgan Chase & Co. | GLASGOW, LANARKSHIRE, United Kingdom
Data Analyst (m/w/d)
@ Collaboration Betters The World | Berlin, Germany
Data Engineer, Quality Assurance
@ Informa Group Plc. | Boulder, CO, United States
Director, Data Science - Marketing
@ Dropbox | Remote - Canada