How to Construct Domain Specific LLM Evaluation Systems: Hamel Husain and Emil Sedgh

AI Engineer
AI Engineer
4.6 هزار بار بازدید - 2 هفته پیش - Many failed AI products share
Many failed AI products share a common root cause: a failure to create robust evaluation systems. Evaluation systems allow you to improve your AI quickly in a systematic way and unlock superpowers like the ability to curate data for fine-tuning. However, many practitioners struggle with how to construct evaluation systems that are specific to their problems. In this talk, we will walk through a detailed example of how to construct domain-specific evaluation systems. Recorded live in San Francisco at the AI Engineer World's Fair. See the full schedule of talks at https://www.ai.engineer/worldsfair/2024/schedule & join us at the AI Engineer World's Fair in 2025! Get your tickets today at https://ai.engineer/2025 About Hamel Hamel Husain started working with language models five years ago when he led the team that created CodeSearchNet, a precursor to GitHub CoPilot. Since then, he has seen many successful and unsuccessful approaches to building LLM products. Hamel is also an active open source maintainer and contributor of a wide range of ML/AI projects. Hamel is currently an independent consultant. About Emil Emil is CTO at Rechat, where he leads the development of Lucy, an AI personal assistant designed to support real estate agents.
2 هفته پیش در تاریخ 1403/06/29 منتشر شده است.
4,650 بـار بازدید شده
... بیشتر