MLOps.community

LLM Evaluation with Arize AI's Aparna Dhinakaran // #210

Feb 9, 2024
The podcast discusses the complexities of Language Model evaluation, the use of open-source versus private models, and the urgency of getting models into production. It also explores the challenges of evaluating LLM outcomes and highlights the importance of prompt engineering. Additionally, it emphasizes the need to quickly get ML models into production for identifying bottlenecks and setting up metrics.
Ask episode
Chapters
Transcript
Episode notes