The Real Python Podcast cover image

Measuring Bias, Toxicity, and Truthfulness in LLMs With Python

The Real Python Podcast

00:00

Introduction

This chapter introduces the topic of measuring the quality of large language models (LLMs), discusses techniques and tools for evaluating LLMs with Python, and covers methods to measure bias, toxicity, and hallucinations in LLMs.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app