The Real Python Podcast cover image

Measuring Bias, Toxicity, and Truthfulness in LLMs With Python

The Real Python Podcast

CHAPTER

Introduction

This chapter introduces the topic of measuring the quality of large language models (LLMs), discusses techniques and tools for evaluating LLMs with Python, and covers methods to measure bias, toxicity, and hallucinations in LLMs.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner