80,000 Hours Podcast cover image

#184 – Zvi Mowshowitz on sleeping on sleeper agents, and the biggest AI updates since ChatGPT

80,000 Hours Podcast

00:00

Language Training and Model Bias

Training models on specific language inputs can result in better compatibility due to the limited data availability, presenting a challenge for gathering sufficient data. Models trained on internet data tend to exhibit bias towards a left libertarian position consistently, as this reflects the predominant content on the internet. Chinese language training faces difficulties in training unbiased models due to limited success in data collection.

Play episode from 02:00:48
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app