Super Data Science: ML & AI Podcast with Jon Krohn cover image

Super Data Science: ML & AI Podcast with Jon Krohn

771: Gradient Boosting: XGBoost, LightGBM and CatBoost, with Kirill Eremenko

Apr 2, 2024
Machine learning expert Kirill Eremenko discusses decision trees, random forests, and the top gradient boosting algorithms: XGBoost, LightGBM, and CatBoost. Topics include advantages of XGBoost, LightGBM efficiency, and CatBoost's handling of categorical variables.
01:55:27

Episode guests

Podcast summary created with Snipd AI

Quick takeaways

  • Gradient boosting iteratively improves predictions by focusing on residuals and gradients.
  • XGBoost prioritizes speed in gradient boosting with efficient enhancements.

Deep dives

Transition to Gradient Boosting from Previous Methods

Unlike random forest, gradient boosting focuses on iteratively improving predictions by predicting errors and chaining models together. Adaboost, a precursor, adapted to model errors selectively. In gradient boosting, each model in the ensemble predicts the gradient of the loss function, optimizing directionally based on residuals.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode