Adventures in Machine Learning cover image

Apache Spark Integration and Platform Execution for ML - ML 073

Adventures in Machine Learning

00:00

Spark ML vs Spark RDD?

Spark's default operation mode is if it can be distributed. It knows where each part of the data exists on each executor, it knows which rows are where. But there's not enough justification for the effort required to implement that in a high-level API.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app