In this podcast, Niels Bantilan, Chief Machine Learning Engineer at Union, discusses the role of infrastructure in ML leveraging open source. Topics covered include data quality, schema definition, integrating tools like Polars, tracking data quality over time, generative DevOps, reproducibility in MLOps, and navigating edge cases.
Read more
AI Summary
AI Chapters
Episode notes
auto_awesome
Podcast summary created with Snipd AI
Quick takeaways
Orchestration tools like Flyte can unify open-source ML tools into a single platform for scalable data processing, experiment tracking, versioning, and monitoring.
Maintaining data quality and reproducibility in ML workloads requires tracking data lineage, managing artifact reproducibility, and considering an organization's ML practice maturity.
Leveraging ML at the infrastructure layer can optimize resource allocation by predicting requirements based on data set size and model architecture.
Deep dives
Pandera: A Data Quality Tool and its Creation
Neal's Bentilin, Chief ML Engineer at Union AI, discusses the creation of Pandera, a data quality tool. Pandera was developed to address issues such as data type errors, formatting problems, and null values in data frames. It allows users to define and visualize data schemas, making it easier to understand and work with complex data transformations. Although initially a personal project, Pandera gained traction and has since merged with Union. Neal's also discusses the integration of Pandera with Polars, highlighting the benefits of the lazy frame concept in Polars for efficient data transformations. Integration with Polars has been well received by the community.
Challenges in Data Quality and Maturity Levels in ML
Neal's explores the challenges faced in maintaining data quality and ensuring reproducibility in machine learning workloads. He highlights the importance of tracking data lineage, managing artifact reproducibility, and considering the maturity level of an organization's ML practice. Neal's emphasizes that while the vision of AGI and advanced LLMs is exciting, organizations must prioritize practicality and pragmatism, focusing on their current needs and technical limitations. He highlights the ongoing evolution in ML ops and the need for orchestration layers to address scalability, iteration speed, and developer experience, while allowing flexibility and customization.
The Intersection of LLMs and Infrastructure
Neal's delves into the usage of LLMs at the infrastructure layer, beyond text-to-terraform applications. He discusses the potential of leveraging ML to predict resource requirements and optimize workload execution, considering parameters such as data set size and model architecture. Neal's emphasizes the need for empirical testing and research to explore these possibilities, as well as the role of orchestrators in collecting and utilizing metadata to enhance ML infrastructure and enable better resource allocation.
OpenAI's Dev Day and its Relation to MLOps
Neal's shares his thoughts on OpenAI's Dev Day and its implications for the MLOps space. While acknowledging the potential of LLMs and Token-based architectures, Neal's sees the current focus of OpenAI as orthogonal to the practical challenges faced by ML practitioners and organizations. He emphasizes the need for a balance between speculative visions of AGI and the immediate requirements of production-level ML operations, highlighting the fundamental role of reproducibility, resource isolation, and maturity in real-world ML practices.
The Value of Declarative Approaches and Balance of Imperative Reasoning
Neal's discusses the value of declarative approaches in both ML and real-life interactions. He highlights the need for trust, autonomy, and collaboration in working with individuals who can understand and execute declarative instructions. Neal's also reflects on the importance of striking a balance between declarative and imperative reasoning in order to address edge cases, exceptions, and adaptability. He draws parallels between this approach and the LLM space, where models navigate and reason based on declared instructions to achieve desired outcomes.
MLOps podcast #197 with Niels Bantilan, Chief Machine Learning Engineer at Union, The Role of Infrastructure in ML Leveraging Open Source brought to us by Union.
// Abstract
When we start out building and deploying models in a new organization, life is simple: all I need to do is grab some data, iterate on a model that fits the data well and performs reasonably well on some held-out test set. Then, if you’re fortunate enough to get to the point where you want to deploy it, it’s fairly straightforward to wrap it in an app framework and host it on a cloud server. However, once you get past this stage, you’re likely to find yourself needing:
More scalable data processing framework
Experiment tracking for models
Heavier duty CPU/GPU hardware
Versioning tools to link models, data, code, and resource requirements
Monitoring tools for tracking data and model quality
There’s a rich ecosystem of open-source tools that solves each of these problems and more: but how do you unify all of them together into a single view? This is where orchestration tools like Flyte can help. Flyte not only allows you to compose data and ML pipelines, but it also serves as “infrastructure as code” so that you can leverage the open-source ecosystem and unify purpose-built tools for different parts of the ML lifecycle on a single platform. ML systems are not just models: they are the models, data, and infrastructure combined.
// Bio
Niels is the Chief Machine Learning Engineer at Union.ai, and core maintainer of Flyte, an open-source workflow orchestration tool, author of UnionML, an MLOps framework for machine learning microservices, and creator of Pandera, a statistical typing and data testing tool for scientific data containers. His mission is to help data science and machine learning practitioners be more productive.
He has a Masters in Public Health with a specialization in sociomedical science and public health informatics, and prior to that a background in developmental biology and immunology. His research interests include reinforcement learning, AutoML, creative machine learning, and fairness, accountability, and transparency in automated systems.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: https://github.com/cosmicBboy, https://union.ai/Flyte: https://flyte.org/
MLOps vs ML Orchestration // Ketan Umare // MLOps Podcast #183 - https://youtu.be/k2QRNJXyzFg
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Niels on LinkedIn: https://www.linkedin.com/in/nbantilan/
Timestamps:
[00:00] Niels' preferred coffee
[00:17] Takeaways
[03:45] Shout out to our Premium Brand Partner, Union!
[04:30] Pandera
[08:12] Creating a company
[14:22] Injecting ML for Data
[17:30] ML for Infrastructure Optimization
[22:17] AI Implementation Challenges
[24:25] Generative DevOps movement
[28:27] Pushing Limits: Code Responsibility
[29:46] Orchestration in OpenAI's Dev Day
[34:27] MLOps Stack: Layers & Challenges
[42:45] Mature Companies Embrace Kubernetes
[45:29] Horizon Challenges
[47:24] Flexible Integration for Resources
[49:10] MLOps Reproducibility Challenges
[53:14] MLOps Maturity Spectrum
[57:48] First-Class Citizens in Design
[1:00:16] Delegating for Efficient Collaboration
[1:04:55] Wrap up
Get the Snipd podcast app
Unlock the knowledge in podcasts with the podcast player of the future.
AI-powered podcast player
Listen to all your favourite podcasts with AI-powered features
Discover highlights
Listen to the best highlights from the podcasts you love and dive into the full episode
Save any moment
Hear something you like? Tap your headphones to save it with AI-generated key takeaways
Share & Export
Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more
AI-powered podcast player
Listen to all your favourite podcasts with AI-powered features
Discover highlights
Listen to the best highlights from the podcasts you love and dive into the full episode