MLOps.community

Demetrios
undefined
Apr 12, 2021 • 59min

Machine Learning at Atlassian // Geoff Sims // Coffee Session#34

Coffee Sessions #34 with Geoff Sims of Atlassian, Machine Learning at Atlassian. //Abstract As one of the world's most visible software companies, Atlassian's vast data and deep product suite pose an interesting MLOps challenge, and we're grateful to Geoff for taking us behind the curtain. //Bio Geoff is a Principal Data Scientist at Atlassian, the software company behind Jira, Confluence & Trello. He works with the product teams and focuses on delivering smarter in-product experiences and recommendations to our millions of active users by using machine learning at scale. Prior to this, he was in the Customer Support & Success division, leveraging a range of NLP techniques to automate and scale the support function. Prior to Atlassian, Geoff has applied data science methodologies across the retail, banking, media,  and renewable energy industries. He began his foray into data science as a research astrophysicist, where he studied astronomy from the coldest & driest location on Earth: Antarctica. --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/ Connect with Geoff on https://www.linkedin.com/in/geoff-sims-0a37999b/ Timestamps: [00:00] Introduction to Geoff Sims [01:20] Geoff's background [04:00] Evolution of ML Ecosystem in Atlassian [06:50] Figure out by necessity [08:47] Machine Learning not priority number one and disconnected to MLOps [11:53] Atlassian being behind or advanced? [16:38] Serious switch of Atlassian around machine learning [17:47] What data org did it come from? [20:00] Consolidation of the stack [21:21] Tooling - blessing and curse [24:37] Tackling play out [29:38] Staying on the same page [30:48] Priority of needs [31:55] How did it evolve? [35:12] Where is Atlassian now? [40:21] "Architecturally, Tecton is very very similar (to ours), it was just way more mature." [41:17] What unleashed you to do now? [41:36] "The biggest thing is independence from a data science perspective. Less reliance and less dependence on an army of engineers to help deploy features and models." [44:25] Have you bought other tools? [45:43] "At any given time, there's something that's a bottleneck. Look where the bottleneck is, then fix it and move on to the next thing."   [48:20] Atlassian bringing a model into production [50:01] "When we undertake whatever the project is, its days or weeks to go to a prototype rather than months or quarters." [53:10] "Conceptually, you're struggling walking towards that place because that's the place you want to be. If that's your problem, that's good. That's the promised land." [54:45] "Using our own tools is paramount because we are customers as well. So we see and feel the pain which helps us identify the problems and understand them."
undefined
Apr 9, 2021 • 59min

MLOps Community 1 Year Anniversary! // Demetrios Brinkmann, David Aponte & Vishnu Rachakonda // MLOps Meetup #59

MLOps community meetup #59! Last Wednesday was the celebration of the MLOps Community 1 Year Anniversary! This has been a conversion of Demetrios Brinkmann, David Aponte and Vishnu Rachkonda! //Abstract Over the past year Demetrios, David and Vishnu have interviewed many of the top names in MLOps. During this time they have been able to apply these learnings at their jobs and see what works for them. In this one year anniversary meetup the three of them will discuss some of the most impacting advice they have received in the last year and how they have put it into practice. //Bio Demetrios Brinkmann At the moment Demetrios is immersing himself in Machine Learning by interviewing experts from around the world in the weekly MLOps.community meetups. Demetrios is constantly learning and engaging in new activities to get uncomfortable and learn from his mistakes. He tries to bring creativity into every aspect of his life, whether that be analyzing the best paths forward, overcoming obstacles, or building lego houses with his daughter. David Aponte David is one of the organizers of the MLOps Community. He is an engineer, teacher, and lifelong student. He loves to build solutions to tough problems and share his learnings with others. He works out of NYC and loves to hike and box for fun. He enjoys meeting new people so feel free to reach out to him! Vishnu Rachakonda Vishnu is the operations lead for the MLOps Community and co-hosts the MLOps Coffee Sessions podcast. He is a machine learning engineer at Tesseract Health, a 4Catalyzer company focused on retinal imaging. In this role, he builds machine learning models for clinical workflow augmentation and diagnostics in on-device and cloud use cases. Since studying bioengineering at Penn, Vishnu has been actively working in the fields of computational biomedicine and MLOps. In his spare time, Vishnu enjoys suspending all logic to watch Indian action movies, playing chess, and writing. ----------- Connect With Us ✌️-------------    Join our Slack community:  https://go.mlops.community/slack Follow us on Twitter:  @mlopscommunity Sign up for the next meetup:  https://go.mlops.community/register Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/ Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/ Timestamps: [01:07] Big shoutout to everybody that's in these meetups! [02:03] Big shoutout to Ivan Nardini for leading the Engineering Labs and to everyone who took part in the Engineering Labs! [02:26] Big shoutout to Charlie You leading the Reading Group and to everyone who takes part in it! [02:39] Big shoutout to everyone who takes part in the Office Hours! [02:49] Big shoutout to the people who are helping with shaping the website! [03:34] Thanks to all the people in Slack! Laszlo, Ariel, and people answering Slack questions. [04:10] Big thanks to all our Sponsors FiddlerAI, Algorithmia, and Tecton!   [06:13] David's Background [08:08] Vishnu's Background [09:55] High-Level Points [15:57] Starting small [24:05] Over-optimization - the root of all evil [26:42] Keeping text deck open [36:45] Missing from current MLOps tooling [48:00] How to communicate in these data products?
undefined
Apr 6, 2021 • 46min

MLOps Investments // Sarah Catanzaro // Coffee Session #33

Coffee Sessions #33 with Sarah Catanzaro of Amplify Partners, MLOps Investments. //Bio Sarah Catanzaro is a Partner at Amplify Partners, where she focuses on investing in and advising high potential startups in machine intelligence, data management, and distributed systems. Her investments at Amplify include startups like RunwayML, Maze Design, OctoML, and Metaphor Data among others. Sarah also has several years of experience defining data strategy and leading data science teams at startups and in the defense/intelligence sector including through roles at Mattermark, Palantir, Cyveillance, and the Center for Advanced Defense Studies. //We had a wide-ranging discussion with Sarah, three takeaways stood out: The relationship between unstructured data and structured data is due for change. In most settings, you have some form of structured data (i.e. a metadata table) and unstructured data (i.e. images, text, etc.) Managing the relationship between these forms of data can constitute the bulk of MLOps. Because of this difficulty, Sarah forecasted new tooling arising to make data management easier. Academic benchmarks suffer from a lack of transparency on production/industry use cases. In conversation with Andrew Ng, Sarah shared her lesson that despite all the blame industry professionals place on academics for narrowly optimizing to benchmarks with little practical meaning, they also share the blame for making it difficult to create meaningful benchmarks. Companies are loath to share realistic data and the true context in which ML has to operate. MLOps is due for consolidation, especially as companies adopt platform-driven strategies. As many of you all know, there are tons and tons of MLOps tools out there. As more companies address these challenges, Sarah predicted that many of the point solutions would start to be consolidated into larger platforms. // Other Links https://amplifypartners.com/team/sarah/ https://projectstoknow.amplifypartners.com/ml-and-data https://twitter.com/sarahcat21/status/1360105479620284419 --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/ Connect with Sarah on LinkedIn: https://www.linkedin.com/in/sarah-catanzaro-9770b98/ Timestamps: [00:00] Introduction to Sarah Catanzaro [02:07] Sarah's background in tech [06:00] Staying engineer oriented despite being an investment firm [08:50] Tools you wished you had earlier in your career [12:36] 2 Motives of ML Engineers and ML Platform Team [16:36] Open-sourcing [21:29] Startup focus on resources [23:57] Playout of open-source project [27:32] Consolidation [33:18] Finding solutions [36:18] Evolution of MLOps industry in the coming years [42:36] Frameworks   [43:14] Structure data sets available to researchers. Meaningful advances of deep learning applied to structure data as well.
undefined
Apr 4, 2021 • 53min

Model Watching: Keeping Your Project in Production // Ben Wilson // MLOps Meetup #58

MLOps community meetup #58! Last Wednesday we talked to Ben Wilson, Practice Lead Resident Solutions Architect, Databricks. Model Monitoring Deep Dive with the author of Machine Learning Engineering in Action. It was a pleasure getting to talk to Ben about difficulties in monitoring in machine learning. His expertise obviously comes from experience and as he said a few times in the meetup, I learned the hard way over 10 years as a data scientist so you don't have to! Ben was also kind enough to give us a 35% off promo code for his book! Use the link: http://mng.bz/n2P5 //Abstract A great deal of time is spent building out the most effectively tuned model, production-hardened code, and elegant implementation for a business problem. Shipping our precious and clever gems to production is not the end of the solution lifecycle, though, and many-an-abandoned projects can attest to this. In this talk, we will discuss how to think about model attribution, monitoring of results, and how (and when) to report those results to the business to ensure a long-lived and healthy solution that actually solves the problem you set out to solve. //Bio Ben Wilson has worked as a professional data scientist for more than ten years. He currently works as a resident solutions architect at Databricks, where he focuses on machine learning production architecture with companies ranging from 5-person startups to global Fortune 100. Ben is the creator and lead developer of the Databricks Labs AutoML project, a Scala-and Python-based toolkit that simplifies machine learning feature engineering, model tuning, and pipeline-enabled modelling. He's the author of Machine Learning Engineering in Action, a primer on building, maintaining, and extending production ML projects. //Takeaways Understanding why attribution and performance monitoring is critical for long-term project success Borrowing hypothesis testing, stratification for latent confounding variable minimization, and statistical significance estimation from other fields can help to explain the value of your project to a business Unlike in street racing, drifting is not cool in ML, but it will happen. Being prepared to know when to intervene will help to keep your project running. ----------- Connect With Us ✌️------------- Join our Slack community:  https://go.mlops.community/slack Follow us on Twitter:  @mlopscommunity Sign up for the next meetup:  https://go.mlops.community/register Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Ben on LinkedIn: www.linkedin.com/in/benjamin-wilson-arch/ Timestamps: [00:00] Introduction to Ben Wilson [00:11] Ben's background in tech [03:40] Human aspect of Machine Learning in MLOps [05:51] MLOps is an organizational problem [09:27] Fragile Models [12:36] Fraud Cases [15:21] Data Monitoring [18:37] Importance of knowing what to monitor for [22:00] Monitoring for outliers [24:16] Staying out of Alert Hell [29:40] Ground Truth [31:25] Model vs Data Drift on Ground Truth Unavailability [34:25] Benefit to monitor system or business level metrics [38:20] Experiment in the beginning, not at the end [40:30] Adaptive windowing [42:22] Bridge the gap [46:42] What scarred you really bad?
undefined
Mar 26, 2021 • 56min

A Missing Link in the ML Infrastructure Stack // Josh Tobin // MLOps Meetup #57

MLOps community meetup #57! Last Wednesday we talked to Josh Tobin, Founder, Stealth-Stage Startup. // Abstract: Machine learning is quickly becoming a product engineering discipline. Although several new categories of infrastructure and tools have emerged to help teams turn their models into production systems, doing so is still extremely challenging for most companies. In this talk, we survey the tooling landscape and point out several parts of the machine learning lifecycle that are still underserved. We propose a new category of tool that could help alleviate these challenges and connect the fragmented production ML tooling ecosystem. We conclude by discussing similarities and differences between our proposed system and those of a few top companies. // Bio: Josh Tobin is the founder and CEO of a stealth machine learning startup. Previously, Josh worked as a deep learning & robotics researcher at OpenAI and as a management consultant at McKinsey. He is also the creator of Full Stack Deep Learning (fullstackdeeplearning.com), the first course focused on the emerging engineering discipline of production machine learning. Josh did his PhD in Computer Science at UC Berkeley advised by Pieter Abbeel. // Other Links https://josh-tobin.com course.fullstackdeeplearning.com ----------- Connect With Us ✌️-------------    Join our Slack community:  https://go.mlops.community/slack Follow us on Twitter:  @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Josh on LinkedIn: https://www.linkedin.com/in/josh-tobin-4b3b10a9/ Timestamps: [00:00] Introduction to Josh Tobin [01:18] Background of Josh into tech [08:27] We're you guys behind the Rubik's Cube? [09:26] Rubik's Cube Project [09:51] "Research is meant to show you what's possible to solve." [11:07] "That's one of the things that's started to change and I think the MLOps world is maybe a part of that. What I'm excited about this is that people are focusing on the impact of their models." [13:18] Insights on Testing [17:11] Evaluation Store [18:33] "Production Machine Learning is data-driven products that have predictions in the loop." [23:40] Analyzing and moving forward [24:02] "My medium term mindset how machine learning is created is that is there's still gonna be humans involved but humans will be more efficient by tools." [25:50] Is there a market for this? [27:40] "The long tale of machine learning use cases is becoming part of every products and service more or less the companies create but it's the same way the software part of the products and services the companies create these days. It's going to create an enormous amount of value." [30:09] Talents [32:52] Organizational by-ends and knowledge [35:16] Tools used for Evaluation Store 39:59] Difference from Monitoring Tool [42:10] Who is the right person to interact in Evaluation Store? [50:05] Technical challenges of Apple and Tesla [53:30] "As Machine Learning use cases are getting more and more complicated, higher and higher dimensional data, bigger and bigger models, larger training sets many companies would need in order to continually improve their systems over time."
undefined
Mar 23, 2021 • 52min

The Godfather Of MLOps // D. Sculley // MLOps Coffee Sessions #32

Coffee Sessions #32 with D. Sculley of Google, The Godfather Of MLOps. //Bio D is currently a director in Google Brain, leading research teams working on robust, responsible, reliable and efficient ML and AI. In his time at Google, D worked on nearly every aspect of machine learning, and have led both product and research teams including those on some of the most challenging business problems. // Links to D. Sculley's Papers ML Test Score: https://research.google/pubs/pub46555/ Machine Learning: The high-interest credit card of technical debt https://research.google/pubs/pub43146/ Google Scholar: https://scholar.google.com/citations?user=l_O64B8AAAAJ&hl=en --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/ Connect with Vishnu on LinkedIn: https://www.linkedin.com/in/vrachakonda/ Connect with D. Sculley on LinkedIn: https://www.linkedin.com/in/d-sculley-90467310/ Timestamps: [00:00] Introduction to D. Sculley [00:40] Biggest Papers were written by D for Machine Learning [02:08] What's changed since you wrote those papers? [02:56] "No 1, there is an MLOps community." [04:38] Old best practices [05:12] "The fact that there are jobs titled MLOps, this is different than it was 5 or 6 years ago." [06:30] Machine Learning Systems then and now [07:08] "There wasn't the level of general infrastructure that was looking to offer the large scale integrated solutions."    [07:57] ML Test Score [11:09] "The Test Score was really written for situations where you don't care about one prediction. You care about millions or billions of predictions per day." [12:27] "In the end, it's not about the score. It's about the process of asking the questions making sure that each of the important questions that you're asking yourself, you have a good answer to."   [13:04] What else is needed in the Test Score? [14:36] Stratified testing   [17:05] Counterfactual testing [18:34] Boundaries [19:15] Dark ages [20:27] How do you try in Triage? [21:10] "Reliability is important. There are no small mistakes. If there are errors, they're going to get spotted and publicised. They're going to impact user's lives. The bar is really high and it's worth the effort to ensure strong reliability." [23:11] How do you build that interest stress test? [24:39] "I believe that stress test is going to look like a useful way to encode expert knowledge about domain areas." [25:37] How do I bring robustness? [26:47] "Because we don't know how to specify the behaviour in advance, testing the behaviour that we wanted to have is a fundamentally hard problem." [27:22] Underspecification Paper [30:58] "It's important to be evaluating models on this auto domain stress test and make sure that we understand the implications of what we're thinking about while we are in deployment land." [32:27] Principal challenges in productionizing Machine Learning [34:57] "As we expose our models to more specifics, this means there are more potential places our models might be exhibiting unexpected or undesirable behaviour." [42:37] Splintering of ML Engineering [46:00] Communities shaping the MLOps sphere [46:42] "It's much better to have one large community than three smaller communities because of those edufacts." [47:47] Concept of technical debt in machine learning. [49:28] "The good idea is to tend to make their way into the community if they are in a form that people can digest and share."
undefined
Mar 19, 2021 • 1h 5min

Operationalizing Machine Learning at a Large Financial Institution // Daniel Stahl // MLOps Meetup #56

MLOps community meetup #56! Last Wednesday we talked to  Daniel Stahl, Head of Data and Analytic Platforms, Regions Bank. // Abstract: The Data Science practice has evolved significantly at Regions, with a corresponding need to scale and operationalize machine learning models. Additionally, highly regulated industries such as finance require a heightened focus on reproducibility, documentation, and model controls.  In this session with Daniel Stahl, we will discuss how the Regions team designed and scaled their data science platform using DevOps and MLOps practices.  This has allowed Regions to meet the increased demand for machine learning while embedding controls throughout the model lifecycle.  In the 2 years since the data science platform has been onboarded, 100% of data products have been successfully operationalized. // Bio: Daniel Stahl leads the ML platform team at Regions Bank and is responsible for tooling, data engineering, and process development to make operationalizing models easy, safe, and compliant for Data Scientists.   Daniel has spent his career in financial services and has developed novel methods for computing tail risk in both credit risk and operational risk, resulting in peer-reviewed publications in the Journal of Credit Risk and the Journal of Operational Risk. Daniel has a Masters in Mathematical Finance from the University of North Carolina Charlotte.      Daniel lives in Birmingham, Alabama with his wife and two daughters. ----------- Connect With Us ✌️-------------    Join our Slack community:  https://go.mlops.community/slack Follow us on Twitter:  @mlopscommunity Sign up for the next meetup:  https://go.mlops.community/register Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Dan on LinkedIn: https://www.linkedin.com/in/daniel-stahl-6685a52a/ Timestamps: [00:00] Introduction to Ben Wilson [00:11] Ben's background in tech [01:17] "How do you do what I have always done pretty well which is being as lazy as possible in order to automate things that I hate doing. So I learned about Regression Problems." [03:40] Human aspect of Machine Learning in MLOps [05:51] MLOps is an organizational problem [09:27] Fragile Models [12:36] Fraud Cases [15:21] Data Monitoring [18:37] Importance of knowing what to monitor for [22:00] Monitoring for outliers [24:16] Staying out of Alert Hell [29:40] Ground Truth [31:25] Model vs Data Drift on Ground Truth Unavailability [34:25] Benefit to monitor system or business level metrics [38:20] Experiment in the beginning, not at the end [40:30] Adaptive windowing [42:22] Bridge the gap [46:42] What scarred you really bad?
undefined
Mar 12, 2021 • 58min

How to Avoid Suffering in Mlops/Data Engineering Role // Igor Lushchyk // MLOps Meetup #55

MLOps community meetup #55! Last Wednesday we talked to Igor Lushchyk, Data Engineer, Adyen.   // Abstract: Building Data Science and Machine Learning platforms at a scale-up. Having the main difficulty in finding correct processes and basically being a toddler who learns how to walk on a steep staircase. The transition from homegrown platform to open source solutions, supporting old solutions and maturing them with making data scientists happy.   // Bio: Igor is a software engineer with more than 10 years of experience. With a background in bioinformatics, he even started PhD but didn't finish it. As a data engineer, Igor has been working for the last 6 or 7 years, or maybe more - because he was doing almost the same data engineering stuff but his position was named differently. Igor has been doing a lot of MLOps in 4-5 years now. He doesn't know what he was doing more then - Data Engineering or MLOps. And that’s how this topic came about.   ----------- Connect With Us ✌️-------------    Join our Slack community:  https://go.mlops.community/slack Follow us on Twitter:  @mlopscommunity Sign up for the next meetup:  https://go.mlops.community/register Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Igor on LinkedIn: https://www.linkedin.com/in/igor-lushchyk/ Timestamps: [00:00] Introduction to Igor Lushchyk [02:05] Igor's background in tech [07:42] Tips you can pass on [11:05] How these tools work and how they play together and what is underneath? [13:18] Dedicated MLOps team [13:55] Central Data Infrastructure Section [16:57] Transfer over to open-source [20:24] If you don't plan for production from the beginning, then it's going to be painful trying to go from POC to production. [22:08] Ho do you handle data lineage? [25:09] You chose that back in the day but you're regretting it. [26:34] "Try to use tools which solve 80% of your use cases and maybe 20% you'll have the suffering but at least it's not 100% suffering." [27:27] Friction points [28:53] Interaction with Data Scientists [29:21] "We have alignment sessions. We have different levels of representations. We share our progress." [32:42] Build verse by decisions [34:04] When to build or grab an open-source tool [35:51] Build your own or buy open-source? [37:11] Certain maturity and a certain number of engineers [38:11] Startup to go with open-source [40:14] Correct transition process [40:56] "There are no other ways but to communicate with data scientists. Your team needs to have a close loop for future priorities, what to take with you and what to leave behind." [44:51] What to use in monitoring piece [45:36] Prometheus and Grafana [48:07] Do you automatic retriggering monitoring of Models set up? [51:55] Hardware for on Prim model training [52:38] "Machine Learning model prediction is a spear bomb." [53:55] War or horror stories [54:15] "Guys, don't do context switching!" [55:54] "I won't say that Adyen is a company that allows you to make mistakes but you can make mistakes."
undefined
Mar 5, 2021 • 58min

Product Management in Machine Learning // Laszlo Sragner // MLOps Meetup #54

MLOps community meetup #54! Last Wednesday we talked to Laszlo Sragner, Founder, Hypergolic. // Abstract: How my experience in quant finance and software engineering influenced how we ran ML at a London Fintech Startup. How to solve business problems with incremental ML? What's the difference between academic and industrial ML? // Bio: Laszlo worked as a quant researcher at multiple investment managers and as a DS at the world's largest mobile gaming company. As Head of Data Science at Arkera, he drove the company's data strategy delivering solutions to Tier 1 investment banks and hedge funds. He currently runs Hypergolic (hypergolic.co.uk) an ML Consulting company helping startups and enterprises bring the maximum out of their data and ML operations. // Takeaways Continuous evaluation and monitoring is indistinguishable in a well set up product team. Separation of concerns (SE, ML, DevOps, MLOps) is very important for smooth operation, low friction team coordination/communication is key. To be able to iterate business features into models you need a modelling framework that can express these which is usually a DL package. DS-es are well motivated to go more technical because they see the rewards of it. All well run (from DS perspective) startups in my experience do the same. // Other Links Free eBook about MLPM: https://machinelearningproductmanual.com/ Lightweight MLOps Python package: https://hypergol.ml/ Blog: laszlo.substack.com ----------- Connect With Us ✌️-------------    Join our Slack community:  https://go.mlops.community/slack Follow us on Twitter:  @mlopscommunity Sign up for the next meetup:  https://go.mlops.community/register   Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Laszlo on LinkedIn: https://www.linkedin.com/in/laszlosragner/ Timestamps: [00:00] Introduction to Laszlo Sranger [02:15] Laszlo's Background [09:18] Being a Quant, then influenced, what you were doing with the Investment Banks? [12:24] Do you think this can be applied in different use cases or specific to what you are doing? [14:41] Do you have any thoughts of a potentially highly opinionated person? [16:54] Product management in Machine Learning [24:59] You have to be at a large company or you have to have a large team? [26:38] What are your thoughts on MLOps products helping with product management for ML? Is it an overreach or scope creep? [32:00] In the messy world of startups due to the big cost of an MVP for NLP is RegEx which means to user feedbacks it's incorporated by tweaking RegEx? [33:04] Does the ensemble recent models more than older models? If so, what is the decay rate of weights for older models? [35:40] Since the iterative management model is generic enough for most ML projects, which component of it can be easily generalized and tools built for version control? [36:38] Topic Extraction: What type of model do you train for that task? [52:55] Thoughts on Notebooks [53:34] "I don't hate notebooks. Let's be clear about that. I put it this way, notebooks are whiteboards. You don't want your whiteboards to be your output because it's a sketch of your solution. You want the purest solution."
undefined
Mar 2, 2021 • 1h 4min

MLOps Engineering Labs Recap // Part 2 // MLOps Coffee Sessions #31

This is a deep dive into the most recent MLOps Engineering Labs from the point of view of Team 3.   // Diagram Link:   https://github.com/dmangonakis/mlops-lab-example-yelp   --------------- ✌️Connect With Us ✌️ ------------- Join our slack community: https://go.mlops.community/slack Follow us on Twitter: @mlopscommunity Sign up for the next meetup: https://go.mlops.community/register   Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/ Connect with Laszlo on LinkedIn https://www.linkedin.com/in/laszlosragner/ Connect with Artem on LinkedIn: https://www.linkedin.com/in/artem-yushkovsky/ Connect with Paulo on LinkedIn: https://www.linkedin.com/in/paulo-maia-410874119/ Connect with Dimi on LinkedIn:

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app