AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
Coffee Sessions #48 with Julien Chaumond, Tour of Upcoming Features on the Hugging Face Model Hub.
//Abstract
Julien Chaumond’s Tour of Upcoming Features on the Hugging Face Model Hub. Our MLOps community guest in this episode is Julien Chaumond the CTO of Hugging Face - every data scientist’s favorite NLP Swiss army knife.
Julien, David, and Demetrios spoke about many topics including:
Infra for hosting models/model hubs
Inference widgets for companies with CPUs & GPUs (for companies)
Auto NLP which trains models
“Infrastructure as a service”
// Bio
Julien Chaumond is Chief Technical Officer at Hugging Face, a Brooklyn and Paris-based startup working on Machine learning and Natural Language Processing, and is passionate about democratizing state-of-the-art AI/ML for everyone.
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with David on LinkedIn: https://www.linkedin.com/in/aponteanalytics/
Connect with Julien on LinkedIn: https://www.linkedin.com/in/julienchaumond/
Timestamps:
[00:00] Introduction to Julien Chaumond
[01:57] Julien's background in tech
[04:35] "I have this vision of building a community where the greatest people in AI can come together and basically invent the future of Machine Learning together."
[04:55] What is Hugging Face?
[06:17] "We have the goal of bridging the gap between research and production on actual production use cases."
[06:45] Start of open-source in Hugging Face
[07:50] Chatbox experiment (reference resolution system) - linking pronouns to the subjects of sentences
[10:20] From a project to a company
[11:46] "The goal was to explore in the beginning."
[11:57] Importance of platform
[14:25] "Transfer learning is an efficient way of Machine Learning. Providing your platform around change that people want to start from pre-trained model and fine-tune them into the specific use case is something that can be big so we built some stuff to help people do that."
[15:35] Narrowing down the scope of service to provide
[16:27] "We have some vision of what we want to build but a lot of it is the small incremental improvements that we bring to the platform. I think it's the natural way of building stuff nowadays because Machine Learning is moving so fast."
[20:00] Model Hubs
[22:37] "We're guaranteeing that we don't build anything that introduces any lagging to Hugging Face because we're using Github. You'll have that peace of mind."
[26:31] Storing model artifacts
[27:00] AWS - cache - stored to an edge location all around the globe
[28:39] Inference widgets powering
[27:17] "For each model on the model hub we try to ensure that we have the metadata about the model to be able to actually run it."
[32:11] Deploying infra function
[32:38] "Depending on the model and library, we optimize the custom containers to make sure that they run as fast as possible on the target hardware that we have."
[34:59] "Machine Learning is still pretty much hardware dependent."
[36:11] Hardware usage
[39:04] "CPU is super cheap. If you are able to run Berks served with a 1-millisecond on CPU because you have powerful optimizations, you don't really need GPUs anymore. It's cost-efficient and energy-efficient."
[40:30] Challenges of Hugging Face and what you learned
[41:10] "It may sound like a super cliche but the team that you assembled is everything."
[43:22] War stories in Hugging Face
[44:12] "Our goal is more forward-looking to be helpful as much as we can to the community."
[48:25] Hugging Face accessibility