AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
Using Smaller Micro LLMs and Minimizing Costs
The speakers discuss the use of smaller micro LLMs and the computational requirements involved. They also talk about minimizing costs by utilizing CPUs for running models and GPUs for training, running on Kubernetes and multi-cloud/on-premises solutions.