AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
Optimizing Large Language Models with Wafers in High-Performance Computing
The chapter delves into the unique characteristics of wafer technology and its applications in large language models and high-performance computing. It explores fine-tuning large models, pre-training from scratch, and benefits in computational fluid dynamics due to low latency memory access. The discussion also includes partnerships with Qualcomm for efficient deployment and ongoing research on sparsification and pruning for enhanced performance.