

801: Merged LLMs Are Smaller And More Capable, with Arcee AI's Mark McQuade and Charles Goddard
8 snips Jul 16, 2024
Mark McQuade and Charles Goddard from Arcee AI discuss merging LLMs efficiently, using MergeKit and evolutionary algorithms. They explore commercial applications, compare MoE vs. MoA, and highlight the advantages of smaller language models. The podcast also covers the Spectrum Project for efficient training and the future of SLMs.
Chapters
Transcript
Episode notes
1 2 3 4 5 6 7 8 9
Intro
00:00 • 4min
Exploring Model Merging for Enhanced Neural Networks
03:39 • 12min
MergeKit: Empowering Organizations with Custom Language Models and Evolutionary Model Merging
15:57 • 7min
Innovative Model Merging Using Evolutionary Algorithms
23:08 • 7min
Benefits of Model Merging and Case Studies on Smaller Merged Models
30:01 • 4min
Value of Merged Large Language Models in the Enterprise
34:28 • 7min
Efficient Training of Language Models with Sparse Upcycling and MergeKit
41:16 • 23min
Evolution of Models: Small and Powerful
01:04:06 • 9min
Enhancing Language Models Through Model Merging and Innovative Training Approaches
01:13:22 • 4min