AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
Adapters in Fine-Tuning Pre-Trained Transformers
This chapter discusses the concept of adapters in fine-tuning pre-trained transformers. It explains how adapters, small matrices of weights inserted into the linear layers of the transformer, can reduce the number of parameters and make the process more memory-efficient and faster. It also touches upon the challenges of dealing with large models and introduces the concept of proximal policy optimization in reinforcement learning.