LessWrong (Curated & Popular) cover image

"LoRA Fine-tuning Efficiently Undoes Safety Training from Llama 2-Chat 70B" by Simon Lermen & Jeffrey Ladish.

LessWrong (Curated & Popular)

00:00

The Impact of Future Model Weight Releases and the Introduction of QloRA Technique

This chapter discusses the debate on future model weight releases and their impact. It introduces the QloRA technique, a low-rank adaptation method that reduces memory and compute requirements through quantization. The chapter also explores the effectiveness of the technique in reducing refusals from the model's outputs and compares the refusal rates of different models.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app