
"LoRA Fine-tuning Efficiently Undoes Safety Training from Llama 2-Chat 70B" by Simon Lermen & Jeffrey Ladish.
LessWrong (Curated & Popular)
00:00
The Impact of Future Model Weight Releases and the Introduction of QloRA Technique
This chapter discusses the debate on future model weight releases and their impact. It introduces the QloRA technique, a low-rank adaptation method that reduces memory and compute requirements through quantization. The chapter also explores the effectiveness of the technique in reducing refusals from the model's outputs and compares the refusal rates of different models.
Transcript
Play full episode