Super Data Science: ML & AI Podcast with Jon Krohn

674: Parameter-Efficient Fine-Tuning of LLMs using LoRA (Low-Rank Adaptation)

4 snips
Apr 28, 2023
Learn about Parameter-Efficient Fine-Tuning with LoRA, Atta-Lora, and optimization techniques for large language models. Discover how to reduce trainable parameters and memory usage while adapting fine-tuning in specific model sections for efficiency.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

PEFT Fixes Cost And Forgetting

  • Fine-tuning entire LLMs on small data is expensive and causes catastrophic forgetting.
  • Parameter-efficient fine-tuning (PEFT) fixes cost and preserves original capabilities while improving small-data performance.
ADVICE

Use Hugging Face PEFT Library

  • Use Hugging Face's PEFT library to implement parameter-efficient methods without deep linear algebra expertise.
  • These methods store checkpoints in megabytes and avoid retraining full model weights.
INSIGHT

LoRA's Low-Rank Trick

  • LoRA inserts low-rank decomposition matrices into transformer layers while freezing original weights.
  • This reduces trainable parameters ~10,000x and memory footprint about 3x, sometimes improving performance.
Get the Snipd Podcast app to discover more snips from this episode
Get the app