Рет қаралды 2,532
In this sequel to my previous video ( • Summarizing legal docu... ), I delve into optimizing the fine-tuning of a Google FLAN-T5 model for legal text summarization. The focus is on employing QLoRA for parameter-efficient fine-tuning. All it takes is a few extra lines of simple code in your existing script.
This methodology allows us to train the model with remarkable cost efficiency, utilizing even modest GPU instances, which I demonstrate on AWS with Amazon SageMaker. Tune in for a detailed exploration of the technical nuances behind this process.
⭐️⭐️⭐️ Don't forget to subscribe to be notified of future videos ⭐️⭐️⭐️
- Original model: huggingface.co/google/flan-t5...
- LoRA model: huggingface.co/juliensimon/fl...
- Dataset: huggingface.co/datasets/billsum
- Notebook: gitlab.com/juliensimon/huggin...
Follow me on Medium at / julsimon or Substack at julsimon.substack.com.