LoRA & QLoRA Fine-tuning Explained In-Depth

  Рет қаралды 23,415

Entry Point AI

Entry Point AI

5 ай бұрын

🎁 Join our Skool community: www.skool.com/entry-point-ai
In this video, I dive into how LoRA works vs full-parameter fine-tuning, explain why QLoRA is a step up, and provide an in-depth look at the LoRA-specific hyperparameters: Rank, Alpha, and Dropout.
0:26 - Why We Need Parameter-efficient Fine-tuning
1:32 - Full-parameter Fine-tuning
2:19 - LoRA Explanation
6:29 - What should Rank be?
8:04 - QLoRA and Rank Continued
11:17 - Alpha Hyperparameter
13:20 - Dropout Hyperparameter
Ready to put it into practice? Try LoRA fine-tuning at www.entrypointai.com

Пікірлер: 47
@DanielTompkinsGuitar
@DanielTompkinsGuitar 4 ай бұрын
Thanks! This is among the clearest and most concise explanations of LoRA and QLoRA. Really great job.
@steve_wk
@steve_wk 5 ай бұрын
I've watched a couple other of your videos - you're a very good teacher - thanks for doing this.
@user-os2rb3lx7h
@user-os2rb3lx7h 5 ай бұрын
I have been using thiese techniques for a while now without having a good understanding of each of the prameters. Thanks for giving a good overview of both the techniques and the papers
@naevan1
@naevan1 28 күн бұрын
I love this video man. watched it at least 3 times and came back to it before a job interview also. Please do more tutorials /explanations !
@VerdonTrigance
@VerdonTrigance 4 ай бұрын
It was incredible and very helpful video. Thank you man!
@drstrangeluv1680
@drstrangeluv1680 2 ай бұрын
I loved the explanation! Please make more such videos!
@Sonic2kDBS
@Sonic2kDBS 3 күн бұрын
Some nice details here. Keep on.
@SantoshGupta-jn1wn
@SantoshGupta-jn1wn 4 ай бұрын
great video, i think the best explanation i've seen on this, i'm also really confused about why they picked the rank and alpha that they did.
@anujlahoty8022
@anujlahoty8022 Ай бұрын
Loved the contnt! Simply explained no BS.
@varun_skywalker
@varun_skywalker 4 ай бұрын
This is really helpful, Thank you!!
@SanjaySingh-gj2kq
@SanjaySingh-gj2kq 5 ай бұрын
Good explanation of LoRA and QLoRA
@user-wr4yl7tx3w
@user-wr4yl7tx3w 2 ай бұрын
This is really well presented
@titusfx
@titusfx 3 ай бұрын
🎯 Key Takeaways for quick navigation: 00:00 🤖 *Introduction to Low Rank Adaptation (LoRA) and QLoRA* - LoRA is a parameter-efficient fine-tuning method for large language models. - Explains the need for efficient fine-tuning in the training process of large language models. 02:29 🛡️ *Challenges of Full Parameter Fine-Tuning* - Full parameter fine-tuning updates all model weights, requiring massive memory. - Limits fine-tuning to very large GPUs or GPU clusters due to memory constraints. 04:19 💼 *How LoRA Solves the Memory Problem* - LoRA tracks changes to model weights instead of directly updating all parameters. - It uses rank-one matrices to efficiently calculate weight changes. 06:11 🎯 *Choosing the Right Rank for LoRA* - Rank determines the precision of the final output table in LoRA fine-tuning. - For most tasks, rank can be set lower without sacrificing performance. 08:12 🔍 *Introduction to Quantized LoRA (QLoRA)* - QLoRA is a quantized version of LoRA that reduces model size without losing precision. - It exploits the normal distribution of parameters to achieve compression and recovery. 10:46 📈 *Hyperparameters in LoRA and QLoRA* - Discusses hyperparameters like rank, alpha, and dropout in LoRA and QLoRA. - The importance of training all layers and the relationship between alpha and rank. 13:30 🧩 *Fine-Tuning with LoRA and QLoRA in Practice* - Emphasizes the need to experiment with hyperparameters based on your specific data. - Highlights the ease of using LoRA with integrations like Replicate and Gradient.
@louisrose7823
@louisrose7823 Ай бұрын
Great video!
@SergieArizandieta
@SergieArizandieta Ай бұрын
wow I'm noobie in this field n I been testing fine-tunen my own chatbot with differents techniques, n I found a lot of stuff, but It's not commonly find a some explanation to understand the main reason of the use of it, ty a lot < 3
@stutters3772
@stutters3772 Ай бұрын
This video deserves more likes
@nafassaadat8326
@nafassaadat8326 16 күн бұрын
can we use QLoRA in a simple ML model like CNN for image classification ?
@markironmonger223
@markironmonger223 5 ай бұрын
This was wonderfully educational and very easy to follow. That either it makes you a great educator or me an idiot :P Regardless, thank you.
@EntryPointAI
@EntryPointAI 5 ай бұрын
let's both say it's the former and call it good! 🤣
@YLprime
@YLprime 2 ай бұрын
Dude u look like the lich king with those blue eyes
@practicemail3227
@practicemail3227 Ай бұрын
True. 😅 He should be in acting career ig.
@EntryPointAI
@EntryPointAI Ай бұрын
You mean Lich King looks like me I think 🤪
@RafaelPierre-vo2rq
@RafaelPierre-vo2rq 2 ай бұрын
Awesome explanation! Which camera you use?
@EntryPointAI
@EntryPointAI 2 ай бұрын
Thanks, it’s a Canon 6d Mk II
@kunalnikam9112
@kunalnikam9112 Ай бұрын
In LoRA, Wupdated = Wo + BA, where B and A are decomposed matrices with low ranks, so i wanted to ask you that what does the parameters of B and A represent like are they both the parameters of pre trained model, or both are the parameters of target dataset, or else one (B) represents pre-trained model parameters and the other (A) represents target dataset parameters, please answer as soon as possible
@EntryPointAI
@EntryPointAI 29 күн бұрын
Wo would be the original model parameters. A and B multiplied together represent the changes to the original parameters learned from your fine-tuning. So together they represent the difference between your final fine-tuned model parameters and the original model parameters. Individually A and B don't represent anything, they are just intermediate stores of data that save memory.
@kunalnikam9112
@kunalnikam9112 29 күн бұрын
@@EntryPointAI got it!! Thank you
@TheBojda
@TheBojda Ай бұрын
Nice video, congrats! LoRA is about fine-tuning, but is it possible to use it to compress the original matrices to speed up inference? I mean decompose the original model's original weight matrices to products of low-rank matrices to reduce the number of weights.
@rishiktiwari
@rishiktiwari Ай бұрын
I think you mean distillation with quantisation?
@EntryPointAI
@EntryPointAI Ай бұрын
Seems worth looking into, but I couldn't give you a definitive answer on what the pros/cons would be. Intuitively I would expect it could reduce the memory footprint but that it wouldn't be any faster.
@TheBojda
@TheBojda Ай бұрын
@@rishiktiwari Ty. I learned something new. :) If I understand well, this is a form of distillation.
@rishiktiwari
@rishiktiwari Ай бұрын
​@@TheBojdaCheers mate! Yes, in distillation there is student-teacher configuration and the student tries to be like teacher with less parameters (aka. weights). This can also be combined with quantisation to reduce memory footprint.
@ArunkumarMTamil
@ArunkumarMTamil 26 күн бұрын
how is Lora fine-tuning track changes from creating two decomposition matrix?
@EntryPointAI
@EntryPointAI 21 күн бұрын
The matrices are multiplied together and the result is the changes to the LLM's weights. It should be explained clearly in the video, it may help to rewatch.
@ArunkumarMTamil
@ArunkumarMTamil 21 күн бұрын
@EntryPointAI My understanding: Orignal weight = 10 * 10 to form a two decomposed matrices A and B let's take the rank as 1 so, The A is 10 * 1 and B is 1 * 10 total trainable parameters is A + B = 20 In Lora even without any dataset training if we simply add the A and B matrices with original matric we can improve the accuracy slighty And if we use custom dataset in Lora the custom dataset matrices will captured by A and B matrices Am I right @EntryPointAI?
@EntryPointAI
@EntryPointAI 17 күн бұрын
@@ArunkumarMTamil Trainable parameters math looks right. But these decomposed matrices will be initialized as all zeroes so adding them without any custom training dataset will have no effect.
@ecotts
@ecotts Ай бұрын
LoRa (Long Range) is a physical proprietary radio communication technique that uses a spread spectrum modulation technique derived from chirp spread spectrum. It's a low powered wireless platform that has become the de facto wireless platform of Internet of Things (IoT). Get your own acronym! 😂
@EntryPointAI
@EntryPointAI Ай бұрын
Fair - didn’t create it, just explaining it 😂
@Ian-fo9vh
@Ian-fo9vh 5 ай бұрын
Bright eyes
@vediodiary1754
@vediodiary1754 2 ай бұрын
Oh my god your eyes 😍😍😍😍everybody deserves hot teacher😂❤
@nabereon
@nabereon 3 ай бұрын
Are you trying to hypnotize us with those eyes 😜
@DrJaneLuciferian
@DrJaneLuciferian 4 ай бұрын
I wish people would actually share links to papers they reference...
@EntryPointAI
@EntryPointAI 4 ай бұрын
LoRA: arxiv.org/abs/2106.09685 QLoRA: arxiv.org/abs/2305.14314 Click "Download PDF" in top right to view the actual papers.
@DrJaneLuciferian
@DrJaneLuciferian 4 ай бұрын
@@EntryPointAI Thank you, that's kind. I did already go look it up. Sorry I was frustrated. It's very common for people to forget to putlikes to papers in show note :^)
@TR-707
@TR-707 4 ай бұрын
Ahh very interesting thank you! *goes to fine tune pictures of anime girls*
@coco-ge4xg
@coco-ge4xg 15 күн бұрын
omg I always distracted by his blue eyes😆and ignoring what his talking
LoRA explained (and a bit about precision and quantization)
17:07
Did you find it?! 🤔✨✍️ #funnyart
00:11
Artistomg
Рет қаралды 120 МЛН
ISSEI funny story😂😂😂Strange World | Magic Lips💋
00:36
ISSEI / いっせい
Рет қаралды 192 МЛН
Edge Is The Key for Long-Term Trading Success
9:33
2BETRADING
Рет қаралды 67
Finetuning LLM- LoRA And QLoRA Techniques- Krish Naik Hindi
26:54
Krish Naik Hindi
Рет қаралды 2,3 М.
Diffie-Hellman Key Exchange: How to Share a Secret
9:09
Spanning Tree
Рет қаралды 42 М.
What is Apache Kafka®?
11:42
Confluent
Рет қаралды 335 М.
AI’s Hardware Problem
16:47
Asianometry
Рет қаралды 614 М.
👎Главный МИНУС планшета Apple🍏
0:29
Demin's Lounge
Рет қаралды 469 М.
#miniphone
0:18
Miniphone
Рет қаралды 11 МЛН
AMD больше не конкурент для Intel
0:57
ITMania - Сборка ПК
Рет қаралды 481 М.
What percentage of charge is on your phone now? #entertainment
0:14
How Neuralink Works 🧠
0:28
Zack D. Films
Рет қаралды 31 МЛН