Fine-tuning Large Language Models (LLMs) | w/ Example Code

  Рет қаралды 358,912

Shaw Talebi

Shaw Talebi

Күн бұрын

Пікірлер: 316
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
UPDATE: Someone pointed out that the fine-tuned model here is overfitting, so I created an improved example that uses transfer learning: kzbin.info/www/bejne/aoKrmGt_rZ2KjbM 👉More on LLMs: kzbin.info/aero/PLz-ep5RbHosU2hnz5ejezwaYpdMutMVB0 -- References [1] Deeplearning.ai Finetuning Large Langauge Models Short Course: www.deeplearning.ai/short-courses/finetuning-large-language-models/ [2] arXiv:2005.14165 [cs.CL] (GPT-3 Paper) [3] arXiv:2303.18223 [cs.CL] (Survey of LLMs) [4] arXiv:2203.02155 [cs.CL] (InstructGPT paper) [5] PEFT: Parameter-Efficient Fine-Tuning of Billion-Scale Models on Low-Resource Hardware: huggingface.co/blog/peft [6] arXiv:2106.09685 [cs.CL] (LoRA paper) [7] Original dataset source - Andrew L. Maas, Raymond E. Daly, Peter T. Pham, Dan Huang, Andrew Y. Ng, and Christopher Potts. 2011. Learning Word Vectors for Sentiment Analysis. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pages 142-150, Portland, Oregon, USA. Association for Computational Linguistics.
@salamsoftyt
@salamsoftyt 4 ай бұрын
@salamsoftyt
@salamsoftyt 4 ай бұрын
I
@beaux2572
@beaux2572 11 ай бұрын
Honestly the most straightforward explanation I've ever watched. Super excellent work Shaw. Thank you. It's so rare to find good communicators like you!
@ShawhinTalebi
@ShawhinTalebi 11 ай бұрын
Thanks, glad it was clear 😁
@yannicbrose
@yannicbrose 3 ай бұрын
Fully agree!
@JaishreeramCoder
@JaishreeramCoder 7 ай бұрын
You have explained this so clearly, that even a novice in NLP can understand it.
@srinivasguptha9538
@srinivasguptha9538 9 ай бұрын
One thing that really standout for me is not using Google Colab for explanation. Explaining all code without scrolling helps the audience better grasp the content as it goes with the flow without waiting for the code to execute and helps the audience to remember where the variables were defined and all. Great approach and thanks for the amazing content!
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Thanks, that's good feedback! I'll keep this in mind for future videos.
@junjieya
@junjieya 11 ай бұрын
A very clear and straightforward video explaining finetuning.
@ShawhinTalebi
@ShawhinTalebi 11 ай бұрын
Glad it was clear :)
@lukaboljevicboljevic
@lukaboljevicboljevic 9 ай бұрын
Such a great video. This is the first one I watched from you. You explain everything so nicely, and in my opinion you provided just the right amount of information - not too little, so it doesn't feel superficial and you feel like you've learned something, but not too much, so that you can take what you've learned and read more about it yourself if you're interested. Looking forward to seeing more of your content!
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Great to hear! Glad it was clear :)
@balubalaji9956
@balubalaji9956 4 ай бұрын
Hey KZbin algorithm , I loved this video . suggest me more of them
@ShawhinTalebi
@ShawhinTalebi 4 ай бұрын
Here's the series playlist: kzbin.info/aero/PLz-ep5RbHosU2hnz5ejezwaYpdMutMVB0
@BRK326
@BRK326 Ай бұрын
Nice job, efficient , straight forward, enough detail for learning big picture without confusion.
@egemenklc2515
@egemenklc2515 3 ай бұрын
Hey KZbin, I really liked this kind of machine learning and fine tuning topics. Please recommend me more of these.
@ShawhinTalebi
@ShawhinTalebi 3 ай бұрын
Glad you liked it :)
@checkdgt
@checkdgt 9 ай бұрын
Just came to this video from HF and I have to say, I love they way you describe this! Thanks for the great video!
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Great to hear! Thanks for watching :)
@Josia-p5m
@Josia-p5m 2 ай бұрын
You're a fantastic communicator - this was very helpful. I would love to see more walkthroughs like this that have a good balance of theory, math and python.
@ShawhinTalebi
@ShawhinTalebi 2 ай бұрын
More to come!
@fakharmursaleen9889
@fakharmursaleen9889 3 ай бұрын
Hands down one of the most best explanations on youtube keep it up homie
@ayyanarjayabalan
@ayyanarjayabalan 6 ай бұрын
Excellent way of teaching. Keep doing this kind of good work.
@jonathanleroy269
@jonathanleroy269 2 ай бұрын
This video series is underrated. Loved it, thank you!
@soudaminipanda
@soudaminipanda 5 ай бұрын
Thanks for these great videos. I really love the fact that you focus on building an intuitive understanding as opposed to throwing jargons. Could you please start a langchain series?
@ShawhinTalebi
@ShawhinTalebi 5 ай бұрын
Great suggestion! I'll add that to my list. In the meantime, I have a code tutorial using LangChain here: kzbin.info/www/bejne/ZpTJaKmwgsSXkJIsi=Iyc1GzQ7g-A_n5Rr&t=1359
@Random-bq8qc
@Random-bq8qc 5 ай бұрын
Thanks for the well defined video because it helped me prepare my proposal related to this topic.
@sakarsalunke3245
@sakarsalunke3245 Ай бұрын
Indepth overview of fine tuning!
@saraesshaimi
@saraesshaimi 7 ай бұрын
excellent simple explanation to the point. Love it !
@Bboreal88
@Bboreal88 9 ай бұрын
My next question after this video would be on how to pack this fine-tuned model into a UI and deploy.
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Great question. I discussed how to create a chat interface with Hugging Face + Gradio in a previous video: kzbin.info/www/bejne/oJLRYWqdodSUiMk
@AbdulademAljamel-n5d
@AbdulademAljamel-n5d Жыл бұрын
I was struggling to understand some details, before this video, thanks a lot
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Great to hear. I’m glad it helped!
@alokkale2718
@alokkale2718 Ай бұрын
Bro you are the best instructor
@ShawhinTalebi
@ShawhinTalebi Ай бұрын
Thank you, glad it was clear 🙏
@yoffel2196
@yoffel2196 Жыл бұрын
Wow dude, just you wait, this channel is gonna go viral! You explain everything so clearly, wish you led the courses at my university.
@ShawhinTalebi
@ShawhinTalebi 11 ай бұрын
Thanks for the kind words! Maybe one day 😉
@sreeramch
@sreeramch 9 ай бұрын
Thank you for the detailed explaination line by line. Finally a place, I can rely on with working example
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Glad it was helpful!
@melliott117
@melliott117 8 ай бұрын
Really great content. I love your balance of details and overview. It’s made it easy for me as a newcomer who is interested in details. My only criticism/advice is that you edit to remove silence. This is great for minimizing pauses mid sentence. But it would be helpful to have slightly more time at the end of each thought/point. Pausing for that extra 0.25 seconds at the end of a coherent teaching point helps greatly.
@ShawhinTalebi
@ShawhinTalebi 8 ай бұрын
Thanks, that's good feedback! I do get a bit heavy-handed with the edits 😅
@EigenA
@EigenA 8 ай бұрын
Great video, I wanted to hear further discussion on mitigation techniques for overfitting. Thanks for making the video!
@thehousehusbandcn5074
@thehousehusbandcn5074 11 ай бұрын
You are the man! No BS, just good useful info
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Thanks, glad it was helpful 😁
@azizhassouna9919
@azizhassouna9919 6 ай бұрын
Im really gratful for youre work , you really help me when I had no one to ask .
@scifithoughts3611
@scifithoughts3611 10 ай бұрын
Great video Shaw! It was a good balance between details and concepts. Very unusual to see this so well done. Thank you.
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Glad you enjoyed it!
@arunshrestha791
@arunshrestha791 7 ай бұрын
Clear Explanation, Amazing
@rma1563
@rma1563 4 ай бұрын
I needed to know how parameter efficient finetuning works to finetune a voice encoder for emotion detection task. This video helped me a lot. I used LoRA for it. Thanks ❤
@ShawhinTalebi
@ShawhinTalebi 4 ай бұрын
Glad it was helpful!
@alikarooni9713
@alikarooni9713 10 ай бұрын
Even though this was high level instruction, it was perfect. I can continue from here. Thanks Shahin jan!
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Glad it helped!
@MannyBernabe
@MannyBernabe 9 ай бұрын
Excellent walk-thru. Thank you, Shaw!I was getting errors on the new model. Switching the device worked for me. # Check if CUDA is available and set the device accordingly device = 'cuda' if torch.cuda.is_available() else 'cpu' model.to(device) # Move the model to the appropriate device (GPU or CPU)
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Thanks Manny! That's a good note, I wasn't able to test the code on a non-Mac machine.
@alex70301
@alex70301 11 ай бұрын
Best video on llm fine tuning. Very concise and informative.
@ShawhinTalebi
@ShawhinTalebi 11 ай бұрын
Thanks! Glad you liked it :)
@dendi1076
@dendi1076 10 ай бұрын
this channel is going to hit 6 figure subscribers at this rate
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
I hope so 😅
@lotusluthor
@lotusluthor 6 ай бұрын
Finally something that I can grok!!!! Fantastic tutorial!
@salmaelbarbori579
@salmaelbarbori579 9 ай бұрын
Clear and straightforward to the point, thanks a lot for making this valuable content accessible on ytb💡
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Happy to help!
@bruceyin603
@bruceyin603 2 ай бұрын
Very good explain about fine tune, thanks Shaw
@janjiavicii1328
@janjiavicii1328 2 ай бұрын
very explicit explanation. Thanks!
@SaharAbo-y5d
@SaharAbo-y5d 6 ай бұрын
Your style of conveying information is wonderful. Good luck to you
@rubencabrera8519
@rubencabrera8519 Жыл бұрын
This was one of the best videos on this topic, really nice man, keep going.
@ShawhinTalebi
@ShawhinTalebi 11 ай бұрын
Thanks! Glad it was clear :)
@saadati
@saadati Жыл бұрын
Amazing video Shawhin. It was quite easy to follow and stuff were clearly explained. Thank you so much,
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Thanks! I'm glad it was clear and helpful
@Throwingness
@Throwingness 10 ай бұрын
Very good. Very fast and also easy to follow. As far as future content, keep us posted about how to do LoRA on quantized models. How can the future be anything but LoRA on quantized models?!?!?!?
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Thanks, glad you liked it. Video coming this quarter on exactly that!
@ajeethsuryash5123
@ajeethsuryash5123 4 ай бұрын
Very informative video. Thanks for sharing
@mookiejapan7351
@mookiejapan7351 9 ай бұрын
Wow! Amazing make-up! If it wasn't for the voice, I wouldn't believe this is actually David Cross!
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Haha, I was wearing jean shorts while filming this 😂
@Akshatgiri
@Akshatgiri 9 ай бұрын
This is gonna come handy. Thanks for breaking it down
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Happy to help!
@richardpinter9218
@richardpinter9218 Жыл бұрын
Fantastic video. Thanks for the upload. Keep up the good work, you're awesome 😎
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Thanks, I’m glad you liked it 😁
@swaykao1703
@swaykao1703 16 күн бұрын
thank you! I like your vedios, they help me a lot!
@hoseinmirhoseini6013
@hoseinmirhoseini6013 3 ай бұрын
Thank you for the great content.
@arthurs6405
@arthurs6405 8 ай бұрын
This was beautifully described. I wish you had provided a Linux alternative for the "model.to('mps/cpu'). I have a linux workstation and a p100 gpu. Also, you did not include the means to save your newly trained model. I think most of us students would appreciate knowing how to save the model locally and to huggingface. Thanks for your efforts.
@ShawhinTalebi
@ShawhinTalebi 8 ай бұрын
I do fine-tuning on a Linux machine here: kzbin.info/www/bejne/jqHSfHVpqtiggcU
@jasoncole3253
@jasoncole3253 Жыл бұрын
Well done, even if I already knew all this shit it was really nice to listen to your clear explanation
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
lol! Glad you enjoyed it :)
@RealNikolaus
@RealNikolaus Жыл бұрын
Very good video and explanation!
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Glad it helped!
@totalcooljeff
@totalcooljeff Жыл бұрын
Random question i how do you edit you audio clips together to make them so seamless because idk where to mate them. And great video by the way 👍
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
I use iMovie :)
@upadisetty
@upadisetty 9 ай бұрын
Best video i saw. thanks a ton for sharing. glad i found right place
@zeusgamer5860
@zeusgamer5860 Жыл бұрын
HI Shaw, amazing video - very nicely explained! Would be great if you could also do a video (with code examples) for Retrieval Augmented Generation as an alternative to fine-tuning :)
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Great suggestion. I have a few follow-up use cases planned out and RAG will definitely be part of it.
@BamiCake
@BamiCake 11 ай бұрын
​@@ShawhinTalebimaybe also how to fine tune openai model too?
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Just dropped! kzbin.info/www/bejne/aoOkp32qaMuKpas
@tgyawali
@tgyawali 9 ай бұрын
I found you in youtube just today. Your presentation style, quality of content is very good. Keep up the great work. I am very passionate about AI technology in general, have been trying to conduct basic trainings to undergraduate college students and would love to connect to collaborate if you are interested. Thank you for doing this!
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Thanks for watching! Glad it was clear :) Feel free to set up a call if you like: calendly.com/shawhintalebi
@tgyawali
@tgyawali 9 ай бұрын
@@ShawhinTalebi Thank you. I will set up some time to connect.
@Wei-KuoLi
@Wei-KuoLi 7 ай бұрын
Very good & simple showcase, thanks
@amparoconsuelo9451
@amparoconsuelo9451 Жыл бұрын
Understood. The codes were very helpful. They were not constantly scrolling and panning. But please display the full code and mention the Python version and system configuration, including folders, etc.
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Good to hear! All code and env files are available here: github.com/ShawhinT/KZbin-Blog/tree/main/LLMs/fine-tuning
@Mastin70
@Mastin70 7 ай бұрын
Fantastic explanation.
@rbrowne4255
@rbrowne4255 Жыл бұрын
Fantastic job on this overview, as for other videos, I don't see many videos on Inference scaling, i.e requirements for concurrency, latency etc...what are the hardware requirements i.e number of GPUs per systems or number of systems, etc
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
I'm glad it was helpful :) That's is a great suggestion. I will add it to my list. Thank you!
@kevon217
@kevon217 10 ай бұрын
Excellent walkthrough
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
🙏
@simplyshorts748
@simplyshorts748 8 ай бұрын
Great video! I love good explainations
@raffelravionaldo463
@raffelravionaldo463 2 ай бұрын
Thank you, is a nice video and you have a clear explanation, actually I try to do this with GPT Neo model (EleutherAI/gpt-neo-1.3B) but when I do training, the Training Loss always have no log values and Validation Loss always NaN (when do it with BERT or distilbert, is run perfectly), Do you have any suggestions or reading resources to fix this?
@ShawhinTalebi
@ShawhinTalebi 2 ай бұрын
Hard to say without seeing the code. Perhaps you can share more info here: www.shawhintalebi.com/contact
@ITforGood
@ITforGood 11 ай бұрын
Thanks Shaw, very helpful.
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Glad it was helpful!
@Kevin.Kawchak
@Kevin.Kawchak 7 ай бұрын
Thank you for the discussion
@yb3134
@yb3134 9 ай бұрын
Very well explained
@pawan3133
@pawan3133 7 ай бұрын
Thanks for the beautifully explanation!! When you said, for PEFT "we augment the model with additional parameters that are trainable", how do we add these parameters exactly? Do we add a new layer? Also, when we say "%trainable parameters out of total parameters", doesn't that mean that we are updating a certain % of original parameters?
@chilupuriharshitha7261
@chilupuriharshitha7261 2 ай бұрын
Hey, It was good but did your Model also take a lot of time to get fine-tuned while you were Pluging everything to the Trainer Class as for me it's taking nearly i lost the time it is taking but 10 epochs are getting trained for every 0.05 it/s.
@ShawhinTalebi
@ShawhinTalebi 2 ай бұрын
Are you running this locally?
@younespiro
@younespiro Жыл бұрын
amazing video, very well explained
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Glad it helped! Thanks :)
@bitschips
@bitschips 9 ай бұрын
So educative, thanks a lot!
@PrinceSinghal-j8m
@PrinceSinghal-j8m 9 ай бұрын
i was amazing ....thanks for uploading Shaw
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Thanks, happy to help!
@payam-bagheri
@payam-bagheri Жыл бұрын
Great video, Shawhin!
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Thanks, glad you enjoyed it!
@machireddyshyamsunder987
@machireddyshyamsunder987 9 ай бұрын
Thankyou very much it is really very useful .
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
Happy to help!
@vicenteenriquemachacaarced2103
@vicenteenriquemachacaarced2103 6 ай бұрын
Thank you so much for that video. ¿From where did you get TRAINING loss metrics? In the console and in the trainer_state.json, I only see evaluation metrics.
@ShawhinTalebi
@ShawhinTalebi 6 ай бұрын
Great question. The training loss is predefined as a property of the base model, so no need to redefine that explicitly.
@ramp2011
@ramp2011 10 ай бұрын
Excellent..... Thank you for sharing
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
My pleasure, glad you liked it!
@evan7306
@evan7306 6 ай бұрын
Thank you for your great tutorial! What I don't understand is how to use the fine tuned model as an API so we can use it on website. Do you have any tutorial about that?
@ShawhinTalebi
@ShawhinTalebi 6 ай бұрын
Great question. I haven't covered that yet, but a common approach is to containerize the model using Docker and equipping it with an API using a library like FastAPI. Then you can host that on a local server or via a cloud provider.
@goinsgroove
@goinsgroove 4 ай бұрын
Loving the info! Do you have a video on self-supervised training? I want to train a llm to write in my style.
@ShawhinTalebi
@ShawhinTalebi 4 ай бұрын
Not yet! But I think I'll do this for my next video because of your comment :)
@goinsgroove
@goinsgroove 4 ай бұрын
@@ShawhinTalebi AWSOME!
@tintumarygeorge9309
@tintumarygeorge9309 Жыл бұрын
Thank you, Keep up the good work
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Thanks, happy to help!
@heatherbrm
@heatherbrm 7 ай бұрын
here, you earned this: 👑
@ShawhinTalebi
@ShawhinTalebi 7 ай бұрын
Thanks 🤴
@FalahgsGate
@FalahgsGate 6 ай бұрын
thanks for sharing this video ....but Is the technique of fine-tuning model for the custom dataset better than the technique of using the RAG system for LLM Apps. or reverse
@ShawhinTalebi
@ShawhinTalebi 6 ай бұрын
While the best choice of technique will depend on the specific use case, here's the intuition I've gathered. RAG is great when you want the AI to have access to specific knowledge. Fine-tuning is great when you want the AI to provide responses in a particular format and style. Additionally, the techniques are not mutually exclusive, so they can be combined to potentially provide even better results.
@lucycaffrey-maffei6666
@lucycaffrey-maffei6666 2 ай бұрын
thank you sooo much for this content, it's so helpful and clear!!! if i don't use LoRa when fine-tuning and instead just specify `Trainer(model=AutoModelForSequenceClassification.from_pretrained( model_checkpoint, num_labels=2)` (which is what huggingface does in most of its documentation), what type of parameter training is that performing? retraining all of the parameters?
@ShawhinTalebi
@ShawhinTalebi 2 ай бұрын
Yes, that will do full retraining!
@keithhickman7399
@keithhickman7399 Жыл бұрын
Shaw, terrific job explaining very complicated ideas in an approachable way! One question - are there downsides to combining some of the approaches you mentioned, say, prompt engineering + fine-tuning + RAG to optimize output...how would that compare to using one of the larger OOTB LLMs with hundreds of billions of params?
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Great question. The biggest consideration is the tradeoff between cost and performance. On one side you can use an LLM OOTB (e.g. ChatGPT) which costs nothing and has some baseline performance. One the other side you can build a custom system using all the bells and whistles (e.g. fine-tuning, PE, and RAG) which will likely perform much better than ChatGPT but comes at significantly greater cost. Hope that helps!
@Bboreal88
@Bboreal88 9 ай бұрын
This feature could already be available on KZbin for creators. Perhaps, you could refine a chatbot that can automatically respond to comments using Gemini. It could even learn to respond based on your videos, eliminating the need for you to upload anything or messing with fine-tuning.
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
It is to some extent, as we get response recommendations in the creator studio. Using multimodal models might takes this to the next level!
@jijie133
@jijie133 6 ай бұрын
Great video. Thank you.
@RocktCityTim
@RocktCityTim 10 ай бұрын
Hi Shaw - this answered so many questions about specializing an LLM in concise terms, thanks! One question that I'm running up against is physical machine abilities (CPU Speed/Cores, System Memory, GPU cores and memory, and storage speeds. In my case, I have a 32/64 core/thread Epyc CPU on PCIE4.0 MB with 128GB of DDR4 RAM and a PNY/NVIDIA RTX A5000 with 24GB DDR5 VRAM and 8192 CUDA cores dedicated to ML/AI (video is via a separate RTX A2000 GPU). With that info, what should I be looking at as a starting point that will take full advantage of those specs in local mode?
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Wow that's a lot of firepower. While I'm less knowledgeable about the ML engineering side of things, I'd suggest checking out DeepSpeed: github.com/microsoft/DeepSpeed. They have several resources on training/running large models efficiently.
@RocktCityTim
@RocktCityTim 10 ай бұрын
@@ShawhinTalebi Thatnks for the pointer. And thinks for all of your output. I've picked up some great information.
@liaoyixu6882
@liaoyixu6882 4 ай бұрын
Very amazing video!!! I have one question: when I use your code to fine-tune the model with my own dataset, but since my dataset is too large it leads to memory error (not gpu memory) when I read the dataset, what should I do to avoid this issue? Can I read and fine-tune in a small batch?
@ShawhinTalebi
@ShawhinTalebi 4 ай бұрын
You can try reducing the batch size. Also happy to help troubleshoot via office hours: calendly.com/shawhintalebi/office-hours
@naehalmulazim
@naehalmulazim 7 ай бұрын
Greetings! Really nice tutorial! THANK YOU for including Lora! I need to train an Llm on a higher level language we wrote in C++, to produce our code. It's all private infrastructure. Time isnt an issue but I'd like to do it locally on a mac m2 if I can and was considering Lora on a tiny llm. Is this going to be possible?
@ShawhinTalebi
@ShawhinTalebi 7 ай бұрын
While I haven't done that myself, that is surely possible. The challenge I've run into is that many open-source models don't work so easily on Mac, but I plan to figure it out and many video about it.
@ShawhinTalebi
@ShawhinTalebi 3 ай бұрын
Update: I got a fine-tuning example working locally on my M1 Mac Mini :) Here's a similar example running on my Mac: kzbin.info/www/bejne/aYGsopuah9-brqc
@samadhanpawar6554
@samadhanpawar6554 Жыл бұрын
Can you recommend any course where i can learn to build llm from scratch and fine-tune in depth
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Paul Iusztin has some good content on that. Hands-on-llms: github.com/iusztinpaul/hands-on-llms More resources: www.pauliusztin.me/
@harshanaru1501
@harshanaru1501 11 ай бұрын
Such a great video ! Wondering how self supervised fine tuning works. Is there any video available on that ?
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Thanks! I found this on self-supervised fine-tuning: kzbin.info/www/bejne/i6SYi3mmZpuDl7c
@naevan1
@naevan1 Жыл бұрын
Hey dude nice video. I think I'll try to find tuned Lamma to detect phrases and subsequently classify tweets - but multiclass classification. Hope it works ,I guess I'll transfer the csv to the prompt you mentioned like alpaca was done and see if it works
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Thanks! Sounds like a fun project :)
@devtest202
@devtest202 8 ай бұрын
Hi thanks!! A question for a model in which I have more than 2,000 pdfs. Do you recommend improving the handling of vector databases? When do you recommend fine tunning and when do you recommend vector database
@ShawhinTalebi
@ShawhinTalebi 8 ай бұрын
Great question! Generally, fine-tuning and RAG have different strengths. Fine-tuning is great when you want to endow the model with a particular style or to tailor completions for a particular use case, while RAG is good to provide the model with specialized and specific knowledge.
@xugefu
@xugefu 6 ай бұрын
Thanks!
@ShawhinTalebi
@ShawhinTalebi 6 ай бұрын
Thanks for the support! Glad it was helpful :)
@elrecreoadan878
@elrecreoadan878 Жыл бұрын
Would a botpress with a vector kb connected to chatgpt would be enough for Q&A ? When fine tuning starts to be neededvand is there an inexpensive way to do it with no or low code? Thank you!
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
This depends on the use case. However, taking a quick-and-easy no code approach to start is never a bad idea. It typically gives you a sense of how sophisticated approaches will pan out. Fine-tuning will come into play when the "quick-and-easy" starts to becomes too inconvenient (or expensive) due to the scale of the solution. Hope that helps!
@NateKrueger805
@NateKrueger805 11 ай бұрын
Nicely done!
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Thanks!
@KaptainLuis
@KaptainLuis 11 ай бұрын
So nice video thank you soooo much!!❤
@ShawhinTalebi
@ShawhinTalebi 10 ай бұрын
Happy to help 😁
@amanpreetsingh8100
@amanpreetsingh8100 Жыл бұрын
This was a great video. I have one question though. In the LoRA demonstration in your video(at ~14 minutes) you mention this operation (W0 + BA)x = h(x), in this how the sum (W0 + BA) is possible, as W0 has dimentions d*k, and output of operation BA would have the dimentions r*r. This matrix sum is not mathematiaclly possible. So can you elaborate more on this...
@ShawhinTalebi
@ShawhinTalebi Жыл бұрын
Good question! The math works out here because B is d x r and A is r x k, therefore BA will be d x k.
@amanpreetsingh8100
@amanpreetsingh8100 Жыл бұрын
@@ShawhinTalebi 👍
@SolomonManuelraj
@SolomonManuelraj 11 ай бұрын
Nice Video. I need your help to clarify my doubt. When we do the PEFT based finetuning, the final finetuned model size (in KBs/GBs) will increase by the additional parameters ( base model size + additional parameters size) . In this case base model size will be lesser and final finetuned model size will be more. Deploying the final finetuned model in the edge devices will be more difficult because of the limited edge device resources. Are there any way adapters / LoRA can help in reducing the final finetuned model memory size so that easily we can deploy the final model in the edge devices? Your insights will be helpful. Currently i am working in the vision foundation model deployment in the edge device where i am finding it difficult to deploy because of vision foundation model memory size and inference speed.
@ShawhinTalebi
@ShawhinTalebi 11 ай бұрын
Great question. PEFT methods like LoRA only reduce the number of trainable parameters not the total number of parameters. And to your point, the storage requirements actually increase in the case of LoRA! To reduce the final model size, you will need to fine-tune a smaller base model. Hope that helps!
@aldotanca9430
@aldotanca9430 11 ай бұрын
Very clear, thanks!
@ShawhinTalebi
@ShawhinTalebi 11 ай бұрын
Thanks Aldo!
@nandkumarghatge5152
@nandkumarghatge5152 3 ай бұрын
great video.
@vitola1111
@vitola1111 9 ай бұрын
Great video! Is the process for fine tuning a stable diffusion model the same? I think if you make a vid on that itd get a lot of views as well.
@ShawhinTalebi
@ShawhinTalebi 9 ай бұрын
I haven't worked with stable diffusion models before, so I don't now, but that would be a great video. Thanks for the suggestion!
@charismaowojoameh7681
@charismaowojoameh7681 7 ай бұрын
When trying to create a Ai model that generates airticle for a particular niche, is it best to gather airtcle on that niche and Fine-tune it or use open ai knowledge base just giving it some prompts.
@ShawhinTalebi
@ShawhinTalebi 7 ай бұрын
Good question. This depends how you are trying to generate the article. If you have a clear structure for how the articles should be written, you can go far with an off-the-shelf model + RAG. However, if the article format is not so rigid (but you have lots of examples), fine-tuning may work best.
How to Build an LLM from Scratch | An Overview
35:45
Shaw Talebi
Рет қаралды 270 М.
How to Improve LLMs with RAG (Overview + Python Code)
21:41
Shaw Talebi
Рет қаралды 77 М.
When Cucumbers Meet PVC Pipe The Results Are Wild! 🤭
00:44
Crafty Buddy
Рет қаралды 58 МЛН
كم بصير عمركم عام ٢٠٢٥😍 #shorts #hasanandnour
00:27
hasan and nour shorts
Рет қаралды 10 МЛН
Players vs Pitch 🤯
00:26
LE FOOT EN VIDÉO
Рет қаралды 135 МЛН
Каха и лужа  #непосредственнокаха
00:15
Fine-Tuning BERT for Text Classification (Python Code)
23:24
Shaw Talebi
Рет қаралды 5 М.
RAG vs Fine Tuning
8:21
Cloud Journeys
Рет қаралды 1,5 М.
RAG vs. Fine Tuning
8:57
IBM Technology
Рет қаралды 62 М.
EASIEST Way to Fine-Tune LLAMA-3.2 and Run it in Ollama
17:36
Prompt Engineering
Рет қаралды 46 М.
"okay, but I want Llama 3 for my specific use case" - Here's how
24:20
QLoRA-How to Fine-tune an LLM on a Single GPU (w/ Python Code)
36:58
Transformers (how LLMs work) explained visually | DL5
27:14
3Blue1Brown
Рет қаралды 3,7 МЛН
Training Your Own AI Model Is Not As Hard As You (Probably) Think
10:24
Steve (Builder.io)
Рет қаралды 602 М.
When Cucumbers Meet PVC Pipe The Results Are Wild! 🤭
00:44
Crafty Buddy
Рет қаралды 58 МЛН