Finetune LLMs to teach them ANYTHING with Huggingface and Pytorch | Step-by-step tutorial

  Рет қаралды 19,660

Neural Breakdown with AVB

Neural Breakdown with AVB

Күн бұрын

Пікірлер: 35
@prashlovessamosa
@prashlovessamosa 3 ай бұрын
This is so top notch education great thanks 🙏👍
@avb_fj
@avb_fj 3 ай бұрын
Thanks! Glad you enjoyed it!
@saimahmed1095
@saimahmed1095 3 ай бұрын
just by seeing title make me to subscribe. Really needed Thanks bhai.
@avb_fj
@avb_fj 2 ай бұрын
Haha you are welcome bhai! Glad to have you as a sub. :)
@JaysThoughts-q5e
@JaysThoughts-q5e 3 ай бұрын
Fun video. You explained it well.
@tessdejaeghere6972
@tessdejaeghere6972 2 ай бұрын
Great explanation!
@kenchang3456
@kenchang3456 3 ай бұрын
Great consumable explanation, thanks. Nice change of glasses BTW.
@avb_fj
@avb_fj 3 ай бұрын
Haha thanks for noticing the glasses! I totally did not accidentally sit on my old pair.
@zeelthumar
@zeelthumar 3 ай бұрын
This is crazyyy man keep it up
@avb_fj
@avb_fj 3 ай бұрын
Thanks!
@PedroGabriel-jr3rb
@PedroGabriel-jr3rb 3 ай бұрын
Invaluable content, thanks!
@Urvil_Panchal
@Urvil_Panchal 3 ай бұрын
The content is Top Notch ❤‍🔥 , easy to understand. What's your linkedin bro?
@mbpiku
@mbpiku 3 ай бұрын
Thanks. Great video.
@boogati9221
@boogati9221 3 ай бұрын
Hey AVB, great video! Do you have plans on making a video for finetuning embeddings? I believe this topic is extremely useful for custom RAG pipelines.
@avb_fj
@avb_fj 3 ай бұрын
Great idea. I didn’t have plans to make this, but sounds like a super important topic. I’ll add it to my TODO list!
@boogati9221
@boogati9221 3 ай бұрын
Legit
@ratand4619
@ratand4619 3 ай бұрын
damn i just came across this video and ive subscribed already. I come from arts background but ai has got all my focus lately, and i believe your channel would be the greatest help for me and people like me. Just one view, if you could explain technical points with the simplest explaination taking simplest examples, it'll be of a great help for people like me who don't come from this background. Thousands of people woulf benefit from your amazing content.
@avb_fj
@avb_fj 3 ай бұрын
Thank you so much for your kind words! Welcome to the channel! Regarding your feedback - I think because this topic was pretty intense, I tried to keep the technical points as concise as possible. I already have some older videos that describe attention, transformers, and LLMs, in more intuitive ways, so I opted to point the viewers towards those videos depending on what they are looking for. That said, I am definitely still working on my skills of distilling down technical points to it's simplest forms, and I hope to keep on improving one video at a time!
@Rohitkumar-vg5xi
@Rohitkumar-vg5xi 6 күн бұрын
Great but i have a requirement where i have a bunch of pdfs,txt,json files etc. I want to train the model and ask questions to get the content .How can i do ?
@avb_fj
@avb_fj 6 күн бұрын
In most cases fine tuning isn’t necessary. You should look into RAG pipelines first. Basically, you extract some relevant snippets out of your files given the input query, and pass these snippets into the LLM to extract an answer. I have a video covering best practices for RAGs, I’ll leave a link here: The RAG Visual Breakdown - The Ultimate guide to building powerful LLM pipelines! kzbin.info/www/bejne/hXnLkIZ4rreMo7M
@ashmin.bhattarai
@ashmin.bhattarai 3 ай бұрын
What is the system spec (RAM / VRAM / GPU) you have used for fine tuning 1B Model?
@avb_fj
@avb_fj 3 ай бұрын
Great question. Something I should have mentioned in the video. This is MacBook Pro M2 16GB ram. If you have CUDA gpus, you could leverage better quantization. For my machine, I was able to train with batch size of 8 in float32… the sequence lengths were around 250 on average for this task. Honestly, if I were working on a real project I’ll rent GPU on the cloud and train there after prototyping locally. Since this is a YT video and it’s for education, I decided to not go into cloud servers.
@ashmin.bhattarai
@ashmin.bhattarai 3 ай бұрын
@@avb_fj If so, one could easy replicate this exact project on free colab or kaggle..
@sukantdebnath4463
@sukantdebnath4463 2 ай бұрын
Can you share the notebook as open source pls?
@PaneerPakoda-d9e
@PaneerPakoda-d9e 3 ай бұрын
Hello sir how to actually create custom tokenizers for like let's say Odia language?
@avb_fj
@avb_fj 3 ай бұрын
I would check if any of the existing open source LLMs have Oriya words/alphabets in them. If not, I’d look at things like “Byte pair encoding” and “Subword tokenizers”. Huggingface has a Tokenizers library that provides useful APIs for these. If you find a pretrained model where a fair amount of Odia words are in the vocab then that’ll be pretty great. If they are not present, then the job becomes quite difficult, coz you’d either need to add new vocab tokens into existing models, or train your own Odia LM from scratch.
@PaneerPakoda-d9e
@PaneerPakoda-d9e 3 ай бұрын
@@avb_fj thanks for the info🙏🏻
@yingpang2437
@yingpang2437 17 күн бұрын
could you share the notebook
@yingpang2437
@yingpang2437 17 күн бұрын
it is hard to follow without code.
@VarshneyKashish
@VarshneyKashish Ай бұрын
good but i found this video little confusing and many more things in a single video
@mohsenghafari7652
@mohsenghafari7652 3 ай бұрын
thanks . please share code.
@avb_fj
@avb_fj 3 ай бұрын
As mentioned in the video, all the code produced in this video and all other videos in the channel are shared on my Patreon. www.patreon.com/c/NeuralBreakdownwithAVB
@ErkanER1227
@ErkanER1227 3 ай бұрын
I wish we didn't have to pay to access the materials 😢
@avb_fj
@avb_fj 3 ай бұрын
Email me at neural.avb@gmail.com
The RAG Visual Breakdown - The Ultimate guide to building powerful LLM pipelines!
17:27
Fine-tuning Large Language Models (LLMs) | w/ Example Code
28:18
Shaw Talebi
Рет қаралды 404 М.
1% vs 100% #beatbox #tiktok
01:10
BeatboxJCOP
Рет қаралды 67 МЛН
СИНИЙ ИНЕЙ УЖЕ ВЫШЕЛ!❄️
01:01
DO$HIK
Рет қаралды 3,3 МЛН
黑天使被操控了#short #angel #clown
00:40
Super Beauty team
Рет қаралды 61 МЛН
LoRA explained (and a bit about precision and quantization)
17:07
How to Improve LLMs with RAG (Overview + Python Code)
21:41
Shaw Talebi
Рет қаралды 98 М.
From Attention to Generative Language Models - One line of code at a time!
47:09
Neural Breakdown with AVB
Рет қаралды 4,8 М.
Transformers (how LLMs work) explained visually | DL5
27:14
3Blue1Brown
Рет қаралды 4,7 МЛН
EASIEST Way to Fine-Tune a LLM and Use It With Ollama
5:18
warpdotdev
Рет қаралды 278 М.
RAG vs. Fine Tuning
8:57
IBM Technology
Рет қаралды 151 М.
QLoRA-How to Fine-tune an LLM on a Single GPU (w/ Python Code)
36:58
Complete DSPy Tutorial - Master LLM Prompt Programming in 8 amazing examples!
34:42
Neural Breakdown with AVB
Рет қаралды 29 М.