Encoder-Only Transformers (like BERT) for RAG, Clearly Explained!!!

  Рет қаралды 19,849

StatQuest with Josh Starmer

StatQuest with Josh Starmer

Күн бұрын

Пікірлер: 68
@statquest
@statquest Ай бұрын
Support StatQuest by buying my books The StatQuest Illustrated Guide to Machine Learning, The StatQuest Illustrated Guide to Neural Networks and AI, or a Study Guide or Merch!!! statquest.org/statquest-store/
@tuananhvo9506
@tuananhvo9506 Ай бұрын
Complete this new one I may have been roughly watching all of the videos of StatQuest already. Deeply invested in the channel for the last few months, I feel much more confident in my quest to get the first AI related job. Massive thanks Josh for relentlessly bringing the right intuition for the mass of us!!
@statquest
@statquest Ай бұрын
Good luck with that first job!
@PradeepKumar-hi8mr
@PradeepKumar-hi8mr Ай бұрын
Wowww! Glad to have you back, Sir. Awesome videos 🎉
@statquest
@statquest Ай бұрын
Thank you!
@NottoriousGG
@NottoriousGG Ай бұрын
Such a cleverly disguised master of the craft. 🙇
@statquest
@statquest Ай бұрын
bam! :)
@Kimgeem
@Kimgeem Ай бұрын
So excited to watch this later 🤩✨
@statquest
@statquest Ай бұрын
future bam! :)
@mbeugelisall
@mbeugelisall Ай бұрын
Just the thing I’m learning about right now!
@statquest
@statquest Ай бұрын
bam! :)
@nossonweissman
@nossonweissman Ай бұрын
Yay!!! ❤❤ I'm starting it now and saving to remember to finish later. Also, I'm requesting a video on Sparse AutoEncoders (used in Anthropic's recent research). They seem super cool and I have a basic idea on how they work, but I'd to see a "simply explained" version of them.
@statquest
@statquest Ай бұрын
Thanks Nosson! I'll keep that topic in mimd.
@free_thinker4958
@free_thinker4958 Ай бұрын
You're the man ❤️💯👏 thanks for everything you do here to spread that precious knowledge 🌹 we hope if you could possibly dedicate a future video to talk about multimodal models (text to speech, speech to speech etc...) ✨
@statquest
@statquest Ай бұрын
I'll keep that in mind!
@davidlu1003
@davidlu1003 Ай бұрын
And thx for the courses. They are great!!!!😁😁😁
@statquest
@statquest Ай бұрын
Glad you like them!
@tcsi_
@tcsi_ Ай бұрын
100th Machine Learning Video 🎉🎉🎉
@statquest
@statquest Ай бұрын
Yes! :)
@THEMATT222
@THEMATT222 Ай бұрын
Noice 👍 Doice 👍Ice 👍
@davidlu1003
@davidlu1003 Ай бұрын
I love you, I will keep going and learn the other courses of yours if they are always free. keep them free please, I will always be your fan.😁😁😁
@statquest
@statquest Ай бұрын
Thank you, I will!
@kamal9294
@kamal9294 Ай бұрын
Nice explanation, if the next topic is about rag or reinforcement learning , i will be happier (or even object detection, object tracking).
@statquest
@statquest Ай бұрын
I guess you didn't get to 16:19 where I explain how RAG works...
@kamal9294
@kamal9294 Ай бұрын
@statquest bro but in LinkedIn I saw many rag types and some retrieval techniques using advanced dsa(like HNSW). That's why I asked.
@statquest
@statquest Ай бұрын
@@kamal9294 Those are just optimizations, which will change every month. However, the fundamental concepts will stay the same and are described in this video.
@kamal9294
@kamal9294 Ай бұрын
@@statquest Now I am clear, thank you!.
@thegimel
@thegimel Ай бұрын
Great instructional video, as always, StatQuest! You mentioned in the video that the training task for these networks is next word prediction, however, models like BERT have only self-attention layers so they have "bidirectional awareness". They are usually trained on masked language modeling and next sentence prediction, if I recall correctly?
@statquest
@statquest Ай бұрын
I cover how a very basic word embedding model might be trained in order to illustrate its limitations - that it doesn't take position into account. However, the video does not discuss how an encoder-only transformer is trained. That said, you are correct, an encoder-only transformer uses masked language modeling.
@etgaming6063
@etgaming6063 Ай бұрын
This video came just in time, trying to make my own RoBERTa model and have been struggling understanding how they work under the hood. Not anymore!
@statquest
@statquest Ай бұрын
BAM!
@swarupdas8043
@swarupdas8043 16 күн бұрын
What can be better to learn ML when we have a teacher like you. Thanks for all the effort you have put into. I would buy if you have any Udemy courses covering ML stuff. Please let me know
@statquest
@statquest 15 күн бұрын
I have a book coming out in the next few weeks about all these neural network videos with Pytorch tutorials
@iamumairjaffer
@iamumairjaffer Ай бұрын
Well explained ❤❤❤
@statquest
@statquest Ай бұрын
Thanks!
@rishidixit7939
@rishidixit7939 29 күн бұрын
Very Beautifully Explained as Always. It takes a great amount of intuitive understanding and talent to explain a relatively tougher topic in such an easy way. I just had some doubts - 1. In case of context aware embeddings of a Sentence of a Doc are the individual Embeddings of the tokens averaged. Does this have something to do with the CLS token ? 2. Like a Variational Autoencoder helps in understanding the intricate patterns of images and then creates its own latent space , can BERT (or any similar model) do that for Vision task (or are they only suitable for NLP Tasks) 3. Are Knowledge Graphs made using BERT ? Any help on these will be appreciated . Thank You again for the Awesome Explanation
@statquest
@statquest 29 күн бұрын
1. The CLS token is specifically used for classification problems and I talk about how it works in my upcoming book. That said, if you embed a whole sentence, then you can average the output values. 2. Transformers work great with images and image classificaiton. 3. I don't know.
@tonym4926
@tonym4926 Ай бұрын
Are you planning to add this video to neutral network/ deep learning playlist?
@statquest
@statquest Ай бұрын
yes! Just did.
@aryasunil9041
@aryasunil9041 Ай бұрын
Great Video, When is the Neural Networks book coming out? Very eager for it
@statquest
@statquest Ай бұрын
Early january. Bam! :)
@barackobama7757
@barackobama7757 Ай бұрын
Hello StatQuest. I was hoping if you could make a video on PSO (Particle Swarm Optimisation) Will really help! Thank you, amazing videos as always!
@statquest
@statquest Ай бұрын
I'll keep that in mind.
@draziraphale
@draziraphale Ай бұрын
Great explanation
@statquest
@statquest Ай бұрын
Thanks!
@alecollins01
@alecollins01 Ай бұрын
THANK YOU
@statquest
@statquest Ай бұрын
double bam! :)
@nathannguyen2041
@nathannguyen2041 Ай бұрын
Did math always come easy to you? Also how did you study? Do math topics stay in your mind e.g., fancy integral tricks in probability theory, or dominated convergence, etc?
@statquest
@statquest Ай бұрын
Math was never easy for me and it's still hard. I just try to break big equations down into small bits that I can plug numbers into and see what happens to them. And I quickly forget most math topics unless I can come up with a little song that will help me remember.
@benjaminlucas9080
@benjaminlucas9080 Ай бұрын
Have you done anything on vision tranformers? or can you?
@statquest
@statquest Ай бұрын
I'll keep that in mind. They are not as fancy as you might guess.
@noadsensehere9195
@noadsensehere9195 Ай бұрын
good
@statquest
@statquest Ай бұрын
Thanks!
@aihsdiaushfiuhidnva
@aihsdiaushfiuhidnva Ай бұрын
not many people outside the know knows about bert it seems
@statquest
@statquest Ай бұрын
yep.
@SuperRobieboy
@SuperRobieboy Ай бұрын
Great video, encoders are very interesting in applications like vector search or down-stream prediction tasks (my thesis!). I'd love to see a quest on positional encoding, but perhaps generalised to not just word positions in sentences but also pixel positions in an image or graph connectivity. Image and graph transformers are very cool and positional encoding is too often only discussed for the text-modality. Would be a great addition to educational ML content on KZbin ❤
@statquest
@statquest Ай бұрын
Thanks! I'll keep that in mind.
@dailygrowth7967
@dailygrowth7967 15 күн бұрын
PIZZA GREAT!❤
@statquest
@statquest 13 күн бұрын
:)
@epberdugoc
@epberdugoc Ай бұрын
Actually is, LA PIZZA ES MAGNÍFICA!! ha ha
@statquest
@statquest Ай бұрын
:)
@Apeiron242
@Apeiron242 Ай бұрын
Thumbs down for using the robot voice.
@statquest
@statquest Ай бұрын
Noted
@ChargedPulsar
@ChargedPulsar Ай бұрын
Another bad video, promises simplicity dives right into graphs with no background or explanation.
@statquest
@statquest Ай бұрын
Noted
@Austinlorenzmccoy
@Austinlorenzmccoy Ай бұрын
@@ChargedPulsar the video is great, visualization helps people capture context more Maybe cause i have read about it before but it sure explains better But if you feel you do better, create the content and share so we dive in too
Tensors for Neural Networks, Clearly Explained!!!
9:40
StatQuest with Josh Starmer
Рет қаралды 199 М.
Transformer Neural Networks, ChatGPT's foundation, Clearly Explained!!!
36:15
StatQuest with Josh Starmer
Рет қаралды 794 М.
ССЫЛКА НА ИГРУ В КОММЕНТАХ #shorts
0:36
Паша Осадчий
Рет қаралды 8 МЛН
Counter-Strike 2 - Новый кс. Cтарый я
13:10
Marmok
Рет қаралды 2,8 МЛН
NEW Transformer for RAG: ModernBERT
17:07
Discover AI
Рет қаралды 4,6 М.
AdaBoost, Clearly Explained
20:54
StatQuest with Josh Starmer
Рет қаралды 798 М.
Word Embedding and Word2Vec, Clearly Explained!!!
16:12
StatQuest with Josh Starmer
Рет қаралды 363 М.
A Number to the Power of a Matrix - Numberphile
16:45
Numberphile
Рет қаралды 197 М.
LangChain Advanced RAG - Two-Stage Retrieval with Cross Encoder (BERT)
14:21
Coding Crash Courses
Рет қаралды 12 М.
The Man Who Solved the World’s Most Famous Math Problem
11:14
Newsthink
Рет қаралды 1,2 МЛН
ROC and AUC, Clearly Explained!
16:17
StatQuest with Josh Starmer
Рет қаралды 1,6 МЛН
Decoder-Only Transformers, ChatGPTs specific Transformer, Clearly Explained!!!
36:45
StatQuest with Josh Starmer
Рет қаралды 144 М.
RAG vs. Fine Tuning
8:57
IBM Technology
Рет қаралды 115 М.
Rotary Positional Embeddings: Combining Absolute and Relative
11:17
Efficient NLP
Рет қаралды 41 М.