Encoder-Only Transformers (like BERT), Clearly Explained!!!

  Рет қаралды 10,692

StatQuest with Josh Starmer

StatQuest with Josh Starmer

Күн бұрын

Пікірлер: 56
@statquest
@statquest 12 күн бұрын
Support StatQuest by buying my book The StatQuest Illustrated Guide to Machine Learning or a Study Guide or Merch!!! statquest.org/statquest-store/
@NottoriousGG
@NottoriousGG 17 сағат бұрын
Such a cleverly disguised master of the craft. 🙇
@statquest
@statquest 5 сағат бұрын
bam! :)
@PradeepKumar-hi8mr
@PradeepKumar-hi8mr 12 күн бұрын
Wowww! Glad to have you back, Sir. Awesome videos 🎉
@statquest
@statquest 12 күн бұрын
Thank you!
@tcsi_
@tcsi_ 12 күн бұрын
100th Machine Learning Video 🎉🎉🎉
@statquest
@statquest 12 күн бұрын
Yes! :)
@THEMATT222
@THEMATT222 9 күн бұрын
Noice 👍 Doice 👍Ice 👍
@nossonweissman
@nossonweissman 12 күн бұрын
Yay!!! ❤❤ I'm starting it now and saving to remember to finish later. Also, I'm requesting a video on Sparse AutoEncoders (used in Anthropic's recent research). They seem super cool and I have a basic idea on how they work, but I'd to see a "simply explained" version of them.
@statquest
@statquest 12 күн бұрын
Thanks Nosson! I'll keep that topic in mimd.
@kamal9294
@kamal9294 9 күн бұрын
Nice explanation, if the next topic is about rag or reinforcement learning , i will be happier (or even object detection, object tracking).
@statquest
@statquest 9 күн бұрын
I guess you didn't get to 16:19 where I explain how RAG works...
@kamal9294
@kamal9294 9 күн бұрын
@statquest bro but in LinkedIn I saw many rag types and some retrieval techniques using advanced dsa(like HNSW). That's why I asked.
@statquest
@statquest 9 күн бұрын
@@kamal9294 Those are just optimizations, which will change every month. However, the fundamental concepts will stay the same and are described in this video.
@kamal9294
@kamal9294 9 күн бұрын
@@statquest Now I am clear, thank you!.
@free_thinker4958
@free_thinker4958 12 күн бұрын
You're the man ❤️💯👏 thanks for everything you do here to spread that precious knowledge 🌹 we hope if you could possibly dedicate a future video to talk about multimodal models (text to speech, speech to speech etc...) ✨
@statquest
@statquest 12 күн бұрын
I'll keep that in mind!
@etgaming6063
@etgaming6063 12 күн бұрын
This video came just in time, trying to make my own RoBERTa model and have been struggling understanding how they work under the hood. Not anymore!
@statquest
@statquest 11 күн бұрын
BAM!
@barackobama7757
@barackobama7757 9 күн бұрын
Hello StatQuest. I was hoping if you could make a video on PSO (Particle Swarm Optimisation) Will really help! Thank you, amazing videos as always!
@statquest
@statquest 8 күн бұрын
I'll keep that in mind.
@davidlu1003
@davidlu1003 4 күн бұрын
And thx for the courses. They are great!!!!😁😁😁
@statquest
@statquest 4 күн бұрын
Glad you like them!
@thegimel
@thegimel 12 күн бұрын
Great instructional video, as always, StatQuest! You mentioned in the video that the training task for these networks is next word prediction, however, models like BERT have only self-attention layers so they have "bidirectional awareness". They are usually trained on masked language modeling and next sentence prediction, if I recall correctly?
@statquest
@statquest 12 күн бұрын
I cover how a very basic word embedding model might be trained in order to illustrate its limitations - that it doesn't take position into account. However, the video does not discuss how an encoder-only transformer is trained. That said, you are correct, an encoder-only transformer uses masked language modeling.
@Kimgeem
@Kimgeem 10 күн бұрын
So excited to watch this later 🤩✨
@statquest
@statquest 10 күн бұрын
future bam! :)
@tonym4926
@tonym4926 12 күн бұрын
Are you planning to add this video to neutral network/ deep learning playlist?
@statquest
@statquest 12 күн бұрын
yes! Just did.
@aryasunil9041
@aryasunil9041 8 күн бұрын
Great Video, When is the Neural Networks book coming out? Very eager for it
@statquest
@statquest 7 күн бұрын
Early january. Bam! :)
@davidlu1003
@davidlu1003 4 күн бұрын
I love you, I will keep going and learn the other courses of yours if they are always free. keep them free please, I will always be your fan.😁😁😁
@statquest
@statquest 4 күн бұрын
Thank you, I will!
@mbeugelisall
@mbeugelisall 12 күн бұрын
Just the thing I’m learning about right now!
@statquest
@statquest 12 күн бұрын
bam! :)
@nathannguyen2041
@nathannguyen2041 12 күн бұрын
Did math always come easy to you? Also how did you study? Do math topics stay in your mind e.g., fancy integral tricks in probability theory, or dominated convergence, etc?
@statquest
@statquest 12 күн бұрын
Math was never easy for me and it's still hard. I just try to break big equations down into small bits that I can plug numbers into and see what happens to them. And I quickly forget most math topics unless I can come up with a little song that will help me remember.
@benjaminlucas9080
@benjaminlucas9080 10 күн бұрын
Have you done anything on vision tranformers? or can you?
@statquest
@statquest 10 күн бұрын
I'll keep that in mind. They are not as fancy as you might guess.
@draziraphale
@draziraphale 12 күн бұрын
Great explanation
@statquest
@statquest 12 күн бұрын
Thanks!
@iamumairjaffer
@iamumairjaffer 12 күн бұрын
Well explained ❤❤❤
@statquest
@statquest 12 күн бұрын
Thanks!
@aihsdiaushfiuhidnva
@aihsdiaushfiuhidnva 5 күн бұрын
not many people outside the know knows about bert it seems
@statquest
@statquest 5 күн бұрын
yep.
@epberdugoc
@epberdugoc 12 күн бұрын
Actually is, LA PIZZA ES MAGNÍFICA!! ha ha
@statquest
@statquest 12 күн бұрын
:)
@noadsensehere9195
@noadsensehere9195 11 күн бұрын
good
@statquest
@statquest 11 күн бұрын
Thanks!
@SuperRobieboy
@SuperRobieboy 12 күн бұрын
Great video, encoders are very interesting in applications like vector search or down-stream prediction tasks (my thesis!). I'd love to see a quest on positional encoding, but perhaps generalised to not just word positions in sentences but also pixel positions in an image or graph connectivity. Image and graph transformers are very cool and positional encoding is too often only discussed for the text-modality. Would be a great addition to educational ML content on KZbin ❤
@statquest
@statquest 12 күн бұрын
Thanks! I'll keep that in mind.
@Apeiron242
@Apeiron242 12 күн бұрын
Thumbs down for using the robot voice.
@statquest
@statquest 12 күн бұрын
Noted
@ChargedPulsar
@ChargedPulsar 12 күн бұрын
Another bad video, promises simplicity dives right into graphs with no background or explanation.
@statquest
@statquest 12 күн бұрын
Noted
@Austinlorenzmccoy
@Austinlorenzmccoy 10 күн бұрын
@@ChargedPulsar the video is great, visualization helps people capture context more Maybe cause i have read about it before but it sure explains better But if you feel you do better, create the content and share so we dive in too
Tensors for Neural Networks, Clearly Explained!!!
9:40
StatQuest with Josh Starmer
Рет қаралды 192 М.
Decoder-Only Transformers, ChatGPTs specific Transformer, Clearly Explained!!!
36:45
StatQuest with Josh Starmer
Рет қаралды 136 М.
Accompanying my daughter to practice dance is so annoying #funny #cute#comedy
00:17
Funny daughter's daily life
Рет қаралды 23 МЛН
Миллионер | 3 - серия
36:09
Million Show
Рет қаралды 2,2 МЛН
Turn Off the Vacum And Sit Back and Laugh 🤣
00:34
SKITSFUL
Рет қаралды 8 МЛН
Why 4d geometry makes me sad
29:42
3Blue1Brown
Рет қаралды 984 М.
Visualizing transformers and attention | Talk for TNG Big Tech Day '24
57:45
New divisibility rule! (30,000 of them)
26:51
Stand-up Maths
Рет қаралды 377 М.
Transformer Neural Networks, ChatGPT's foundation, Clearly Explained!!!
36:15
StatQuest with Josh Starmer
Рет қаралды 754 М.
AI can't cross this line and we don't know why.
24:07
Welch Labs
Рет қаралды 1,3 МЛН
10 weird algorithms
9:06
Fireship
Рет қаралды 1,3 МЛН
How are holograms possible? | Optics puzzles 5
46:24
3Blue1Brown
Рет қаралды 1,3 МЛН
Accompanying my daughter to practice dance is so annoying #funny #cute#comedy
00:17
Funny daughter's daily life
Рет қаралды 23 МЛН