BERT Research - Ep. 1 - Key Concepts & Sources

  Рет қаралды 94,997

ChrisMcCormickAI

ChrisMcCormickAI

Күн бұрын

Пікірлер: 64
@zhou7yuan
@zhou7yuan 4 жыл бұрын
Significance [0:08] Research Posts [1:27] BERT Mountain [3:02] can we skip over LSTMs [5:14] BERT Paper [5:50] BERT Repo [7:15] BERT Announcement Post [7:40] Attention is All You Need (Transformer) [8:11] The Annotated Transformer [8:42] Jay Alammar's Posts [10:28] Sequence Models on Coursera [11:23] Next Up [13:19]
@AbdelhakMahmoudi
@AbdelhakMahmoudi 4 жыл бұрын
Hi Chris, I like the way you explain things. I like visual explanations and the BERT's mountain was "all I need" :D !, thanks a lot.
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Thanks, appreciate the encouragement!
@leliaglass1568
@leliaglass1568 4 жыл бұрын
thanks for making this video, I am enjoying the series. I would especially like to see hands-on demos as Jupyter notebooks! :)
@thalanayarmuthukumar5472
@thalanayarmuthukumar5472 4 жыл бұрын
A very no nonsense way of representing the work you are doing. It felt like I was with you and studying with you. Thanks. I am planning to go through the rest of your videos, in my journey to learn BERT
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Thanks so much Thalanayar! I'm so glad the videos are helping you on your BERT journey! :D
@davidz6828
@davidz6828 4 жыл бұрын
Hi Chris, I read your articles on BERT before and have learned a ton from them. Can't believe you have videos as well. Thanks for sharing the knowledge!
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
That's what I like to hear :) Thanks, Yikai!
@keshavramaswamy6217
@keshavramaswamy6217 4 жыл бұрын
You sir, are a legend in your own right! Keep up all this work you are doing! At some point it will be helpful if you can put a guide to effective science writing like you do! :)
@abeersalam1623
@abeersalam1623 8 ай бұрын
Sir, I'm new to this field, my research topic is about automatically evaluating essay answers using Bert what should I learn in advance so that I pick up only the main points related only to my research in order not to be distracted by too much information and could please give me your email I want to consult you And thank you
@chuanjiang6931
@chuanjiang6931 Жыл бұрын
What is the difference between attention and self-attention?
@binwangcu
@binwangcu 3 жыл бұрын
10:16 "all these sound very discouraging" - says Chris :)
@JJ_eats_wings
@JJ_eats_wings 4 жыл бұрын
Hahaha bursted into laughter at 3:47. Chris you are exactly right - I started researching BERT and then just keeps bouncing from topic to topic. (as a beginner to deep NNs)
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Nice :) Glad I'm not the only one!
@nana-xf7dx
@nana-xf7dx 2 жыл бұрын
Your explanation is super clear and I like the Bert mountain which shows what I need to understand first .
@ChrisMcCormickAI
@ChrisMcCormickAI 2 жыл бұрын
Thanks!
@chronicfantastic
@chronicfantastic 4 жыл бұрын
Great video, thanks for this looking forward to the rest of the series!
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Thanks, Luke!
@riasingh2558
@riasingh2558 4 жыл бұрын
Hi Chris, Firstly, thanks a lot for writing the most comprehensive blog post, extremely helpful. I have been following it to understand BERT more closely. Secondly, besides creating word and sentence vectors by using different pooling strategies and layers, could you please extend the blog post by showing how to compute the word attentions and their respective positions? Thanks!
@prakashkafle454
@prakashkafle454 3 жыл бұрын
Token indices sequence length is longer than the specified maximum sequence length for this model (1312 > 512). Running this sequence through the model will result in indexing errors . Get this error message while doing news classification .
@syedhamza3314
@syedhamza3314 Жыл бұрын
Hi Chris, absolutely amazing series on Transformers. I have a question regarding how transformers handle the variable length inputs. Suppose I set the max_lenght for my sequences to be 32 and feed the input_id and attnetion_mask only for 32 tokens during training, given some tokens can be padded tokens since each sequence won't be exactly of 32 lengths. Now if we talk about bert the default max_lenght is 512 tokens so my question is does the transformer implicitly add 512-32 padded tokens to calculate MHA on 512 tokens as it will not attend to the tokens with the padded token ID? If that's the case then are we not updating the parameters directly attach to the remaining 512-32 positional vectors?
@learn2know79
@learn2know79 Жыл бұрын
Excellent work... it's very informative, especially the prerequisite domain knowledge area. Waiting to see more from you
@kingeng2718
@kingeng2718 4 жыл бұрын
Nice Job, Thanks a lot for sharing
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Thanks, King!
@mahdiamrollahi8456
@mahdiamrollahi8456 3 жыл бұрын
So, Bert is a model different from other language models like word2vec or Glove, right?
@akbarghurbal
@akbarghurbal 4 жыл бұрын
Thanks a lot for your videos. It's almost the end of 2020 and still there are no books on Amazon about BERT!
@aanchalagarwal6886
@aanchalagarwal6886 4 жыл бұрын
Hey, the link to your blog page is throwing a 404: Page Not Found error. Could you please help me with the problem
@CristianTraina
@CristianTraina 2 жыл бұрын
Really great content! Does anyone know how can I contact Chris? I need to ask permittion to use and quote some of his work
@flamingflamingo4021
@flamingflamingo4021 4 жыл бұрын
Do you have a playlist for all the episodes regarding BERT? It'd be really organized and helpful.
@akhilsebastian3804
@akhilsebastian3804 3 жыл бұрын
Hi Chris, I am back here in your first video again after an year. I guess this time I ll be able to follow you better.
@praveenchalampalem4038
@praveenchalampalem4038 4 жыл бұрын
Wonderful Explanation Chris!!!!
@tobiasgiesemann2180
@tobiasgiesemann2180 3 жыл бұрын
Hi Chris, thanks so much for the video. I actually got stuck on the same line in the BERT paper, where it says "we will omit an exhaustive explanation". From there I went down the BERT mountain and finally got to your video, so thanks a lot for picking me up on the way. Looking forward to the rest of the series!
@Ramm165
@Ramm165 4 жыл бұрын
Hi Chris thanks for the wonderful video. I would like to know if the topics covered in the ebook are different from videos or not .Thank you
@AbdennacerAyeb
@AbdennacerAyeb 3 жыл бұрын
Thank you a lot. You are making it easier for us to understand hard topics..
@rabirajbanerjee3872
@rabirajbanerjee3872 3 жыл бұрын
Awesome series, I have a basic idea about how Attention Mechanism work but this builds on the concepts
@vinayreddy8683
@vinayreddy8683 4 жыл бұрын
Like the way you teach. Not many people are teaching nlp, so it's good to have a person like you. Btw, 1000th subscriber.
@swapnil9047
@swapnil9047 4 жыл бұрын
HI Chris, Great video! Do you have a medium /twitter channel to follow your latest works in Data Science?
@kooshatahmasebipour690
@kooshatahmasebipour690 3 жыл бұрын
Feeling so lucky to find your website, resources, and channel. Thanks a lot!
@akshayklr057
@akshayklr057 4 жыл бұрын
I would appreciate if you could cover other models as well, these tutorials are good for a noob to start with.
@mourady.650
@mourady.650 4 жыл бұрын
Hello Chris, thanks for this beautiful series. You described the training tasks as fake/bogus tasks. I prefer to name them proxy tasks - as in proxy war, but for good purposes. :) What do you think?
@aytuncun6910
@aytuncun6910 4 жыл бұрын
Hi Chris, thanks for the post. Feeling lucky I've found your videos. Currently, I'm going through what you've been through basically. Can't wait to watch the whole series. Have you tried Google's Natural Questions challenge yet? Thanks again.
@viiids
@viiids 4 жыл бұрын
I understand RNNs, LSTM, Bidirectional LSTMs and Attention I still found the BERT paper hard to read and had the exact same feeling of the mountain you drew. This video and the subsequent one is getting me much more confident about BERT, hoping to watch the 3rd video in the morning. Thanks for this contribution, your explanation is very concise.
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Glad I'm not the only one! Thanks for your comment :)
@mahadevanpadmanabhan9314
@mahadevanpadmanabhan9314 4 жыл бұрын
What an amazing effort.Super.
@hanman5195
@hanman5195 4 жыл бұрын
Hi Chris, This is really amazing explanation. Can you please help me how to use this Bert model with lime to explain model ?
@geo2073
@geo2073 4 жыл бұрын
great content Chris!
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Thanks George!
@adityasoni121
@adityasoni121 4 жыл бұрын
Cool Video Chris!
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Thanks!
@kraken1350
@kraken1350 4 жыл бұрын
could I use BERT model on language like Arabic?
@oluwapelumiabimbola3280
@oluwapelumiabimbola3280 4 жыл бұрын
Yes. Checkout Multi lingual Bert
@bitbyte8177
@bitbyte8177 4 жыл бұрын
What a great video! You earned a new subscriber.
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Thanks! Glad to have you :D
@azizbenothman5374
@azizbenothman5374 4 жыл бұрын
I gave you the 800th like, good work
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Cool, thanks Aziz!
@8g8819
@8g8819 4 жыл бұрын
Hi, Please keep going on this (hands on)series, I'm pretty sure you will help lots of people out there!!
@ChrisMcCormickAI
@ChrisMcCormickAI 4 жыл бұрын
Thanks giavo, I'll keep them coming! Anything in particular that you'd like to see explained? Thanks!
@8g8819
@8g8819 4 жыл бұрын
@@ChrisMcCormickAI I think BERT research is perfect for now. Soon there gonna be a wide application of BERT in NLP area and a research like this one is perfect for who wants to understand all of it's aspects ( + general aspects as Word Embeddings, what exactly is Attention mechanism...). It would be great to talk about how can we adopt BERT to a certain domain with domain specific terms... Also, personally I would like to understand how to use BERT in order to compute similarity between 2 documents(tried already Cosine Similarity measure based on TF-IDF, Chi square, Keygraphs based keywords but still not happy with results) Thanks again!
@felipeacunagonzalez4844
@felipeacunagonzalez4844 4 жыл бұрын
Thank you sir!
@田英俊
@田英俊 4 жыл бұрын
Thank you!
@jingyingwang767
@jingyingwang767 4 жыл бұрын
OMG, that BERT Mountain picture at the beginning is exactly what I've been conceptualizing!! I love this series of videos! Thanks a lot!
@mail2ak
@mail2ak 3 жыл бұрын
Its a very engaging way of learning a new topic :) love it
BERT Research - Ep. 2 - WordPiece Embeddings
28:04
ChrisMcCormickAI
Рет қаралды 71 М.
The Power of Habit : A Roadmap to Changing Your Life
7:50
UPGRAID Habits
Рет қаралды 15
when you have plan B 😂
00:11
Andrey Grechka
Рет қаралды 67 МЛН
From Small To Giant Pop Corn #katebrush #funny #shorts
00:17
Kate Brush
Рет қаралды 71 МЛН
FlashAttention - Tri Dao | Stanford MLSys #67
58:58
Stanford MLSys Seminars
Рет қаралды 28 М.
BERT Research - Ep. 4 - Inner Workings I
22:13
ChrisMcCormickAI
Рет қаралды 28 М.
BERT Research - Ep. 3 - Fine Tuning - p.2
30:11
ChrisMcCormickAI
Рет қаралды 41 М.
BERT Explained!
13:22
Connor Shorten
Рет қаралды 85 М.
Transformers, explained: Understand the model behind GPT, BERT, and T5
9:11
AI, Machine Learning, Deep Learning and Generative AI Explained
10:01
IBM Technology
Рет қаралды 270 М.
BERT Research - Ep. 6 - Inner Workings III - Multi-Headed Attention
18:40
What is BERT and how does it work? | A Quick Review
8:56
AssemblyAI
Рет қаралды 51 М.
when you have plan B 😂
00:11
Andrey Grechka
Рет қаралды 67 МЛН