No video

NLP Demystified 14: Machine Translation With Sequence-to-Sequence and Attention

  Рет қаралды 13,207

Future Mojo

Future Mojo

Күн бұрын

Пікірлер: 41
@BuddingAstroPhysicist
@BuddingAstroPhysicist Жыл бұрын
First of all thanks a lot for these videos , I think they are one of the best on the internet. I have one doubt at 22:15 isn't the input to the scoring function should be h1,s1 instead of h1,s0 for calculating attention for the second output?
@futuremojo
@futuremojo Жыл бұрын
Yep! Nice catch. That's a mistake in the diagram. It should be s1.
@BuddingAstroPhysicist
@BuddingAstroPhysicist Жыл бұрын
@@futuremojo Ok thought so , Thanks a lot again :)
@byotikram4495
@byotikram4495 Жыл бұрын
@futuremojo In the similar context I want to ask a question. So in this period, you are showing how model will generate outputs in the inference time right ? Since from earlier examples we know that we have to apply teacher forcing while training. Am I correct ?
@ueihgnurt
@ueihgnurt 5 ай бұрын
My god this video is genius.
@futuremojo
@futuremojo 2 жыл бұрын
Timestamps 00:00:00 Seq2Seq and Attention 00:00:37 Seq2Seq as a general problem-solving approach 00:02:17 Translating language with a seq2seq model 00:05:53 Machine translation challenges 00:09:07 Effective decoding with Beam Search 00:13:04 Evaluating translation models with BLEU 00:16:23 The information bottleneck 00:17:56 Overcoming the bottleneck with Attention 00:22:39 Additive vs Multiplicative Attention 00:26:47 [DEMO] Neural Machine Translation WITHOUT Attention 00:50:59 [DEMO] Neural Machine Translation WITH Attention 01:04:53 Attention as information retrieval
@klausschmidt982
@klausschmidt982 2 жыл бұрын
I love your clear and succinct explanations. I really appreciate the effort you put in these videos. Thank you.
@futuremojo
@futuremojo 2 жыл бұрын
Thank you, Klaus. I'm glad you're getting value from it.
@vipulmaheshwari2321
@vipulmaheshwari2321 Жыл бұрын
In the future, when I launch my own company, your series will be an essential foundation for NLP. Your teaching is top-notch, presentations are engaging, and your exceptional clarity in explanations truly stands out! YOU ARE BEST! Long Live Brother
@futuremojo
@futuremojo 11 ай бұрын
Thank you!
@ungminhhoai4510
@ungminhhoai4510 Жыл бұрын
khóa học của bạn đúng là khai sáng cho những người muốn bắt đầu học ML
@MadhukaraPhatak-xb4op
@MadhukaraPhatak-xb4op Жыл бұрын
Really love this series. Thank you for sharing theses videos and notebook.
@cheridhanlissassi8716
@cheridhanlissassi8716 Жыл бұрын
Thank you a lot for sharing. Words can't express my gratitude now. The explanation, the illustration are very good. Wish you all the best and thanks once again.
@johnbarnes7485
@johnbarnes7485 Жыл бұрын
Loving this series.
@futuremojo
@futuremojo Жыл бұрын
Thanks, John. Working hard on the last module on transformers.
@johnbarnes7485
@johnbarnes7485 Жыл бұрын
@@futuremojo Great!
@mahmoudreda1083
@mahmoudreda1083 Жыл бұрын
You are the BEST, Thank you.
@horoshuhin
@horoshuhin 2 жыл бұрын
I can't express how great this series on NLP is. Every video is like a Christmas present. I'm really interested in how you approached your learning about NLP. What have you found helped you along the way? thank you Nitin
@futuremojo
@futuremojo 2 жыл бұрын
Thanks for the kind, motivating message! My approach to NLP is similar to most technical things. The top things: 1) Use multiple resources (books, videos, blog posts, etc). We're in a golden age of autodidactism and we can blend multiple resources to get a cohesive picture of a subject. And we don't even need to go through every resource in its entirety. Perhaps this book explains one concept more intuitively than the other, but the other one fleshes out the math better, and this other resource shows one approach to implementing the concept. Another benefit is that multiple resources act like a voting mechanism. A *lot* of materials out there, particularly the blog posts, have conflicting facts. Trying to get at what's true and what's wrong really forces one to dig deep at the fundamentals. 2) Start at the right level of difficulty, then expand outwards. In my case, it was important to cover both theory and practice, so I'd start with a resource that I could understand. If it was too difficult, I looked for something more practical. Once I was comfortable with that, I revisited the more theoretical stuff, but just enough to serve the goal of helping others gain a solid grounding in the subject. And the more I learned, the further I could expand. 3) Lean into the pain. Learning this stuff was often frustrating because so much out there is hand-wavy or opaque or doesn't answer "why?". Implementing this stuff was even more painful and frustrating. But I kept going because I knew I would eventually get it. And it's important that one believes one can learn anything if one persists long enough. 4) Few things solidify your understanding and act as a reality check than teaching it to others. 5) And of course, actually putting things into practice through code.
@rabailkamboh8857
@rabailkamboh8857 Жыл бұрын
thats a brilliant way to explain these difficult topics. Thankyou so much . Also please make a video on transformer models for neural machine translation. thats quite a hot topic .much needed also
@futuremojo
@futuremojo Жыл бұрын
Thank you, Rabail. Transformers (along with pretraining and fine-tuning) is coming up next. Sign up at the course site for updates.
@exxzxxe
@exxzxxe 5 ай бұрын
Very well done! Thanks.
@anujsolanki5588
@anujsolanki5588 7 ай бұрын
Best channel
@CSKdataLab
@CSKdataLab Жыл бұрын
Your videos are one of the best. Flow of topics and concise language. But, when explaining attention theory and code, two terms are beings used interchangeably --> "encoder output sequence" (y1,y2,y3 and so on) and "encoder hidden states" (h1, h2, h3 and so on). This is creating lot of confusion, making it difficult to follow along the tutorial .... maybe something wrong with my understanding. Can you please make a similar series explaining "Generative Models for Images" and "Reinforcement Learning".
@aliabasnezhad7872
@aliabasnezhad7872 Жыл бұрын
great playlist! are you planning to add more videos to this playlist? thanks!
@futuremojo
@futuremojo Жыл бұрын
Thanks, Ali. Yep, there is one more module that's going to be released this month. It's going to cover transformers, pre-training, and transfer learning. We'll go over transformers in depth, code one from scratch, and then learn how to use pre-trained transformers for our own projects. Sign up for updates on the course homepage at nlpdemystified.org.
@aliabasnezhad7872
@aliabasnezhad7872 Жыл бұрын
@@futuremojo Great, looking forward to it!
@SaikatDas-jd9xd
@SaikatDas-jd9xd 9 ай бұрын
Hi there! Loved the series on NLP. Can you please share any link or resource on how to code up the accuracy function like you did with loss? I would like to calculate accuracy of the epochs.
@curdyco
@curdyco Жыл бұрын
Why the padding (0) is converted to at 34:57 ???? The index for is 1 when i print source_tonkenizer.word_index so why 0 is converted to ? Doea this mean that 0 and 1 index both are reserved for ???
@jenibites
@jenibites Жыл бұрын
How does it can translate “my name is udi” if it never saw the word Udi”
@futuremojo
@futuremojo Жыл бұрын
If it's a word-level model, it can't. If your model uses characters or subwords, *maybe* it can translate the individual components of the word to get something sensible, but there needs to be enough data for that to happen. Names, especially from low-resource languages, are a special case which are hard to translate without prior exposure.
@jenibites
@jenibites Жыл бұрын
@@futuremojo what about CopyNet? Could it help?
@futuremojo
@futuremojo Жыл бұрын
@@jenibites I don't know. I don't know what CopyNet does and haven't looked at it. How would a system translate the Hungarian name "Andras" to its English-equivalent of "Andrew" without data? And in practical cases, I imagine one wouldn't want to translate the name at all but keep it the same.
@ateyashuborna1554
@ateyashuborna1554 Жыл бұрын
hey, the video is really amazing, however, I was hoping if you could share how to implement and use the BLEU score in your model?
@amparoconsuelo9451
@amparoconsuelo9451 Жыл бұрын
Where, when and how can I download the corresponding source code your clear explanation of NLP together with the libraries? I am halfway towards watching all your videos I have downloaded in my cell phone. I will watch them again with the source code.
@futuremojo
@futuremojo Жыл бұрын
The notebooks are here: github.com/nitinpunjabi/nlp-demystified
@amparoconsuelo9451
@amparoconsuelo9451 Жыл бұрын
@@futuremojo Thanks.
@sebastianbejarano350
@sebastianbejarano350 Жыл бұрын
UnicodeDecodeError: 'charmap' codec can't decode byte 0x81 in position 4706: character maps to :(
@sebastianbejarano350
@sebastianbejarano350 Жыл бұрын
Adding encoding='utf-8' to open solved it
@ristaaryantiwi3795
@ristaaryantiwi3795 Жыл бұрын
when I try to execute method translator_trainer.fit() NameError: name 'encoder' is not defined
@onlysainaa5764
@onlysainaa5764 9 ай бұрын
what is this model accuracy? or bleu score? how to solve it brother?
Они так быстро убрались!
01:00
Аришнев
Рет қаралды 3,1 МЛН
Идеально повторил? Хотите вторую часть?
00:13
⚡️КАН АНДРЕЙ⚡️
Рет қаралды 18 МЛН
Sequence-to-Sequence (seq2seq) Encoder-Decoder Neural Networks, Clearly Explained!!!
16:50
StatQuest with Josh Starmer
Рет қаралды 181 М.
Лекция. Внимание (Attention)
38:51
Deep Learning School
Рет қаралды 15 М.
This is why Deep Learning is really weird.
2:06:38
Machine Learning Street Talk
Рет қаралды 382 М.
What are Transformer Models and how do they work?
44:26
Serrano.Academy
Рет қаралды 113 М.
How Google Translate Works - The Machine Learning Algorithm Explained!
15:03
What is NLP & How Does It Work?  Neuro Linguistic Programming Basics
27:01
Bay Area NLP And Hypnosis
Рет қаралды 962 М.
[1hr Talk] Intro to Large Language Models
59:48
Andrej Karpathy
Рет қаралды 2,1 МЛН
Они так быстро убрались!
01:00
Аришнев
Рет қаралды 3,1 МЛН