Lecture 6.2: Latent variable model
26:06
Lecture 1.3: Autoencoders
11:01
Жыл бұрын
Lecture 1.1: Neural networks
14:46
Lecture 12.1 Self-attention
22:30
4 жыл бұрын
Lecture 12.2 Transformers
18:08
4 жыл бұрын
Lecture 11.3: World Models
24:27
4 жыл бұрын
Lecture 11.1: Deep Q-Learning
33:06
4 жыл бұрын
Lecture 10.3: ARM & Flows
32:47
4 жыл бұрын
Lecture 10.2: ARM & Flows
33:42
4 жыл бұрын
Lecture 10.1: ARM & Flows
2:20
4 жыл бұрын
Lecture 9.2: The REINFORCE algorithm
25:14
Lecture 9.3: Gradient Estimation
31:29
Lecture 8.3: Graph Neural Networks
21:27
Lecture 8.2: Graph and node embedding
23:34
Lecture 8.1a: Introduction - Graphs
10:42
Lecture 7.2 Implicit models: GANs
23:42
Lecture 5.5 ELMo, Word2Vec
22:59
4 жыл бұрын
Пікірлер
@marcserraortega8772
@marcserraortega8772 17 сағат бұрын
Perfect explanation!!
@msukhar
@msukhar 19 сағат бұрын
Great lectures. I have a question though. Why on 21:34, in the multi-variable chain rule calculation of partial derivatives (of loss) by h we use partial derivatives (of loss) by y, and do not use partial derivatives (of loss) by V, which we calculated on the previous step? I'm sure it's a silly question :)
@msukhar
@msukhar 16 сағат бұрын
Never mind! h & V are independent inputs for y, hence dl/dh is not dependent on dl/dV. Great leactures!
@thehigheststateofsalad
@thehigheststateofsalad 2 ай бұрын
You should do more. Your videos are far clearer than anything else on the market.
@JP-xm3qf
@JP-xm3qf 2 ай бұрын
This was perfect, thank you!
@vitaliy_dushepa
@vitaliy_dushepa 2 ай бұрын
Nice presentation.
@vitaliy_dushepa
@vitaliy_dushepa 3 ай бұрын
Very good!
@vitaliy_dushepa
@vitaliy_dushepa 3 ай бұрын
Very good lecture!
@m1hdi333
@m1hdi333 3 ай бұрын
Pure gold! Thank you.
@HackDealerz
@HackDealerz 3 ай бұрын
thank you for actually explaining the filter operations and the effect on the output dimensions :)
@alexandrehmf
@alexandrehmf 3 ай бұрын
Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. "Imagenet classification with deep convolutional neural networks." Advances in neural information processing systems. 2012
@alexandrehmf
@alexandrehmf 3 ай бұрын
LeCun, Y.; Boser, B.; Denker, J. S.; Henderson, D.; Howard, R. E., Hubbard, W.; Jackel, L. D. (1989). "Backpropagation Applied to Handwriten Zip Code Recognition"
@metehkaya96
@metehkaya96 3 ай бұрын
Perfect explanation, but don't we have softmax operation in the practical SA just like simple SA? I could not see softmax in the representations of practical SA (18:42) unlike simple SA (05:16).
@rachadlakis1
@rachadlakis1 4 ай бұрын
Thanks for the information
@DhaneshKasinathanlove
@DhaneshKasinathanlove 5 ай бұрын
Thanks for the course
@igorras-ff7oe
@igorras-ff7oe 5 ай бұрын
Thank you for this video!
@vitaliy_d
@vitaliy_d 6 ай бұрын
Very useful lecture. Thanks for sharing! It may be a small typo (at 18:25), should be l = loss(output,target) # "output" instead of "input"
@MrCobraTraders
@MrCobraTraders 6 ай бұрын
I didn't understand why adding gaussian noise to image does affect the accuracy of discriminative model. (I think it doesn't) In reality models are robust to noise
@ShivangiTomar-p7j
@ShivangiTomar-p7j 6 ай бұрын
Absolutely amazing! Thank you so much!!
@nirajrajai3116
@nirajrajai3116 7 ай бұрын
Loved it! Clear explanations with simple examples.
@Mars.2024
@Mars.2024 7 ай бұрын
Finally i have intuitive view of seld_attention . Thank you😇
@vesk4000
@vesk4000 9 ай бұрын
This is exceptionally well explained. I'm a student at TU Delft and this really helped me understand how to speed up my code, and why it works. Thanks a lot!
@olileveque
@olileveque 9 ай бұрын
Absolutely amazing series of videos! Congrats!
@abhilashbalachandran7160
@abhilashbalachandran7160 10 ай бұрын
Very well explained
@prateekpatel6082
@prateekpatel6082 11 ай бұрын
i dont understand why we have summation in conditionals , should that be a product instead of summation ?
@prateekpatel6082
@prateekpatel6082 Жыл бұрын
quite bad explanation , just repeating slides text
@MariemStudiesWithMe
@MariemStudiesWithMe Жыл бұрын
the approach of the noise filter presented in the video can cause neuron saturation i guess because having high weighted input with maximize the output of the sigmoid function.. which is not desirable
@润龙廖
@润龙廖 Жыл бұрын
Thanks for your sharing! Nice and clear video!
@nadeem1969100
@nadeem1969100 Жыл бұрын
Currently I am doing research work on TCN
@37kuba
@37kuba Жыл бұрын
Superb Superb explanation, Superb explanation thanks Superb explanation thanks a Superb explanation thanks a lot! With no predictions now: Wish you all the best and am very grateful for your work.
@Yassinius
@Yassinius Жыл бұрын
Why do these videos have ads on them
@MrOntologue
@MrOntologue Жыл бұрын
Google should rank videos according to the likes and the number of previously viewed videos on the same topics: this should go straight to the top for Attention/Transformer searches because I have seen and read plenty, and this is the first time the QKV as dictionary vs RDBMs made sense; that confusion had been so bad it literally stopped me thinking every time I had to consider Q, or K, or V and thus prevented me grokking the big idea. I now want to watch/read everything by you.
@saurabhmahra4084
@saurabhmahra4084 Жыл бұрын
Watching this video feels like trying to decipher alien scriptures with a blindfold on.
@ucbrowser-q3g
@ucbrowser-q3g Жыл бұрын
Waiting for more lectures..
@scienceprojectsofdccpn3430
@scienceprojectsofdccpn3430 Жыл бұрын
Self attension animationkzbin.info/www/bejne/jabWgnVqa5mkg7ssi=NBxi02yTPzSfMCb6
@soumilbinhani8803
@soumilbinhani8803 Жыл бұрын
Hello can someone explain me this, the key and the values for each iteration wont it be the same, as we compare it to 5:29 , please help me on this
@fredoliveira7569
@fredoliveira7569 Жыл бұрын
Best explanation ever! Congratulations and thank you!
@sergionic1821
@sergionic1821 Жыл бұрын
what the dimensions of first conv filter in AlexNet - is it 11x11x 1 or it 11x11x 3 ?
@linux2650
@linux2650 11 ай бұрын
its 11x11x 3, operating on those three channels.
@adrielcabral6634
@adrielcabral6634 Жыл бұрын
I loved u explanation !!!
@zadidhasan4698
@zadidhasan4698 Жыл бұрын
You are a great teacher.
@somerset006
@somerset006 Жыл бұрын
Great lecture, thanks!
@somerset006
@somerset006 Жыл бұрын
Really good series of mini-lectures, thanks!
@erdemozkol9049
@erdemozkol9049 Жыл бұрын
Brilliant content and explanation! It's unfortunate that the fourth part of this lecture series was never published, this realization has left me very sad in 2023. :(
@davealsina848
@davealsina848 Жыл бұрын
love this two times
@davealsina848
@davealsina848 Жыл бұрын
Loved this, thanks a lot now I undersand better this things and feel more confident to jump into the code part.
@mahmoudebrahimkhani1384
@mahmoudebrahimkhani1384 Жыл бұрын
Such a clear explanation! Thank you!
@小孟滴儿丫
@小孟滴儿丫 Жыл бұрын
Thanks for the great explanation! Just one question, if simple self-attention has no parameters, how can we expect it to learn? it is not trainable.
@Isomorphist
@Isomorphist Жыл бұрын
Is this ASMR?
@xiaoweidu4667
@xiaoweidu4667 Жыл бұрын
good tutorial
@senthil2sg
@senthil2sg Жыл бұрын
Better than the Karpathy explainer video. Enough said!
@HiHi-iu8gf
@HiHi-iu8gf Жыл бұрын
holy shit, been trying to wrap my head around self-attention for a while, but it all finally clicked together with this video. very well explained, very good video :)