Adam Optimizer for Neural Network || Lesson 15 || Deep Learning || Learning Monkey ||

  Рет қаралды 6,280

Learning Monkey

Learning Monkey

Күн бұрын

Пікірлер: 9
@saswatam
@saswatam 6 ай бұрын
Sir in the final Equation: Wt = Wt-1 - Alpha * m^t / sqrt(V^t + ε)), here we are representing gt which was dl/dw earlier as m^t / sqrt(V^t + ε))?, because in the same previous Equation: Wt = Wt-1 - Alpha *gt. Where Alpha = Learning Rate
@babakdianati6515
@babakdianati6515 Жыл бұрын
Thanks for the nice video
@muhammadumarbello6574
@muhammadumarbello6574 9 ай бұрын
Very nice explanation,but how can I get a detail pdf on Adam
@OnTastySpots
@OnTastySpots 2 жыл бұрын
Wow! Thank you for the detailed explanation! I'm wondering do we have m_t, v_t for each parameter? For example, f(w1,w2,w3) will need 3 m_t's and 3 v_t's in each timestep.
@ayushx4831
@ayushx4831 11 ай бұрын
Send link of that T shirt 🙂
@AkshayRakate
@AkshayRakate 3 жыл бұрын
What is alpha in the Adam equation? else a brilliant explanation !
@LearningMonkey
@LearningMonkey 3 жыл бұрын
Please watch our Ada delta video. Alpha is learning rate
@bangarrajumuppidu8354
@bangarrajumuppidu8354 3 жыл бұрын
super sir
@mohammedmunavarbsa573
@mohammedmunavarbsa573 4 жыл бұрын
super bro
these are the only perfect squares
12:39
Michael Penn
Рет қаралды 2,1 М.
REAL or FAKE? #beatbox #tiktok
01:03
BeatboxJCOP
Рет қаралды 18 МЛН
Aggvent Calendar Day 18
3:36
Andy Math
Рет қаралды 10 М.