Bro I'm watching this years after Graduating, props for the excellent explanation
@anuradhadas87954 жыл бұрын
This is pure gold !
@AJ-et3vf Жыл бұрын
Best explanation indeed for NAGD. I completely understood the derivation especially the look ahead term.
@jatayubaxi45534 жыл бұрын
Great explanation sir ! especially the last 5 minutes of the explanation.
@ductn9 Жыл бұрын
Excellent explanation! Thank you.
@muratcan__225 жыл бұрын
man you are really a good teacher thank you
@anubhavgupta81464 жыл бұрын
Dude it was an excellent explanation, which i was looking for ofcourse.👍
@Ankurkumar146805 жыл бұрын
Amazing video, nicely explained
@jahanvi9429 Жыл бұрын
You explain things soo clearly
@newbie8051 Жыл бұрын
Notes to self : Nestorov Momentum is just look ahead GD, we first move by the amount suggested by the history, then calculate the gradient and make the final move.
@aashishmalhotra2 жыл бұрын
Wow awesome why to go to Coursera edx and to del with accent when our professors are so good 😂 thank you sir
@GiiWiiDii5 жыл бұрын
Very, very good! Thanks:)
@chetankhaire2 жыл бұрын
Very good explanation😀
@AnkitKumar-sz3by Жыл бұрын
best explaination
@sumankumarChoudhury5 жыл бұрын
Awesome lecture
@tanujguhathakurta4133 жыл бұрын
what and awesome explanation.. now I gotta subscribe to one fourth labs
@김수연-r8c5 жыл бұрын
Thank you !
@PatatjesDora2 жыл бұрын
Great vid
@rubenpole15704 жыл бұрын
Nice explanation!
@sivabonthada10762 жыл бұрын
Holy smoke!!
@sagaradoshi Жыл бұрын
Excellent explanation . I have one observation with respect to the content in slide and code you shared. From the slide initially wlook_ahead = wt - gamma * update t-1. But in code v_w = gamma * prev_v_w i.e., wlook_ahead = gamma * prevW .. so getting confused here..
@AlejandraRodriguezTi4 жыл бұрын
Such a great video! Thanks a lot!. If you could share a github with the code and the simulation, it would be really nice, to be able to try it faster, without rewriting. I would also appreciate if you guys could show how exactly is "update " vector. formed. And how is the NAG seen in a computer graph within the whole NN.
@vishnudevk37372 жыл бұрын
At 1:11, shouldn't it be w(t-1) = w(t) + gamma*update(t-1) - eta*gradient @NPTEL
@aboudaladdin86043 жыл бұрын
best
@muhammadiqbalbazmi92754 жыл бұрын
Amazing video first time on NPTEL.
@ankitfparmar3 жыл бұрын
24k Gold for free.
@Irmbaes4 жыл бұрын
Sorry, this explanation is too vague to make sense: it reveals nothing about how nor why it works. For instance, the method works provably well only for very specific values of gamma and eta, while your explanation would "prove" that the method works for a large range of values of these parameters.