In this lecture we will gain more insights into the Loss function of Generative Adversarial Networks #adversarial#generative#deeplearning
Пікірлер: 62
@mrtandon52784 жыл бұрын
I invested 40k in Udacity Deep Learning Nanodegree program, followed by 32k investment in applied ai course. But nobody bothers themselves to teach GAN mathematics concepts. On the other side, a great teacher digs deep in the ocean just to make their student understand what actually GAN is. You left no stone unturned. Thanks for sharing your knowledge and for your efforts.
@wolfisraging5 жыл бұрын
Literally the best explanation of gans on whole KZbin
@prateekthakur40605 жыл бұрын
I am with you man. One of the best explanation on youtube.
@nisanchhetri11623 жыл бұрын
@@prateekthakur4060 do watch this then you will retake this statement. kzbin.info/www/bejne/eoKxf4CfdrVootk
@nisanchhetri11623 жыл бұрын
do watch this then you will retake this statement. kzbin.info/www/bejne/eoKxf4CfdrVootk
@anshumishra41933 жыл бұрын
@@nisanchhetri1162, it is in Chinese. The content looks excellent, but it's tough to follow just looking at the subtitles.
@superaluis4 жыл бұрын
Only lecture on KZbin with the advanced concepts all throughly explained to the detail. Subscribed
@rachitverma9088 Жыл бұрын
Your attention to detail regarding the concepts is truly remarkable. Thanks a lot for making this excellent series!
@user-ky7bi9cl2z4 жыл бұрын
Only KZbin channel which gives proper insights to Deeplearning.
@MLDawn5 жыл бұрын
are you for real?! Keep it up please! It is like you know where I had problem understanding it, and you dissect that very part beautifully!
@shineshine95995 жыл бұрын
True...the best explanation...thank u so much!!
@NeerajAithani5 жыл бұрын
It's been 10 months I implemented GAN and it was easy. I tried to read paper didn't understand. finally, your lectures helped me to understand GANs fully, probably the best explanation on whole internet :).
@Gauravkr00715 жыл бұрын
neeraj how do u implement it, i am searching for how to implement it on keras or any other librray
@NeerajAithani5 жыл бұрын
@@Gauravkr0071Dewangan there are plenty of resources on the internet to implement it. If you are a beginner I would suggest you first get comfortable with the library you are using and build some small models. After you feel comfortable then go to advance topics like GAN.
@Gauravkr00715 жыл бұрын
@@NeerajAithani ok sir , just one thing , there is so much maths involved in it , but when i see the code of any GAN , there is no such maths involved in it, in the code
@hiteshnagothu8874 жыл бұрын
I really appreciate all the series explained so well without giving up on the foundation of Maths, of course, it's all the game of Maths! Looking forward for more such content.
@puspaksahu334210 ай бұрын
Love from IISc pure maths ! salute , no one cares to go into mathematical detail as yours on utube
@elyseemanimpiregasana21172 жыл бұрын
I don't know what I can say. But always in deep learning, If I have a presentation related to the topics he explained, I have to take the course. After that, everyone takes me as a specialist in the topic. you really deserve respect.
@bosepukur5 жыл бұрын
fantastic lecture series ....I am hooked
@shail12aksh4 жыл бұрын
Thoroughly explained the whole concept of GAN...Thanks for explaining GAN in such an understandable manner...
@rahulrai13295 жыл бұрын
Thanks a lot. Very nice and intuitive explanation of GAN and very useful.I request you to Please upload more and more content regarding GAN and latest papers.
@akshitasood64554 жыл бұрын
This explanation makes the concept of GANs so much easier !!! Great Lecture !!!
@vivek655222 жыл бұрын
Thank you so much Prof. You have done a great job explaining. Beauty of the concepts lie in the details of mathematics which you have to explained to the core.
@jitendradhiman58154 жыл бұрын
The intuition behind GAN is very nicely explained. Thank you sir for bringing such informative videos to us.
@advaittilak6143 жыл бұрын
I wish I could personally meet and thank you Alhad Sir.
@doublesami4 жыл бұрын
I'm very motivated sir after reading about you. Following your lectures from a long time . Your methodology is best one . Best wishes Just keep motivated us by providing quality literature reviews
@fawadkhan29554 жыл бұрын
This is great. Liked and Subscribed. Please keep making more videos like this on the most common deep-learning techniques. Suggestion: CycleGAN
@aashwinsharma18592 жыл бұрын
The efforts just to explain small concepts and proofs.... 🙏
@RAJANKUMAR-mi1ib5 жыл бұрын
wow....you made me understand each and every concept....keep the good work going!
@ssshukla264 жыл бұрын
Nobody explained this ... Not even my college professor ... Thank you sir ..
@ahmedheakl51813 жыл бұрын
You are the best youtuber ever
@soumambanerjee18163 жыл бұрын
You have the power to generate Netflix alike addictive content in deep learning ❤️
@baabakasadi54407 ай бұрын
Beautiful explanation
@msic13764 жыл бұрын
Sir, this is the best explanation of GANs on youtube by far! Only one doubt: why at 12:05 we take the derivative only of the argument of the integral and not of the whole integral?
@ashj11654 жыл бұрын
Started loving maths again with these explanations!
@gauravsahani24994 жыл бұрын
Best Explaining Sir, Thankyou so much for this Playlist!
@kwippo4 жыл бұрын
3:17 In equation 2, is it supposed to be D(G(z)) instead of D(G(x))?
@sugamtyagi1014 жыл бұрын
Thank you so much for going into the details.
@solitaryreaper13634 жыл бұрын
That's a really good explanation sir, thank you so much.
@MuhammadSaleem-hc6ng Жыл бұрын
Good Insight
@mariasargsyan51702 жыл бұрын
Hi, thanks for explaining the change of variables in here. However, it is generally not true that the G is an invertible function, especially in the context of Deep Neural Networks, right or am I missing something?
@CarlosGarcia-hs8yg4 жыл бұрын
best video ever. Thanks a lot man!
@karimmache40182 жыл бұрын
Very interesting lecture, but please make your notation clear.
@aakanksha78773 жыл бұрын
thanks sir, you made it relatiively simple and effective.
@prateekthakur25192 жыл бұрын
After reading so many blogs, papers, videos and paying for courses... finally I found the best explanation!!! Thank you so much! Just one question while take derivative of integrad why did we not consider the derivative of PDFs Pdata(x) and Pg(x) ?
@elyseemanimpiregasana21172 жыл бұрын
You know to change the output we use gradient descent. In that case, we differentiate the loss with respect to the parameters of the Generator and Discriminator. No need for a probability derivative, I hope it is understandable, for any misunderstanding let me know.
@nadarasarbahavan62694 жыл бұрын
Amazing explanation. I am telling everybody
@bSharpHacker3 жыл бұрын
What happened to the x's inside the integrals of equation 1, after combining 1 & 2?
@amarnathjagatap23394 жыл бұрын
Best explanation sir.. Thank you sir
@ayanmaiti17894 жыл бұрын
i did not get how to convert instance to sample? that is how we arrived at the expectation terms ? any pointers to good reference?
@superaluis4 жыл бұрын
I too got a little bit confused on this part... If you find on internet any explanation about this part please share here.
@pareshupadhyay2973 жыл бұрын
Thank you for the amazing explanation .I have one query that is-- You said that random variable x has a distribution Pdata(x) and at time 9:25 you said that x=G(z), but this only be true after many iteration, my question is why this assumption is true at the starting.
@elyseemanimpiregasana21172 жыл бұрын
I think you fix it as a target to achieve since our wish is to generate output similar to the existing data that will fool the discriminator.
@ankurgupta28064 жыл бұрын
what can be the intuitive explanation of the change between the two integrals?
@pratik99485 жыл бұрын
can u make a video on gan loss functions that is which loss function to be used for generator and disriminatoor and also combine model which to use and why it is hard to train gan thanks in advance
@ivanrezic5 жыл бұрын
Hello, great video! Can you please point me to the proof of that transformation formula used at 7:29?
@AhladKumar5 жыл бұрын
see Dependent variables and change of variables at the following link en.wikipedia.org/wiki/Probability_density_function
@bosepukur5 жыл бұрын
In equation 2 should it not be D(G(x)) ----> D (G(z))