This is beautifully explained Jeremy! From real basics to some of the most complicated state of the art models we have today. Bravo.
@TTTrouble2 жыл бұрын
Wow this is such a treasure to have freely available and I am so thankful that you put this out for the community. Many many thanks good sir, your work towards educating the masses about AI and Machine Learning is so very much appreciated. 🎉❤
@flavioferlin31272 ай бұрын
Such a brilliant, humble, altruistic, knowledgeable person is a Godsend. Thank you Sir! Every time you watch this you can’t help but appreciate his way of teaching and being inspired to follow the outstanding work.
@numannebuni2 жыл бұрын
I absolutely love the style in which this is explained. Thank you very much!
@howardjeremyp2 жыл бұрын
Glad you like it!
@mamotivated2 жыл бұрын
Liberating the world with this quality of education
@gilbertobatres-estrada5119 Жыл бұрын
I am so glad you took your time to correct the math mistake! Great work! And thank you for your mission of teaching us new findings in AI and deep learning 🙏
@MuhammadJaalouk Жыл бұрын
Thank you so much for this insightful video. The lecture breaks down complex ideas into segments that are very easy to comprehend.
@AIBites Жыл бұрын
This is a nicely thought-through course. Amazing Jeremy! :)
@chyldstudios2 жыл бұрын
Wonderful, I was waiting for these series of videos. Bravo!
@ItzGanked2 жыл бұрын
I thought I was going to have to wait until next year, thank you for making this content accessible
@asheeshmathur Жыл бұрын
Outstanding, the best description so far. God Bless Jeremy. Excellent service to curious souls.
@kartikpodugu Жыл бұрын
🙏🙏🙏 Amazing information. I knew bits and pieces, now I know the entire picture.
@cybermollusk Жыл бұрын
You might want to put this series into a playlist. I see you have playlists for all your other courses.
@marko.p.radojcic7 ай бұрын
I am getting KZbin premium just! to download this series. Thank you!
@sushilkhadka8069 Жыл бұрын
Excellent intiution. You're doing the huge service to humanity
@peregudovoleg9 ай бұрын
At 1:13:20 aren't we supposed to add derivatives to pixel values since we are maximizing P? Unless, since P is binary and it looks like a classification problem, we are going to get negative logits, then deducting seems ok (not touching the sign). Great course!
@sotasearcher11 ай бұрын
28:36 - I'm here in February '24, where they are good enough to do it in 1 go with SDXL-Turbo / ADD (Nov '23) :)
@danielhabibioАй бұрын
2024, the fundamentals are still sound. Thank you Jeremy and team!
@akheel_khan Жыл бұрын
Undoubtedly an accessible and insightful guide
@208muppallahindu56 ай бұрын
Thank you , Jeremy Howard for teaching me concepts of diffusion.
@ricardocalleja2 жыл бұрын
Awesome material! Thank you very much for sharing
@howardjeremyp2 жыл бұрын
My pleasure!
@ghpkishore Жыл бұрын
That math correction was very essential to me. Coming from a mechanical background, I knew something was off, but then thought I didn't know enough about DL to figure out what it is, and that I was on the wrong. With the math correction, it clicked, and was something I knew all along.Thanks.
@sushilkhadka8069 Жыл бұрын
at 1:56:50 I'm having hard time understanding the cost function. I think we need to maximise ( Green Summation - Red Summation ) , for that reason we can't call it a cost function because cost functions are usually minimised. Please correct me If I'm wrong
@Omunamantech2 ай бұрын
Marvellous, Beautiful, Loved It :D
@ayashiyumi Жыл бұрын
Ottimo video. Continua a pubblicare altre cose del genere.
@johngrabner2 жыл бұрын
Very informative video. Thank you for taking the time to produce.
@SadAnecdote2 жыл бұрын
Thanks for the early release
@rashulsel Жыл бұрын
Amazing video and really easy to follow up with the topics. Its neat how different research is coming together to build something more efficient and promising. So future of AI is how models fit together?
@HoomanRafrafАй бұрын
Thanks for the awesome elaboration. There's one thing that was not clear to me in the process and I wasn't quite able to find the answer by googling: Why do we need CLIP and can't use a regular text encoder? As far as I can see, in this architecture what matters for the role of the text encoder is to give similar embeddings for similar contexts, e.g. "a beautiful swan" and "such a lovely swan" should have similar embeddings. But why should this embedding also be similar to the embeddings for images of the swan?
@SubhadityaMukherjee2 жыл бұрын
YAY its hereeee. My excitement!!
@mariuswuyo87422 жыл бұрын
Very an excellent course, I would like to ask a question that the noise N(0,0.1) is added equally to each pixel or to the whole image at 1:21:51? These two are equivalent?
@user-wf3bp5zu3u2 жыл бұрын
Different per pixel! You’re drawing a vector of random noise samples then reshaping it into an image, so you get many values but all from a dist with low variance. The python random numbers let you sample in the shape of an image directly, so you don’t need to manually reshape. But that’s just for convenience
@atNguyen-gt6nd Жыл бұрын
Thank you so much for your lectures.
@kirak Жыл бұрын
Wow this helped me a lot. Thank you!
@edmondj.2 жыл бұрын
I love you, its so clear as usual, i owed you embeddings, now i owe you diffusion too.
@edmondj.2 жыл бұрын
Please open a tipee
@useless_deno4 ай бұрын
Amazingly explained !
@pranavkulkarni64892 жыл бұрын
Thank you for great explanation .. I just wanted to know ans to 'what is U-net?' could not understand where is it used in whole process ? I mean what I could not get is what is the difference between VAE (Autoencoder) and an Unet
@tildebyte Жыл бұрын
During *training*, you pass an actual image into the VAE ENcoder (to reduce the amount of data you have to deal with), which then passes the latent it produces on to the UNet, which does the learning involving noising/denoising the latent. During *inference* ("generating"), the UNet (after a lot of other stuff happens :D) passes out a denoised latent to the VAE DEcoder, which then produces an actual image
@adityagupta-hm2vs13 күн бұрын
Why do we predict the noise and not the number as the output itself ? Is it because loss calculation is easier with that ?
@ramansarabha871 Жыл бұрын
Thanks a ton! have been waiting.
@howardjeremyp Жыл бұрын
Hope you like it!
@pankajsinghrawat10568 ай бұрын
since we want to increase the probability of our image being a digit, we should "add" and not "substract" the grad of probability wrt to img. Is this right? or am I missing something ?
@homataha56262 жыл бұрын
Can I ask a video of these model that is used for colonization?
@andrewimanuel2838 Жыл бұрын
where can I find the latest recommended cloud computing resource?
@tinkeringengr2 жыл бұрын
Thanks -- great lecture!
@mikhaeldito2 жыл бұрын
Released already??
@jonatan01i Жыл бұрын
Good thing is that with git we can go back to the state of the code as of (11/20).10.2022
@ДмитроПавличко-п9д2 жыл бұрын
Thank you for this lecture
@tildebyte Жыл бұрын
I've been working on/with diffusion models (and before that VQGANs!) for years now, so I'm pretty familiar (from the end-user/theoretical POV, not so much the math/code side heh) with samplers/schedulers - this is the first time I've conceived of them as optimizers, and that seems like a *really* fertile area to research. Have you (or anyone else, for that matter) made any progress in this direction? It's (not too surprisingly) VERY hard to prompt today's search engines to find anything to do with "denoise|diffusion|schedule|sample|optimize" and NOT come up with dozens of either HuggingFace docs pages, or pages w.r.t. Stable DIffusion ROFL
@edwardhiscoke4712 жыл бұрын
Already out, wow. Then I'd better push on with part 1!
@super-eth84782 жыл бұрын
THANKS 🙏🏻🙏🏻
@rubensmau2 жыл бұрын
Thanks, very clear.
@kawalier1 Жыл бұрын
Jerremy, Adam has eps, SGD momentum
@sotasearcher11 ай бұрын
52:12 - The upside down triangle is "nabla", "del" is the squiggly d that goes before each partial derivative. Also, I'm jealous of people who started calculus in high school lol
@sotasearcher11 ай бұрын
Nevermind! Getting to the next section edited in lol
@sotasearcher11 ай бұрын
1:04:12 wait you still mixed them up 😅 At this rate with your following, you're going to speak it into existence though lol. Math notation ultimately is whatever everyone agrees upon, so I could see it being possible.
@sotasearcher11 ай бұрын
1:05:22 - While I'm being nit-picky - Right-side-up triangle is called "delta", and just means change, not necessarily small
@gustavojuantorena Жыл бұрын
👏👏👏
@mohdil1232 жыл бұрын
Awesome
@AndrewRafas2 жыл бұрын
There is a small (not that important) correction: when you talk about 16384 bytes of latents, it is 16384 numbers, which are 65536 bytes in fact.
@sambitmukherjee17139 ай бұрын
Each number is 4 bytes because it's a float32 precision?
@flavioferlin31273 ай бұрын
Wow!
@susdoge376710 ай бұрын
gold
@TiagoVello2 жыл бұрын
UHUUUL ITS OUT
@JingyingAIEducation Жыл бұрын
I was wondering if I could give some suggestions, you spent 20 mins to explain the course materials and different people. Why not start from main fun play first, then later introduce more about the course materials. People will lose interest to listen to 20 mins course materials.
@yufengchen49442 жыл бұрын
Great! I can only see 2019 version of Part 2, look foward to see the new Part 2 course available!
@yufengchen49442 жыл бұрын
Looks like the part 2 2022 webpage is still not public right? or I didn't find the way?
@michaelnurse90892 жыл бұрын
I thought I was going to have to wait until next year, thank you for making this content accessible