Lesson 9: Deep Learning Foundations to Stable Diffusion

  Рет қаралды 152,750

Jeremy Howard

Jeremy Howard

Күн бұрын

Пікірлер: 72
@rajahx
@rajahx Жыл бұрын
This is beautifully explained Jeremy! From real basics to some of the most complicated state of the art models we have today. Bravo.
@TTTrouble
@TTTrouble 2 жыл бұрын
Wow this is such a treasure to have freely available and I am so thankful that you put this out for the community. Many many thanks good sir, your work towards educating the masses about AI and Machine Learning is so very much appreciated. 🎉❤
@flavioferlin3127
@flavioferlin3127 2 ай бұрын
Such a brilliant, humble, altruistic, knowledgeable person is a Godsend. Thank you Sir! Every time you watch this you can’t help but appreciate his way of teaching and being inspired to follow the outstanding work.
@numannebuni
@numannebuni 2 жыл бұрын
I absolutely love the style in which this is explained. Thank you very much!
@howardjeremyp
@howardjeremyp 2 жыл бұрын
Glad you like it!
@mamotivated
@mamotivated 2 жыл бұрын
Liberating the world with this quality of education
@gilbertobatres-estrada5119
@gilbertobatres-estrada5119 Жыл бұрын
I am so glad you took your time to correct the math mistake! Great work! And thank you for your mission of teaching us new findings in AI and deep learning 🙏
@MuhammadJaalouk
@MuhammadJaalouk Жыл бұрын
Thank you so much for this insightful video. The lecture breaks down complex ideas into segments that are very easy to comprehend.
@AIBites
@AIBites Жыл бұрын
This is a nicely thought-through course. Amazing Jeremy! :)
@chyldstudios
@chyldstudios 2 жыл бұрын
Wonderful, I was waiting for these series of videos. Bravo!
@ItzGanked
@ItzGanked 2 жыл бұрын
I thought I was going to have to wait until next year, thank you for making this content accessible
@asheeshmathur
@asheeshmathur Жыл бұрын
Outstanding, the best description so far. God Bless Jeremy. Excellent service to curious souls.
@kartikpodugu
@kartikpodugu Жыл бұрын
🙏🙏🙏 Amazing information. I knew bits and pieces, now I know the entire picture.
@cybermollusk
@cybermollusk Жыл бұрын
You might want to put this series into a playlist. I see you have playlists for all your other courses.
@marko.p.radojcic
@marko.p.radojcic 7 ай бұрын
I am getting KZbin premium just! to download this series. Thank you!
@sushilkhadka8069
@sushilkhadka8069 Жыл бұрын
Excellent intiution. You're doing the huge service to humanity
@peregudovoleg
@peregudovoleg 9 ай бұрын
At 1:13:20 aren't we supposed to add derivatives to pixel values since we are maximizing P? Unless, since P is binary and it looks like a classification problem, we are going to get negative logits, then deducting seems ok (not touching the sign). Great course!
@sotasearcher
@sotasearcher 11 ай бұрын
28:36 - I'm here in February '24, where they are good enough to do it in 1 go with SDXL-Turbo / ADD (Nov '23) :)
@danielhabibio
@danielhabibio Ай бұрын
2024, the fundamentals are still sound. Thank you Jeremy and team!
@akheel_khan
@akheel_khan Жыл бұрын
Undoubtedly an accessible and insightful guide
@208muppallahindu5
@208muppallahindu5 6 ай бұрын
Thank you , Jeremy Howard for teaching me concepts of diffusion.
@ricardocalleja
@ricardocalleja 2 жыл бұрын
Awesome material! Thank you very much for sharing
@howardjeremyp
@howardjeremyp 2 жыл бұрын
My pleasure!
@ghpkishore
@ghpkishore Жыл бұрын
That math correction was very essential to me. Coming from a mechanical background, I knew something was off, but then thought I didn't know enough about DL to figure out what it is, and that I was on the wrong. With the math correction, it clicked, and was something I knew all along.Thanks.
@sushilkhadka8069
@sushilkhadka8069 Жыл бұрын
at 1:56:50 I'm having hard time understanding the cost function. I think we need to maximise ( Green Summation - Red Summation ) , for that reason we can't call it a cost function because cost functions are usually minimised. Please correct me If I'm wrong
@Omunamantech
@Omunamantech 2 ай бұрын
Marvellous, Beautiful, Loved It :D
@ayashiyumi
@ayashiyumi Жыл бұрын
Ottimo video. Continua a pubblicare altre cose del genere.
@johngrabner
@johngrabner 2 жыл бұрын
Very informative video. Thank you for taking the time to produce.
@SadAnecdote
@SadAnecdote 2 жыл бұрын
Thanks for the early release
@rashulsel
@rashulsel Жыл бұрын
Amazing video and really easy to follow up with the topics. Its neat how different research is coming together to build something more efficient and promising. So future of AI is how models fit together?
@HoomanRafraf
@HoomanRafraf Ай бұрын
Thanks for the awesome elaboration. There's one thing that was not clear to me in the process and I wasn't quite able to find the answer by googling: Why do we need CLIP and can't use a regular text encoder? As far as I can see, in this architecture what matters for the role of the text encoder is to give similar embeddings for similar contexts, e.g. "a beautiful swan" and "such a lovely swan" should have similar embeddings. But why should this embedding also be similar to the embeddings for images of the swan?
@SubhadityaMukherjee
@SubhadityaMukherjee 2 жыл бұрын
YAY its hereeee. My excitement!!
@mariuswuyo8742
@mariuswuyo8742 2 жыл бұрын
Very an excellent course, I would like to ask a question that the noise N(0,0.1) is added equally to each pixel or to the whole image at 1:21:51? These two are equivalent?
@user-wf3bp5zu3u
@user-wf3bp5zu3u 2 жыл бұрын
Different per pixel! You’re drawing a vector of random noise samples then reshaping it into an image, so you get many values but all from a dist with low variance. The python random numbers let you sample in the shape of an image directly, so you don’t need to manually reshape. But that’s just for convenience
@atNguyen-gt6nd
@atNguyen-gt6nd Жыл бұрын
Thank you so much for your lectures.
@kirak
@kirak Жыл бұрын
Wow this helped me a lot. Thank you!
@edmondj.
@edmondj. 2 жыл бұрын
I love you, its so clear as usual, i owed you embeddings, now i owe you diffusion too.
@edmondj.
@edmondj. 2 жыл бұрын
Please open a tipee
@useless_deno
@useless_deno 4 ай бұрын
Amazingly explained !
@pranavkulkarni6489
@pranavkulkarni6489 2 жыл бұрын
Thank you for great explanation .. I just wanted to know ans to 'what is U-net?' could not understand where is it used in whole process ? I mean what I could not get is what is the difference between VAE (Autoencoder) and an Unet
@tildebyte
@tildebyte Жыл бұрын
During *training*, you pass an actual image into the VAE ENcoder (to reduce the amount of data you have to deal with), which then passes the latent it produces on to the UNet, which does the learning involving noising/denoising the latent. During *inference* ("generating"), the UNet (after a lot of other stuff happens :D) passes out a denoised latent to the VAE DEcoder, which then produces an actual image
@adityagupta-hm2vs
@adityagupta-hm2vs 13 күн бұрын
Why do we predict the noise and not the number as the output itself ? Is it because loss calculation is easier with that ?
@ramansarabha871
@ramansarabha871 Жыл бұрын
Thanks a ton! have been waiting.
@howardjeremyp
@howardjeremyp Жыл бұрын
Hope you like it!
@pankajsinghrawat1056
@pankajsinghrawat1056 8 ай бұрын
since we want to increase the probability of our image being a digit, we should "add" and not "substract" the grad of probability wrt to img. Is this right? or am I missing something ?
@homataha5626
@homataha5626 2 жыл бұрын
Can I ask a video of these model that is used for colonization?
@andrewimanuel2838
@andrewimanuel2838 Жыл бұрын
where can I find the latest recommended cloud computing resource?
@tinkeringengr
@tinkeringengr 2 жыл бұрын
Thanks -- great lecture!
@mikhaeldito
@mikhaeldito 2 жыл бұрын
Released already??
@jonatan01i
@jonatan01i Жыл бұрын
Good thing is that with git we can go back to the state of the code as of (11/20).10.2022
@ДмитроПавличко-п9д
@ДмитроПавличко-п9д 2 жыл бұрын
Thank you for this lecture
@tildebyte
@tildebyte Жыл бұрын
I've been working on/with diffusion models (and before that VQGANs!) for years now, so I'm pretty familiar (from the end-user/theoretical POV, not so much the math/code side heh) with samplers/schedulers - this is the first time I've conceived of them as optimizers, and that seems like a *really* fertile area to research. Have you (or anyone else, for that matter) made any progress in this direction? It's (not too surprisingly) VERY hard to prompt today's search engines to find anything to do with "denoise|diffusion|schedule|sample|optimize" and NOT come up with dozens of either HuggingFace docs pages, or pages w.r.t. Stable DIffusion ROFL
@edwardhiscoke471
@edwardhiscoke471 2 жыл бұрын
Already out, wow. Then I'd better push on with part 1!
@super-eth8478
@super-eth8478 2 жыл бұрын
THANKS 🙏🏻🙏🏻
@rubensmau
@rubensmau 2 жыл бұрын
Thanks, very clear.
@kawalier1
@kawalier1 Жыл бұрын
Jerremy, Adam has eps, SGD momentum
@sotasearcher
@sotasearcher 11 ай бұрын
52:12 - The upside down triangle is "nabla", "del" is the squiggly d that goes before each partial derivative. Also, I'm jealous of people who started calculus in high school lol
@sotasearcher
@sotasearcher 11 ай бұрын
Nevermind! Getting to the next section edited in lol
@sotasearcher
@sotasearcher 11 ай бұрын
1:04:12 wait you still mixed them up 😅 At this rate with your following, you're going to speak it into existence though lol. Math notation ultimately is whatever everyone agrees upon, so I could see it being possible.
@sotasearcher
@sotasearcher 11 ай бұрын
1:05:22 - While I'm being nit-picky - Right-side-up triangle is called "delta", and just means change, not necessarily small
@gustavojuantorena
@gustavojuantorena Жыл бұрын
👏👏👏
@mohdil123
@mohdil123 2 жыл бұрын
Awesome
@AndrewRafas
@AndrewRafas 2 жыл бұрын
There is a small (not that important) correction: when you talk about 16384 bytes of latents, it is 16384 numbers, which are 65536 bytes in fact.
@sambitmukherjee1713
@sambitmukherjee1713 9 ай бұрын
Each number is 4 bytes because it's a float32 precision?
@flavioferlin3127
@flavioferlin3127 3 ай бұрын
Wow!
@susdoge3767
@susdoge3767 10 ай бұрын
gold
@TiagoVello
@TiagoVello 2 жыл бұрын
UHUUUL ITS OUT
@JingyingAIEducation
@JingyingAIEducation Жыл бұрын
I was wondering if I could give some suggestions, you spent 20 mins to explain the course materials and different people. Why not start from main fun play first, then later introduce more about the course materials. People will lose interest to listen to 20 mins course materials.
@yufengchen4944
@yufengchen4944 2 жыл бұрын
Great! I can only see 2019 version of Part 2, look foward to see the new Part 2 course available!
@yufengchen4944
@yufengchen4944 2 жыл бұрын
Looks like the part 2 2022 webpage is still not public right? or I didn't find the way?
@michaelnurse9089
@michaelnurse9089 2 жыл бұрын
I thought I was going to have to wait until next year, thank you for making this content accessible
@Beyondarmonia
@Beyondarmonia 2 жыл бұрын
Thank you 🙏
Lesson 10: Deep Learning Foundations to Stable Diffusion, 2022
1:49:14
Why Does Diffusion Work Better than Auto-Regression?
20:18
Algorithmic Simplicity
Рет қаралды 420 М.
Леон киллер и Оля Полякова 😹
00:42
Канал Смеха
Рет қаралды 4,7 МЛН
黑天使被操控了#short #angel #clown
00:40
Super Beauty team
Рет қаралды 61 МЛН
СИНИЙ ИНЕЙ УЖЕ ВЫШЕЛ!❄️
01:01
DO$HIK
Рет қаралды 3,3 МЛН
Cat mode and a glass of water #family #humor #fun
00:22
Kotiki_Z
Рет қаралды 42 МЛН
Practical Deep Learning for Coders: Lesson 1
1:22:56
Jeremy Howard
Рет қаралды 386 М.
Visualizing transformers and attention | Talk for TNG Big Tech Day '24
57:45
AI can't cross this line and we don't know why.
24:07
Welch Labs
Рет қаралды 1,5 МЛН
Transformers (how LLMs work) explained visually | DL5
27:14
3Blue1Brown
Рет қаралды 4,3 МЛН
The U-Net (actually) explained in 10 minutes
10:31
rupert ai
Рет қаралды 134 М.
Learn Machine Learning Like a GENIUS and Not Waste Time
15:03
Infinite Codes
Рет қаралды 358 М.
Genius Machine Learning Advice for 10 Minutes Straight
9:46
Data Sensei
Рет қаралды 101 М.
How Stable Diffusion Works (AI Image Generation)
30:21
Gonkee
Рет қаралды 162 М.
Леон киллер и Оля Полякова 😹
00:42
Канал Смеха
Рет қаралды 4,7 МЛН