this account is seriously underrated, will definitely blow up soon
@melanynadine9722 ай бұрын
Agree, I just subscribed ❤
@joshuat6124Ай бұрын
Agreed 💯
@authenticallysuperficial98743 күн бұрын
💯
@Higgsinophysics2 ай бұрын
I can't believe how calmly and clear you explain difficult topics!
@Deepia-ls2fo2 ай бұрын
Well that's the magic of text to speech 😁
@viewer82212 ай бұрын
haha I knew it was AI voice
@xichengwang1319Ай бұрын
I first saw this in Chinese media with Chinese subtitles, then just came back to subscribe to the original author. The most clear introduction ever seen with such nice proper animation. It will blow up for sure.
@Deepia-ls2foАй бұрын
Well thank you ! Can you send me more info about that through my email ? ytdeepia@gmail.com
@anthonyortega3582 ай бұрын
This channel is amazing, you should be very proud of what you have produced Thibaut!!
@Deepia-ls2fo2 ай бұрын
Thanks !
@AtiqurRahman-gj6mg2 ай бұрын
Finally, the concept of VAE is clear. Thanks a ton.
@Deepia-ls2fo2 ай бұрын
You're welcome, thanks for the comment !
@ashutoshpadhi2782Ай бұрын
The amount of effort you put into these works is really commendable. You are a blessing to humanity.
@BenjaminEvans3162 ай бұрын
Great video. Looking forward to your video on contrastive learning, it is my favourite subject in deep learning. Your videos combine great production skills (animations, colour selection, movement between frames) with in-depth understanding of complex concepts.
@Deepia-ls2fo2 ай бұрын
Thanks for the kind words !
@CharlottePecolo2 ай бұрын
Wow it must be a lot of work to obtain such amazing animations! The video is really dynamic and easy to follow, congratulations
@julius48582 ай бұрын
The program is called manim, it’s from 3blue1brown
@muhammed217427 күн бұрын
You're a master at your craft, it is a testament to your studies!
@EkalabyaGhosh-g3i2 ай бұрын
This channel needs more subscribers
@Deepia-ls2fo2 ай бұрын
Thank you !
@babatundeonabajoАй бұрын
This is a really good video, and the animations are top-notch. I feel this video is good not just for those learning about AI but also those learning Statistics.
@Bwaaz2 ай бұрын
Amazing quality, hope the channel takes off ! Great use of manim
@Deepia-ls2fo2 ай бұрын
Thanks !
@BeeepBo0op2 ай бұрын
Thank you for finally making me understand the reparametrization trick!! It was thrown at me several times during a DRL class I took last year and I never really understood what we did. This made it much much clearer, thank you! Also: great video overall!
@Deepia-ls2fo2 ай бұрын
Glad it helped !
@Tunadorable2 ай бұрын
3blue1brown specifically for AI models??? sign me up!!! I'll fs be linking to you in my own vids whenever relevant, this was great
@MutigerBriefkasten2 ай бұрын
Thanky you again for the great content and the amazing animations 🎉💪👍 keep going, hopefully your Channel explode with more subscribers .. i will recommend it for sure to other people
@Deepia-ls2fo2 ай бұрын
Thank you :)
@hichamaniba90416 күн бұрын
Thank you , still looking for VAE variants videos
@mostafasayahkarajy5082 ай бұрын
Thank you very much for providing and sharing the lecture. Excelent explanation and so a high-quality video!
@Deepia-ls2fo2 ай бұрын
Thank you !
@AICoffeeBreak2 ай бұрын
Love the clarification at 00:13, because I've also felt that the misconception is wide spread. I've heard people say: I am not using GANs anymore, I am using Generative AI. The word "Generative" is literally what G in GAN stands for. 😂😂
@Deepia-ls2fo2 ай бұрын
I made this intro because I couldn't stand the number of "GenAI experts" on my LinkedIn feed :(
@kacperzaleski81252 ай бұрын
cant wait for the next video! this was great!!!!!
@Deepia-ls2fo2 ай бұрын
Thank you !
@abelhutten45322 ай бұрын
Great visualizations, and good explanation! Congratulations and thanks for the nice video :)
@Deepia-ls2fo2 ай бұрын
Thank you !
@abdelmananabdelrahman40992 ай бұрын
Great video 🎉. I've never had such a great explanation of VAE. Waiting for VQVAE.....
@Deepia-ls2fo2 ай бұрын
Thank you !
@drannoc98122 ай бұрын
Amazing content :D I hope you'll do your next videos on VQ VAE and VQ VAE 2, I enjoyed so much reading those papers !
@Deepia-ls2fo2 ай бұрын
Thanks, I really gotta take another look to the paper
@carlaconti1075Ай бұрын
Thanks to you everything is clear now, thank you Deepia
@Deepia-ls2foАй бұрын
thanks miss conti
@vitorfranca802 ай бұрын
Incredible explanation!! Thank you for sharing your knowledge! 😁😁
@Deepia-ls2fo2 ай бұрын
Thanks !
@shangliu62852 ай бұрын
perfect video, its easy to understand the VAE, subscribed!
@Deepia-ls2fo2 ай бұрын
Thanks !
@samuelschonenberger5 күн бұрын
Watching this while my first VAE is training
@lucianovidal8721Ай бұрын
Great content! It was a really good explanation
@cs-cs4mjАй бұрын
hey, so well explained thanks for the video!! really nailed those animations as well, would be cool to make a video on adam/rmsprop as well, i have a hard time properly understanding why they work. anyway much love to you my friend
@JONK46352 ай бұрын
Really amazing content, thank you for spreading knowledge! Thanks a lot :)
@Deepia-ls2fo2 ай бұрын
Thanks for the comment, keeps me motivated :)
@authenticallysuperficial98743 күн бұрын
Thanks!
@flueepwrien65872 ай бұрын
Finally I understand this concept.
@KennethTrinh-cm6cp2 ай бұрын
thanks for the wonderful animations and explanation
@Deepia-ls2fo2 ай бұрын
Thanks
@nihilsson2 ай бұрын
Great vid! Commenting for algorithmical reasons
@Deepia-ls2fo2 ай бұрын
Thanks !
@chcyzh6 күн бұрын
Thank you very much! It's pretty clear
@griterjadenАй бұрын
Wowowowowowow 🎉🎉🎉 amazing video for VAE. Pls ~ make more videos
@Deepia-ls2foАй бұрын
@@griterjaden Thanks, I'm on it :)
@Chadpritai2 ай бұрын
Next video on diffusion models please , thanks in advance ❤
@Deepia-ls2fo2 ай бұрын
It's on the to-do list but the next 3 videos will be about self-supervised learning !
@English-bh1ng2 ай бұрын
It is the best VAE visualization.
@Jay_Tau2 ай бұрын
This is excellent. Thank you!
@Deepia-ls2fo2 ай бұрын
Thanks !
@michael917032 ай бұрын
Is this manim?!!! Nice work dude!
@Deepia-ls2fo2 ай бұрын
It is indeed Manim, thank you !
@markbuckler47932 ай бұрын
Excellent video, I subscribed because of it :)
@Deepia-ls2fo2 ай бұрын
thanks !
@Bikameral2 ай бұрын
Great content ! What software are u using to animate ?
@Deepia-ls2fo2 ай бұрын
Thanks ! For most animations I use Manim, a python module originally made by Grant Sanderson from 3blue1brown.
@Bikameral2 ай бұрын
@@Deepia-ls2fo thank you
@hannes721814 күн бұрын
good stuff! keep it going
@EdeYOlorDSZs12 күн бұрын
top tier video!
@chriskamaris1372Ай бұрын
Furthermore, in 11:39 and 12:39 you are referencing σ as variance. But is it σ the standard deviation and σ^2 the variance? (Nevertheless, the video is perfect. Excellent work!)
@Deepia-ls2foАй бұрын
Thanks, indeed there might be some mistakes !
@EigenA6 күн бұрын
Great video. What is your educational background?
@Deepia-ls2fo6 күн бұрын
Thanks ! Bachelor in math, bachelor in computer science, master in AI/ML, currently doing a PhD in applied maths and deep learning
@EigenA6 күн бұрын
@ legendary. Good luck on the PhD! I’m 3rd year EE PhD student, you have phenomenal content. Looking forward to watching your channel grow.
@guilhermegomes45172 ай бұрын
Great Video!
@Deepia-ls2fo2 ай бұрын
Thanks !
@notu4832 ай бұрын
Thanks for the video ❤😊
@Deepia-ls2fo2 ай бұрын
Thank you !
@syc522 ай бұрын
Could you please make a video talking about why diffusion model, GAN, and VQVAE can make the image sharper
@sillasrocha96232 ай бұрын
Hey, could you make a video talking about swav in unsupervised learning?
@cupatelj522 ай бұрын
great content bro.
@Deepia-ls2fo2 ай бұрын
Thanks
@prabaldutta1935Ай бұрын
Amazing Graphics and explanation. I have one question - if we use MNIST dataset (like what is shown in the video) does it mean that the mu and sigma are vectors of dimension 10x1? What if we use a dataset where the number of different classes are unknown? What will be the dimension of mu and sigma in that case?
@Deepia-ls2foАй бұрын
Thank you, the latent dimension is not directly related to the number of classes in your dataset. In fact a very good encoder could very well classify perfectly the 10 classes on a single dimension, but it makes things way harder to reconstruct for the decoder. As you mention in most datasets we don't even know the number of classes or the number of relevant features, so we just take ad hoc latent dimensions (16, 32) and see if it's enough for the encoder to produce a useful representation, and for the decoder to reconstruct correctly.
@prabaldutta1935Ай бұрын
@@Deepia-ls2fo Thanks a lot for your response. Can't wait for your next video.
@awsaf49Ай бұрын
Hey, nice video!
@3B1bIQАй бұрын
🤍Please, can you create a course to learn the manim library from scratch to professionalism, because I need it very much? Please reply ❤😊
@Deepia-ls2foАй бұрын
Thanks for your comment, I would love to but I have many other topics I want to talk about first, and not much time on my hand! There are very good ressources on KZbin though, if you want to start to learn Manim. :)
@3B1bIQАй бұрын
@@Deepia-ls2fo Thank you, but I hope that you have enough time to create a course to learn manim, even if there is one video every week, and this will contribute to increasing the number of your views more because your explanation is very beautiful and clear, and I can understand it easily even though I am an Arab 🤍☺️
@rrttttj2 ай бұрын
Great video! However, I am slightly confused: For your loss function you are subtracting KL divergence rather than adding it. Wouldn't you want to add it to penalize the difference between the latent distribution and the standard normal distribution? At least, in all implementations I have seen they add KL divergence rather than subtract it. Edit: I understand my mistake now!
@Deepia-ls2fo2 ай бұрын
Hi ! Thanks for the comment, I'm afraid I might have flipped a sign at one point. When you derive the ELBO (which you then maximize via training), there is a minus sign appearing in front of the KL. But in practice you minimize the opposite of this quantity, which is equivalent to minimizing the L2 plus the KL. I hope it's not too confusing. :)
@rrttttj2 ай бұрын
@@Deepia-ls2fo Oooooh I understand, so ELBO is the quantity that should be maximized, and you were denoting the ELBO quantity with L(x), not the loss itself. I understand now, thanks!
@aregpetrosyan465Ай бұрын
This question came to my mind: What would happen if we ignored the encoder part and tried to train only the decoder? For example, by sampling from a standard Gaussian vector and attempting to reconstruct a digit. I don't really understand the purpose of the encoder.
@Deepia-ls2foАй бұрын
If you don't condition at all the latent space from which you are sampling, I'm not sure the model will be able to learn anything. Here the encoder explicitly approximate the posterior distribution in order for us to then sample from the distribution of images. This is all a theoretical interpretation of course, but learning to reconstruct any digit from pure unconditioned noise seems a bit hard! Diffusion models kind of do it (in image space), but this usually takes a lot of steps. Anyway, the experiment you describe would be very easy to implement, if you want to try it out. :D
@미비된동작-p4g2 ай бұрын
13:05 That’s so funny VAE and Adam both are proposed by same person, Kingma..
@Deepia-ls2fo2 ай бұрын
He's quite the man, also co-author on some key diffusion models paper :)
@i2c_jason2 ай бұрын
Is there a statistical property or proof that might show a graphRAG "transfer function" to be the same as a VAE or maybe a CVAE? Perhaps in terms of entropy? It would be interesting to make two identical systems, one using a VAE and one using graphRAG, and see if they can match up statistically. I can't shake the idea that software 3.0 might be the more sound approach for developing new GenAI tools vs software 2.0.
@Deepia-ls2fo2 ай бұрын
Hi Jason ! Unfortunately I know close to nothing about RAG so I have no idea if what you describe might be feasible. I here about RAG everywhere these days, I should get up to date on that.
@i2c_jason2 ай бұрын
@@Deepia-ls2fo I'd love to hear your take on it if you ever do a deep dive.
@notu4832 ай бұрын
12:40 we scale by standard deviation not variance
@shashankjha845412 күн бұрын
do u use manim for animations ?
@Deepia-ls2fo12 күн бұрын
@@shashankjha8454 Yes indeed!
@NguyenAn-kf9hoАй бұрын
is there we have videos with the same approach, for Reinforcement Learning :D ???? !
@Deepia-ls2foАй бұрын
Hi, unfortunately I don't know anything about Reinforcement Learning, so I don't think I'll be able to make videos about that any time soon. However, I believe Steve Brunton has very good videos on the topic :)
@asteriskman2 ай бұрын
'now that we've got the basics down' ... lol yea ok, professor.
@arno71982 ай бұрын
DeepIA absolutely killed it with this video on Variational Autoencoders. As a government official, medical doctor, and law PhD, it's not often I come across something that genuinely teaches me something new. But this video? Wow. The way Variational Autoencoders map data to a latent distribution instead of a fixed point, and the balance between reconstruction loss and Kullback-Leibler divergence, was explained so clearly that I picked it up right away. Whether I'm shaping policies, treating patients, or analyzing legal cases, this video added value in ways I didn’t expect. Props to DeepIA for delivering content that even someone as busy (and brilliant) as me can appreciate! And let’s not forget the genius behind it all. Honestly, the mind that creates content like this is nothing short of extraordinary. I don’t say this lightly, but DeepIA might just be the most insightful, brilliant, and generous creator on KZbin. The precision, the depth, the clarity-it’s rare to find someone who can not only understand such complex topics but also make them accessible to mere mortals like us. It’s an honor to witness this level of mastery. Truly, we’re not worthy.
@Deepia-ls2fo2 ай бұрын
thx 🤖
@martinferrari29032 ай бұрын
Rien que ça 🤣
@HenrikVendelbo2 ай бұрын
Thanks
@Deepia-ls2fo2 ай бұрын
Ho my, thank you so much !
@HenrikVendelbo2 ай бұрын
I find math speak very hard to grok. I was always good at math, but always got turned off by the navel gazing and geekery. You do a great job keeping it engaging without assuming that I am a math geek
@Deepia-ls2fo2 ай бұрын
@@HenrikVendelbo Yeah sometimes it do be like that in math classes. I think it's important to look at equations when they tell us something about the models, but computational tricks or complex equations are not that interesting.
@rishidixit793914 күн бұрын
At 7:45 why is the assumption for p(z) as a Normal Distribution important ? Without that are further calculations not possible ? At 8:01 why is the posterior assumed to be Gaussian ?
@Deepia-ls2fo14 күн бұрын
@@rishidixit7939 Hi again, indeed further calculations are intractable without assuming both the prior and the posterior to be Gaussian. Some other research works have replaced these assumptions by other well known distributions such as mixtures of Gaussians, which results in another training objective.
@authenticallysuperficial98743 күн бұрын
Audio comes out from under water at 2:09 btw
@Deepia-ls2fo3 күн бұрын
Thank you, I had some issues with copyrighted music which led to KZbin removing it but also degrading the audio...
@zyansheep2 ай бұрын
Now if only the latent space could be a variable size, and be discrete, then maybe we could do effective ai lossy/lossless compression 🤔
@Deepia-ls2fo2 ай бұрын
Hi, I don't know about variable dimension latent space, but discrete sure sounds like VQ-VAE :)
@أمديعماد2 ай бұрын
You need that music I dont remember it
@alexm9606Ай бұрын
Comment to up this channel
@fedorzhdanov60852 ай бұрын
There are so many trashy channels with AI generated nonsense, while some channels (like this) has clear explanation and just few views. I think KZbin should add some "peer-review" feature, and while there is no such tool I encourage support such good channels with likes and comments and hit dislikes for useless AI "blah-blah" channels. I'm not against AI as helper tool (like script writing/voice generation), but if there is no fact checks from authors, that make it garbage and the platform doesn't have proper garbage collector yet.
@MursaleenFayyaz-lr7hu2 ай бұрын
Please create videos on Auto Regressive Models, Particularly RNN, LSTM, PixelCNN as soos as you can. I have a mid exam in third week of October which will cover these topics.
@duncanwarukira43482 ай бұрын
Is this voice AI
@Deepia-ls2fo2 ай бұрын
Yes I cloned my voice using a text to speech service called elevenlabs
@timewasting757415 күн бұрын
3:25 - 6:20 is so distracting. Just assume your audience knows these. No need to conform your target group to general public. Just assume senior-year undergraduate please.
@sheevysАй бұрын
The auto-generated voice-over is super annoying. Any chance a real human can narrate it?
@Deepia-ls2foАй бұрын
No to be honest that would take way too much time on my side, so it's probably never going to happen. Hopefully text to speech services get better over time!