The Softmax : Data Science Basics

  Рет қаралды 52,865

ritvikmath

ritvikmath

Күн бұрын

Пікірлер: 116
@wennie2939
@wennie2939 3 жыл бұрын
I really love how you progress step by step instead of directly throwing out the formulas! The best video on KZbin on the Softmax! +1
@birajkoirala5383
@birajkoirala5383 4 жыл бұрын
tutorials with boards noww...nice one dude...underrated channel I must say!
@ritvikmath
@ritvikmath 4 жыл бұрын
Much appreciated!
@MrDullBull
@MrDullBull 3 жыл бұрын
agreed. greetings from russia!
@DFCinBE
@DFCinBE Жыл бұрын
For a non-mathematician like myself, this was crystal clear, thanks very much!
@akum007
@akum007 3 ай бұрын
I love it! The choice of examples and the way you explain ... just bravo!
@ritvikmath
@ritvikmath 3 ай бұрын
Thank you! 😃
@debapriyabanerjee8486
@debapriyabanerjee8486 3 жыл бұрын
This is excellent! I saw your video on the sigmoid function and both of these explain the why behind their usage.
@ritvikmath
@ritvikmath 3 жыл бұрын
Glad it was helpful!
@omniscienceisdead8837
@omniscienceisdead8837 2 жыл бұрын
the person who is going to be responsible for me kick starting my ML journey with a good head on my shoulders, thank you ritvik, very enlightening
@okeuwechue9238
@okeuwechue9238 10 ай бұрын
Thnx. Very clear explanation of the rationale for employing exponential fns instead of linear fns
@ritvikmath
@ritvikmath 10 ай бұрын
Great to hear!
@marcusakiti7608
@marcusakiti7608 2 жыл бұрын
Awesome stuff. Searched this video because I was trying to figure out why the scores/sum scores approach wouldn't work and you addressed it first thing. Great job.
@iraklisalia9102
@iraklisalia9102 3 жыл бұрын
What a great explanation! Thank you very much. The why do we choose this formula versus this formula explanation is truly makes everything clear. Thank you once again :)
@MORE2Clay
@MORE2Clay 3 жыл бұрын
The introduction to softmax which explains why softmax exists helped me a lot understanding it
@suparnaprasad8187
@suparnaprasad8187 3 ай бұрын
This channel is literally the best.
@ritvikmath
@ritvikmath 3 ай бұрын
Thanks!
@zvithaler9443
@zvithaler9443 2 жыл бұрын
Great explenations, your addition of the story to the objects really help understanding the material
@ekaterinakorneeva4792
@ekaterinakorneeva4792 Жыл бұрын
Thank you!!! This is so much clearer and straighter than 2 20-minutes videos on Softmax from "Machine Learning with Python-From Linear Models to Deep Learning" from MIT! To be fair, the latter explains multiple perspectives and is also good in its sense. But you deliver just the most importaint first bit of what is softmax and what are all these terms are about.
@ritvikmath
@ritvikmath Жыл бұрын
Glad it helped!
@ManpreetKaur-ve5gw
@ManpreetKaur-ve5gw 3 жыл бұрын
The only video I needed to understand the SOFTMAX function. Kudos to you!!
@zafarnasim9267
@zafarnasim9267 2 жыл бұрын
Woooow ,really liked our teaching approach, awesome!
@vamshi755
@vamshi755 4 жыл бұрын
Now i know why lot of your videos answers WHY question. You give importance to application not the theory alone. concept is very clear. thanks
@johnlabarge
@johnlabarge Ай бұрын
This is a glorious explanation.
@somteezle1348
@somteezle1348 4 жыл бұрын
Wow...teaching from first principles...I love that!
@ritvikmath
@ritvikmath 4 жыл бұрын
Glad you liked it!
@karimamakhlouf2411
@karimamakhlouf2411 2 жыл бұрын
An excellent and straightforward way of explaining. So helpful! Thanks a lot :)
@rizkysyahputra98
@rizkysyahputra98 4 жыл бұрын
Clearest explanation about softmax.. thank you
@ritvikmath
@ritvikmath 4 жыл бұрын
Glad it was helpful!
@grzegorzchodak
@grzegorzchodak Жыл бұрын
Great explanation! Easy and helpful!
@YAlsadah
@YAlsadah 2 жыл бұрын
What an amazing, simple explanation. thank you!
@cobertizo
@cobertizo 4 жыл бұрын
I came for the good-looking teacher but stayed for the really clear an good explanation.
@salmans1224
@salmans1224 4 жыл бұрын
awesome man..your videos make me less anxious about math..
@ritvikmath
@ritvikmath 4 жыл бұрын
You can do it!
@MLDawn
@MLDawn 4 жыл бұрын
please note that the outputs of Softmax are NOT probabilities but are interpreted as probabilities. This is an important distinction! The same goes for the Sigmoid function. Thanks
@shiyuyuan7958
@shiyuyuan7958 3 жыл бұрын
Very clear explained , thank you, subscribed
@michael88704
@michael88704 2 жыл бұрын
I like the hierarchy implied by the indices on the S vector ;)
@serdarufukkara7109
@serdarufukkara7109 4 жыл бұрын
thank you very much, you are very good at teaching, very well prepared!
@AIPeerAcademy
@AIPeerAcademy Жыл бұрын
Thank you so much. I now understand why exp is used instead of simple calc.😊
@ritvikmath
@ritvikmath Жыл бұрын
Of course!
@bryany7344
@bryany7344 3 жыл бұрын
1:14, how is it a single dimensional for sigmoid? Shouldn't it be two dimensions?
@vahegizhlaryan5052
@vahegizhlaryan5052 3 жыл бұрын
well after applying sigmoid you get only one probability p (the other one you can calculate as 1-p) so actually you only need one number in case of sigmoid
@maralazizi
@maralazizi 4 ай бұрын
Another great content, thank you so much!
@ritvikmath
@ritvikmath 4 ай бұрын
My pleasure!
@masster_yoda
@masster_yoda 10 ай бұрын
Great explanation, thank you!
@fatemehsefishahpar3626
@fatemehsefishahpar3626 3 жыл бұрын
How great was this video! thank you
@debaratiray2482
@debaratiray2482 3 жыл бұрын
Awesome explanation.... thanks !!!
@nehathakur8221
@nehathakur8221 4 жыл бұрын
Thanks for such intuitive explanation Sir :)
@karimomrane7556
@karimomrane7556 2 жыл бұрын
I wish you were my teacher haha great explanation :D Thank you so much ♥
@Nova-Rift
@Nova-Rift 4 жыл бұрын
You're amazing. great teacher
@aFancyFatFish
@aFancyFatFish 3 жыл бұрын
Thank you very much, clear and helpful to me as a beginer😗
@eliaslara6964
@eliaslara6964 4 жыл бұрын
Dude! I really love you.
@anandiyer5361
@anandiyer5361 2 жыл бұрын
@ritwikmath want to understand why you chose the subscript N to describe the features; they should be S_1..S_M isn't it?
@bittukumar-rv6rx
@bittukumar-rv6rx 2 ай бұрын
Thank s for uploading ❤❤❤
@jackshaak
@jackshaak 3 жыл бұрын
Just great! Thanks, man.
@ritvikmath
@ritvikmath 3 жыл бұрын
You're welcome!
@seojun2599
@seojun2599 Жыл бұрын
How to dealing with high Xi values? I got 788, 732 for Xi value, and if I exp(788) it gives error bcs it exp results near to infinity
@bhp72
@bhp72 3 ай бұрын
really enjoyed that. thanks!
@ManishaGupta-rj3bq
@ManishaGupta-rj3bq 3 ай бұрын
Great tutorials!
@ritvikmath
@ritvikmath 3 ай бұрын
Glad you like them!
@burger_kinghorn
@burger_kinghorn Ай бұрын
How does this relate to multinomial logistic regression?
@igoroliveira5463
@igoroliveira5463 3 жыл бұрын
Could you do a video about the maxout unit? I read it on Goodfellow's Deep Learning book, but I did not grasp the intuition behind it clearly.
@yingchen8028
@yingchen8028 4 жыл бұрын
more people should watch this
@d_b_
@d_b_ Жыл бұрын
Maybe this was explained in a past video, but why is "e" chosen over any other base (like 2 or 3 or pi)...
@zahra_az
@zahra_az 3 жыл бұрын
that was so much sweet and inspiring
@zacharydan7236
@zacharydan7236 4 жыл бұрын
Solid video, subscribed!
@kausshikmanojkumar2855
@kausshikmanojkumar2855 Жыл бұрын
Absolutely beautiful.
@jeeezsh4704
@jeeezsh4704 3 жыл бұрын
You teach better than my grad school professor 😂
@diegosantosuosso806
@diegosantosuosso806 Жыл бұрын
Thanks Professor!
@ヨママ
@ヨママ 4 жыл бұрын
Thank you so much! You made it very clear :)
@dikshanegi1028
@dikshanegi1028 Жыл бұрын
Keep going buddy
@tm0209
@tm0209 Жыл бұрын
What does dP_i/dS_j = -P_i * P_j mean and how did you get it? I understand dP_i/dS_i because S_i is a single variable. But dP_i/DS_j is a whole set of variables (Sum(S_j) = S_1 + S_2 ... S_n) rather than a single one. How are you taking a derivative of that?
@mrahsanahmad
@mrahsanahmad 3 жыл бұрын
I am new to Data Sceince. However, why would a model output 100, 101 and 102 as three outputs unless the input had similarity to all three classes. Even in our daily lives, we would ignore 2 dollar variance on $100 think but complain if something which was originally free but now costs 2 dollars. Question is, why would we give up the usual practice and use some fancy transformation function here ?
@oligneflix6798
@oligneflix6798 3 жыл бұрын
bro you're a legend
@dragolov
@dragolov 2 жыл бұрын
Bravo! + Thank you very much!
@ridhampatoliya4680
@ridhampatoliya4680 4 жыл бұрын
Very clearly explained!
@sukursukur3617
@sukursukur3617 4 жыл бұрын
3:18 very good teacher
@jasonokoro8400
@jasonokoro8400 2 жыл бұрын
I don't understand *why* it's weird that 0 maps to 0 or why we need the probability to be the same for a constant shift...
@kavitmehta9143
@kavitmehta9143 4 жыл бұрын
Awesome Brother!
@ZimoNitrome
@ZimoNitrome 3 жыл бұрын
good video
@shreyasshetty6850
@shreyasshetty6850 3 жыл бұрын
Holy shit! That makes so much sense
@kausshikmanojkumar2855
@kausshikmanojkumar2855 Жыл бұрын
Beautiful!
@evagao9701
@evagao9701 4 жыл бұрын
hi there, what is the meaning of the square summation?
@suyashdixit682
@suyashdixit682 2 жыл бұрын
Yet again an Indian dude is saving me!
@ritvikmath
@ritvikmath 2 жыл бұрын
Lol 😂
@johnginos6520
@johnginos6520 4 жыл бұрын
Do you do one on one tutoring?
@mmm777ization
@mmm777ization 3 жыл бұрын
4:00 I thank you have express it in a wrong way you wanted to say that we need to go into depth and not just focus on the application that is the façade which here's deriving formula
@azinkatiraee6684
@azinkatiraee6684 Жыл бұрын
a clear explanation!
@ritvikmath
@ritvikmath Жыл бұрын
Glad you think so!
@markomarkus8560
@markomarkus8560 4 жыл бұрын
Nice video
@peterniederl3662
@peterniederl3662 4 жыл бұрын
Very helpful!!! Thx!
@yuchenzhao6411
@yuchenzhao6411 4 жыл бұрын
Very good video
@ritvikmath
@ritvikmath 4 жыл бұрын
Thanks!
@Fat_Cat_Fly
@Fat_Cat_Fly 4 жыл бұрын
👍🏻👍🏻👍🏻👍🏻👍🏻👍🏻
@evgenyv5687
@evgenyv5687 3 жыл бұрын
Hey, thank you for a great video! I have a question: in your example, you said that probabilities between 0,1 and 2 should not be different from 100, 101, and 102. But in the real world, the scale which is used to assess students makes difference and affects probabilities. The difference between 101 and 102 is actually smaller than between 1 and 2, because in the first case the scale is probably much smaller, so the difference between scores is more significant. So wouldn't a model need to predict different probabilities depending on the assessment scale?
@EW-mb1ih
@EW-mb1ih 3 жыл бұрын
same question!
@imingtso6598
@imingtso6598 3 жыл бұрын
My point of view is that the softmax scenario is different from sigmoid scenario. In the sigmoid case, we need to capture the changes in relative scale because subtle changes around the 1/2 prob. point result in significant prob. changes(turns the whole thing around, drop out or not); whereas in the softmax case, there are more outputs and our goal is to select the very case which is most likely to happen, so we are talking about an absolute amount rather than a relative amount(final judge). I guess that's why ritvik said" change in constant shouldn't change our model'.
@ltang
@ltang 3 жыл бұрын
Oh.. softmax is for multiple classes and sigmoid is for two classes. I get that your i here is the class. In the post below though, is their i observations and k the classes? stats.stackexchange.com/questions/233658/softmax-vs-sigmoid-function-in-logistic-classifier
@wduandy
@wduandy 4 жыл бұрын
Amazing!
@brendanamuh5683
@brendanamuh5683 2 жыл бұрын
thank you so much !!
@ayeddie6788
@ayeddie6788 2 жыл бұрын
PRETTY GOOD
@tsibulsky4900
@tsibulsky4900 Жыл бұрын
Thanks 👍
@ritvikmath
@ritvikmath Жыл бұрын
No problem 👍
@matgg8207
@matgg8207 2 жыл бұрын
what a shame that this dude is not a professor!!!!!!!!
@joelpaddock5199
@joelpaddock5199 Жыл бұрын
Hello Boltzmann distribution we meet again, cool nickname
@hezhu482
@hezhu482 4 жыл бұрын
thank you!
@anishbabus576
@anishbabus576 4 жыл бұрын
Thank you
@y0n1n1x
@y0n1n1x 2 жыл бұрын
thanks
@jkhhahahhdkakkdh
@jkhhahahhdkakkdh 4 жыл бұрын
Very different from how *cough* Siraj *cough* explained this lol
@QiyuanSong
@QiyuanSong Жыл бұрын
Why do I need to go to school?
@gestucvolonor5069
@gestucvolonor5069 4 жыл бұрын
I knew things were about to go down when he flipped the pen.
@mrahsanahmad
@mrahsanahmad 3 жыл бұрын
are you crazy. the moment he did that, I knew it would be fun listening to him. He was focused. Like he said, theory is relevant only in context of practicality.
@srl2017
@srl2017 2 жыл бұрын
god
@suryatejakothakota7742
@suryatejakothakota7742 4 жыл бұрын
Binod stop ads
@fintech1378
@fintech1378 Жыл бұрын
minute 11-12.30 you are not very clear and going too fast
@ritvikmath
@ritvikmath Жыл бұрын
hey thanks for the feedback, will work on it
Matrix Norms : Data Science Basics
9:57
ritvikmath
Рет қаралды 58 М.
The Sigmoid : Data Science Basics
11:34
ritvikmath
Рет қаралды 41 М.
Война Семей - ВСЕ СЕРИИ, 1 сезон (серии 1-20)
7:40:31
Семейные Сериалы
Рет қаралды 1,6 МЛН
요즘유행 찍는법
0:34
오마이비키 OMV
Рет қаралды 12 МЛН
Why Do Neural Networks Love the Softmax?
10:47
Mutual Information
Рет қаралды 67 М.
Entropy (for data science) Clearly Explained!!!
16:35
StatQuest with Josh Starmer
Рет қаралды 649 М.
Backpropagation : Data Science Concepts
19:29
ritvikmath
Рет қаралды 41 М.
Softmax Function Explained In Depth with 3D Visuals
17:39
Elliot Waite
Рет қаралды 40 М.
Softmax Regression (C2W3L08)
11:48
DeepLearningAI
Рет қаралды 166 М.
But what is a neural network? | Deep learning chapter 1
18:40
3Blue1Brown
Рет қаралды 18 МЛН
Terence Tao at IMO 2024: AI and Mathematics
57:24
AIMO Prize
Рет қаралды 701 М.
Loss Functions : Data Science Basics
16:40
ritvikmath
Рет қаралды 35 М.
Softmax - What is the Temperature of an AI??
8:34
MarbleScience
Рет қаралды 15 М.
Google's 9 Hour AI Prompt Engineering Course In 20 Minutes
20:17
Tina Huang
Рет қаралды 115 М.