This was exactly the baby step I needed to get me on my way with entropy. Far too many people try to explain it by going straight to the equation. There's no intuition in that. Brilliant explanation. I finally understand it.
@jankinsics5 жыл бұрын
Sean Walsh feel the same way.
@user-or7ji5hv8y4 жыл бұрын
how does one make something so complicated into something so intuitive that others can finally see the picture. your explanation itself is an amazing feat.
@josephbolton80924 ай бұрын
an amazing teacher is an invaluable thing
@AlexMcClung977 жыл бұрын
Excellent explanation, very clear and concise! I have always pondered the significance of the log in cross-entropy loss function. The explanation (particularly: "products are small and volatile, sums are good") completely clears this up.
@effemmkay4 жыл бұрын
I have been scared of delving into entropy in detail for so long because the first time I studied it, it wasn’t a good experience. All I want to say is THANK YOU!!!!!! I should have been supplementing the udacity ND lesson videos with these since the beginning.
@freemanguess86346 жыл бұрын
With great knowledge comes low entropy
@SerranoAcademy6 жыл бұрын
Hahaaa, love it!!!
@fantomraja91375 жыл бұрын
lol
@hyperduality28384 жыл бұрын
@@SerranoAcademy Repetition (redundancy) is dual to variation -- music. Certainty is dual to uncertainty -- the Heisenberg certainty/uncertainty principle. Syntropy (prediction) is dual to increasing entropy -- the 4th law of thermodynamics. Randomness (entropy) is dual to order (predictability) -- "Always two there are" -- Yoda.
@B2T7RID2QGLEHH5UZFB0T3 жыл бұрын
And low entropy is easier to rig
@lani03 жыл бұрын
You win
@carnivalwrestler6 жыл бұрын
Luis, you are such an incredibly gifted teacher and so meticulous in your explanations. Thank you for your hard work.
@RyanJensenEE2 жыл бұрын
Good video! Minor correction of calculations: at 5:50, the probability of getting the same configuration is 0.25. This is because there are only 4 possible configurations of the balls (there is only one blue ball, and only four slots, so only 4 places the blue ball can be). This can also be calculated by selecting red balls first multiplying 0.75 * 0.66667 * 0.5 = 0.25. Similarly, at 6:58, the probability is 1/6 because there are 6 possible configurations. We can calculate the probability by multiplying (2/4) * (1/3) = (2/12) = (1/6) ~= 0.166667.
@elmoreglidingclub30304 жыл бұрын
Excellent! Great explanation. Enjoyable video (except YT’s endless, annoying ads). Thank you for composing and posting.
@TheGenerationGapPodcast3 жыл бұрын
Confession: I was a math kiddy; I know to use it but I often missed the deeper meaning and intuition. Your videos are turning me into a math hacker.
@Asli_Dexter7 жыл бұрын
i wish i had this lecture during college examination.....still it's nice to finally understand the intuition behind the formulas i already knew.
@pixboi6 жыл бұрын
Teaching should be like this, from practice to theory - no the other way around!
@dyutinrobin10 ай бұрын
Thank you so much. This was the only video in youtube that clarified all my doubts regarding the topic of entropy.
@NoOne-uz4vs5 жыл бұрын
I'm studying Decision Tree (Machine Learning Algorithm) and it uses Entropy to efficiently build the tree. I finally understand the details. Thank you!!
@drakZes5 жыл бұрын
Great work. Compared to my textbook you explained it 100 times better, Thank you.
@jackallread Жыл бұрын
Thanks for the relationship between knowledge and entropy, that was very helpful. Your explanation of statistics is also good! Though, I am only half way through the video at this point, I will finish it! Thanks
@123liveo6 жыл бұрын
2nd time I found this video and loved it both times. Much better description than the prof at the uni I am at!!!
@eprabhat7 жыл бұрын
Luis, You have a great way of explaining. At times , I like your videos more than even some highly rated professors
@sdsa007 Жыл бұрын
Wow! Awesome, so books and encyclopedias and biographies of Shannon to understand what you just clearly explained! Thank You!
@msctube454 жыл бұрын
I needed this video to get me up to speed on entropy. Great job Luis!
@Skandar00075 жыл бұрын
That moment when you realize you don't need to search for another video because you got it from the first time. What I'm trying to say is Thank You!
@ketlebelninja5 жыл бұрын
This was one of the best explanations on entropy. Thanks
@sasthra31592 жыл бұрын
Great clarity. Have never got this idea about the Shannon Entropy. Thank you. Great work!
@Bvic36 жыл бұрын
At 13:44 it's not 0.000488 but 0.00006103515 ! There is a computation error. The entropy is correct, 1.75.
@SerranoAcademy5 жыл бұрын
Thank you for the correction! Yes, you're right.
@SenhorMsandiFelipe3 жыл бұрын
Gracias. Muito claro Senhor. I have been struggling to wrap my head around this and you just made it easy. Thank you.
@Johncowk5 жыл бұрын
You made a mistake/approximation by saying the entropy is equal to the number of question needed to be asked in order to find out which letter it is. If I do a scenario with only three letters, all equiprobable, the entropy is about 1.59 but the average number of question needed to find out the correct letter is about 1.66. Your presentation gives a great way to gain an intuitive feeling about the entropy, but maybe you should include a small disclaimer on this point.
@victorialeigh27263 жыл бұрын
Hola Luis, estupendo, espectacular, excelente!
@mau_lopez6 жыл бұрын
What a great explanation ! I wish I had a teacher like you Luis, everything wold be way easier ! Thanks a lot
@patricklemaire2256 жыл бұрын
Great video! Now I understand what Claude Shannon discovered and how useful and essential maths are in Computer Science.
@AJK5444 жыл бұрын
your explain is perfect. Even though I am not good at listening english. I can understand everything :)
@dianafarhat947910 ай бұрын
Can you make a part 2 with the full proof, not just the intuition behind the formula? Your explanation's amazing & would love to see a part 2.
@jordyb4862 Жыл бұрын
I find sum(p*log(p^-1)) more intuitive. Inverse p (i.e. 1/P) is the ratio of total samples to this sample. If you ask perfect questions you'll ask log(1/p) questions. Entropy is then the sum of these values, each multiplied by the probability of each, which is how much it contributes to the total entropy.
@hanaelkhalifa26304 жыл бұрын
Thank you for excellent explanation of entropy concept first... Then reach to final equation step-by-step it is really good and simple way
@shekelboi6 жыл бұрын
Thanks a lot Luis, just had an exam about this Wednesday and your video helped me a lot to understand the whole concept.
@haimmadmon35314 жыл бұрын
Very good explanation - hope to hear more of your videos
@Vuvuzella164 жыл бұрын
This video is helping to keep me floating in my Data Science course; thank you so much for your time!
@therealsachin7 жыл бұрын
The best explanation about Shannon entropy that I have ever heard. Thanks!
@kleberloayza78395 жыл бұрын
hi Luis, nice to meet you, I am reading the book of Deep learning of Ian Godfellow, and I needed to view your video for understand the chapter, 3.13 information theory. thanks very much.
@mehmetzekeriyayangn37825 жыл бұрын
You are the best.Such a great explanation.Better than lots of text books.
@poxyu_was_here7 жыл бұрын
Easy and Great explanation! Thank you very much, Luis
@mulangonando2942 Жыл бұрын
I love the explanation of the negative sign in the Entropy Equation many people wonder
@JohnsonChen-t9r5 жыл бұрын
It's very helpful for me to introduce the concept of entropy to students. Thank you for your clear presentation of entropy.
@TheZilizopendwa3 жыл бұрын
Excellent presentation for an otherwise complex concept.
@SixStringTheory67 жыл бұрын
Wow ..... I wish more people could teach like you this is so insightful
@eka22135 жыл бұрын
So, after watching the video, the entropy for giving you thumbs up and subcribing to your channel was 0 - i.e. great explanation!
@amperro3 жыл бұрын
I watched it straight through. Very good.
@christinebraun96105 жыл бұрын
Great explanation. But I think what’s still missing is an explanation of why we use log base 2....didn’t quite get that
@olivercopleston5 жыл бұрын
In the last minute of the video, he explains that using Log base 2 corresponds to the level of a decision tree, which is the number of questions you'd have to ask to determine a value.
@MatheusSilva-dragon6 жыл бұрын
Wow, thank you, man. I needed that information! There are many ways to teach the same stuff! That number of question stuff is great! It's good to have more than one way to measure something!
@clarakorfmacher73944 жыл бұрын
Great Video! I really liked the intuitive approach. My professors was waaaay messier.
@rajudey16734 жыл бұрын
Really, you have given us outstanding information.
@RenanCostaYT4 жыл бұрын
Great explanation, greetings from Brazil!
@hanaizdihar43684 жыл бұрын
What a great explanation! And so i subscribed😊
@subhashkonda50007 жыл бұрын
Its always hard to understand the equations but u made it so simple :-)
@patriciof.calatayud98613 жыл бұрын
I think that the Huffman compression that you use and the end of the video is near the entropy value but not exactly the same
@aryamahima34 жыл бұрын
Thank you so much for a such a easy explanation...respect from india...
@karinasakurai98675 жыл бұрын
Brilliant lecture! I learn so much with this explanation. Thanks from Brazil :)
@Dennis128695 жыл бұрын
Best explanation I found so far
@Darnoc-sudo3 жыл бұрын
Very nice video. Insightful, inutuitive and very well explained. Thank you!
@tilugulilwa4 жыл бұрын
Superb step by step explanation
@MH_HD6 жыл бұрын
This is the best explanation I have come across for a long time, Can you please answer how can we use entropy to find the uncertainty of a naive Bayesian classifier with let's say 4 feature variables and a binomial class variable?
@pkittali7 жыл бұрын
Lovely explanation...Superb
@YoussefAhmed-uv7ti5 жыл бұрын
Actually, there is something wrong here. the entropy and information in information theory are representing the same thing which is how much information we will get after decoding the random message, so in case of the balls in the box if all are the same color we have no information after decoding the message as its probability to be red =1 hence low entropy and low information.
@emrahyener4023 жыл бұрын
Thanks for this perfect explanation 👏👏👏👍
@nijunicholas6315 жыл бұрын
Thanks..Got the intuition behind Entropy
@KayYesYouTuber4 жыл бұрын
Superb explanation. I like your teaching style. Thank you very much :-)
@ravikumar376 Жыл бұрын
Sir good explanation thank you very much. But at sequence3 hot to get 8/8log2. 1/4 result is 2
@logosfabula7 жыл бұрын
Luis, you really are a great communicator. Looking forward to your other explanations.
@francismcguire68845 жыл бұрын
Best instructor there is! Thanks
@kingshukbanerjee7486 жыл бұрын
very lucid explanation - excellent, intuitive build-up to Shannon's theorem from scratch
@kasraamanat54533 жыл бұрын
best, as always ❤️ thank you Luis❤️
@nassimbahri5 жыл бұрын
For the first time in my life i understand the real meaning of the Entropy
@amitkumarmaiti53924 жыл бұрын
Great Intuition Luis
@amatya.rakshasa3 жыл бұрын
Is there a construction or characterization or description of how to ask the smartest questions every time ?
@user-or7ji5hv8y4 жыл бұрын
wow, another great and insightful presentation . really helps to build intuition
@VC-zo9mt3 жыл бұрын
I know this may be easier for others to understand, but could you show an explanation of the actual symbols of this formula and show an example of numbers plugged in to see which numbers go where. I am not familiar with Log other than it's related to exponents. The minus aspect of it is also unfamiliar.
@themightyquinn1002 жыл бұрын
At 13:34 the product does not equal 0.000488. It is approximately 0.000061035. You are missing the last 1/8 factor.
@cariboux23 жыл бұрын
Luis, Thank you so much for this brilliant elucidation of information theory & entropy. Merely as an avocation, I have been toying around with a pet evolutionary theory about belief systems and societies. In order to test it - if that is even possible - I felt I needed to develop some sort of computer program as a model. Since I have very little programming experience and only mediocre math skills, I have been teaching myself both (with a lot of help from the web). It was purely by accident that I stumbled upon Claude Shannon and information theory, and I immediately became fascinated with the topic, and have a hunch that it may somehow be relevant to my own research. Regardless, I am now interested in it for its own sake. I had a an ephemeral understanding of how all the facets (probability, logs, choices, etc.) were all related mathematically, but it wasn't until after watching your video that I believe I fully grok the concept. At one point early on, I found myself shouting, "if he brings up yes/no questions, I know I understand this!" And then you did. It was such a wonderful moment for someone who finds math so challenging, and it is greatly appreciated! I shall check out your other videos later. You're a very good teacher!
@Faustus_de_Reiz2 жыл бұрын
For your work, I would look into some of the work by Loet Leydesdorf.
@cariboux22 жыл бұрын
@@Faustus_de_Reiz Thank you! I shall.
@scherwinn6 жыл бұрын
Very clever explanation of mighty ENTROPY.
@justinphilpott2 ай бұрын
Great video, thanks!
@xThomas19954 жыл бұрын
Thank you for the very good video. Easiest to understand so far.
@yhat3146 жыл бұрын
Lovely job Luis! Very very good!
@namename64355 жыл бұрын
You explanation was crystal clear, if possible share some real time examples of data mining where entropy, gini index are used
@paulstevenconyngham78806 жыл бұрын
this is a really great explanation, thanks so much for sharing mate!
@carlitos53364 жыл бұрын
Excelente explicación! Gracias por compartirla.
@micahdelaurentis65513 жыл бұрын
you killed it. Great video
@sosoboy775 жыл бұрын
Best video this week
@jaeimp2 жыл бұрын
Excellent job, Luis! Plain and simple: the log base 2 gives the number of bifurcations to arrive at the answer, and the probability of the answer serves to temper down the chaos introduced into the system by very rare events. Genius!
@rolfbecker45124 жыл бұрын
Thank you very much for this beautiful and clear explanation!
@bismeetsingh3525 жыл бұрын
That was highly intuitive, thank you, sir, I appreciate the effort behind this.
@蔡小宣-l8e3 жыл бұрын
十分谢谢! Thank you very much, Luis.
@shakeelurrahman1846 Жыл бұрын
thanks a lot for such a beautiful explanation..!
@scottsara1235 жыл бұрын
Easy and excellent explain, Please do for loss and cost function as well (convex)
@hyperduality28384 жыл бұрын
Syntropy is dual to increasing entropy -- The 4th law of thermodynamics! Thesis is dual to anti-thesis -- The time independent Hegelian dialectic. Schrodinger's cat: Alive (thesis, being) is dual to not alive (anti-thesis, non being) -- Hegel's cat. Syntropy is the process of optimizing your predictions to track targets or teleological physics. Teleological physics (syntropy) is dual to non teleological physics (entropy, information).
@Omsip123 Жыл бұрын
Very well explained, thank you
@paulinagc69863 жыл бұрын
Thank you so much. You are such a good teacher, really :D :D :D
@jonathanfrancis4 жыл бұрын
Wow. Amazing video.
@bhupeshrao23594 жыл бұрын
As you said the game is to have red,red,red,blue. but for first case we have all reds. hence probability of winning in first case should be 1*1*1*0(prob of blue in first bucket). How did you calculate (1*1*1*1). please explain?
@miguelfernandosilvacastron32795 жыл бұрын
Thank you. Nice, concise explanation.
@meshackamimo19456 жыл бұрын
Hi. Thanks a million times for simplifying a very complicated topic. Kindly find time n post a simplified tutorial on mcmc.... I am overwhelmed by your unique communication skills. Markov chain Monte Carlo. God bless you.
@TheGenerationGapPodcast3 жыл бұрын
Help us smash Markov chain Monte Carlo
@YugoGautomo5 жыл бұрын
Hi Luis, Thanks for your explanation. I guess you're wrong in minute 6.29 and 7.41. I think P winning for P bucket 1 should be 0, since there were no Blue Balls in the bucket as expected outcome of the game. should be R, R, R, B. Am I right?