In times of Covid, I hope this makes young people realize why older people are so important. Long live Prof Strang.
@justpaulo4 жыл бұрын
In times of Covid I hear in the background of the class someone sneezing and nose-blowing and it gives me the chills ...
@somadityasantra55724 жыл бұрын
He is human equivalent of God
@godfreypigott3 жыл бұрын
@@somadityasantra5572 Are you saying he doesn't exist?
@somadityasantra55723 жыл бұрын
@@godfreypigott U are assuming that I mean God does not exist. But how can u prove or disprove that?
@godfreypigott3 жыл бұрын
@@somadityasantra5572 There is no "god" - that is a given. So by saying he is the "human equivalent of god" you are saying that he doesn't exist.
@EduardoGarcia-tv2fc4 жыл бұрын
I'd say without any doubt that Professor Strang is the best Algebra professor in the entire world. I'm sure he has helped tons of students all around the world to understand the beauty of algebra
@lizijian70905 жыл бұрын
Long live your kindly,mild professor
@Pablo-y2k6bАй бұрын
31:20 intuitive explanation of how the norm choice effects the minimization problem was eye opening to me
@andrewmeowmeow3 жыл бұрын
What a smart and humble person! Long live Prof. Strang!
@deeptendusantra6704 жыл бұрын
After reading so many texts finally some actual geometric interpretation of L1 and L2 ...he explains it so beautifully.Came here only to understand definition but his charsima made me watch whole 50 mins
@Forced24 жыл бұрын
Exactly the same for me
@denys222224 жыл бұрын
Ahah I am in the same situation.
@minoh15433 жыл бұрын
the same for me 2222222
@pondie53812 жыл бұрын
EXACTLY the same!!!
@JulieIsMe8244 жыл бұрын
Best linear algebra course ever! Best wishes for Prof. Strang's health during this horrible pandemic
@rogiervdw4 жыл бұрын
Teaching norms with their R2 pictures is just brilliant. So much insight, even emerging while teaching (sparsity of L1 optimum: it's on the axis!!). An absolute joy to watch & learn from
@abdulghanialmasri55502 жыл бұрын
This man does not stop giving, many thanks.
@bilyzhuang92425 жыл бұрын
LONG LIVE PROFESSOR STRANG!!!!!
@abdowaraiet21694 жыл бұрын
"You start from the origin and you blow up the norm until you get a point on the line that satisfies your constraint, and because you are blowing up the norm, when it hit first, that's the smallest blow up possible, that's min, that's the guy that minimize" (31:23-31:42) that's 2-D optimization in a nutshell...clear and simple, thanks very much Professor Strang..
@ShwetankT4 жыл бұрын
3D as well, no?
@KirtiDhruv4 жыл бұрын
This lecture needs to reach more people asap. Total respect for the Professor!
@atulsrmcem Жыл бұрын
I'm currently reading Calculus by Dr. Strang. One of the best books on the subject I have ever come across.
@asifahmed18013 жыл бұрын
After passing the linear algebra course, i was kind of disappointed no need to see your lecture again . but for data analysis u came again in a HD resolution. So glad to see you professor .
@arnaud50333 жыл бұрын
Probably, this has been said before, so forgive me if I repeat someone else's words. I acknowledge here, that professor Strang is a good pedagogue. I learnt some math over the years. I completely support the use of the geometrical visualization of some properties, as it is a learning need. I can say that for me it is easy to see how to derive properties like the one he gave for the assignment on the Frobenius norm. I say this, because I may not be the only one thinking it and I wanted to tell those people that there is more to math here. Only recently, I understood the huge degree of humility and teaching wit that it takes one to pass one's knowledge along. It requires to pretend or to honestly feel you are no better than any of your students. For instance, as I could witness here, Pr Strang shared with his students the latest cool research topics as if they were his colleagues, he thanked them for contributing to the course by giving out some answers. That's what allows him to successfully challenge them in solving some assignments, like the Frobenius norm - SVD problem. All of it is summarized by Gilbert himself at the very end in 48:12, when he explains his view of his relationship with the students (such as "We have worked to do!", an honest use of the pronoun "we" by the lecturer). This 48 min long lecture, honestly impressed me in this regard. Today, I had the privilege of a double lecture: one in math (that could have been compressed to 15 min, since most proofs were skipped) and one in being a better passer of knowledge (that could be extended to 10+ years). Hat off!
@georgesadler78303 жыл бұрын
DR. Strang, thank you explaining and analyzing Norms. I understand this lecture from start to finish.
@diysumit3 жыл бұрын
Love this man, thanks MIT for looking out for us!
@supersnowva6717 Жыл бұрын
This lecture just brought my understanding of norms to a whole new level! Thank you so much Professor Strang!
@karthikeyakethamakka2 жыл бұрын
27:07 minimizing something with a constraint Lagrangian Formulation.
@wangxiang20442 жыл бұрын
Frobenius norm squared = trace of (A transpose times A) = sum of eigenvalues of (A transpose times A) = sum of squares of singular values
@ashutoshpatidar3288 Жыл бұрын
Feeling so emotional watching him teaching at the age of 84😢
@sebah19912 жыл бұрын
The reason I went from hating math to loving math (especially linear algebra) is Gilbert Strang. What an incredible teacher.
@hieuphamngoc62583 жыл бұрын
He is such a sweet man and a genius teacher at the same time
@naterojas92724 жыл бұрын
I highly recommend doing the Frobenius norm proof he mentions. It is elegant and uses some nice properties of linear algebra. If you took 18.06 (or watched the lectures) using the column & row picture of matrix multiplication really helps. I'll finalize my proof and post a link - hopefully I didn't make a mistake ;)
@naterojas92724 жыл бұрын
Maybe I shouldn't post a link... I wouldn't want anyone enrolled in 18.065 to copy it... Hmm......
@xingjieli30693 жыл бұрын
Great point on comparing matrix Nuclear norm with vector L1 norm, which tends to find the most sparse winning vector. I guess the matrix Nuclear norm may tend to find 'least' weights during the optimization.
@xc2530 Жыл бұрын
35:00 matrix norm
@jonahansen5 жыл бұрын
Dang - He's good!
@mgh2565 жыл бұрын
come on.... He is Gilbert Strang......
@lololamize5 жыл бұрын
Does anyone know something more concrete about the Srebro results? Have they been verified already? How general are they? 44:54
@karthikeyakethamakka2 жыл бұрын
The Largest Singular Value is the same as largest Eigen Value for a fully connected layer which is also called as spectral Normalization.
@shvprkatta3 жыл бұрын
Prof Strang...my respects sir...
@HoangLe-rk2ke3 жыл бұрын
Protect him with all cost MIT
@fredsmith8944 жыл бұрын
I love Linear Algebra!
@Leeidealist4 жыл бұрын
I love him so much I don't believe in God but I prayed for his health
@RohithBhattaram4 жыл бұрын
Good Video about Norms. Thank you Prof.
@shaurovdas58424 жыл бұрын
At 32:41 when professor Strang says L2 norm of a matrix is 'sigma1', what does he mean by sigma1?
@oscarys4 жыл бұрын
Hi Shaurov. He is referring to the largest singular value in the SVD of A
@sergiohuaman60844 жыл бұрын
@44:00 by now Prof. Strang should know that nothing is ever taken out of the tape haha.
@jxw71963 жыл бұрын
This man is brilliant!
@oscarlu99193 жыл бұрын
12:12 prof: it's just exploded in importance. me: I just burst in laugh :)
@nuclearrambo31673 ай бұрын
I would use laplace's succession rule in coin flipping problem
@prajwalchoudhary48243 жыл бұрын
Is L0 norm is not convex ????
@ZeroManifold Жыл бұрын
yes, cause the origin point is excluded
@filialernotina10604 жыл бұрын
Can someone explain to me when we should use Frobenius norm and when we should use the nuclear norm ?
@HardLessonsOfLife3 жыл бұрын
Why is L half not a good norm? Why the P is restricted to be >= 1 instead of just p >0?
@namimmsadeghi84753 жыл бұрын
perfect. God bless you...
@wernerhartl20693 жыл бұрын
Max ||Ax||/||x||=Max ||Akx||/||kx||=Max ||Ax||/||x||, ||x||=1. So you can think of a unit circle ||x||=1 with ||Ax|| plotted on the radius which might look like a circle and ellipse with a point on the Ellipse being at Max ||Ax||/||x||.
@TeejK4 жыл бұрын
Holy crap this is a good lecture
@eyobgizaw83623 жыл бұрын
how would the shape look like for p, 1
@godfreypigott3 жыл бұрын
Between the diamond and the circle.
@raphaelambrosiuscosteau8294 жыл бұрын
How do we actually see what sigma1 is the maximum blow-up factor and what v1 is the vector what gets blown up the most? Because i initially thought it would be the first eigenvector, and then it would make sense, but then i realised what sigma is not an eigenvalue after professor said it and i'm struggling a bit with imagening what's happening here
@justpaulo4 жыл бұрын
Recall the picture Prof. Strang draw when explaining SVD. Here's a refresher (in slide #25): ocw.mit.edu/resources/res-18-010-a-2020-vision-of-linear-algebra-spring-2020/videos/MITRES_18_010S20_LA_Slides.pdf As Prof. Strang mentioned, U and V only perform rotation or possible reflection of x, which does not changes the norm of x. It is Sigma that is responsible for stretching and among those sigma1 is the biggest and it is therefore the "maximum blow-up factor". I hope this helps.
@zkhandwala5 жыл бұрын
Compelling lecture (as always), but I'm unsettled about one thing: much of it is based on the fact that the first singular vector of A is the maximizing x in the definition of ||A||2. However, this fact just seems to be mentioned without proof or argument, and accordingly it doesn't feel as though the proof that ||A||2 = sigma1 is complete. Thoughts?
@gordongustafson27995 жыл бұрын
I agree. I can give a proof sketch: 1. A = U Σ V^t by the SVD. 2. To maximize ||Σy|| for a unit vector y, we would choose y to have all 0's except for a 1 in the position multiplying the largest value in the diagonal matrix Σ, which is sigma1. This effectively scales every component of y by sigma1 (all the other components are 0). Any other choice of y results in some component of y being scaled by a value less than sigma1, and no component scaled by more than sigma1. 3. U is orthonormal, so ||Uz|| = ||z|| 4. 1 and 3 give us ||Ax|| = ||U Σ V^t x|| = ||Σ V^t x|| 5. Assume ||x|| = 1. V^t is orthonormal, so ||V^t x|| = 1. 6. Thus, the maximizing value of x satisfies V^t x = y for the y we found in step 2. 7. This gives x = v1, and ||Ax|| = sigma1. 8. Since the L2 norm of A is the maximum value of ||Ax||/||x|| over all x's, the L2 norm of A is sigma1 (small leap here, but straightforward)
@yupm15 жыл бұрын
Whenever I make money I will donate!
@jaimelima24205 жыл бұрын
You will make a lot of money man. In Wall Street, perhaps!
@freeeagle60742 жыл бұрын
When you earn 20 dollars, you can donate 1/2 dollars. When you earn 20,000 dollars, you can donate 100 dollars. When you earn 2 billion, you'll leave here and forget about donation for ever.
@SimmySimmy5 жыл бұрын
I've watched the first 6 videos without difficulty, but I'm confused by the definition and geometric meaning of different norms. Could anyone please tell me which textbook I should read to help me understand? Thanks for your helping!
@turdferguson34005 жыл бұрын
Rewatch the videos, and maybe you'll get it! It has worked for me!
@darkwingduck425 жыл бұрын
Linear Algebra and Learning from Data by Gilbert Strang!
@karthikeyakethamakka2 жыл бұрын
40:10 F norm
@phuongnamphan3355 жыл бұрын
why sigma 1 is the largest singular value ? Why it's position relate to largest or not ? I dont understand
@BorrWick5 жыл бұрын
Yes, singular values are ordered based on size
@csl13845 жыл бұрын
Is there a link to the notes Prof. Strang keeps alluding to?
@NolanZewariligon5 жыл бұрын
@Bob Mama There aren't any lecture notes on that link.
@matthewpublikum31142 жыл бұрын
Are the notes available somewhere?
@mitocw2 жыл бұрын
There are no lectures notes available for this course; that is because the book (Strang, Gilbert. Linear Algebra and Learning from Data. Wellesley-Cambridge Press, 2018. ISBN: 9780692196380) is basically the lecture notes for the course. See the course on MIT OpenCourseWare for more info and materials at: ocw.mit.edu/18-065S18. Best wishes on your studies!
@nazarm6215 Жыл бұрын
so is a sigmoid a norm or a norm is a sigmoid?
@zeke-hc3rc3 жыл бұрын
thank you
@vyvo14735 жыл бұрын
Why ||A||2 = max ||Ax||2/||x||2? Can someone help me explain? :(
@sricharanbattu45025 жыл бұрын
That is actually the definition of matrix norm, induced by a vector
@tanweermahdihasan41195 жыл бұрын
@Rich Caputo shouldn't there be a l2 norm constraint for x? say, ||x|| = 1.
@myoung14454 жыл бұрын
It's a definition rather than a result
@obsiyoutube48284 жыл бұрын
sure big professor
@MrFurano2 жыл бұрын
43:49 The "actual humans" statement is still on the tape 🤣
@sanjaykrkk4 жыл бұрын
Awesome!
@sumanchaudhary87575 жыл бұрын
can somebody provide lecture notes of this course?
@mitocw5 жыл бұрын
Course materials are available on MIT OpenCourseWare at: ocw.mit.edu/18-065S18. Best wishes on your studies!
@NolanZewariligon5 жыл бұрын
@@mitocw There aren't any lecture notes on that link.
@sukuya4 жыл бұрын
github.com/ws13685555932/18.065_lecture_notes are some summary notes till lecture 14.
@NolanZewariligon5 жыл бұрын
He forgot to finish PCA.
@СергейКумейко-й8г4 жыл бұрын
I can give available information . The lecture here connects the optimization problem with eigenvectors. But sorry , the lecture in Russian))) kzbin.info/www/bejne/jWatfYaBmNqUh9E
@bpc15704 жыл бұрын
How about referring to Andrew Ng lecture at cs229, which is not in Russian for English speakers
@justpaulo4 жыл бұрын
kzbin.info/www/bejne/m6qVgXhrrc5sY6M
@NolanZewariligon4 жыл бұрын
@@justpaulo @BP C MVPs.
@charliehou95533 жыл бұрын
Long live!
@quanyingliu71685 жыл бұрын
The phenomenon he mentioned in the first 5 minutes is a very interesting psychological question. Is it about the sequential effects of decision making? Anyone knows the field? Please feel free to share some papers. Thank you.
@mdrasel-gh5yf4 жыл бұрын
Multi Arm Bandits problem?
@intoeleven5 жыл бұрын
may I ask what the p mean?
@ethereal1m5 жыл бұрын
it's the mode of norm
@avapwhl4 жыл бұрын
This is some nice chalk
@bird92 жыл бұрын
well... How old are these students ?
@YNRUIZ693 жыл бұрын
Cool vid
@thetedmang5 жыл бұрын
I didn't get the part where he minimized the L2 norm geometrically, why was it that particular point?
@yuchenzhao64114 жыл бұрын
L2 norm for a vector is the distance from origin. Since the candidate vectors have to be on the constraint line, the problem (find a vector that subject to the constraint minimize L2 norm) became "which point on that line has smallest distance to the origin".