Principal Component Analysis (PCA) 1 [Python]

  Рет қаралды 44,904

Steve Brunton

Steve Brunton

4 жыл бұрын

This video describes how the singular value decomposition (SVD) can be used for principal component analysis (PCA) in Python (part 1).
Book Website: databookuw.com
Book PDF: databookuw.com/databook.pdf
These lectures follow Chapter 1 from: "Data-Driven Science and Engineering: Machine Learning, Dynamical Systems, and Control" by Brunton and Kutz
Amazon: www.amazon.com/Data-Driven-Sc...
Brunton Website: eigensteve.com
This video was produced at the University of Washington

Пікірлер: 27
@NiKogane
@NiKogane 2 жыл бұрын
Thank you so much for providing all of this knowledge online for free !
@tomlane6590
@tomlane6590 3 жыл бұрын
A brilliant set of videos. Thank you so much.
@nguyenvan-hau9577
@nguyenvan-hau9577 4 жыл бұрын
Beautiful code!
@yenunadeesaselviento
@yenunadeesaselviento 3 жыл бұрын
The code cuts off at the edge of the video. Where can we download it. Thanks for sharing this!
@subramaniannk3364
@subramaniannk3364 4 жыл бұрын
Great lecture Steve! You explained that "u" in svd represents principal direction, "sigma" represents loading. What does "v" represents ?
@sambroderick5156
@sambroderick5156 3 жыл бұрын
There’s a whole series a lectures explaining this (and a book.
@kanacaredes
@kanacaredes 3 жыл бұрын
excellent video!!! Thks
@Eigensteve
@Eigensteve 3 жыл бұрын
You are welcome!
@muhammadmuneeburrahman1262
@muhammadmuneeburrahman1262 2 жыл бұрын
You said in the video that each row of X will represent an example/or record, and column will represent the feature. In your code, X.shape = (2, 1000) where each column represent one data point and. B is passed to the SVD with same shape. Hence the VT matrix size is (2,1000) which means that there are 1000 principle Componenets, that is not possible of 2D data??? Am I right or wrong?? Please explain?
@jbhsmeta
@jbhsmeta 4 жыл бұрын
Hi Mr. Steve, I have one question, why are you dividing the "B by np.sqrt(nPoints)" U, S, VT = np.linalg.svd(B/np.sqrt(nPoints),full_matrices=0) dividing mean centered data by sqrt of no.of data -?? Could not understand.
@melvinlara6151
@melvinlara6151 3 жыл бұрын
Actually i have the exact same question. Could you figure it out?
@JoaoVitorBRgomes
@JoaoVitorBRgomes 3 жыл бұрын
@@melvinlara6151 I didn't see the whole lecture yet, but I guess B is data with mean =0 and np.sqrt(nPoints) probably is the standard deviation (variance squared). So he first standardize the data then he applies SVD ...
@melvinlara6151
@melvinlara6151 3 жыл бұрын
@@JoaoVitorBRgomes hey! actually i figured the same thing out. But, thank you;
@JoaoVitorBRgomes
@JoaoVitorBRgomes 3 жыл бұрын
@@melvinlara6151 no problem Melvin Lara, I am a student of Data Science too. If you have a kaggle profile and want to exchange knowledge my alias is " topapa .
@anirbanbhattacharjee8093
@anirbanbhattacharjee8093 6 ай бұрын
In PCA literature, the covarience matrix (B*)B is normalized by nPoints (or the bessel correction (nPoints -1), but doesn't matter here because nPoints is large). So if you normalize B by np.sqrt(nPoints) instead, B* also gets normalized by np.sqrt(nPoints) and you end up getting the (B*)B normalized by nPoints
@EladM8a
@EladM8a 3 жыл бұрын
Why the division in B/np.sqrt(nPoints)?
@anirbanbhattacharjee8093
@anirbanbhattacharjee8093 6 ай бұрын
In PCA literature, the covarience matrix B*B is normalized by nPoints (or the bessel correction (nPoints -1), but doesn't matter here because nPoints is large). So if you normalize B by np.sqrt(nPoints) instead, B* also gets normalized by np.sqrt(nPoints) and you end up getting the C normalized by nPoints
@anirbanbhattacharjee8093
@anirbanbhattacharjee8093 6 ай бұрын
where C = (B*)B, & B* is the transpose of B
@sheiladespard8861
@sheiladespard8861 3 жыл бұрын
I tried to download the code from the website, but Python code folder includes only Matlab code :(
@NiKogane
@NiKogane 2 жыл бұрын
Hi, it was corrected - I downloaded it today !
@1PercentPure
@1PercentPure 7 ай бұрын
i kneel............................................................
@charlespatterson8412
@charlespatterson8412 4 жыл бұрын
I would prefer to do this in my head because I can visualize it and move it around. I am not a mathematician but many of these are terms for things I am already familiar with. Perhaps I should have kept my TRS80 and took Bill's Class at Juanita High. I decided to concentrate on 'Salmon Enhancement' and 'European History' instead. It's probably just as well, I find writing code quite boring because I am more into Concepts... "Keep up the Good work!"
@saitaro
@saitaro 4 жыл бұрын
Math is fully about concepts. And how would you visualize something that is higher than 3 dimension?
@charlespatterson8412
@charlespatterson8412 4 жыл бұрын
@@saitaro Extrapolation
@user-iiii234a5gc
@user-iiii234a5gc 4 жыл бұрын
add a time term? or 4dimension more is exist just at theorical expression
@yaseenmohammad9600
@yaseenmohammad9600 4 жыл бұрын
this technique is generally used when large amounts of higher dimensional data are there. like in image processing for example if u take 50(50*50) images it will become 50,2500 dimensional data resulting in covariance of 2500*2500 matrix where pca is used to extract eigen faces. now i don't think there are people who can solve eigen value equation for 2500 * 2500 matrix in head
@charlespatterson8412
@charlespatterson8412 4 жыл бұрын
@@yaseenmohammad9600 Maybe if the variables are 'round' enough I could 'take a shot' at it...
Principal Component Analysis (PCA) 2 [Python]
7:56
Steve Brunton
Рет қаралды 26 М.
PCA Analysis in Python Explained (Scikit - Learn)
16:11
Ryan Nolan Data
Рет қаралды 1,3 М.
Increíble final 😱
00:37
Juan De Dios Pantoja 2
Рет қаралды 110 МЛН
Osman Kalyoncu Sonu Üzücü Saddest Videos Dream Engine 170 #shorts
00:27
МАМА И STANDOFF 2 😳 !FAKE GUN! #shorts
00:34
INNA SERG
Рет қаралды 3,9 МЛН
Principal Component Analysis (PCA)
13:46
Steve Brunton
Рет қаралды 365 М.
Singular Value Decomposition (SVD): Mathematical Overview
12:51
Steve Brunton
Рет қаралды 378 М.
StatQuest: Principal Component Analysis (PCA), Step-by-Step
21:58
StatQuest with Josh Starmer
Рет қаралды 2,8 МЛН
Principal Component Analysis (PCA)
6:28
Visually Explained
Рет қаралды 191 М.
Principal Component Analysis (PCA)
26:34
Serrano.Academy
Рет қаралды 400 М.
Principal Component Analysis (PCA) - easy and practical explanation
10:56
Principal Component Analysis (PCA) using Python (Scikit-learn)
19:56
Michael Galarnyk
Рет қаралды 200 М.
Robust Principal Component Analysis (RPCA)
22:11
Steve Brunton
Рет қаралды 68 М.
Хотела заскамить на Айфон!😱📱(@gertieinar)
0:21
Взрывная История
Рет қаралды 4,4 МЛН
ИГРОВОВЫЙ НОУТ ASUS ЗА 57 тысяч
25:33
Ремонтяш
Рет қаралды 274 М.