PYTORCH COMMON MISTAKES - How To Save Time 🕒

  Рет қаралды 55,390

Aladdin Persson

Aladdin Persson

Күн бұрын

Пікірлер: 109
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Here is the outline for the video, let me know which ones you think I missed: 0:00 - Introduction 0:21 - 1. Didn't overfit batch 2:45 - 2. Forgot toggle train/eval 4:47 - 3. Forgot .zero_grad() 6:15 - 4. Softmax when using CrossEntropy 8:09 - 5. Bias term with BatchNorm 9:54 - 6. Using view as permute 12:10 - 7. Incorrect Data Augmentation 14:19 - 8. Not Shuffling Data 15:28 - 9. Not Normalizing Data 17:28 - 10. Not Clipping Gradients 18:40 - Which ones did I miss?
@seanbenhur
@seanbenhur 3 жыл бұрын
Shape mismatch error!
@caoviethainam9363
@caoviethainam9363 3 жыл бұрын
save the model and not to rerun the whole shit.
@eshtaranyal3011
@eshtaranyal3011 2 жыл бұрын
CUDA OOM error
@vijayak7308
@vijayak7308 2 жыл бұрын
I'm getting error : RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.FloatTensor) should be the same. I have assigned the input and model to 'cuda'. could you throw some light on this.
@FaisalAES
@FaisalAES 4 жыл бұрын
I honestly didn’t expect this video to be this professional and informative judging by the thumbnail and title
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Haha, thank you :)
@Hoxle-87
@Hoxle-87 3 жыл бұрын
So much good info here. I’ve been doing ML for 5 years n it is always good to review the basics every now n then.
@sagnikroy6405
@sagnikroy6405 3 жыл бұрын
This channel doesn't provide the basic tutorials which are there in the documentations and that's why it's very awesome. Thanks for your genuine content :D
@MrCmon113
@MrCmon113 Жыл бұрын
Some of my favourites are breaking the computational graph (e.g. using numpy functions instead of pytorch ones) or backpropagating somewhere you shouldn't. Or getting dimensionalities wrong and getting screwed over by Numpy''s automatic broadcasting. Or in general not looking for existing Pytorch functions and reinventing the wheel over and over again.
@xl0xl0xl0
@xl0xl0xl0 2 жыл бұрын
My fun mistake - added a ReLU in the last layer (before CrossEntropyLoss) - the model trains poorly for a while, then just stops training (once all logits have been driven below zero).
@jim_79
@jim_79 Жыл бұрын
Very useful tips for a novice like me Thank you
@user-or7ji5hv8y
@user-or7ji5hv8y 4 жыл бұрын
These practical tips are really useful.
@igordemidion9912
@igordemidion9912 4 жыл бұрын
Common mistakes for me: Getting confused with tensor dimensions (as a new guy you can spend plenty of time before harnessing the power of unsqueeze()) Forgetting .cuda() or .to(device) Getting confused with convnet dimensions after conv layer is applied Not attempting to balance or disbalance the dataset on purpose, which can be useful etc. Love your videos man, they've helped me alot.
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Those are some great things to keep in mind! Thank you, I appreciate you taking the time to comment
@saminchowdhury7995
@saminchowdhury7995 3 жыл бұрын
I made all these mistakes when I was newbie at Pytorch and still do it now sometimes This is a very helpful video
@siyuancheng9575
@siyuancheng9575 2 жыл бұрын
Many many many thanks to your video! The contents are all gold to newbie pytorch user and such a great guide!
@vanerk_
@vanerk_ Жыл бұрын
Applying different augmentations to the same batch, for example when training GANs and applying random flip.
@ogito999
@ogito999 3 жыл бұрын
The batch norm and bias evaluation difference is probably due to the randomness inherent in initializing 2 sets of biases instead of just 1.
@sheldonsebastian7232
@sheldonsebastian7232 3 жыл бұрын
This video is pure gold
@ceo-s
@ceo-s Жыл бұрын
You are the best! Just fixed few things
@not_a_human_being
@not_a_human_being 4 жыл бұрын
`seed=0` and not usual "42" - finally non-total-nerds are getting into the field! 😅Great Vid btw, please keep making more!
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Real Computer Scientists already know the answer to life, the universe and everything. That's trivial so we just use start index as 0. Mathematicians are a bit behind so they use 1.
@hadjdaoudmomo9534
@hadjdaoudmomo9534 3 жыл бұрын
Extremely helpful! thanks a lot!!
@MorisonMs
@MorisonMs 3 жыл бұрын
1:52 Low loss doesn't mean overfitting (I agree it's a good idea to run on small dataset at first don't get me wrong)
@deepshankarjha5344
@deepshankarjha5344 4 жыл бұрын
aladdin is the best. never l was not able to understand pytorch until this video series
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Thanks for the kind words man:)
@lakeguy65616
@lakeguy65616 3 жыл бұрын
great video, very informative! thank you!
@ashishbhatnagar8682
@ashishbhatnagar8682 4 жыл бұрын
Very helpful tips. Thanks a lot.
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Thank you for the comment :)
@wolfisraging
@wolfisraging 4 жыл бұрын
I don't think we need to shuffle the validation or test set right? Cuz there we will only be making the predictions and calculating our metrics like loss and accuracy, which are totally unaffected whether you shuffle or not. Plz do correct me if I'm wrong, thanks.
@AladdinPersson
@AladdinPersson 4 жыл бұрын
You're absolutely right, there's no need to shuffle the test set, so that was a mistake on my part. Good catch! :)
@georgepap9510
@georgepap9510 6 ай бұрын
I think you have an error in the check_accruracy function. You need to put the scores(since the are just the logits from a Linear layer) first in a softmax layer and then calculate the argmax .Am i missing something?
@john832-w1e
@john832-w1e 4 жыл бұрын
a7la Great Video 3alek !!!!! !!!! ya gamed!!!!!!
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Thank youuu :)
@neelabhmadan6820
@neelabhmadan6820 4 жыл бұрын
Spot on. 🙌
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Thanks :)
@boquangdong
@boquangdong 3 жыл бұрын
number 5. Bias term with BatchNorm. Can you explain more to me in this comment?
@homataha5626
@homataha5626 2 жыл бұрын
why did we need a DecoderBlock and a Decoder class? why no block for encoder?
@martimchaves9734
@martimchaves9734 3 жыл бұрын
Nice one mate
@Han-ve8uh
@Han-ve8uh 2 жыл бұрын
Could you clarify at 7:03 how does softmax on softmax lead to vanishing gradient?
@ali_nawaz_khattak
@ali_nawaz_khattak 2 жыл бұрын
@Aladdin Persson how to same thing using keras?
@coolz4ravs457
@coolz4ravs457 Жыл бұрын
How do you pad the mnist dataset by 2?
@mohdkashif7295
@mohdkashif7295 3 жыл бұрын
can i use torch.clamp for clipping gradient instead of torch.nn.utils.clip_grad_norm
@SaiPrabanjan
@SaiPrabanjan 3 жыл бұрын
Can you explain sir how to solve "CUDA out of memory" error in fastai package in pytorch. I am a beginner in fastai package and pytorch in general. Thanks for your great content sir.
@AladdinPersson
@AladdinPersson 3 жыл бұрын
Most commonly because you don't have enough vram on your gpu, i.e you're running too large batch_size or too large of a model
@TeachAI-UZ
@TeachAI-UZ 3 жыл бұрын
As far as I know, it is not advised to shuffle the validation (testing) data. Anyone experimented with this, too?
@wosleepy
@wosleepy 2 жыл бұрын
Shuffle will affect learning process. We are actually not learning during the val/test phase, so I guess it won't effect the accuracy. Hope it helps.
@pankajshinde475
@pankajshinde475 4 жыл бұрын
You're pytorch skills are just amazing, are you a phd student? 🤔BTW bow down to your pytorch skills ✌🙇‍♂️
@AladdinPersson
@AladdinPersson 4 жыл бұрын
No, masters student :)
@bpmsilva
@bpmsilva 4 жыл бұрын
@@AladdinPersson, your videos are awesome! Where did you learn all of that? You should do a video about your learning experience
@sasaglamocak2846
@sasaglamocak2846 2 жыл бұрын
Please tell us, how to learn PyTorch...
@736939
@736939 2 жыл бұрын
When I deploy the model, shoud i also use model.eval() ?
@sahil-7473
@sahil-7473 4 жыл бұрын
Great Vid! One more doubts. What's the exactly difference between torch.nn.Conv1d and torch.nn.functional.conv1d? Both seems to be present equally. That confusing me😅
@AladdinPersson
@AladdinPersson 4 жыл бұрын
For nn.modules you need to initialize them in the init function, for functional they are "stateless" and you need to manually set the weights. Basically functional has things without parameters/weights (and you would need to set weights manually). You can read more on the forum: discuss.pytorch.org/t/what-is-the-difference-between-torch-nn-and-torch-nn-functional/33597/6
@feravladimirovna1044
@feravladimirovna1044 4 жыл бұрын
when it comes to permute I am out of the space!!! lol!
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Yeah unfortunately my explanation wasn't very good there, just remember if you need to switch some axes of your tensor, use permute not view
@DanielWeikert
@DanielWeikert 4 жыл бұрын
In Tensorflow we often divide only be 255 to normalize. Would that be possible in Pytorch as well? (Would probably save time so we do not have to figure out mean and std) Thanks
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Just doing ToTensor() will divide by 255 so it gets in the range [0,1], but it's been shown to be better if you also do the additional step of obtaining zero mean and std 1 so it gets in the range [-1, 1]
@DanielWeikert
@DanielWeikert 4 жыл бұрын
@@AladdinPersson Thanks and how do you dertermine the mean and std. Is it like torch.mean(mydataset) ?
@lau.m.7698
@lau.m.7698 3 жыл бұрын
Hi! Love your channel! I have a question, what if the data that you want to normalize is not an image but a vector (a sequence of numbers)? What do you think would be the best type of normalization? I've tought about max-min norm that also set the data into [0,1] range but would it be necessary to use normalize with respect the mean and std? Thanks!!!
@bassemkaroui4914
@bassemkaroui4914 3 жыл бұрын
Always normalize using mean and std, because if your inputs are always positive then the gradients of the first layer connected to your inputs will always be either positive or negative (depending on the sign of the upcoming gradients) which essentially mean the weights of that layer are all updated in the same direction (either all increase or decrease) and this makes training a bit tricky using zigzag paths
@donfeto7636
@donfeto7636 29 күн бұрын
15:26 i don't think it is good to shuffle the test data, if you want to compare models bassed on test you need should not shuffle the test.
@jonatan01i
@jonatan01i 4 жыл бұрын
9:23 I think it is the gradient of the two layers' biases that are equal. If so, isn't having a bias in the conv layer and another one in the batch layer equivalent to having bias only in one of them but multiplying its gradient by 2?
@AladdinPersson
@AladdinPersson 4 жыл бұрын
From my understanding if we're first running it through a bias (and let's say every node activation gets raised +1) then running it through BatchNorm is going to remove this regardless and therefore it was completely irrelevant of having the bias. So I guess it's not a big deal but it's just an irrelevant parameter. I follow your point that the gradients are equal for the two layers but I don't follow when you say multiplying the gradient by 2
@jonatan01i
@jonatan01i 4 жыл бұрын
@@AladdinPersson Oh, you are right! I was not aware of the fact while I wrote the comment that the bias of the convolutional layer will be removed first by the batchnorm layer and after running the BN will we just add the bias of the BN layer. For some reason I thought that we add the BN's bias right after we've added the conv's bias. In that case would be the gradients of the two bias terms be the same. There comes from the factor of 2. But I was completely wrong about how we do batchnorm, so I was completely wrong. Then, the difference of the loss after training with and without the conv's bias term could be because of numerical reasons, couldn't it?
@dataaholic
@dataaholic 4 жыл бұрын
At 16:06 Why we pass the mean and std_dev as tuple ? I'm new to Deep learning and today i train a CNN on MNIST. After watching your video I change it to tuple and got a better accuracy and after training. Can you please tell me why this happens? Thanks in Advance and sorry for pasting the logs here in comments . --> with: transforms.Normalize(mean_gray, stddev_gray) Epoch: 1/10, Train(loss, accuracy): 1.058, 64.758, Test(loss, accuracy): 0.128, 96.480 Epoch: 2/10, Train(loss, accuracy): 0.349, 88.157, Test(loss, accuracy): 0.063, 98.310 Epoch: 3/10, Train(loss, accuracy): 0.160, 95.342, Test(loss, accuracy): 0.049, 98.410 Epoch: 4/10, Train(loss, accuracy): 0.117, 96.635, Test(loss, accuracy): 0.046, 98.570 Epoch: 5/10, Train(loss, accuracy): 0.094, 97.348, Test(loss, accuracy): 0.041, 98.660 ---------------------------------------------------------------------------------------------------------------------------- --> with: transforms.Normalize((mean_gray, ), (stddev_gray,)) Epoch: 1/10, Train(loss, accuracy): 0.439, 89.382, Test(loss, accuracy): 0.056, 98.290 Epoch: 2/10, Train(loss, accuracy): 0.120, 96.565, Test(loss, accuracy): 0.041, 98.780 Epoch: 3/10, Train(loss, accuracy): 0.087, 97.508, Test(loss, accuracy): 0.036, 98.840 Epoch: 4/10, Train(loss, accuracy): 0.077, 97.803, Test(loss, accuracy): 0.041, 98.890 Epoch: 5/10, Train(loss, accuracy): 0.068, 98.065, Test(loss, accuracy): 0.039, 98.930
@AladdinPersson
@AladdinPersson 4 жыл бұрын
I think if it's only for a single channel it shouldn't matter, did you make sure to set the seed etc so that the results are comparable?
@dataaholic
@dataaholic 4 жыл бұрын
@@AladdinPersson yeah, it is for one channel only and no I didn't use any seeding . I try again with seeding.
@dw61w
@dw61w 3 жыл бұрын
Is the intro made with manim?
@konstantin6482
@konstantin6482 3 жыл бұрын
What you didn't understand is that by typing in the mean and the variance for your normalization error you introduced a bias and that's why the performance has risen. Read "Learning from data". Awesome video otherwise, thanks 👍
@Dan-uf2vh
@Dan-uf2vh 3 жыл бұрын
Can someone explain to me how I should manage Dropout layers, considering I am using batch state - action - reward? I don't understand how just setting a mode.train() would work out. In my view, the Dropout layers would have to drop the same way when performing backpropagation, on the batch. Am I wrong? Is there something I am missing and how could I synchronize them if required and possible. Or do they just average out somehow?
@TeachAI-UZ
@TeachAI-UZ 3 жыл бұрын
You are right when you said the dropout (randomly) drops particular neurons in a layer based on the probability defined by an engineer. However, you do not perform backpropagation when validating (testing) your model. Specifically, model.eval(), which turns your model into testing mode, does not backpropagate; consequently, it does not to use dropout.
@Dan-uf2vh
@Dan-uf2vh 3 жыл бұрын
@@TeachAI-UZ there are training methods which REQUIRE a history in order to organize outputs and rewards and I would assume they need the exact form being used; for example q-learning with epsilon-greedy; if everything changes between the saved input-output-reward state and the moment you do the backpropagation, then I see no way that could work out
@Ip_man22
@Ip_man22 3 жыл бұрын
Really helpful! Thank you so much.
@MorisonMs
@MorisonMs 3 жыл бұрын
Why on the test dataset you perform shuffling?
@AladdinPersson
@AladdinPersson 3 жыл бұрын
That was a mistake! :)
@1chimaruGin0_0
@1chimaruGin0_0 4 жыл бұрын
Great work, as always! I used this to normalize images. I want to know is this good? loader = torch.utils.data.DataLoader(train_set, batch_size=16, shuffle=True) def mean_and_std(loader): mean = 0. std = 0. nb_samples = 0. for data,_ in loader: batch_samples = data.size(0) data = data.view(batch_samples, data.size(1), -1) mean += data.mean(2).sum(0) std += data.std(2).sum(0) nb_samples += batch_samples mean /= nb_samples std /= nb_samples print(mean) print(std)
@AladdinPersson
@AladdinPersson 4 жыл бұрын
The way you're calculating mean seems good but I think with the standard deviation there's a bit of a mistake. Since standard deviation isn't a linear operation you cannot do std += (batch_std_here). Doing in this way you will not obtain the real standard deviation. I made a video on it to show how you would do it which you can check out. Although your way will probably work just fine even with a minor flaw with the std :)
@1chimaruGin0_0
@1chimaruGin0_0 4 жыл бұрын
Thanks you
@gomeincraft
@gomeincraft 4 жыл бұрын
Great thanks from Russia. Really love your videos. In a very short time a got PyTorch essentials with the help of yours. So many models have been understood and implemented with your help. Keep it going buddy!!!!
@AladdinPersson
@AladdinPersson 4 жыл бұрын
To hear that makes me very happy, thank you :)
@_RMSG_
@_RMSG_ 2 жыл бұрын
I'm not understanding how Normalizing data doesn't hurt accuracy With time series data, if I take in a set of numbers, and then normalize those numbers, don't I get a normalized output instead of an accurate output?
@mihneaandreescu9922
@mihneaandreescu9922 3 жыл бұрын
AMAZING VIDEO, THANKS VERY VERY MUCH!!!
@ShakirKhan-th7se
@ShakirKhan-th7se 2 жыл бұрын
What is the best way to get input from different folders with different numbers of images each?
@vslaykovsky
@vslaykovsky 3 жыл бұрын
So what is the point of normalizing val/test data?
@sulavojha8322
@sulavojha8322 4 жыл бұрын
Extremely informative as always. Thank you !
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Appreciate your comment
@bassemkaroui4914
@bassemkaroui4914 3 жыл бұрын
Always normalize inputs using mean and std, because if your inputs are always positive then the gradients of the first layer connected to your inputs will always be either positive or negative (depending on the sign of the upcoming gradients) which essentially mean the weights of that layer are all updated in the same direction (either all increase or decrease) and this makes training a bit tricky using zigzag paths
@vanerk_
@vanerk_ Жыл бұрын
Can you please elaborate on why they would move in positive or negative directions? Imagine that your inputs are positive, you pass them through a Linear layer, then apply BatchNorm, then Linear again, then CrossEntropy. It is not obvious why the grads would be positive if chain rule would change signs for some direction derivatives.
@talha_anwar
@talha_anwar 3 жыл бұрын
the mistake I made is not putting my model in a function when doing cross-validation. in each fold it retrain on previous model
@ngawangchoeda3551
@ngawangchoeda3551 2 жыл бұрын
Do we really have to normalize the data in initial data transformation, if we use BatchNorm2d layer in our model architecture, because both would perform an identical task.
@shaikrasool1316
@shaikrasool1316 4 жыл бұрын
One word i can say, The best.. Thank you so much 😀
@AladdinPersson
@AladdinPersson 4 жыл бұрын
🙏🙏
@ali_nawaz_khattak
@ali_nawaz_khattak 2 жыл бұрын
can I do it for a custom dataset? if yes can you share code snippets for helping purposes?
@moorchini
@moorchini 3 жыл бұрын
That is a perfect video, really thankful. Would you plz tell me what's the best way to get the Accuracy of multiclass classification?
@monisprabu1174
@monisprabu1174 3 жыл бұрын
Great work bro do more pytorch vids keep it up!!!!!
@emanuelhuber4312
@emanuelhuber4312 2 жыл бұрын
The first tip just led me to the solution. Thanks!
@henrikvendelbo1117
@henrikvendelbo1117 3 жыл бұрын
Which of these are taken care of in lightning trainer?
@saurrav3801
@saurrav3801 4 жыл бұрын
Nice video bro....😇😇 More pytorch videos....about Layers, activation func,optimizers, etc...... Dont know really where to use which layer and activation func... 1. How to find mean and std of RGB ? 2. Is it possible to use batchnorm1d in linear layer ?
@AladdinPersson
@AladdinPersson 4 жыл бұрын
Thank you for your comment! 1. I made a video on it just now :) 2. I think so, but I haven't used this
@floriansommer1094
@floriansommer1094 3 жыл бұрын
Thank you so much, you really safed me a lot of time :)
@pawnagon4874
@pawnagon4874 3 жыл бұрын
These videos are always so fire, thank you sir
@tarakapoor5085
@tarakapoor5085 3 жыл бұрын
Thank you for this super helpful video! Do you have to do transformations and normalize your data (if it is images), or can you just feed in the pixel array without transformation/normalization?
@doyu_
@doyu_ 2 жыл бұрын
Should we leave shuffle=False for test_loader since the order of test data basically doesn't affect on test result, even for non time series?
@ИванЖарский-к9э
@ИванЖарский-к9э Жыл бұрын
Yes, it is unnecessary to shuffle when you test your model, so set shuffle=False when testing
Pytorch TensorBoard Tutorial
30:17
Aladdin Persson
Рет қаралды 35 М.
Please Help This Poor Boy 🙏
00:40
Alan Chikin Chow
Рет қаралды 11 МЛН
Mean Average Precision (mAP) Explained and PyTorch Implementation
27:32
7 PyTorch Tips You Should Know
17:12
Edan Meyer
Рет қаралды 21 М.
Pytorch ResNet implementation from Scratch
30:25
Aladdin Persson
Рет қаралды 94 М.
PyTorch Hooks Explained - In-depth Tutorial
22:08
Elliot Waite
Рет қаралды 35 М.
Autoencoder In PyTorch - Theory & Implementation
30:00
Patrick Loeber
Рет қаралды 68 М.
Building a Neural Network with PyTorch in 15 Minutes | Coding Challenge
20:34
Has Generative AI Already Peaked? - Computerphile
12:48
Computerphile
Рет қаралды 992 М.