Here is the outline for the video, let me know which ones you think I missed: 0:00 - Introduction 0:21 - 1. Didn't overfit batch 2:45 - 2. Forgot toggle train/eval 4:47 - 3. Forgot .zero_grad() 6:15 - 4. Softmax when using CrossEntropy 8:09 - 5. Bias term with BatchNorm 9:54 - 6. Using view as permute 12:10 - 7. Incorrect Data Augmentation 14:19 - 8. Not Shuffling Data 15:28 - 9. Not Normalizing Data 17:28 - 10. Not Clipping Gradients 18:40 - Which ones did I miss?
@seanbenhur4 жыл бұрын
Shape mismatch error!
@caoviethainam93633 жыл бұрын
save the model and not to rerun the whole shit.
@eshtaranyal30112 жыл бұрын
CUDA OOM error
@vijayak73082 жыл бұрын
I'm getting error : RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.FloatTensor) should be the same. I have assigned the input and model to 'cuda'. could you throw some light on this.
@igordemidion99124 жыл бұрын
Common mistakes for me: Getting confused with tensor dimensions (as a new guy you can spend plenty of time before harnessing the power of unsqueeze()) Forgetting .cuda() or .to(device) Getting confused with convnet dimensions after conv layer is applied Not attempting to balance or disbalance the dataset on purpose, which can be useful etc. Love your videos man, they've helped me alot.
@AladdinPersson4 жыл бұрын
Those are some great things to keep in mind! Thank you, I appreciate you taking the time to comment
@Hoxle-873 жыл бұрын
So much good info here. I’ve been doing ML for 5 years n it is always good to review the basics every now n then.
@FaisalAES4 жыл бұрын
I honestly didn’t expect this video to be this professional and informative judging by the thumbnail and title
@AladdinPersson4 жыл бұрын
Haha, thank you :)
@sagnikroy64053 жыл бұрын
This channel doesn't provide the basic tutorials which are there in the documentations and that's why it's very awesome. Thanks for your genuine content :D
@MrCmon113 Жыл бұрын
Some of my favourites are breaking the computational graph (e.g. using numpy functions instead of pytorch ones) or backpropagating somewhere you shouldn't. Or getting dimensionalities wrong and getting screwed over by Numpy''s automatic broadcasting. Or in general not looking for existing Pytorch functions and reinventing the wheel over and over again.
@gomeincraft4 жыл бұрын
Great thanks from Russia. Really love your videos. In a very short time a got PyTorch essentials with the help of yours. So many models have been understood and implemented with your help. Keep it going buddy!!!!
@AladdinPersson4 жыл бұрын
To hear that makes me very happy, thank you :)
@emanuelhuber43122 жыл бұрын
The first tip just led me to the solution. Thanks!
@mihneaandreescu99223 жыл бұрын
AMAZING VIDEO, THANKS VERY VERY MUCH!!!
@user-or7ji5hv8y4 жыл бұрын
These practical tips are really useful.
@sulavojha83224 жыл бұрын
Extremely informative as always. Thank you !
@AladdinPersson4 жыл бұрын
Appreciate your comment
@pawnagon48743 жыл бұрын
These videos are always so fire, thank you sir
@siyuancheng95752 жыл бұрын
Many many many thanks to your video! The contents are all gold to newbie pytorch user and such a great guide!
@jim_79 Жыл бұрын
Very useful tips for a novice like me Thank you
@saminchowdhury79953 жыл бұрын
I made all these mistakes when I was newbie at Pytorch and still do it now sometimes This is a very helpful video
@xl0xl0xl02 жыл бұрын
My fun mistake - added a ReLU in the last layer (before CrossEntropyLoss) - the model trains poorly for a while, then just stops training (once all logits have been driven below zero).
@ceo-s Жыл бұрын
You are the best! Just fixed few things
@sheldonsebastian72323 жыл бұрын
This video is pure gold
@monisprabu11743 жыл бұрын
Great work bro do more pytorch vids keep it up!!!!!
@floriansommer10943 жыл бұрын
Thank you so much, you really safed me a lot of time :)
@Ip_man223 жыл бұрын
Really helpful! Thank you so much.
@moorchini4 жыл бұрын
That is a perfect video, really thankful. Would you plz tell me what's the best way to get the Accuracy of multiclass classification?
@lakeguy656163 жыл бұрын
great video, very informative! thank you!
@shaikrasool13164 жыл бұрын
One word i can say, The best.. Thank you so much 😀
@AladdinPersson4 жыл бұрын
🙏🙏
@bassemkaroui49143 жыл бұрын
Always normalize inputs using mean and std, because if your inputs are always positive then the gradients of the first layer connected to your inputs will always be either positive or negative (depending on the sign of the upcoming gradients) which essentially mean the weights of that layer are all updated in the same direction (either all increase or decrease) and this makes training a bit tricky using zigzag paths
@vanerk_ Жыл бұрын
Can you please elaborate on why they would move in positive or negative directions? Imagine that your inputs are positive, you pass them through a Linear layer, then apply BatchNorm, then Linear again, then CrossEntropy. It is not obvious why the grads would be positive if chain rule would change signs for some direction derivatives.
@vanerk_ Жыл бұрын
Applying different augmentations to the same batch, for example when training GANs and applying random flip.
@ogito9993 жыл бұрын
The batch norm and bias evaluation difference is probably due to the randomness inherent in initializing 2 sets of biases instead of just 1.
@wolfisraging4 жыл бұрын
I don't think we need to shuffle the validation or test set right? Cuz there we will only be making the predictions and calculating our metrics like loss and accuracy, which are totally unaffected whether you shuffle or not. Plz do correct me if I'm wrong, thanks.
@AladdinPersson4 жыл бұрын
You're absolutely right, there's no need to shuffle the test set, so that was a mistake on my part. Good catch! :)
@martimchaves97343 жыл бұрын
Nice one mate
@deepshankarjha53444 жыл бұрын
aladdin is the best. never l was not able to understand pytorch until this video series
@AladdinPersson4 жыл бұрын
Thanks for the kind words man:)
@ashishbhatnagar86824 жыл бұрын
Very helpful tips. Thanks a lot.
@AladdinPersson4 жыл бұрын
Thank you for the comment :)
@MorisonMs3 жыл бұрын
1:52 Low loss doesn't mean overfitting (I agree it's a good idea to run on small dataset at first don't get me wrong)
@john832-w1e4 жыл бұрын
a7la Great Video 3alek !!!!! !!!! ya gamed!!!!!!
@AladdinPersson4 жыл бұрын
Thank youuu :)
@saurrav38014 жыл бұрын
Nice video bro....😇😇 More pytorch videos....about Layers, activation func,optimizers, etc...... Dont know really where to use which layer and activation func... 1. How to find mean and std of RGB ? 2. Is it possible to use batchnorm1d in linear layer ?
@AladdinPersson4 жыл бұрын
Thank you for your comment! 1. I made a video on it just now :) 2. I think so, but I haven't used this
@not_a_human_being4 жыл бұрын
`seed=0` and not usual "42" - finally non-total-nerds are getting into the field! 😅Great Vid btw, please keep making more!
@AladdinPersson4 жыл бұрын
Real Computer Scientists already know the answer to life, the universe and everything. That's trivial so we just use start index as 0. Mathematicians are a bit behind so they use 1.
@MorisonMs3 жыл бұрын
Why on the test dataset you perform shuffling?
@AladdinPersson3 жыл бұрын
That was a mistake! :)
@tarakapoor50854 жыл бұрын
Thank you for this super helpful video! Do you have to do transformations and normalize your data (if it is images), or can you just feed in the pixel array without transformation/normalization?
@neelabhmadan68204 жыл бұрын
Spot on. 🙌
@AladdinPersson4 жыл бұрын
Thanks :)
@georgepap95107 ай бұрын
I think you have an error in the check_accruracy function. You need to put the scores(since the are just the logits from a Linear layer) first in a softmax layer and then calculate the argmax .Am i missing something?
@ShakirKhan-th7se2 жыл бұрын
What is the best way to get input from different folders with different numbers of images each?
@henrikvendelbo11173 жыл бұрын
Which of these are taken care of in lightning trainer?
@Han-ve8uh3 жыл бұрын
Could you clarify at 7:03 how does softmax on softmax lead to vanishing gradient?
@1chimaruGin0_04 жыл бұрын
Great work, as always! I used this to normalize images. I want to know is this good? loader = torch.utils.data.DataLoader(train_set, batch_size=16, shuffle=True) def mean_and_std(loader): mean = 0. std = 0. nb_samples = 0. for data,_ in loader: batch_samples = data.size(0) data = data.view(batch_samples, data.size(1), -1) mean += data.mean(2).sum(0) std += data.std(2).sum(0) nb_samples += batch_samples mean /= nb_samples std /= nb_samples print(mean) print(std)
@AladdinPersson4 жыл бұрын
The way you're calculating mean seems good but I think with the standard deviation there's a bit of a mistake. Since standard deviation isn't a linear operation you cannot do std += (batch_std_here). Doing in this way you will not obtain the real standard deviation. I made a video on it to show how you would do it which you can check out. Although your way will probably work just fine even with a minor flaw with the std :)
@1chimaruGin0_04 жыл бұрын
Thanks you
@sahil-74734 жыл бұрын
Great Vid! One more doubts. What's the exactly difference between torch.nn.Conv1d and torch.nn.functional.conv1d? Both seems to be present equally. That confusing me😅
@AladdinPersson4 жыл бұрын
For nn.modules you need to initialize them in the init function, for functional they are "stateless" and you need to manually set the weights. Basically functional has things without parameters/weights (and you would need to set weights manually). You can read more on the forum: discuss.pytorch.org/t/what-is-the-difference-between-torch-nn-and-torch-nn-functional/33597/6
@ali_nawaz_khattak2 жыл бұрын
can I do it for a custom dataset? if yes can you share code snippets for helping purposes?
@lau.m.76983 жыл бұрын
Hi! Love your channel! I have a question, what if the data that you want to normalize is not an image but a vector (a sequence of numbers)? What do you think would be the best type of normalization? I've tought about max-min norm that also set the data into [0,1] range but would it be necessary to use normalize with respect the mean and std? Thanks!!!
@bassemkaroui49143 жыл бұрын
Always normalize using mean and std, because if your inputs are always positive then the gradients of the first layer connected to your inputs will always be either positive or negative (depending on the sign of the upcoming gradients) which essentially mean the weights of that layer are all updated in the same direction (either all increase or decrease) and this makes training a bit tricky using zigzag paths
@doyu_2 жыл бұрын
Should we leave shuffle=False for test_loader since the order of test data basically doesn't affect on test result, even for non time series?
@ИванЖарский-к9э Жыл бұрын
Yes, it is unnecessary to shuffle when you test your model, so set shuffle=False when testing
@ngawangchoeda35512 жыл бұрын
Do we really have to normalize the data in initial data transformation, if we use BatchNorm2d layer in our model architecture, because both would perform an identical task.
@mohdkashif72953 жыл бұрын
can i use torch.clamp for clipping gradient instead of torch.nn.utils.clip_grad_norm
@vslaykovsky3 жыл бұрын
So what is the point of normalizing val/test data?
@_RMSG_2 жыл бұрын
I'm not understanding how Normalizing data doesn't hurt accuracy With time series data, if I take in a set of numbers, and then normalize those numbers, don't I get a normalized output instead of an accurate output?
@boquangdong3 жыл бұрын
number 5. Bias term with BatchNorm. Can you explain more to me in this comment?
@SaiPrabanjan3 жыл бұрын
Can you explain sir how to solve "CUDA out of memory" error in fastai package in pytorch. I am a beginner in fastai package and pytorch in general. Thanks for your great content sir.
@AladdinPersson3 жыл бұрын
Most commonly because you don't have enough vram on your gpu, i.e you're running too large batch_size or too large of a model
@7369392 жыл бұрын
When I deploy the model, shoud i also use model.eval() ?
@dataaholic4 жыл бұрын
At 16:06 Why we pass the mean and std_dev as tuple ? I'm new to Deep learning and today i train a CNN on MNIST. After watching your video I change it to tuple and got a better accuracy and after training. Can you please tell me why this happens? Thanks in Advance and sorry for pasting the logs here in comments . --> with: transforms.Normalize(mean_gray, stddev_gray) Epoch: 1/10, Train(loss, accuracy): 1.058, 64.758, Test(loss, accuracy): 0.128, 96.480 Epoch: 2/10, Train(loss, accuracy): 0.349, 88.157, Test(loss, accuracy): 0.063, 98.310 Epoch: 3/10, Train(loss, accuracy): 0.160, 95.342, Test(loss, accuracy): 0.049, 98.410 Epoch: 4/10, Train(loss, accuracy): 0.117, 96.635, Test(loss, accuracy): 0.046, 98.570 Epoch: 5/10, Train(loss, accuracy): 0.094, 97.348, Test(loss, accuracy): 0.041, 98.660 ---------------------------------------------------------------------------------------------------------------------------- --> with: transforms.Normalize((mean_gray, ), (stddev_gray,)) Epoch: 1/10, Train(loss, accuracy): 0.439, 89.382, Test(loss, accuracy): 0.056, 98.290 Epoch: 2/10, Train(loss, accuracy): 0.120, 96.565, Test(loss, accuracy): 0.041, 98.780 Epoch: 3/10, Train(loss, accuracy): 0.087, 97.508, Test(loss, accuracy): 0.036, 98.840 Epoch: 4/10, Train(loss, accuracy): 0.077, 97.803, Test(loss, accuracy): 0.041, 98.890 Epoch: 5/10, Train(loss, accuracy): 0.068, 98.065, Test(loss, accuracy): 0.039, 98.930
@AladdinPersson4 жыл бұрын
I think if it's only for a single channel it shouldn't matter, did you make sure to set the seed etc so that the results are comparable?
@dataaholic4 жыл бұрын
@@AladdinPersson yeah, it is for one channel only and no I didn't use any seeding . I try again with seeding.
@pankajshinde4754 жыл бұрын
You're pytorch skills are just amazing, are you a phd student? 🤔BTW bow down to your pytorch skills ✌🙇♂️
@AladdinPersson4 жыл бұрын
No, masters student :)
@bpmsilva4 жыл бұрын
@@AladdinPersson, your videos are awesome! Where did you learn all of that? You should do a video about your learning experience
@ali_nawaz_khattak2 жыл бұрын
@Aladdin Persson how to same thing using keras?
@DanielWeikert4 жыл бұрын
In Tensorflow we often divide only be 255 to normalize. Would that be possible in Pytorch as well? (Would probably save time so we do not have to figure out mean and std) Thanks
@AladdinPersson4 жыл бұрын
Just doing ToTensor() will divide by 255 so it gets in the range [0,1], but it's been shown to be better if you also do the additional step of obtaining zero mean and std 1 so it gets in the range [-1, 1]
@DanielWeikert4 жыл бұрын
@@AladdinPersson Thanks and how do you dertermine the mean and std. Is it like torch.mean(mydataset) ?
@dw61w3 жыл бұрын
Is the intro made with manim?
@coolz4ravs457 Жыл бұрын
How do you pad the mnist dataset by 2?
@jonatan01i4 жыл бұрын
9:23 I think it is the gradient of the two layers' biases that are equal. If so, isn't having a bias in the conv layer and another one in the batch layer equivalent to having bias only in one of them but multiplying its gradient by 2?
@AladdinPersson4 жыл бұрын
From my understanding if we're first running it through a bias (and let's say every node activation gets raised +1) then running it through BatchNorm is going to remove this regardless and therefore it was completely irrelevant of having the bias. So I guess it's not a big deal but it's just an irrelevant parameter. I follow your point that the gradients are equal for the two layers but I don't follow when you say multiplying the gradient by 2
@jonatan01i4 жыл бұрын
@@AladdinPersson Oh, you are right! I was not aware of the fact while I wrote the comment that the bias of the convolutional layer will be removed first by the batchnorm layer and after running the BN will we just add the bias of the BN layer. For some reason I thought that we add the BN's bias right after we've added the conv's bias. In that case would be the gradients of the two bias terms be the same. There comes from the factor of 2. But I was completely wrong about how we do batchnorm, so I was completely wrong. Then, the difference of the loss after training with and without the conv's bias term could be because of numerical reasons, couldn't it?
@homataha56262 жыл бұрын
why did we need a DecoderBlock and a Decoder class? why no block for encoder?
@sasaglamocak28462 жыл бұрын
Please tell us, how to learn PyTorch...
@talha_anwar3 жыл бұрын
the mistake I made is not putting my model in a function when doing cross-validation. in each fold it retrain on previous model
@TeachAI-UZ3 жыл бұрын
As far as I know, it is not advised to shuffle the validation (testing) data. Anyone experimented with this, too?
@wosleepy2 жыл бұрын
Shuffle will affect learning process. We are actually not learning during the val/test phase, so I guess it won't effect the accuracy. Hope it helps.
@Dan-uf2vh4 жыл бұрын
Can someone explain to me how I should manage Dropout layers, considering I am using batch state - action - reward? I don't understand how just setting a mode.train() would work out. In my view, the Dropout layers would have to drop the same way when performing backpropagation, on the batch. Am I wrong? Is there something I am missing and how could I synchronize them if required and possible. Or do they just average out somehow?
@TeachAI-UZ3 жыл бұрын
You are right when you said the dropout (randomly) drops particular neurons in a layer based on the probability defined by an engineer. However, you do not perform backpropagation when validating (testing) your model. Specifically, model.eval(), which turns your model into testing mode, does not backpropagate; consequently, it does not to use dropout.
@Dan-uf2vh3 жыл бұрын
@@TeachAI-UZ there are training methods which REQUIRE a history in order to organize outputs and rewards and I would assume they need the exact form being used; for example q-learning with epsilon-greedy; if everything changes between the saved input-output-reward state and the moment you do the backpropagation, then I see no way that could work out
@donfeto76362 ай бұрын
15:26 i don't think it is good to shuffle the test data, if you want to compare models bassed on test you need should not shuffle the test.
@feravladimirovna10444 жыл бұрын
when it comes to permute I am out of the space!!! lol!
@AladdinPersson4 жыл бұрын
Yeah unfortunately my explanation wasn't very good there, just remember if you need to switch some axes of your tensor, use permute not view
@konstantin64823 жыл бұрын
What you didn't understand is that by typing in the mean and the variance for your normalization error you introduced a bias and that's why the performance has risen. Read "Learning from data". Awesome video otherwise, thanks 👍