This is gold. Thanks for the thorough and great content!
@DeepFindr2 жыл бұрын
Thanks!!
@robertchamoun79142 жыл бұрын
Thanks!
@jiahao2709 Жыл бұрын
One thing i want to say is that, test data usually only use once as test time, i think it is better call the "test data" as validation data.
@kenbobcorn Жыл бұрын
It's probably worth mentioning you are computing gradients on your test set by not setting torch.no_grad() for the test loop. This series is all about uncertainty so it's important you aren't computing gradients on your test set which leaks into your mu and var values, which in the end is contrary to what you are trying to show.
@DeepFindr Жыл бұрын
Hi! Good remark. But as long as you aren't running back propagation w.r.t. to the test loss it won't leak any information into the model weights. Torch.no_grad is mainly used for memory reasons
@kevinkorfmann87802 жыл бұрын
thanks, that was really helpful! Looking forward to part 3.
@DeepFindr2 жыл бұрын
Thank you! Next part is coming next week :)
@nguyenxuanthanh6988 Жыл бұрын
Brilliant!!! These videos help me a lot in understanding uncertainty. Could you make more videos regarding this topic? Thank you so much.
@shilpimajumder79172 жыл бұрын
Thanks for the video...i learn a lot..Please upload some videos of uncertainty estimation in image classification.
@DeepFindr2 жыл бұрын
Thanks! For image classification the same principles apply. You just have other layers (Conv2d) instead of Linear.
@clima39932 жыл бұрын
What if p(y|x) is not Gaussian? What is y is high dimensional?
@DeepFindr2 жыл бұрын
There are alternative loss function for other distributions like Laplace ect. or you transform the target variable in some way to match a supported distribution. Multidimensionality is no problem for GaussianNNL Loss, it simply apply the calculation per dimesion and averages it.
@clima39932 жыл бұрын
@@DeepFindr Thanks for the helpful reply. A following up question: what about we use conditional generative model to handle aleatoric uncertainty?
@DeepFindr2 жыл бұрын
Sure that's also a reasonable approach. You can learn the data distribution and detect out of distribution samples using generative models.