Lecture 8: CNN Architectures

  Рет қаралды 48,067

Michigan Online

Michigan Online

Күн бұрын

Пікірлер: 19
@JoshuaSaffy
@JoshuaSaffy Жыл бұрын
What's great is I will usually look up a tutorial and it'll say something like "A residual block does this" And I'm like "Great... why?" And this lecture just put everything in perspective. It motivated the problems, showed the purpose of every step, showed how each iteration went to solve the problems of the previous. This is honestly a great lecture.
@nishantkshirsagar2150
@nishantkshirsagar2150 Ай бұрын
This is one of the best lectures on CNN architectures.
@Nihit-n5n
@Nihit-n5n 4 жыл бұрын
what a great documentary on CNN architectures..slides are comprehensive and lecture(Dr Johnson) knows his stuff to extreme
@baskaisimkalmamisti
@baskaisimkalmamisti 3 жыл бұрын
That was a truly very impressive and informative lecture. I wish I could also listen to his summary and comments on the latest developments until 2021 such as NFnet.
@hasan0770816268
@hasan0770816268 3 жыл бұрын
28:26 why use 3x3 kernel in vgg 39:20 Google's way of dealing with kernel size 49:33 resnets
@vaibhavdixit4377
@vaibhavdixit4377 4 ай бұрын
Great resource, thoroughly researched + beautifully curated! Thanks a lot for the teachings!
@VikasKM
@VikasKM 3 жыл бұрын
Amazing lecture explaining the history of convnets from alexNet to resNets and mobileNets and also gives us idea as to which network to use if we are custom designing convnet architecture for our problem. Slides contain ton of information Thank you Justin Johnson
@Jppvv492
@Jppvv492 Жыл бұрын
The formula for calculating the output size of a convolutional layer in a Convolutional Neural Network (CNN) depends on several factors, including: 1. Input size (W_in, H_in): The spatial dimensions (width and height) of the input image or feature map to the convolutional layer. 2. Filter size (F): The spatial dimensions (width and height) of the convolutional filter (kernel). 3. Stride (S): The step size at which the filter is applied to the input. It defines how much the filter is shifted across the input. 4. Padding (P): The number of pixels added to the input on all sides to preserve spatial dimensions after convolution. The formula to calculate the output size (W_out, H_out) of the convolutional layer can be given as: W_out = ((W_in - F + 2 * P) / S) + 1 H_out = ((H_in - F + 2 * P) / S) + 1 If you want to maintain the spatial dimensions (W_in, H_in) of the input after convolution (i.e., no spatial downsampling), you can set the padding as: P = (F - 1) / 2 This formula assumes that the stride S is the same in both the horizontal and vertical directions. If you use different strides for width and height, the formula will change accordingly. It's worth noting that some frameworks and implementations may use slightly different conventions for padding (e.g., 'valid' or 'same' padding), so it's essential to check the documentation and specifications of the specific CNN implementation you are using.
@sardorabdirayimov
@sardorabdirayimov 2 жыл бұрын
Truly amazing lecture to listen many times
@itchainx4375
@itchainx4375 Жыл бұрын
1:08:22 now 1K gpus for several month is common for big giant cooperation
@parkie0517
@parkie0517 Жыл бұрын
Great lecture. Thank you so much
@xiangli1133
@xiangli1133 8 ай бұрын
Thank you so much!
@DED_Search
@DED_Search 3 жыл бұрын
29:50 how is output having the same dimension with input after 2 by 2 max pooling with stride 2 ?
@omerdor6644
@omerdor6644 3 жыл бұрын
halve in the sense of making it half of the value it used to be
@rajivb9493
@rajivb9493 3 жыл бұрын
Depthwise convolution...any references for this..?
@arunmehta8234
@arunmehta8234 3 жыл бұрын
14.14 time : why didn't you include Cin in calculating FLOPS in pool 1?
@JaviOrman
@JaviOrman 3 жыл бұрын
Each pooling operation is done on one input layer at a time, so it's kind of a 2D operation. Each pooling layer downsamples a corresponding input channel.
@АнтонГаничев-щ5ж
@АнтонГаничев-щ5ж 4 жыл бұрын
I believe there is a small error here: AlexNet has 96(not 64) filters in first layer (48+48 = 96). But overall the lecture is awesome.
@anishhui192
@anishhui192 3 жыл бұрын
In the paper, it is 96. But I also find 64 in Pytorch alexnet model pytorch.org/docs/stable/_modules/torchvision/models/alexnet.html#alexnet
Lecture 9: Hardware and Software
1:12:22
Michigan Online
Рет қаралды 27 М.
Lecture 7: Convolutional Networks
1:08:53
Michigan Online
Рет қаралды 57 М.
Каха и дочка
00:28
К-Media
Рет қаралды 3,4 МЛН
Beat Ronaldo, Win $1,000,000
22:45
MrBeast
Рет қаралды 158 МЛН
CNN: Convolutional Neural Networks Explained - Computerphile
14:17
Computerphile
Рет қаралды 864 М.
Lecture 1: Introduction to Deep Learning for Computer Vision
57:56
Michigan Online
Рет қаралды 228 М.
Learn Machine Learning Like a GENIUS and Not Waste Time
15:03
Infinite Codes
Рет қаралды 279 М.
Lecture 14: Visualizing and Understanding
1:12:04
Michigan Online
Рет қаралды 25 М.
CS231n Winter 2016: Lecture 7: Convolutional Neural Networks
1:19:01
Andrej Karpathy
Рет қаралды 169 М.
Lecture 5: Neural Networks
1:02:07
Michigan Online
Рет қаралды 42 М.
Convolutional Neural Networks from Scratch | In Depth
12:56