I have a hard time reading papers because my English isn't very good, but you've been very helpful in explaining it in your videos. Thank you.
@daesoolee10834 жыл бұрын
Interesting. Building a CNN model always depended on my intuition from existing CNN models. I never questioned the significance of each scale-up method. The analysis by disentangling is very helpful to the community. Excellent.
@roymarley51783 жыл бұрын
I know it is quite off topic but do anyone know of a good place to stream new movies online?
Hi! I think you got the resolution scaling wrong. They don't change the input dimensions (from say 224 to 360) but rather increase the number of convolution filters in every convolution, effectively increasing the number of feature maps of the low-level representation of the input at any given point in the model.
@schneeekind4 жыл бұрын
1:07 is the image right? b) and d) should change the figures? how is higher resolution resulting in deeper blocks?
@svm_user4 жыл бұрын
Thanks, great explanation.
@konataizumi58293 жыл бұрын
I feel like the grid search to find alpha, beta and gamma was not elaborated on enough in the paper. Does anyone understand this more deeply, or how one could reproduce it?
@AbdullahKhan-if8fn5 жыл бұрын
Very well explained. Thanks!
@connor-shorten5 жыл бұрын
Thank you!
@Ftur-57-fetr3 жыл бұрын
Superb explanation!!!!
@IvanGoncharovAI5 жыл бұрын
Great explanation!
@connor-shorten5 жыл бұрын
Thank you!!
@gvlokeshkumar4 жыл бұрын
Thank you so much!!!
@shreymishra646 Жыл бұрын
maybe i am a bit stupid on this but there is a mistake at 4:43 and i checked in the paper too w should be equal to -0.07 instead of 0.07 because if assuming asper the video the ratio is 0.07 then say if the flops of the resultant model are half of the target flops then this will become ACC * (0.5)^0.07 ~ ACC * 0.95 which is less than 1 hence penalizing the model (since we need to maximize this, right ? ) which is wrong it should actually support such a model while if we keep w equals to -0.07 then objective fx bcomes Acc * (0.5)^-0.07 ~ Acc * 1.04 I was a bit confused in the begenning of the vide since i didnt read the paper first, but now i am quite certain of it . I am quite surprised no one else noticed it !!!
@anticlementous4 жыл бұрын
Really great video! One thing I don't understand though, is how the scaling works exactly. Are the network dimensions scaled while training and while keeping the weights from the smaller scale or is the entire network retrained from scratch on each scaling? Also, if I do transfer learning with a model pre-trained on efficientnet I could get the benefits of reducing the network size but wouldn't have to run through the same scaling process?
@masoudparpanchi5054 жыл бұрын
a question. by this equation you said : Alpha * (Beta^2) * (gamma^2) = 2 . when I increase Alpha I should decrease two other variables?
@anefuoche10532 жыл бұрын
thank you
@xxRAP13Rxx3 жыл бұрын
At 2:37, shouldn't 2^n more computational resources imply a B^(2n) and a gamma^(2n) increase given the constraint A*B^2*gamma^2 = 2 ?
@xxRAP13Rxx3 жыл бұрын
Also, I'm looking over the actual paper. The chart at 5:32 is a bit different from what I'm seeing. Everything's about the same but every BOLDED Top1 Acc. entry (recorded from their own architecture) has been boosted up a few percentage points to outshine their rival counterparts. I wonder if they updated the paper since you posted this video, or maybe they figure it best to fudge the numbers since this chart is located on the front page of the paper.
@ibropwns5 жыл бұрын
Thanks a lot!
@connor-shorten5 жыл бұрын
Thank you!
@masoudparpanchi5054 жыл бұрын
good explanation
@tashin83124 жыл бұрын
I have a question. For my custom dataset I have used effnet b0-b5 & the results were getting poor each time I am using more complex models. Which means b0 gave best outcome while b5 gave the worst.... image sizes were 2000x1500 ...what could be the reason for that?