This video provides a clear step by step explanation how to get from images to input features for Transformer encoders, which has proven hard to find anywhere else. Thank you.
@PrestonRahim Жыл бұрын
Super helpful. Was very lost on the process from image patch to embedded vector until I watched this.
@DrAIScience7 ай бұрын
Very very very nice explanation!!! I like learning the foundation/origin of the concepts where models are derived..
@crapadopalese Жыл бұрын
10:46 - this is a mistake; the convolution is not equivariant to scaling - if the bird is scaled, the output of the convolution will not be simply a scaling of the original output. That would only be true if you also rescale the filters.
@xXMaDGaMeR Жыл бұрын
amazing lecture, thank you sir!
@ailinhasanpour Жыл бұрын
thanks for sharing , it was extremely helpful 💯
@OpenDataScienceCon Жыл бұрын
Thank you!
@sahil-vz8or Жыл бұрын
you said 196 patches in imagenet data. No of matches will depend on the input image size and the patch size. For eg: if the input image is of 400X400 and patch size of 8X8, then no of patches will be (400X400/8X8) = 50X50 =2500.
@rikki146 Жыл бұрын
20:17 I think the encoder blocks are stacked in parallel fashion rather than sequential?
@SarangBanakhede3 ай бұрын
10:58 Scale Equivariance: Definition: A function is scale equivariant if a scaling (resizing) of the input results in a corresponding scaling of the output. Convolution in CNNs: Standard convolutions are not scale equivariant. This means that if you resize an object in an image (e.g., making it larger or smaller), the CNN may not recognize it as the same object. Convolutional filters have fixed sizes, so they may fail to detect features that are significantly larger or smaller than the size of the filter. Example: If a CNN is trained to detect a small object using a specific filter size, it might struggle to detect the same object when it appears much larger in the image because the filter is not capable of adjusting to different scales. Why is Convolution Not Scale Equivariant? The filters in a CNN have a fixed receptive field, meaning they look for patterns of a specific size. If the size of the pattern changes (e.g., due to scaling), the fixed-size filters may no longer detect the pattern effectively.
@scottkorman4953 Жыл бұрын
What exactly is happening in the self-attention and MLP blocks of the encoder module? Could you describe it in a simplistic way?
@mohammedrakib37368 ай бұрын
Fantastic Video! Really loved the detailed explanation step-by-step.