Just binged the entire playlist, helped me understand the intuitions behind the math. I hope you make more videos :)
@goelnikhils2 жыл бұрын
Thanks a lot Lennart. What a crisp and clear explanation of BERT.
@rickyebay2 жыл бұрын
This is the best explanation on Transformer I have found in the web. Can you doing another set of video for T5 ?
@jeremyyd1258 Жыл бұрын
Excellent video! Thank you!
@JsaintUK2 жыл бұрын
Great video. Are the original word embeddings simple static embeddings? Where do they come from?
@lennartsvensson76362 жыл бұрын
They are "simple static embeddings". It is common to train them along with the other parameters.
@JsaintUK2 жыл бұрын
@@lennartsvensson7636 Okay so they could be something such as Word2vec embeddings? These are then passed into the encoder where they are contextualised?
@lennartsvensson76362 жыл бұрын
@@JsaintUK They could be but more commonly they are trained along all the other network parameters.