This was great. I have seen countless videos on RNNs and LSTMs and nobody explained it by taking a simple example like you did. It was a perfect balance of theory with application.
@hamidrezahosseinkhani59806 күн бұрын
That was incredible! thanks!
@cocoph12 күн бұрын
This is the best explanation of transformers models, please keep going on this channel. There are lots of models still need to explain!
@BrianCarter15 күн бұрын
For a neuroatypical, your background music is too distracting, it would be nice if there wasn’t any
@Boom-em1os16 күн бұрын
thank you
@markthomas243618 күн бұрын
You did a fine job.
@phoenix179919 күн бұрын
Bro I use a setup with 128GB RAM with RTX 4080 16GB, RTX 3060 OC 12GB, RTX 2060 super 8GB on it with 5TB SSD M2 but I use as a open AIR setup for faster cooling. But you cabinet setup looks very efficient and cool. Could you send me the link for it
@tabindahayat349220 күн бұрын
Woah! Exquisite, It's a 15 min video but I spent over an hour taking notes and understanding. You have done a great job, keep it up. Thank you so much! Such explanations are rare. ;)
@alexdaniel7621 күн бұрын
Cool! 👏 Thank you for the video! What about the OS? Maybe Linux? If Linux, which one?
@anamariatiradogonzalez21 күн бұрын
Una estructura árbol de la vida. Kabakah
@anthonybernstein162623 күн бұрын
12:56 isn’t that the other way around (i.e. they queries come from the previous layer and the keys and values from the encoder’s output)?
@guilhermealvessilveira893827 күн бұрын
Excellent
@timothyweakly249627 күн бұрын
I would like to build one but I'm way ignorant on coding and building.
@ashishbhong5901Ай бұрын
it was not just help full but amazing, loved it.
@AravindUkrdАй бұрын
Please create more videos. You are really good.
@Sabumnim666Ай бұрын
For a guy who did a "lot" of research and you want mores cores why not get a thread ripper.
@coolStranger516Ай бұрын
thanks, bro. well explained.
@sweatyninja9755Ай бұрын
How do i fine tune something?
@ayanah4821Ай бұрын
😮
@martinsenuy895Ай бұрын
Hi, super explanatory and easy to follow video! Do you have any updates? Maybe using "cheap" AMDs like 6700xt? haha
@walloouuАй бұрын
i'm in love <3
@aneekeshkumar8199Ай бұрын
The audio kept buggin me, I'd heard it somewhere, then I remembered the Iconic Outros of the Channel Veritasium !!!!!
@sahhaf1234Ай бұрын
in these figures, where are the weights?
@josep1429Ай бұрын
🎯 Key Takeaways for quick navigation: 00:00 *Los transformers están revolucionando el procesamiento del lenguaje natural, superando a modelos anteriores como las redes neuronales recurrentes.* 02:29 *Los transformers introducen una arquitectura basada en atención, permitiendo un acceso potencialmente ilimitado al contexto durante la generación de texto.* 05:04 *La atención múltiple es un módulo clave en los transformers, que permite que cada palabra se relacione con otras en la secuencia de entrada.* 09:24 *La capa codificadora de los transformers utiliza la atención para crear una representación continua de la información de entrada.* 11:01 *Durante la decodificación, se aplica un enmascaramiento para prevenir que el modelo acceda a tokens futuros, asegurando una generación autoregresiva coherente.* 14:39 *Los transformers, al superar las limitaciones de la memoria a corto plazo, son especialmente eficaces para codificar y generar secuencias largas en el procesamiento del lenguaje natural.* Made with HARPA AI
@yashgajjar4838Ай бұрын
Thank you so much! Very well Explained, cleared most of the doubts.
@chriz__3656Ай бұрын
is it possible to build this on raspberry pi 3 plezzz reply 😇
@hussainbhavnagarwala2596Ай бұрын
Can we use CNN instead of RNN here for the classification of MFCC images?
@user-wm8hy8ce2oАй бұрын
bro you made this video before gpt 3 and all the new era of LLMs !!
@azharkhan-tr1wjАй бұрын
best explanation with good visualization
@mohamedibrahimbehery3235Ай бұрын
This is gold, man. Thanks!
@sj55582 ай бұрын
Excellent , I got a clear explanation
@seangai1262 ай бұрын
The i in positional encodings is an index to the dimension of the embedding, not the "timestep"
@muzammilomarzoy66162 ай бұрын
Chad Explanation
@charlesstevens32972 ай бұрын
not much
@abdulazizyaser16162 ай бұрын
thanks a lot, it was a lot of info but very helpful
@adambenghoula2 ай бұрын
where can i find the code of the system ?
@tsunningwah34712 ай бұрын
看進步健康情形不僅是看不見卡巴斯基開心吧就是
@jasonjennings84652 ай бұрын
I have a spare 3080 and want to build a deep learning machine/home server PC. Probably wont be that fast, but I figure it will be good enough for me to further my education.
@amruth25452 ай бұрын
Did u use GPU for training this
@wasifmasood9692 ай бұрын
Hi, many thanks for the great video. One quick question, if I choose AMD Ryzen, would I be able to install CUDA suit on it?
@csvegso2 ай бұрын
Why does the decoder select the token with the maximum probability instead of randomly selecting a token based on the probability distribution?
@BooleanDisorder2 ай бұрын
You have such a sweet and pleasant voice. Thank you, mate, for the good explanation. 😊
@climbersisyphus2 ай бұрын
super helpful. i want to be an ai engineer as well.
@tanveerulmustafa92323 ай бұрын
This explanation is INCREDIBLE!!!
@mohamedboufnichel61873 ай бұрын
great explanation
@ocelotcake93593 ай бұрын
bro gave apple the pass lol worst of all
@JarppaGuru3 ай бұрын
we allready done this decade ago. and it depend on language. if you write and spell same way then its easy but if...english then LOL. different woords spelled same! lol