Transformers, explained: Understand the model behind ChatGPT

  Рет қаралды 4,740

Leon Petrou

Leon Petrou

Күн бұрын

🚀 Learn AI Prompt Engineering: bit.ly/3v8O4Vt
In this technical overview, we dissect the architecture of Generative Pre-trained Transformer (GPT) models, drawing parallels between artificial neural networks and the human brain.
From the foundational GPT-1 to the advanced GPT-4, we explore the evolution of GPT models, focusing on their learning processes, the significance of data in training, and the revolutionary Transformer architecture.
This video is designed for curious non-technical people looking to understand the complexities of GPT models in a way that's easy to understand.
🔗 SOCIAL LINKS:
🌐 Website/Blog: www.futurise.com/
🐦 Twitter/X: / joinfuturise
🔗 LinkedIn: / futurisealumni
📘 Facebook: profile.php?...
📣 Subscribe: www.youtube.com/@leonpetrou?s...
⏰ Timestamps:
0:00 - Intro
0:27 - The Importance of Modeling The Human Brain
1:10 - Basics of Artificial Neural Networks (ANNs)
2:26 - Overview of GPT Models Evolution
3:34 - Training Large Language Models
7:05 - Transformer Architecture
7:45 - Understanding Tokenization
10:19 - Explaining Token Embeddings
17:03 - Deep Dive into Self-Attention Mechanism
18:53 - Multiheaded Self-Attention Explained
19:55 - Predicting the Next Word: The Process
22:33 - De-Tokenization: Converting Token IDs Back to Words
#llm #ml #chatgpt #nvidia #elearning #futurise #promptengineering #futureofwork #leonpetrou #anthropic #claude #claude3 #gemini #openai #transformers #techinsights

Пікірлер: 37
@ravindranshanmugam782
@ravindranshanmugam782 2 ай бұрын
Excellent, went thro' multiple videos on basic understanding of Transformers. This is the best one I could quickly grasp. Effortlessly explained, Well done !!
@LeonPetrou
@LeonPetrou 2 ай бұрын
Thank you Ravindran! I try my best to teach things the same way that I'd like to be taught, which is simple and step-by-step. Let me know what other videos you'd like to see from my channel.
@ravindranshanmugam782
@ravindranshanmugam782 2 ай бұрын
Hi Leon, it would be great if you can make videos on Langchain and its application which are trending now. You can also add topics like Vectordatabase, Embedding, word2vec and so on. Anything on GenAI is hot now in tech space. Thanks.
@ovidioe.cabeza4750
@ovidioe.cabeza4750 9 күн бұрын
Same for me, I am a python backend dev and getting transformer was being tough, but you helped me a lot, thank you!
@programminglover2976
@programminglover2976 2 күн бұрын
thank you so much.. really reallly well explained.
@vj7668
@vj7668 3 күн бұрын
Excellent !!! Thanks for simplifying it. Loved it !
@LeonPetrou
@LeonPetrou 2 күн бұрын
Appreciate that, thank you!
@michaelzap8528
@michaelzap8528 9 күн бұрын
best. Finally i understand how gpt work now. Thanks male, u the champion.
@JohnCohen-ur5hk
@JohnCohen-ur5hk 7 күн бұрын
Very Good Explanation. Thank You
@Clammer999
@Clammer999 24 күн бұрын
Wow, this is one of the easiest to understand video on how transformers work. You also explained very tokens and embeddings which I was searching for. I’m a complete newbie and I kept hearing nuerons and neural networks. Is a neuron a physical device/hardware or it actually an algorithm? And a neural network is not a physical network?
@LeonPetrou
@LeonPetrou 22 күн бұрын
Thank you! Neural networks, and everything explained in this video is all software (except biological neurons which is in a human brain), it is all algorithms. It's basically just code. The hardware that the code runs on usually just requires high processing power / RAM. This can be a CPU or GPU.
@anibeto7
@anibeto7 Ай бұрын
It was indeed a very informative video. It cleared a lot of the important ideas. Thanks a lot.
@wp1300
@wp1300 18 күн бұрын
1:12 ANN 2:26 GPT-1 ~ GPT-4 3:34 LLM 7:09 Transformer architecture 7:45 Tokenization & Detokenization 8:17 Step 1 10:14 Step 2 10:20 Token embeddings 14:48 Step 3 15:10 Position Enbedding 16:58 Step 4 17:17 Self-Attention 18:52 Multi-headed self-attention 19:55 Step 5 20:27 Feed-Forward 22:02 Step 6 22:32 Step 6
@Omniassassin7
@Omniassassin7 2 ай бұрын
This is amazing, thanks a lot man! Quick question, how are the self-attention layers produced? Does the model dynamically “decide” which contextual layer to use depending on the prompt, or is the set of layers learnt during training?
@LeonPetrou
@LeonPetrou 2 ай бұрын
My pleasure man, glad you like it. That's a great question. The structure and behavior of these self-attention layers are determined during the model's training phase, not during inference. Simply put, the model learns which words in a sentence should pay attention to which other words to better understand the sentence's meaning. This learning process is fixed once the model is fully trained.. it does not change or decide on a different structure when it's given new prompts to process.
@karannesh7700
@karannesh7700 8 күн бұрын
thx for this great video !
@LeonPetrou
@LeonPetrou 8 күн бұрын
Appreciate it!
@MotulzAnto
@MotulzAnto 14 күн бұрын
THANK YOU! easy explanation..
@LeonPetrou
@LeonPetrou 13 күн бұрын
Appreciate it!
@sudhanshusaxena8134
@sudhanshusaxena8134 13 күн бұрын
Great explanation.
@LeonPetrou
@LeonPetrou 13 күн бұрын
Thank you very much!
@kamal9991999
@kamal9991999 11 күн бұрын
This video is a lot better one ☝️
@LeonPetrou
@LeonPetrou 10 күн бұрын
Appreciate that!
@abooaw4588
@abooaw4588 2 ай бұрын
Bravo 🇨🇵Dommage que ce très bon niveau de d'explication n'est réservé que pour nous qui comprenons l'anglais. Lecun et Bengio en sont pour beaucoup. Heureusement que le nutshell n'est pas traduit par GPT à la noix!
@LeonPetrou
@LeonPetrou 2 ай бұрын
Merci beaucoup for your thoughtful comment! I'm glad you found the video informative. Your point about language accessibility is very important to us. We're actively exploring options to include subtitles in multiple languages in our future videos to ensure more viewers can benefit from our content.
@Keshi-lz3ef
@Keshi-lz3ef 2 ай бұрын
Great session!
@LeonPetrou
@LeonPetrou 2 ай бұрын
Thank you!
@NavdeepVarshney-ep4ck
@NavdeepVarshney-ep4ck 21 күн бұрын
Sir are u a researcher or ml enthusiast
@LeonPetrou
@LeonPetrou 21 күн бұрын
I'm a ml enthusiast with an engineering background. :)
@dragonwood-hc4sw
@dragonwood-hc4sw 2 күн бұрын
Ed Stafford?
@LeonPetrou
@LeonPetrou 2 күн бұрын
I see it! haha
@d96002
@d96002 8 күн бұрын
not 175 trillion parameters but 1.75 trillion
@LeonPetrou
@LeonPetrou 8 күн бұрын
Thanks for clarifying, my bad.
@MaduraiKallan
@MaduraiKallan 7 күн бұрын
1.76 trillion for GPT 4
@LeonPetrou
@LeonPetrou 6 күн бұрын
indeed, thanks for clarifying!
@saeidnazemi1312
@saeidnazemi1312 3 ай бұрын
What happened to your hair?
@LeonPetrou
@LeonPetrou 3 ай бұрын
New year new me 😂
Generative AI in a Nutshell - how to survive and thrive in the age of AI
17:57
ELE QUEBROU A TAÇA DE FUTEBOL
00:45
Matheus Kriwat
Рет қаралды 27 МЛН
Final increíble 😱
00:39
Juan De Dios Pantoja 2
Рет қаралды 11 МЛН
Transformer Neural Networks, ChatGPT's foundation, Clearly Explained!!!
36:15
StatQuest with Josh Starmer
Рет қаралды 577 М.
The math behind Attention: Keys, Queries, and Values matrices
36:16
Serrano.Academy
Рет қаралды 203 М.
Why Does Diffusion Work Better than Auto-Regression?
20:18
Algorithmic Simplicity
Рет қаралды 144 М.
ChatGPT: 30 Year History | How AI Learned to Talk
26:55
Art of the Problem
Рет қаралды 980 М.
Transformers for beginners | What are they and how do they work
19:59
The Most Important Algorithm in Machine Learning
40:08
Artem Kirsanov
Рет қаралды 236 М.
Imagine This... | When AI Comes for Knowledge Workers
31:20
Boston Consulting Group
Рет қаралды 30 М.
What are Transformer Models and how do they work?
44:26
Serrano.Academy
Рет қаралды 97 М.
Main filter..
0:15
CikoYt
Рет қаралды 219 М.
Carregando telefone com carregador cortado
1:01
Andcarli
Рет қаралды 2,1 МЛН
Теперь это его телефон
0:21
Хорошие Новости
Рет қаралды 1,9 МЛН
How charged your battery?
0:14
V.A. show / Магика
Рет қаралды 3,2 МЛН
How To Unlock Your iphone With Your Voice
0:34
요루퐁 yorupong
Рет қаралды 17 МЛН