How positional encoding in transformers works?

  Рет қаралды 6,949

BrainDrain

BrainDrain

Күн бұрын

Пікірлер: 22
@emitate
@emitate 8 ай бұрын
In my opinion, best explanation so far of positional encoding! Super clear and concise! Thank you very much sir!
@cybermanaudiobooks3231
@cybermanaudiobooks3231 5 ай бұрын
The best explanation of transformer positional encoding on the internet. Awesome video. Thanks!
@atabhatti2844
@atabhatti2844 3 ай бұрын
Great explanation. Short enough. Detailed enough. Enough talking. Enough showing. Loved the examples.
@Wesleykoonps
@Wesleykoonps 6 ай бұрын
I like very concise graphical explanation with the similarity to binary coding and basic linear algebra!
@Adhbutham
@Adhbutham Ай бұрын
Incredible! I have surfed through various resources online and no one got this so accurately. Absolutely spot on explanation.
@marcinstrzesak346
@marcinstrzesak346 4 ай бұрын
I couldn't find anywhere why creators of transformer decied to encode the positions in this way and last minute of your video was what I was looking for. Thanks for good explanation
@markburton5318
@markburton5318 Ай бұрын
It seems the addendum is a 5th requirement. I can’t word this precisely but the positional encoding can be learned easily, that the embedding is only a linear transformation of position. It cannot be an encryption of the token.
@mohammedelsiddig3939
@mohammedelsiddig3939 3 ай бұрын
I'm eternally grateful for this concise explanation, other sources made the positional encoding concept sound so counter-intuitive to grasp
@ea_777
@ea_777 4 ай бұрын
Just when I was about to pull the last hair on top of my head, I came across this video. Beautifully Explained. Thank You !
@JesseNerio
@JesseNerio 5 ай бұрын
Fantastic. This was amazing! Best explanation.
@prabhuramnagarajan1893
@prabhuramnagarajan1893 Ай бұрын
please explain in detail about the linear relation with two encoding. You mathematical proofs, sounds excellent. Please recommend a good book to understand in detail about these concepts.
@simaogoncalves1957
@simaogoncalves1957 4 ай бұрын
Keep these coming!
@temanangka3820
@temanangka3820 3 ай бұрын
1. Why positional encoding is added to the word embedding? Will it changes the semantic value? 2. Why positional encoding use random number produce by sin and cosine... I think it must be simple if we add the one dimension to word embedding storing the position as integer. Why use such a hard, random, and unpredictable algorithm to encode positions!
@gart1609
@gart1609 14 күн бұрын
Why do we need to alternate sine and cosine? It seems like either one on its own should do the job. The only reason I can see for alternations is that this way we can solve the problem of positional encoding with the wavelength twice as short, as opposed to sine or cosine alone. Is that right? Are there other reasons?
@phobosmoon4643
@phobosmoon4643 Ай бұрын
bravo
@wilfredomartel7781
@wilfredomartel7781 Ай бұрын
@temanangka3820
@temanangka3820 3 ай бұрын
How can adding positional encoding to word embedding doesnt change the word semantic meaning? Example: Word embedding of "Cat" is [1, 2, 3], Word embedding of Money is [2, 3, 4]. If the positional encoding is [2, 1, 0] for word "Cat", positional encoding for word "Money" is [1, 0, -1] then the positional encoded of both word is [3, 3, 3] How can "Cat" equal to "Money"?
@BrainDrain9000
@BrainDrain9000 3 ай бұрын
Because positional part is a constant. Token part is stochastic, it changes depending on current token, but positional part remains the same. Imagine that you recorded all embeddings of a 0th token from the whole dataset and you got a map, distribution. If you add some constant, this map will remain the same, but shifted to some other location. And yes, it will not work for two examples, you need sufficient amount of data to prevent confusion.
@temanangka3820
@temanangka3820 2 ай бұрын
@@BrainDrain9000 I see... 🔥 Thank you ✅
@ronin2963
@ronin2963 Ай бұрын
Can you project your speak. You asmr tone is disturbing
@DanielYang-mc6zn
@DanielYang-mc6zn Ай бұрын
This video is already very outdated lol
@suhasbrad4884
@suhasbrad4884 Ай бұрын
How?
Positional Encoding in Transformer Neural Networks Explained
11:54
CodeEmporium
Рет қаралды 41 М.
Positional encodings in transformers (NLP817 11.5)
19:29
Herman Kamper
Рет қаралды 2,9 М.
Magic or …? 😱 reveal video on profile 🫢
00:14
Andrey Grechka
Рет қаралды 88 МЛН
У ГОРДЕЯ ПОЖАР в ОФИСЕ!
01:01
Дима Гордей
Рет қаралды 8 МЛН
How Strong is Tin Foil? 💪
00:26
Preston
Рет қаралды 76 МЛН
SCHOOLBOY. Мама флексит 🫣👩🏻
00:41
⚡️КАН АНДРЕЙ⚡️
Рет қаралды 7 МЛН
AI can't cross this line and we don't know why.
24:07
Welch Labs
Рет қаралды 667 М.
2 Years of My Research Explained in 13 Minutes
13:51
Edan Meyer
Рет қаралды 56 М.
A Very Simple Transformer Encoder for Time Series Forecasting in PyTorch
15:34
Let's Learn Transformers Together
Рет қаралды 7 М.
How 3 Phase Power works: why 3 phases?
14:41
The Engineering Mindset
Рет қаралды 1 МЛН
The KV Cache: Memory Usage in Transformers
8:33
Efficient NLP
Рет қаралды 38 М.
Attention in transformers, visually explained | Chapter 6, Deep Learning
26:10
The Key Equation Behind Probability
26:24
Artem Kirsanov
Рет қаралды 93 М.
How might LLMs store facts | Chapter 7, Deep Learning
22:43
3Blue1Brown
Рет қаралды 500 М.
Magic or …? 😱 reveal video on profile 🫢
00:14
Andrey Grechka
Рет қаралды 88 МЛН