"because love is....I'm becoming like Lex Fridman" -- genius!
@marouanemaachou78753 жыл бұрын
😂😂😂
@nju4153 жыл бұрын
Hahahaha I was about to comment on the same thing
@kimchi_taco3 жыл бұрын
What is meaning of life?
@G12GilbertProduction3 жыл бұрын
Friedman is really subjectionised reductionalist guy.
@konghong38853 жыл бұрын
Schmidhuber: I invented transformer
@mgostIH3 жыл бұрын
Yannic Kilcher Is (not so) Secretely A Fast Paper Reviews System
@yimingqu24033 жыл бұрын
Me a few month ago: I'm not interested in this work so I won't watch it. Me now: Yannic made this so I'll watch it.
@morkovija3 жыл бұрын
Oh my, first time i recognized the author in the paper =0 Yes, its that Jurgen
@scottmiller25913 жыл бұрын
The key, value, and query terminology always felt like someone who knew databases trying to explain something that is not a database. It's like trying to explain welding in terms of nails - it's really more confusing than explanatory.
@АлексейТучак-м4ч3 жыл бұрын
the principle of storing a single piece of data in several entries of a matrix resembles holography holograms store information about light sources in a distributed fashion
@TechyBen3 жыл бұрын
I love when mathematicians, logisticians, programmers/computer scientists and physicians find out one thing is the same as another. :)
@saurabhkulshreshtha89533 жыл бұрын
I know discovering symmetries! :)
@whosthisguythinkheis3 жыл бұрын
Love your videos! On the pixelation, you might have some luck turning the PDF into high res images, importing the pages into something like Krita or Gimp which is free and do your writing on top there. (krita has better pen support though I think)
@morkovija3 жыл бұрын
16:00 Interesting, I was not expecting the concept of arrow of time to appear in ML..
@wqchen35353 жыл бұрын
Thanks for your tutorial! It's very good! But I have a different opinion about what you said in 44:00-46:00 about the "d_dot=2d_key*niu". I think the paper is right. You can simply compare fai_i1(k) and fai_i3(k) when d_key=2. Although fai_11=fai_23, fai_21=fai33, fai_31=fai13, fai_41=fai_43, which is your main concern, they are different among the order of i in [1,4]. Different i corresponds to different dimension of fai(k). So, niu can be chosen from {1, 2, ……, 2d_key-1}, niu_j and niu_2d_key-j is not the same because the different order of i.
@mdmishfaqahmed55233 жыл бұрын
paper headings are amazing these days
@LouisChiaki3 жыл бұрын
The keys sound very similar to what physicists call orthonormal basis. Or, very similar idea used in quantum mechanics.
@etiennetiennetienne3 жыл бұрын
cool video and cool paper! could you provide some intuition (or some link to previous video) why softmax attention can hold exponentially many keys with respect to number of dimensions ? (i think you mentioned it for hopefield networks ?)
@sharannagarajan40898 ай бұрын
This is amazing! I think underrated
@matterhart3 жыл бұрын
I think you can load the pdf into python/acrobat and 2x/3x each page size and then save as pdf or image before dropping into one note.
@siyn0073 жыл бұрын
I would recommend putting the logo and youtube link at the top left or top right since we never try to observe text there
@piratepartyftw3 жыл бұрын
cool paper, clear explanation
@andres_pq3 жыл бұрын
How long back in history does the application of vector products to route information goes? It seems plausible that the key to AGI may already be solved by an obscure paper from the 80-90s.
@Uni1Lab3 жыл бұрын
Just for activate the algorithmes.
@andres_pq3 жыл бұрын
To be honest, I did not quite understood the ML street talk on kernels as well as the videos you make.
@MachineLearningStreetTalk3 жыл бұрын
We will do better 👌🙌
@andres_pq3 жыл бұрын
@@MachineLearningStreetTalk You are great guys!
@siddharthbhargava48573 жыл бұрын
Thank you for the explanation.
@mathematicalninja27563 жыл бұрын
I sometimes feel great that I decided to do masters in maths because I was able to understand this
@bhargav74763 жыл бұрын
I recently picked up 'Mathematics for Machine Learning' by A. Aldo Faisal, I just know calculus at this point. Would I be able to understand maths in this paper after reading that book?
@mathematicalninja27563 жыл бұрын
@@bhargav7476 I don’t know as I haven’t read that book yet
@bublylybub87433 жыл бұрын
this does not require "master" level math to be understood. just saying.
@444haluk3 жыл бұрын
Yannic be like: I am gonna drop transpose sign to mess up with my audience and butcher the math.
@tho2073 жыл бұрын
awesome explanations as always. why don't you just change apps? try pdf expert
@Erotemic3 жыл бұрын
Yannic: It's magic we can just add these outer products to get a database. Me: Wow amazing, how was I never aware of this?! Yannic: oh btw the keys need to be orthogonal. Me: ... so you can only store as many elements as the dimensionality of D... ok... cool, but I'm not nearly as impressed anymore.
@freemind.d27143 жыл бұрын
I actually think the autoregressive self-attention mechanism V(softmax(KT)q) is intuitively easy to understand than Fast Weight Memory Systems (V(KT))q, even though they are basically same concept with different formula
@dr.mikeybee3 жыл бұрын
LOL! Hi, Lex.
@corgirun78924 ай бұрын
oh my god
@dr.mikeybee3 жыл бұрын
I just watched this vlog on writing out the fast network as long-term memory: kzbin.info/www/bejne/jnPKqKRqfsqDgtk
@integralogic3 жыл бұрын
Dense Title.
@dimonenka3 жыл бұрын
these fast weight memory systems look like glorified hypernetworks
@benibachmann92743 жыл бұрын
Yannic „lächerliche Geschwindigkeit“ Kilcher in Hochform!