Пікірлер
@KhazeemasaleemKhazeemasale-o8s
@KhazeemasaleemKhazeemasale-o8s Күн бұрын
Hey khazeema saleem Al here
@GaryGan-US
@GaryGan-US 3 күн бұрын
very concise; what an amazing video.
@AICoffeeBreak
@AICoffeeBreak 3 күн бұрын
Thank you!
@martinkunev9911
@martinkunev9911 4 күн бұрын
I disagree that people are being judged for being bad at math more harshly than for other subjects. Compare "I'm not interested in history" or "I don't like reading" with "I'm bad at math".
@wapsyed
@wapsyed 4 күн бұрын
UMAP rocks! The only problem I see is the explainability of this high dimensionality reduction, which is easily done in PCA. In other words, you can get the best variables to explain the clustering, which is important when you are focusing on variable selection. What do you think?
@LinkhManu
@LinkhManu 4 күн бұрын
You’re the best 👏👏👏
@OlgaIvina
@OlgaIvina 6 күн бұрын
Thank you very much for this thorough, well-curated, and comprehensive review of MAMBA.
@AICoffeeBreak
@AICoffeeBreak 6 күн бұрын
Thank you, for your appreciation! I just saw you on LinkedIn, let's stay connected!
@alexkubiesa9073
@alexkubiesa9073 8 күн бұрын
How are exploration and exploitation abilities or forms of intelligence? To me they're more like competing actions, like going to the shops vs going to the cinema. I am still capable of both actions.
@luise.suelves8270
@luise.suelves8270 9 күн бұрын
sooo well explain, brilliant!
@AICoffeeBreak
@AICoffeeBreak 9 күн бұрын
Thanks!
@Jupiter-Optimus-Maximus
@Jupiter-Optimus-Maximus 9 күн бұрын
Another great video, as usual! This little bean mutant of yours always puts a smile on my face ☺ Is it possible that it is actually an AI? For example, a transformer that converts language information into the facial expressions of the animated bean. That would be so cool 😎 I have a question: I am looking for training methods that are not based on backpropagation. Specifically, I want to avoid running backwards through the NNW again after the forward pass. Do you know of any algorithms like this? Already 2^10 * Thanks in advance 😄
@moeinhasani8718
@moeinhasani8718 10 күн бұрын
Thanks!
@AICoffeeBreak
@AICoffeeBreak 10 күн бұрын
Wow, thank You!
@hannesstark5024
@hannesstark5024 11 күн бұрын
Neat
@maximilianstrasse3904
@maximilianstrasse3904 19 күн бұрын
You were the only lecturer I was able to pass some machine-learning related courses. now i know why :D you really took huge effort while staying nice/helpful/flexible... all the time
@AICoffeeBreak
@AICoffeeBreak 19 күн бұрын
Aww!
@harumambaru
@harumambaru 20 күн бұрын
7:05 Spoiler alert! Are you the only person wearing red?
@AICoffeeBreak
@AICoffeeBreak 20 күн бұрын
Yep! 😅
@harumambaru
@harumambaru 20 күн бұрын
To brew my cup of coffee I need scale with 0.1 gram accuracy and timer to follow best recipe, so much math to achieve one tasty cup of coffee :)
@dtibor5903
@dtibor5903 20 күн бұрын
I was tortured and bullied by math in school. I did engineering of automated systems, it was a nighthmare. I hate math because of the messy notation of variables.
@AICoffeeBreak
@AICoffeeBreak 20 күн бұрын
🫂
@DerPylz
@DerPylz 21 күн бұрын
Creating AI that's actually good at things that are very hard for humans often seems more important; However, tasks that are the easiest for humans can make great benchmarks for AI. The oldest idea in that direction is of course the Turing Test.
@itzhexen0
@itzhexen0 21 күн бұрын
I believe math attacked me once in school. I haven't been the same since.
@Thomas-gk42
@Thomas-gk42 21 күн бұрын
I don´t fear math, I just can´t follow much of the stuff. Now I´m even more frustrated, cause I couldn´t find you on the foto, ok, spatial thinking is connected with math talent, right? 😉Thanks for sharing your thoughts.
@AICoffeeBreak
@AICoffeeBreak 21 күн бұрын
Or maybe you lack training data with views with me from behind? 😜 Tip: I was among the very few people wearing red at the HLF. 😅
@Thomas-gk42
@Thomas-gk42 21 күн бұрын
@@AICoffeeBreak(Ah, jetzt ja) Ok, that´s obvious, 😂Have a nice day.
@AICoffeeBreak
@AICoffeeBreak 21 күн бұрын
@Thomas-gk42 danke, schönen Sonntag!
@autingo6583
@autingo6583 21 күн бұрын
no. obviously you have no idea about the matter and have no clue about psychometrics at all. yes, there is an effect called "subscale scatter" which becomes even more pronounced the higher the iq is you're testing for, but there is such a thing as a general factor to fluid intelligence. this is established fact and no one in their right mind is questioning this.
@DerPylz
@DerPylz 21 күн бұрын
Maths is one of my biggest fears 😮
@AICoffeeBreak
@AICoffeeBreak 21 күн бұрын
🫂
@kenfox9577
@kenfox9577 22 күн бұрын
Congratulations! Your description of being a lecturer giving you both a decent salary and academic freedom sounds very different from what I've heard about US universities. I enjoy Dr. Angela Collier's channel for her insights into physics and academics. She is also a great science communicator so you have that in common too. :) It would be very interesting for you two to have a conversation comparing your experiences. AI might even come up as a spicy side topic.
@aaryanbhagat4852
@aaryanbhagat4852 23 күн бұрын
Great personal and professionl growth
@BogdanOfficalPage
@BogdanOfficalPage 27 күн бұрын
Wow! Just great ❤
@hiramemon6047
@hiramemon6047 28 күн бұрын
Pleaseeeee do image position embeddings as well! This one was so good!
@kumarivin3
@kumarivin3 28 күн бұрын
Thank you so much !! you really super simplified it for any beginner level deep learner to understand
@TheAlexBell
@TheAlexBell Ай бұрын
Good explanation. Most videos on attention focus on how it's implemented, not on the design choices behind it. To my understanding, the goal was to mitigate the computational inefficiencies of RNNs and the spatial limitations of CNNs in order to achieve a universal representation of a sequence. I wanted to clarify one thing: you depicted multiple FFNNs similarly to how RNNs are usually rolled out. Is it just the same one FFNN that takes a single attention-encoded vector as input and predicts the next token from this ONE vector? By the way, what brand is that sweater? Loro Piana? :)
@0x0384
@0x0384 Ай бұрын
Congratulations 🎉
@AICoffeeBreak
@AICoffeeBreak 21 күн бұрын
Thank you!
@ramkumarr1725
@ramkumarr1725 Ай бұрын
Yes, when a person refers to the human brain in comparison to AI, they generally mean the collective intelligence of humanity, rather than the capabilities of an individual brain.
@ramkumarr1725
@ramkumarr1725 Ай бұрын
IT majors in India used to recruit for general intelligence and then make it sparse in the profession, focusing on specialized, repetitive tasks rather than broad skill development.
@jonclement
@jonclement Ай бұрын
Interesting. It's almost like two types of tokens: nodes + edges which can each be compressed to a feature vector. But yes, with positional encoding you're left with "random walk with restart" or a traversal depth. Or one could sum node_vector + edge_vector ~= positional distance. but yeah, more graph solutions coming in the future.
@sadossy-ec7rm
@sadossy-ec7rm Ай бұрын
very helpful, my thanks
@nehzz7433
@nehzz7433 Ай бұрын
Nice work!
@sinaasadiyan
@sinaasadiyan Ай бұрын
Great video
@RutujaKokate-u5h
@RutujaKokate-u5h Ай бұрын
i am so confused
@AntiJew964
@AntiJew964 Ай бұрын
I really like your vids
@catalindemergian2973
@catalindemergian2973 Ай бұрын
Super tare! Succes in continuare
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Mulțumesc!
@vladimirtchuiev2218
@vladimirtchuiev2218 Ай бұрын
I'm interested even more on the generative side, generating large graphs with contained text in them from a prompt, can be useful for modalities which are represented by large graphs. I've yet to see anyone doing this. While you can prompt LLMs to generate small graphs, for larger graphs you see significant performance drops.
@hayatisschon
@hayatisschon Ай бұрын
Beautifully explained by a smart & beautiful person!
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Cheers
@yorailevi6747
@yorailevi6747 Ай бұрын
I need to read this more deeply, I don't understand why would just grafting the parameters willy nilly works
@keeperofthelight9681
@keeperofthelight9681 Ай бұрын
Deep Learning is more of an alchemy than anything, an in-depth thought out plan may not work and sometimes just a hacky way around a solution works a lot better
@cactilio11
@cactilio11 Ай бұрын
Hello Letitia!!! First off, I love your videos! they are helping me a lot throughout my PhD (which, I must add, takes a lot of inspiration from your research too!). Second, I am currently interested in Continual Learning applied to MMLT, so I'm thinking to study some of these multi modal transformers. What do you think? is this an interesting line of research? PD: congrats on your degree!
@jmirodg7094
@jmirodg7094 Ай бұрын
Excellent! need to go deeper that could be a game changer for reasoning, as it makes more sense to reason on a graph rather than on the next token.
@bharanij6130
@bharanij6130 Ай бұрын
Thank you for this video Letitia! As always amazing :=) Side note: Loved the silent Good Bye :)))
@bensimonjoules4402
@bensimonjoules4402 Ай бұрын
Its interesting to see "attention" on graph structures again. I think in the future a more structured knowledge representation may play a role on improving reasoning, as we could leverage logic and rules using engines on them, like compilers aid in code generation.
@sonOfLiberty100
@sonOfLiberty100 Ай бұрын
It would be interesting how much computation this needs
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Do you mean for training or inference? Training is a finetuning setting and you can see performance curves in Figure 4 in the paper. arxiv.org/pdf/2401.07105 Inference costs as much as the base LLM.
@sonOfLiberty100
@sonOfLiberty100 Ай бұрын
@@AICoffeeBreak both things. overall computation. Thank you, I will take a look
@MoritzPlenz
@MoritzPlenz Ай бұрын
Hi, I am Moritz (one of the authors). I don't have much to add to Letitia's reply, but here is another relevant part of the paper, taken from section 4: Being transformers, GLMs have the same computational complexity as their respective LM. For sparse graphs the lGLM could make use of sparse matrix multiplication, making it more efficient than a corresponding LM or gGLM. However, for our experiments this was not necessary.
@hayatisschon
@hayatisschon Ай бұрын
Letitia, you are an amazing teacher/instructor!
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Aww, thank you!
@MariaM-pu4fx
@MariaM-pu4fx Ай бұрын
LOVE IT. Dr any idea on research in this area?
@MariaM-pu4fx
@MariaM-pu4fx Ай бұрын
I understand compleatly nothing. I was focused on this guy aspergery passion. How can I work with such cyborgs :D sorry. I am a hater here but I started questioning my role on the job market after watching it.
@MariaM-pu4fx
@MariaM-pu4fx Ай бұрын
chenged my mind I like the explanation sorry for my ADHD
@declan6052
@declan6052 Ай бұрын
At 13:14 - Is this 'clip guided diffusion' done by adding a term to the loss function or via a different method?
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
It's done by adding an image to the generated image during inference. This extra added image is computed via the gradient with respect to clip's output. It's a bit like deep dream, if you are old enough to know about it.
@SandhyaPatil20
@SandhyaPatil20 Ай бұрын
❤ I am a fan ❤
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thank you!
@yannickpezeu3419
@yannickpezeu3419 Ай бұрын
Do we have an evaluation of the difficulty to learn each human language in terms of flops to achieve a given perplexity ?
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
No, unfortunately. Clearly, this would have completed the study and would have made it less debatable.