Пікірлер
@MatzeU-r8q
@MatzeU-r8q Күн бұрын
One of the BEST tutorials/explanations i have ever seen, thank you really much for that ❤
@AICoffeeBreak
@AICoffeeBreak 20 сағат бұрын
Much appreciated!
@pranavb9768
@pranavb9768 Күн бұрын
you’re cute
@ahsan4life2020
@ahsan4life2020 4 күн бұрын
This is a very valuable video for me. Thank you so much
@AICoffeeBreak
@AICoffeeBreak 4 күн бұрын
For a second I thought you're talking about the three minutes of uncomfortable silence . On the comment notification I don't see in which video the comment was made. 😅 So, I'm glad it is about this one.
@erongjoni3464
@erongjoni3464 5 күн бұрын
Lost it on "Watch the full video in the link below"
@AICoffeeBreak
@AICoffeeBreak 5 күн бұрын
And that is just the finest 3 minute selection of the full 50 minutes awkward silence video. 😏
@foreignconta
@foreignconta 6 күн бұрын
What exactly happened here?🤔
@AICoffeeBreak
@AICoffeeBreak 6 күн бұрын
My editor had fun cutting out moments of awkward silence from my next (1h+) video.
@bmebri1
@bmebri1 7 күн бұрын
I love it; most unawkward facet of you yet😊
@AICoffeeBreak
@AICoffeeBreak 6 күн бұрын
@DaNa-px7ol
@DaNa-px7ol 7 күн бұрын
Wow I didn’t know I can enjoy “awkward” silence 😮
@AICoffeeBreak
@AICoffeeBreak 7 күн бұрын
😂
@Thomas-gk42
@Thomas-gk42 7 күн бұрын
Waiting for the phd result...?
@AICoffeeBreak
@AICoffeeBreak 7 күн бұрын
Haha, fortunately, that one is clear. 😅🎉🤗
@outliier
@outliier 7 күн бұрын
Some would upload this as ASMR
@AICoffeeBreak
@AICoffeeBreak 7 күн бұрын
🎙️
@WhatsAI
@WhatsAI 7 күн бұрын
Haha a classic when recording!
@DerPylz
@DerPylz 7 күн бұрын
I like the part where she doesn't say anything.
@AICoffeeBreak
@AICoffeeBreak 7 күн бұрын
😅
@azmathmoosa4324
@azmathmoosa4324 15 күн бұрын
good concise explanation.
@sohambit9393
@sohambit9393 19 күн бұрын
Damn you like Neffex ❤ Neffex is like 10 % of my life.
@enicay7562
@enicay7562 22 күн бұрын
Thank you Miss Coffee Bean !
@AICoffeeBreak
@AICoffeeBreak 21 күн бұрын
@kristoferkrus
@kristoferkrus 22 күн бұрын
Cool! The simple but effective ideas are the most interesting. Have you checked out Nvidia's normalized transformer (nGPT)? It seems to be one of those cases. The differential transformer (arxiv: 2410.05258) also seems like it could be interesting.
@davide0965
@davide0965 25 күн бұрын
Terrible
@DerPylz
@DerPylz 25 күн бұрын
If you don't like her videos, why do you keep coming back to them just to comment that you didn't like it? Just watch something else.
@AICoffeeBreak
@AICoffeeBreak 26 күн бұрын
03:23 It looks like the MLP takes the representation of JUST the 8th layer, and not also of the previous ones. I just found out from a quick exchange via email with the authors.
@bethany-rp2tq
@bethany-rp2tq 26 күн бұрын
CONGRATULATIONS Letitia !!!! Is there a pdf version where I can read your PhD thesis? Always amazed learning new knowledge from you :)
@AICoffeeBreak
@AICoffeeBreak 26 күн бұрын
Yes, there is! archiv.ub.uni-heidelberg.de/volltextserver/35753/ Thanks for your interest!
@davidrichards1302
@davidrichards1302 29 күн бұрын
Unison's unique features offer promising possibilities for enhancing the FunSearch operational model: Content-addressed code: Unison's hash-based identification of code could streamline FunSearch's program generation and evaluation process. Each generated program would have a unique hash, simplifying tracking and caching of results. Immutable codebase: Unison's immutable data structure approach to codebases could enhance FunSearch's evolutionary process. It would allow for efficient storage and retrieval of program versions without conflicts. No builds: Unison's ability to parse and typecheck definitions once, storing results in a cache, could significantly speed up FunSearch's evaluation phase. This would reduce overhead in assessing generated programs. Easy distributed computing: Unison's content-addressed nature facilitates distributed computation, which could enhance FunSearch's parallelization capabilities. This could allow for more efficient scaling of the search process across multiple machines. First-class documentation: Unison's approach to documentation as executable code could improve FunSearch's ability to generate and evaluate self-documenting programs, potentially leading to more interpretable solutions. Strong typing with inference: Unison's type inference could help FunSearch generate type-safe programs more efficiently, potentially reducing the number of invalid programs generated. These features could potentially make Unison an excellent choice for implementing core components of the FunSearch system, particularly in program generation, evaluation, and distributed processing.
@duzx4541
@duzx4541 Ай бұрын
Hmm, the only thing I dont really understand is why we have to use SIN and COS instead of only one of them D:
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
They are the same with just a phase shift. Rotary embeddings take the idea of phase shifting and turn it up.
@katorea
@katorea Ай бұрын
Loved your explanation! thank you very much!! :D
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
@solsospecial
@solsospecial Ай бұрын
Okay, the approach is “guessing”. Got it.
@KnowledgeSynthesizer
@KnowledgeSynthesizer Ай бұрын
can I ask what app/software tools you use for recording, editing audio and video?
@AICoffeeBreak
@AICoffeeBreak 26 күн бұрын
Hi, it is good old powerpoint for all visualisations. Ms. Coffee Bean comes during editing in Adobe Premiere. :)
@floriankowarsch8682
@floriankowarsch8682 Ай бұрын
It would be definitely interesting to train end to end with contrastive lost. Because right now it is a more like a distillation task that is limited be the teacher model's capability & domain
@dylancope
@dylancope Ай бұрын
"wait a few years"... 9 months later with Genie 2 😂
@AICoffeeBreak
@AICoffeeBreak 26 күн бұрын
Years in ML pass very quickly. 🤣🤣🤣
@rogerthat7190
@rogerthat7190 Ай бұрын
This was very helpful, thank you!
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thanks for the kind words!
@hjups
@hjups Ай бұрын
It's a clever idea, but the paper has many methodological issues which may amount to more hype than substance. Then results are good, but the major gains come from stacking contributions. And of course you have to worry about the extra training cost (DINO is very expensive) and generalization, etc. The peak accuracy shift is also worrisome, indicating that removing the driving force would likely result in an expensive shift / performance degradation when finetuning.
@fast_harmonic_psychedelic
@fast_harmonic_psychedelic Ай бұрын
thats such a good idea. very simple .. i would not have thought to do this. I have added CLIP loss terms on the last layer but not dinoon the 8th layer lol.
@AICoffeeBreak
@AICoffeeBreak 26 күн бұрын
Yes, indeed. 😅 The last layers still need to focus on reconstruction, or rather predicting the noise. Injecting in the first layers helps the network "conceptualise" what the denoising needs to become. The following layers need to focus on predicting that noise.
@EkShunya
@EkShunya Ай бұрын
i missed you, glad to see you back with your great explainers
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
@circuitbreaker7860
@circuitbreaker7860 Ай бұрын
i'm curious how they may scale this with additional external representations. How would one have to change the training approach for e.g. 3 sources? Could one simply co-optimise for all three losses + image-gen or would they interfere with each other? Can't wait to read the followup papers exploring such questions.
@m_ke
@m_ke Ай бұрын
It's not a long term approach, autoregressive generative vision language models are the future.
@LoFiLatentSpaceTools
@LoFiLatentSpaceTools Ай бұрын
lol. JEPA says otherwise 😂
@m_ke
@m_ke Ай бұрын
@ JEPA is not a generative method, I’m saying diffusion models will get replaced by autoregressive ones
@hjups
@hjups Ай бұрын
That probably depends on the application goal. In general, AR isn't a good fit for a modality that is best described by a continuous space with translation invariance. There's even some indication that diffusion / flow matching models may work better for NLP, since it's more parallel and can perform self-correction. If anything, I would say GNN-Diffusion-hybrid models are probably the future.
@m_ke
@m_ke Ай бұрын
@ look up HART, VAR, Infinity and Switti, all recent works that use AR to match or beat diffusion models while being faster to train and way faster at inference. With a decoder transformer you also get a much more natural integration of modalities and easy conditioning / in context learning
@m_ke
@m_ke Ай бұрын
@ other great benefit of AR VLMs for generation is that you get to benefit from all of the advances of multimodal LLMs and share feature representations instead of training a huge denoising model that’s hard to use for other tasks
@theshow3376
@theshow3376 Ай бұрын
Why did nobody think of this earlier!????
@Ali-wf9ef
@Ali-wf9ef Ай бұрын
You are absolutely amazing thanks for this explanation. I read the paper and understood nothing!
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thanks a lot for your appreciation!
@realbenjoyo
@realbenjoyo Ай бұрын
This was really great, never really understood query, key and values before.
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thank you!
@nwokebugoodness4819
@nwokebugoodness4819 Ай бұрын
Congrats! You're inspiring
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thank you
@gettingdatasciencedone
@gettingdatasciencedone Ай бұрын
Great explanation -- loving your videos.The time codes for specific topics is really useful.
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thank you!
@davide0965
@davide0965 Ай бұрын
The begin was clear, then very obscure👎
@muhammadbilalawais2976
@muhammadbilalawais2976 Ай бұрын
Awesome content!
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thank you!
@davide0965
@davide0965 Ай бұрын
Terrible explanation
@KhazeemasaleemKhazeemasale-o8s
@KhazeemasaleemKhazeemasale-o8s Ай бұрын
Hey khazeema saleem Al here
@GaryGan-US
@GaryGan-US Ай бұрын
very concise; what an amazing video.
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thank you!
@martinkunev9911
@martinkunev9911 Ай бұрын
I disagree that people are being judged for being bad at math more harshly than for other subjects. Compare "I'm not interested in history" or "I don't like reading" with "I'm bad at math".
@wapsyed
@wapsyed Ай бұрын
UMAP rocks! The only problem I see is the explainability of this high dimensionality reduction, which is easily done in PCA. In other words, you can get the best variables to explain the clustering, which is important when you are focusing on variable selection. What do you think?
@LinkhManu
@LinkhManu Ай бұрын
You’re the best 👏👏👏
@OlgaIvina
@OlgaIvina Ай бұрын
Thank you very much for this thorough, well-curated, and comprehensive review of MAMBA.
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thank you, for your appreciation! I just saw you on LinkedIn, let's stay connected!
@alexkubiesa9073
@alexkubiesa9073 Ай бұрын
How are exploration and exploitation abilities or forms of intelligence? To me they're more like competing actions, like going to the shops vs going to the cinema. I am still capable of both actions.
@luise.suelves8270
@luise.suelves8270 Ай бұрын
sooo well explain, brilliant!
@AICoffeeBreak
@AICoffeeBreak Ай бұрын
Thanks!
@Jupiter-Optimus-Maximus
@Jupiter-Optimus-Maximus 2 ай бұрын
Another great video, as usual! This little bean mutant of yours always puts a smile on my face ☺ Is it possible that it is actually an AI? For example, a transformer that converts language information into the facial expressions of the animated bean. That would be so cool 😎 I have a question: I am looking for training methods that are not based on backpropagation. Specifically, I want to avoid running backwards through the NNW again after the forward pass. Do you know of any algorithms like this? Already 2^10 * Thanks in advance 😄
@moeinhasani8718
@moeinhasani8718 2 ай бұрын
Thanks!
@AICoffeeBreak
@AICoffeeBreak 2 ай бұрын
Wow, thank You!