One of the BEST tutorials/explanations i have ever seen, thank you really much for that ❤
@AICoffeeBreak20 сағат бұрын
Much appreciated!
@pranavb9768Күн бұрын
you’re cute
@ahsan4life20204 күн бұрын
This is a very valuable video for me. Thank you so much
@AICoffeeBreak4 күн бұрын
For a second I thought you're talking about the three minutes of uncomfortable silence . On the comment notification I don't see in which video the comment was made. 😅 So, I'm glad it is about this one.
@erongjoni34645 күн бұрын
Lost it on "Watch the full video in the link below"
@AICoffeeBreak5 күн бұрын
And that is just the finest 3 minute selection of the full 50 minutes awkward silence video. 😏
@foreignconta6 күн бұрын
What exactly happened here?🤔
@AICoffeeBreak6 күн бұрын
My editor had fun cutting out moments of awkward silence from my next (1h+) video.
@bmebri17 күн бұрын
I love it; most unawkward facet of you yet😊
@AICoffeeBreak6 күн бұрын
@DaNa-px7ol7 күн бұрын
Wow I didn’t know I can enjoy “awkward” silence 😮
@AICoffeeBreak7 күн бұрын
😂
@Thomas-gk427 күн бұрын
Waiting for the phd result...?
@AICoffeeBreak7 күн бұрын
Haha, fortunately, that one is clear. 😅🎉🤗
@outliier7 күн бұрын
Some would upload this as ASMR
@AICoffeeBreak7 күн бұрын
🎙️
@WhatsAI7 күн бұрын
Haha a classic when recording!
@DerPylz7 күн бұрын
I like the part where she doesn't say anything.
@AICoffeeBreak7 күн бұрын
😅
@azmathmoosa432415 күн бұрын
good concise explanation.
@sohambit939319 күн бұрын
Damn you like Neffex ❤ Neffex is like 10 % of my life.
@enicay756222 күн бұрын
Thank you Miss Coffee Bean !
@AICoffeeBreak21 күн бұрын
@kristoferkrus22 күн бұрын
Cool! The simple but effective ideas are the most interesting. Have you checked out Nvidia's normalized transformer (nGPT)? It seems to be one of those cases. The differential transformer (arxiv: 2410.05258) also seems like it could be interesting.
@davide096525 күн бұрын
Terrible
@DerPylz25 күн бұрын
If you don't like her videos, why do you keep coming back to them just to comment that you didn't like it? Just watch something else.
@AICoffeeBreak26 күн бұрын
03:23 It looks like the MLP takes the representation of JUST the 8th layer, and not also of the previous ones. I just found out from a quick exchange via email with the authors.
@bethany-rp2tq26 күн бұрын
CONGRATULATIONS Letitia !!!! Is there a pdf version where I can read your PhD thesis? Always amazed learning new knowledge from you :)
@AICoffeeBreak26 күн бұрын
Yes, there is! archiv.ub.uni-heidelberg.de/volltextserver/35753/ Thanks for your interest!
@davidrichards130229 күн бұрын
Unison's unique features offer promising possibilities for enhancing the FunSearch operational model: Content-addressed code: Unison's hash-based identification of code could streamline FunSearch's program generation and evaluation process. Each generated program would have a unique hash, simplifying tracking and caching of results. Immutable codebase: Unison's immutable data structure approach to codebases could enhance FunSearch's evolutionary process. It would allow for efficient storage and retrieval of program versions without conflicts. No builds: Unison's ability to parse and typecheck definitions once, storing results in a cache, could significantly speed up FunSearch's evaluation phase. This would reduce overhead in assessing generated programs. Easy distributed computing: Unison's content-addressed nature facilitates distributed computation, which could enhance FunSearch's parallelization capabilities. This could allow for more efficient scaling of the search process across multiple machines. First-class documentation: Unison's approach to documentation as executable code could improve FunSearch's ability to generate and evaluate self-documenting programs, potentially leading to more interpretable solutions. Strong typing with inference: Unison's type inference could help FunSearch generate type-safe programs more efficiently, potentially reducing the number of invalid programs generated. These features could potentially make Unison an excellent choice for implementing core components of the FunSearch system, particularly in program generation, evaluation, and distributed processing.
@duzx4541Ай бұрын
Hmm, the only thing I dont really understand is why we have to use SIN and COS instead of only one of them D:
@AICoffeeBreakАй бұрын
They are the same with just a phase shift. Rotary embeddings take the idea of phase shifting and turn it up.
@katoreaАй бұрын
Loved your explanation! thank you very much!! :D
@AICoffeeBreakАй бұрын
@solsospecialАй бұрын
Okay, the approach is “guessing”. Got it.
@KnowledgeSynthesizerАй бұрын
can I ask what app/software tools you use for recording, editing audio and video?
@AICoffeeBreak26 күн бұрын
Hi, it is good old powerpoint for all visualisations. Ms. Coffee Bean comes during editing in Adobe Premiere. :)
@floriankowarsch8682Ай бұрын
It would be definitely interesting to train end to end with contrastive lost. Because right now it is a more like a distillation task that is limited be the teacher model's capability & domain
@dylancopeАй бұрын
"wait a few years"... 9 months later with Genie 2 😂
@AICoffeeBreak26 күн бұрын
Years in ML pass very quickly. 🤣🤣🤣
@rogerthat7190Ай бұрын
This was very helpful, thank you!
@AICoffeeBreakАй бұрын
Thanks for the kind words!
@hjupsАй бұрын
It's a clever idea, but the paper has many methodological issues which may amount to more hype than substance. Then results are good, but the major gains come from stacking contributions. And of course you have to worry about the extra training cost (DINO is very expensive) and generalization, etc. The peak accuracy shift is also worrisome, indicating that removing the driving force would likely result in an expensive shift / performance degradation when finetuning.
@fast_harmonic_psychedelicАй бұрын
thats such a good idea. very simple .. i would not have thought to do this. I have added CLIP loss terms on the last layer but not dinoon the 8th layer lol.
@AICoffeeBreak26 күн бұрын
Yes, indeed. 😅 The last layers still need to focus on reconstruction, or rather predicting the noise. Injecting in the first layers helps the network "conceptualise" what the denoising needs to become. The following layers need to focus on predicting that noise.
@EkShunyaАй бұрын
i missed you, glad to see you back with your great explainers
@AICoffeeBreakАй бұрын
@circuitbreaker7860Ай бұрын
i'm curious how they may scale this with additional external representations. How would one have to change the training approach for e.g. 3 sources? Could one simply co-optimise for all three losses + image-gen or would they interfere with each other? Can't wait to read the followup papers exploring such questions.
@m_keАй бұрын
It's not a long term approach, autoregressive generative vision language models are the future.
@LoFiLatentSpaceToolsАй бұрын
lol. JEPA says otherwise 😂
@m_keАй бұрын
@ JEPA is not a generative method, I’m saying diffusion models will get replaced by autoregressive ones
@hjupsАй бұрын
That probably depends on the application goal. In general, AR isn't a good fit for a modality that is best described by a continuous space with translation invariance. There's even some indication that diffusion / flow matching models may work better for NLP, since it's more parallel and can perform self-correction. If anything, I would say GNN-Diffusion-hybrid models are probably the future.
@m_keАй бұрын
@ look up HART, VAR, Infinity and Switti, all recent works that use AR to match or beat diffusion models while being faster to train and way faster at inference. With a decoder transformer you also get a much more natural integration of modalities and easy conditioning / in context learning
@m_keАй бұрын
@ other great benefit of AR VLMs for generation is that you get to benefit from all of the advances of multimodal LLMs and share feature representations instead of training a huge denoising model that’s hard to use for other tasks
@theshow3376Ай бұрын
Why did nobody think of this earlier!????
@Ali-wf9efАй бұрын
You are absolutely amazing thanks for this explanation. I read the paper and understood nothing!
@AICoffeeBreakАй бұрын
Thanks a lot for your appreciation!
@realbenjoyoАй бұрын
This was really great, never really understood query, key and values before.
@AICoffeeBreakАй бұрын
Thank you!
@nwokebugoodness4819Ай бұрын
Congrats! You're inspiring
@AICoffeeBreakАй бұрын
Thank you
@gettingdatasciencedoneАй бұрын
Great explanation -- loving your videos.The time codes for specific topics is really useful.
@AICoffeeBreakАй бұрын
Thank you!
@davide0965Ай бұрын
The begin was clear, then very obscure👎
@muhammadbilalawais2976Ай бұрын
Awesome content!
@AICoffeeBreakАй бұрын
Thank you!
@davide0965Ай бұрын
Terrible explanation
@KhazeemasaleemKhazeemasale-o8sАй бұрын
Hey khazeema saleem Al here
@GaryGan-USАй бұрын
very concise; what an amazing video.
@AICoffeeBreakАй бұрын
Thank you!
@martinkunev9911Ай бұрын
I disagree that people are being judged for being bad at math more harshly than for other subjects. Compare "I'm not interested in history" or "I don't like reading" with "I'm bad at math".
@wapsyedАй бұрын
UMAP rocks! The only problem I see is the explainability of this high dimensionality reduction, which is easily done in PCA. In other words, you can get the best variables to explain the clustering, which is important when you are focusing on variable selection. What do you think?
@LinkhManuАй бұрын
You’re the best 👏👏👏
@OlgaIvinaАй бұрын
Thank you very much for this thorough, well-curated, and comprehensive review of MAMBA.
@AICoffeeBreakАй бұрын
Thank you, for your appreciation! I just saw you on LinkedIn, let's stay connected!
@alexkubiesa9073Ай бұрын
How are exploration and exploitation abilities or forms of intelligence? To me they're more like competing actions, like going to the shops vs going to the cinema. I am still capable of both actions.
@luise.suelves8270Ай бұрын
sooo well explain, brilliant!
@AICoffeeBreakАй бұрын
Thanks!
@Jupiter-Optimus-Maximus2 ай бұрын
Another great video, as usual! This little bean mutant of yours always puts a smile on my face ☺ Is it possible that it is actually an AI? For example, a transformer that converts language information into the facial expressions of the animated bean. That would be so cool 😎 I have a question: I am looking for training methods that are not based on backpropagation. Specifically, I want to avoid running backwards through the NNW again after the forward pass. Do you know of any algorithms like this? Already 2^10 * Thanks in advance 😄