A Visual Guide to Mixture of Experts (MoE) in LLMs

  Рет қаралды 1,782

Maarten Grootendorst

Maarten Grootendorst

Күн бұрын

Пікірлер: 6
@Mohamed_Shokry
@Mohamed_Shokry Күн бұрын
Thanks for the video! I can tell you put a lot of work into it.
@marloncajamarca2793
@marloncajamarca2793 6 күн бұрын
Best explaination of MoEs I have came across till now! The high-quality explaination of key concepts, production and visuals are superb. Keep up with this amazing work and thanks for sharing this for free Marteen.
@ringpolitiet
@ringpolitiet 7 күн бұрын
Very high production value! Very useful, thanks.
@jacehua7334
@jacehua7334 7 күн бұрын
Always very useful thank you Maarten!
@Deshwal.mahesh
@Deshwal.mahesh 4 күн бұрын
Can we have a fine-tuning or building it from scratch video too 😶
@ekramhossain4601
@ekramhossain4601 7 күн бұрын
Can you also make this kind of video to explain transformer as well? Thanks :)
Visualizing transformers and attention | Talk for TNG Big Tech Day '24
57:45
ТЫ В ДЕТСТВЕ КОГДА ВЫПАЛ ЗУБ😂#shorts
00:59
BATEK_OFFICIAL
Рет қаралды 4,6 МЛН
Turn Off the Vacum And Sit Back and Laugh 🤣
00:34
SKITSFUL
Рет қаралды 6 МЛН
Why Does Diffusion Work Better than Auto-Regression?
20:18
Algorithmic Simplicity
Рет қаралды 379 М.
Which Quantization Method is Right for You? (GPTQ vs. GGUF vs. AWQ)
15:51
Maarten Grootendorst
Рет қаралды 21 М.
Llama 1-bit quantization - why NVIDIA should be scared
6:08
George Xian
Рет қаралды 25 М.
Transformers (how LLMs work) explained visually | DL5
27:14
3Blue1Brown
Рет қаралды 3,8 МЛН
Speculations on Test-Time Scaling (o1)
47:56
Sasha Rush 🤗
Рет қаралды 12 М.
What is Mixture of Experts?
7:58
IBM Technology
Рет қаралды 10 М.
Has Generative AI Already Peaked? - Computerphile
12:48
Computerphile
Рет қаралды 1 МЛН
Top Minds in AI Explain What’s Coming After GPT-4o | EP #130
25:30
Peter H. Diamandis
Рет қаралды 372 М.
ТЫ В ДЕТСТВЕ КОГДА ВЫПАЛ ЗУБ😂#shorts
00:59
BATEK_OFFICIAL
Рет қаралды 4,6 МЛН