Coding Llama 3 from scratch in PyTorch - Part 2

  Рет қаралды 3,016

Prince Canuma

Prince Canuma

Күн бұрын

Пікірлер: 17
@kishoretvk
@kishoretvk 5 ай бұрын
Thanks for committing to the open source and educating people on cutting edge knowledge.
@princecanuma
@princecanuma 5 ай бұрын
Most welcome, it’s my pleasure!
@yoanijosias
@yoanijosias 5 ай бұрын
Very good, can’t wait to see updates to it.
@princecanuma
@princecanuma 5 ай бұрын
You and me both!
@liyanan2004
@liyanan2004 4 ай бұрын
Could you please make a tutorial on vlm, and how it works. Like this series of videos, from scratch.
@princecanuma
@princecanuma 4 ай бұрын
That’s a great idea! 💡 Will do 👌🏽
@spkgyk
@spkgyk 5 ай бұрын
Why do you use 32 bit paged optimzier when the model is being fine-tuned with QLoRA? Surely QLoRA stores the weights in 8bit double quantized form, so using a 32 bit optimizer makes no difference, and the weight updates need to be converted back to 8 bit anyway? Please help me understand this
@princecanuma
@princecanuma 5 ай бұрын
Additionally, 8bit states are dequantized to 32bit for the update anyways. huggingface.co/docs/bitsandbytes/main/en/explanations/optimizers
@spkgyk
@spkgyk 5 ай бұрын
@@princecanuma Thank you for the quick response. With 8-bit optimizers, large models can be finetuned with 75% less GPU memory without losing any accuracy compared to training with standard 32-bit optimizers. The reduced memory requirements means 8-bit optimizers are 4x faster than a standard optimizer, and no hyperparameter tuning is required. Surely this means that using 32 bit just wastes compute power? Please correct me if I'm wrong, I'm really trying to understand the benefits. Is it because training with 32 bit means that despite converting to 8 bit for the weight update, the conversion leads to small accuracy gains?
@princecanuma
@princecanuma 5 ай бұрын
There are no accuracy gains only reduced GPU usage and potentially some extra speed. In terms of speed, I personally didn’t notice any changes. I tested it yesterday and besides reduced GPU usage I noticed that it would take just as long as the 32bit to complete training.
@sergey_a
@sergey_a 5 ай бұрын
Why are there only 3 likes, I put 4 on HF.)
@PaoloTshiyole
@PaoloTshiyole 5 ай бұрын
Your English is nice
@princecanuma
@princecanuma 5 ай бұрын
Thank you very much!
@wilfredomartel7781
@wilfredomartel7781 5 ай бұрын
😊
@wilfredomartel7781
@wilfredomartel7781 5 ай бұрын
😊🎉
@00osmboy
@00osmboy 5 ай бұрын
cool
@princecanuma
@princecanuma 5 ай бұрын
Awesome, I’m happy you liked it :)
Coding Llama 3 from scratch in PyTorch - Part 1
23:59
Prince Canuma
Рет қаралды 4,1 М.
Get started with Gemma Google's NEW open-source LLM model
40:19
Prince Canuma
Рет қаралды 3,2 М.
The IMPOSSIBLE Puzzle..
00:55
Stokes Twins
Рет қаралды 103 МЛН
Who's spending her birthday with Harley Quinn on halloween?#Harley Quinn #joker
01:00
Harley Quinn with the Joker
Рет қаралды 23 МЛН
Make Your RAG Agents Actually Work! (No More Hallucinations)
35:01
Leon van Zyl
Рет қаралды 12 М.
Generative AI in a Nutshell - how to survive and thrive in the age of AI
17:57
Let's Build Llama 3 From Scratch, in Code, Spelled Out
1:00:05
Tunadorable
Рет қаралды 6 М.
Using Ollama To Build a FULLY LOCAL "ChatGPT Clone"
11:17
Matthew Berman
Рет қаралды 257 М.
LoRA & QLoRA Fine-tuning Explained In-Depth
14:39
Entry Point AI
Рет қаралды 47 М.
Patient Can’t Lie For His Own Greater Good | House M.D..
9:38
House M.D.
Рет қаралды 603 М.
Creating an AI Agent with LangGraph Llama 3 & Groq
35:29
Sam Witteveen
Рет қаралды 46 М.
Using Llama Coder As Your AI Assistant
9:18
Matt Williams
Рет қаралды 73 М.
Coding Llama 2 from scratch in PyTorch - Part 3
50:14
Prince Canuma
Рет қаралды 1,3 М.
The IMPOSSIBLE Puzzle..
00:55
Stokes Twins
Рет қаралды 103 МЛН