Llama3: Comparing 8B vs 70B Parameter Models - Which One is Right for You?

  Рет қаралды 5,134

Compute Grid

Compute Grid

Күн бұрын

Пікірлер: 12
@modoulaminceesay9211
@modoulaminceesay9211 2 ай бұрын
i am just grateful to have this running free on my computer.
@heikg
@heikg 8 күн бұрын
Please make sure to clear the screen on the consoles between the different questions if you're doing anything similar in the future. It's not a lot of effort but makes the screen look so much cleaner.
@autoboto
@autoboto Ай бұрын
Since I cant load the 70B its nice to see what I'm missing out on. The side by side compare is great. If you do more tests try to use the same Q&A
@hxxzxtf
@hxxzxtf 4 ай бұрын
🎯 Key points for quick navigation: 00:00 *📹 The video compares the 8B and 70B parameter models of Llama3, focusing on their differences and potential use cases.* 00:41 *💡 Parameters in a neural network refer to the number of learnable weights or connections within the architecture, controlling the model's behavior.* 01:10 *⚖️ A model with more parameters will always perform better compared to one with less parameters, given the same training data.* 02:07 *💻 To run the 70B model, a stronger computer with a good GPU and CPU is required, but comput grid provides access to such resources.* 06:38 *📊 The 70B model has a larger manifest size (39 GB) compared to the 8B model (4.7 GB).* 09:06 *📝 The 70B model can generate more creative and engaging responses due to its increased parameter count.* 11:21 *👥 The 70B model can improve a bad resume by rewriting it in a more professional tone and highlighting relevant skills.* 13:00 *💼 The 70B model can also provide better reasoning through family relationships, understanding complex concepts and providing accurate answers.* Made with HARPA AI
@Aliyah-r7v
@Aliyah-r7v 2 ай бұрын
Hahah your just add this link into Ai and made this😂
@BozesanVlad
@BozesanVlad 29 күн бұрын
Did you showed your ssh password? 5:00
@tsclly2377
@tsclly2377 4 ай бұрын
Retrained Llama3 at 8B to a more specified SML is the way to go. Garbage in, garbage out. Also Llama is tokenized and that will equate to monitized in the future.. They are in this 'for the money'.. Mamba and Jamba maybe a better engine for those that want to remain fully independent.
@McMalhon
@McMalhon Ай бұрын
How did you go about doing that? How much work was the retraining?
@Sasisanju-r5n
@Sasisanju-r5n 18 күн бұрын
What do u mean by tokens, 8b or 70b make a video on explaining them clearly
@faded_taco1816
@faded_taco1816 3 ай бұрын
omg I hate how she is smacking her lips so bad and even worse when you speed up the video. Like what are u doing??? Get a audio compressor or don't put the whole mic down your mouth.
@thelonercoder5816
@thelonercoder5816 Ай бұрын
man this comment made me lol so hard.
Using Clusters to Boost LLMs 🚀
13:00
Alex Ziskind
Рет қаралды 71 М.
Trick-or-Treating in a Rush. Part 2
00:37
Daniel LaBelle
Рет қаралды 46 МЛН
Человек паук уже не тот
00:32
Miracle
Рет қаралды 4,4 МЛН
小路飞还不知道他把路飞给擦没有了 #路飞#海贼王
00:32
路飞与唐舞桐
Рет қаралды 82 МЛН
assignment 6
0:43
Ulug’bek Sharipov
Рет қаралды 3
Qwen Just Casually Started the Local AI Revolution
16:05
Cole Medin
Рет қаралды 68 М.
Why Does Diffusion Work Better than Auto-Regression?
20:18
Algorithmic Simplicity
Рет қаралды 375 М.
host ALL your AI locally
24:20
NetworkChuck
Рет қаралды 1,3 МЛН
How To Run Llama 3.1: 8B, 70B, 405B Models Locally (Guide)
6:21
School of Machine Learning
Рет қаралды 10 М.
LlaMA 3 Architecture & Paper explained step by step
13:50
Aaditya Ura
Рет қаралды 4 М.
Optimize Your AI Models
11:43
Matt Williams
Рет қаралды 14 М.
EASIEST Way to Fine-Tune a LLM and Use It With Ollama
5:18
warpdotdev
Рет қаралды 128 М.
Trick-or-Treating in a Rush. Part 2
00:37
Daniel LaBelle
Рет қаралды 46 МЛН