LoRA Land: How We Trained 25 Fine-Tuned Mistral-7b Models that Outperform GPT-4

  Рет қаралды 5,907

Predibase

Predibase

Күн бұрын

Пікірлер: 8
@JulianHarris
@JulianHarris 6 ай бұрын
Have you guys looked at the next generation of quantisation: eg ternary/1.58 bit quantisation? It’s a different technique to conventional quantisation because you have matrices that only have 0, 1, -1, and you eliminate matrix multiplication almost entirely. The intuition is that the combination may not bring quite as many benefits, but it might be interesting to see how it performs in CPU architectures for instance.
@ofir952
@ofir952 6 ай бұрын
Thanks! How did you manage to remove the surrounding text of the LLM response?
@pieromolino_pb
@pieromolino_pb 6 ай бұрын
It's a side effect of fine-tuning on output that contains only the JSON without tany other text
@ofir952
@ofir952 6 ай бұрын
So, we cannot achieve this without fine-tuning? Llama2 keeps on adding it all the time 🥲@@pieromolino_pb
@jeffg4686
@jeffg4686 6 ай бұрын
Nice !
@tankieslayer6927
@tankieslayer6927 6 ай бұрын
FINE-TUNED MODEL RESPONSE Named Entity Recognition (CoNLL++) {"person": ["Such"], "organization": ["Yorkshire"], "location": [], "miscellaneous": []} Yeah, I am not impressed with the result of this fine-tuning.
@pieromolino_pb
@pieromolino_pb 6 ай бұрын
The input text is: By the close Yorkshire had turned that into a 37-run advantage but off-spinner Such had scuttled their hopes , taking four for 24 in 48 balls and leaving them hanging on 119 for five and praying for rain. Yorkshire in this case is a sports team, so organization is correct, and Such is a a player, so both model's predictions are correct indeed. I'd suggest to try to understand better what is going on next time.
@The_Real_Goodboy_Link
@The_Real_Goodboy_Link 5 күн бұрын
Found the real solution, @tankieslayer6927, click on your icon on the top-right screen here, then settings, advanced settings, delete channel. Then go over to Google and do similarly for your account there. Problem solved!
5 Reasons Why Adapters are the Future of Fine-tuning LLMs
1:01:18
Predibase
Рет қаралды 1,3 М.
Practical Fine-Tuning of LLMs
1:02:18
AI Makerspace
Рет қаралды 2 М.
WORLD BEST MAGIC SECRETS
00:50
MasomkaMagic
Рет қаралды 32 МЛН
when you have plan B 😂
00:11
Andrey Grechka
Рет қаралды 58 МЛН
هذه الحلوى قد تقتلني 😱🍬
00:22
Cool Tool SHORTS Arabic
Рет қаралды 103 МЛН
Virtual Workshop: Fine-tune Your Own LLMs that Rival GPT-4
1:20:06
Fine-Tuning Mistral-7B with LoRA (Low Rank Adaptation)
1:01:16
AI Makerspace
Рет қаралды 4,8 М.
The Most Important Algorithm in Machine Learning
40:08
Artem Kirsanov
Рет қаралды 420 М.
Mixtral8-7B: Overview and Fine-Tuning
34:33
AI Makerspace
Рет қаралды 4 М.
LoRA & QLoRA Fine-tuning Explained In-Depth
14:39
Entry Point AI
Рет қаралды 39 М.
Prompt Engineering Techniques (extended version)
52:38
Jonathan Yarkoni
Рет қаралды 13 М.
Training and deploying open-source large language models
39:53
Niels Rogge
Рет қаралды 16 М.
WORLD BEST MAGIC SECRETS
00:50
MasomkaMagic
Рет қаралды 32 МЛН