Zephyr 7b beta: paper deep dive, code, & RAG

  Рет қаралды 9,079

Sophia Yang

Sophia Yang

7 ай бұрын

Is Zephyr 7b beta the best fine-tuned smaller open-source LLM? It is fine-tuned on the best small open-source foundation model Mistral 7B, outperforms all small models, and even outperforms LLama2-Chat-70B and ChatGPT on multiple benchmarks.
What are the difference between Zephyr 7b alpha and Zephyr 7b beta?
From the author: "The main differences are (a) better dataset filtering of UltraChat, namely fixing grammatical errors and some unwanted responses and (b) training for longer, ie 3 DPO epochs vs 1"
Model: huggingface.co/HuggingFaceH4/...
Technical report: arxiv.org/abs/2310.16944
Colab: colab.research.google.com/git...
LlamaIndex Tweet (colab link in the tweet): / 1718054631413363196
🔔 SUBSCRIBE to my channel: kzbin.info...
⭐ Stay in touch ⭐
📚 DS/ML Book Club: dsbookclub.github.io/
▶ KZbin: / sophiayangds
✍️ Medium: / sophiamyang
🐦 Twitter: / sophiamyang
🤝 Linkedin: / sophiamyang
💚 #ai

Пікірлер: 29
7 ай бұрын
Thanks a lot. Using Zephyr with ollama and it's pretty cool and fast.
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Awesome! Ollama is really nice.
@jstevh
@jstevh 7 ай бұрын
Thanks for that analysis. Like a good overview of a solid paper.
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Thank you!
@taeyangoh7305
@taeyangoh7305 7 ай бұрын
great review, Sophia! I want to give it a try for care robot use case!!
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Thanks! Sounds like a really cool use case. Good luck!
@Hash_Boy
@Hash_Boy 7 ай бұрын
many many thanks
@prestonmccauley43
@prestonmccauley43 7 ай бұрын
Thanks for the information I look forward to giving this model a try though it does seem like we keep producing models every week. They really need to be run for many months to determine how well they work.
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Thanks! Yeah too many models coming out. One of the takeaways is Mistral 7B is a great foundation model to fine-tune on.
@talhayousuf4599
@talhayousuf4599 7 ай бұрын
Your channel is really informative, good thing is that you remain updated with new things coming. Are you a researcher?, do you have any blog? / related to llm research?
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Thanks 🙏 I have a casual blog with book reading notes and other random things: sophiamyang.medium.com/
@AkarshanBiswas
@AkarshanBiswas 7 ай бұрын
I am using this model as my default model currently. It is not perfect but I prompted to according to my liking. I wish the prompt template was a bit simpler. 😅
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Nice!
@DistortedV12
@DistortedV12 7 ай бұрын
It's okay..I tried it with different prompts. Can't wait till they scale it.
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
It's a good 7B model.
@jsalsman
@jsalsman 7 ай бұрын
Do they mention the cost of the GPT-4 calls during the training steps?
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
They did not mention cost in the paper.
@jsalsman
@jsalsman 7 ай бұрын
@@SophiaYangDS Any idea roughly how many API calls they must have used?
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Must be a lot. Just to process and create this UltraFeedback dataset (huggingface.co/datasets/HuggingFaceH4/ultrafeedback_binarized?row=0), it's around 74k x 5 API calls (4 models responses + 1 GPT4 ratings). @@jsalsman
@TheReferrer72
@TheReferrer72 7 ай бұрын
Is this the Model that does not reason or code well? How can this model be useful?
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Still better than llama2 70B on coding and math
@mshonle
@mshonle 7 ай бұрын
Does distilling help to remove overfitting?
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
I'm not sure. You can overfit in distillation.
@mshonle
@mshonle 7 ай бұрын
I wonder how much this cost them to make, in terms of compute? (Or, maybe more relevant: how much would it cost someone else to do the same?)
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
​@@mshonle I think they might be spending most of the compute/API calls in generating the datasets. The actual SFT and DPO might not be that crazy.
@mshonle
@mshonle 7 ай бұрын
Hmm, and those datasets are available?
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
@@mshonle Yes they are public on Hugging Face
@serkhetreo2489
@serkhetreo2489 7 ай бұрын
Sometimes, u discover new channels
@SophiaYangDS
@SophiaYangDS 7 ай бұрын
Thanks for discovering my channel :)
Advanced RAG 01: Small-to-Big Retrieval with LlamaIndex
9:17
Sophia Yang
Рет қаралды 11 М.
GraphRAG: LLM-Derived Knowledge Graphs for RAG
15:40
Alex Chao
Рет қаралды 73 М.
ROCK PAPER SCISSOR! (55 MLN SUBS!) feat @PANDAGIRLOFFICIAL #shorts
00:31
Zephyr 7b Alpha 🎃 As Good As They Say?
13:05
Matthew Berman
Рет қаралды 23 М.
Fine Tuning Mistral v3.0 With Custom Data
6:58
Mosleh Mahamud
Рет қаралды 2,6 М.
Gemini has a Diversity Problem
17:36
Yannic Kilcher
Рет қаралды 52 М.
RAG But Better: Rerankers with Cohere AI
23:43
James Briggs
Рет қаралды 52 М.
Mistral 7B Hype is Totally Justified + AutoGen by Microsoft
8:33
AI Revolution
Рет қаралды 28 М.
Fine-tuning LLMs with PEFT and LoRA
15:35
Sam Witteveen
Рет қаралды 113 М.
How I'd Learn AI (If I Had to Start Over)
15:04
Thu Vu data analytics
Рет қаралды 705 М.
Advanced RAG 01 - Self Querying Retrieval
12:02
Sam Witteveen
Рет қаралды 39 М.