NEW WizardCoder-34B - THE BEST CODING LLM

  Рет қаралды 24,745

Prompt Engineering

Prompt Engineering

Күн бұрын

Well, less than 48 hours since the release of Codellama, there is a new model WizardCoder-Python-34B that beats all previous models on the HumanEval benchmark. We will also cover the Phind model that was able to be the second best model.
#wizardcoder #codellama #llama2
▬▬▬▬▬▬▬▬▬▬▬▬▬▬ CONNECT ▬▬▬▬▬▬▬▬▬▬▬
☕ Buy me a Coffee: ko-fi.com/promptengineering
|🔴 Support my work on Patreon: Patreon.com/PromptEngineering
🦾 Discord: / discord
▶️️ Subscribe: www.youtube.com/@engineerprom...
📧 Business Contact: engineerprompt@gmail.com
💼Consulting: calendly.com/engineerprompt/c...
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
LINKS:
WizardCoder: github.com/nlpxucan/WizardLM/...
Phind Model: www.phind.com/blog/code-llama...
Try it yourself: 47.103.63.15:50085/
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
All Interesting Videos:
Everything LangChain: • LangChain
Everything LLM: • Large Language Models
Everything Midjourney: • MidJourney Tutorials
AI Image Generation: • AI Image Generation Tu...

Пікірлер: 51
@engineerprompt
@engineerprompt 9 ай бұрын
Want to connect? 💼Consulting: calendly.com/engineerprompt/consulting-call 🦾 Discord: discord.com/invite/t4eYQRUcXB ☕ Buy me a Coffee: ko-fi.com/promptengineering |🔴 Join Patreon: Patreon.com/PromptEngineering
@OwnOpinions
@OwnOpinions 10 ай бұрын
"Training Llama2 with Function Calling Capabilities"... I am daily checking your channel if this video gets uploaded. Thanks for all the great content.
@chaithanyachaganti4305
@chaithanyachaganti4305 10 ай бұрын
Been following your channel lately, very informative and insightful analysis give by you on llms. Keep up good work
@vincentjean6756
@vincentjean6756 10 ай бұрын
Yes, amazing. 🎉
@vishnuitsrocking
@vishnuitsrocking 10 ай бұрын
Super excited to see the progress in open-source model! Soon everyone will be a programmer!
@trpultz
@trpultz 10 ай бұрын
Thanks for the great video! Really appreciate the thoroughness (especially mentioning the concerns surrounding possibility of overfitting to the HumanEval test set) and you actually putting the models to the test yourself! Looking forward to seeing where these models are at by the end of the year and looking forward to your future videos on them. 😀
@williammixson2541
@williammixson2541 9 ай бұрын
Outstanding content. Instant sub!
@engineerprompt
@engineerprompt 9 ай бұрын
Thank you 🙏
@sveindanielsolvenus
@sveindanielsolvenus 10 ай бұрын
Awesome stuff! Where is the paper with the updated GPT-4 humaneval scores? I can't find that.
@Yewbzee
@Yewbzee 9 ай бұрын
Very good point regarding HumanEval overfitting
@paraconscious790
@paraconscious790 9 ай бұрын
great video as always, thanks! can you please make a video for installing this model locally on Mac? thanks! the Vicuna video may not work as the text-generation-webui is changed since then.
@Nick_With_A_Stick
@Nick_With_A_Stick 10 ай бұрын
I agree with that. I don’t have proof but I have a feeling a lot of these models are writing good code due to good data in the original training, but they don’t understand the code they are writing. A more complex benchmark would help diagnose that, and I think benchmarks should be remade and harder specifically since we are starting to get close to 100% for some of them. God just seeing how good the 34b coder does makes me wonder how good the unreleased regular llama 34 could’ve done if trained with orca. Such a shame.
@NakedSageAstrology
@NakedSageAstrology 10 ай бұрын
Wonderful video sir thank you, could you please do an instructional video tutorial on how to install this on a Windows PC?
@Artorias920
@Artorias920 10 ай бұрын
amazing updates! End of year prediction: GPT-4 level open source models running on blockchain-esque publicly distributed systems
@jonsantos6056
@jonsantos6056 10 ай бұрын
what would that mean for us - free?
@Artorias920
@Artorias920 10 ай бұрын
@@jonsantos6056 probably something close to it. Something akin to renting your GPU to the network and earning credits that can be redeemed for inference
@AlexanderBukh
@AlexanderBukh 10 ай бұрын
No need for blockchain, too much overhead and only needed when everyone wants to steal or manipulate data. As for the distributed inference, they already do that, see Petals. Actually, llama.cpp also has means for distributed computational (MPI)
@STDFme
@STDFme 9 ай бұрын
I am looking for a llm supporting C/C++, is it suited, too?
@bmoore813
@bmoore813 10 ай бұрын
Do you have any example of how I could run the WizardCoder-Python-34B-V1.0 on my a100? I ask cause that site is so slow and its takes three minutes to get a reply
@3lbios
@3lbios 10 ай бұрын
is the non-python version going to be quantized?
@theresalwaysanotherway3996
@theresalwaysanotherway3996 10 ай бұрын
interestingly enough, people are taking the 34B coding model and fine-tuning it for other purposes like RP or tool usage now (airoboros-34B and samantha-34B being the first ones). It may be worth testing out the 34Bs ability to solve non coding logic problems, to see if the coding knowledge generalises to better overall logic.
@VR_Wizard
@VR_Wizard 10 ай бұрын
Very interesting. I woulf love to see the code interpreter compared to the open source .odels and to GPT-4 base model. Havent seen these comparisonst. I havent used code unterpreter yet so i am not sure it is usefull for creating code but I thought thats one of its strenghts so should be interesting to see how much better code interpreter is.
@engineerprompt
@engineerprompt 10 ай бұрын
That’s actually a good point. Let me do a comparison
@sunfisoft
@sunfisoft 9 ай бұрын
does it recognize my project files and respond accordingly and is it necessary to have gpu i dont have gpu
@SlyNine
@SlyNine 6 ай бұрын
33 billion fits on my 4090 34 normally doesn't launch.
@user-fm5yy2sx2p
@user-fm5yy2sx2p 10 ай бұрын
What is web-ui? is text-generation-webui?
@lsp0
@lsp0 9 ай бұрын
Why are none of these papers touching the HumanEval+ benchmark?
@temp911Luke
@temp911Luke 10 ай бұрын
Has anyone tried this in koboldCPP or gpt4all using CPU only in terms of speed ?
@erikjohnson9112
@erikjohnson9112 10 ай бұрын
5:53 Isn't this a fail because it uses .sort()? That would be an advanced function/methods. Only direct looping and conditionals are allowed.
@engineerprompt
@engineerprompt 10 ай бұрын
You are right, that’s a great catch. Should have caught that.
@chirwatra
@chirwatra 7 ай бұрын
Do you have a video about how to build or deploy WizardCoder-34B app?
@engineerprompt
@engineerprompt 7 ай бұрын
I don't but you could use streamlit or check out databutton.com. Its a good platform for quickly deploying streamlit apps
@BrandosLounge
@BrandosLounge 10 ай бұрын
anywhere we can try this model online? Don't hvae the resources to run locally
@engineerprompt
@engineerprompt 10 ай бұрын
Yes, there is a link in the video description. Responses are slow though
@jonsantos6056
@jonsantos6056 10 ай бұрын
can i dl this onto local machine?
@vincentjean6756
@vincentjean6756 10 ай бұрын
Yes.
@bradgalaxy8419
@bradgalaxy8419 10 ай бұрын
phind, since its a search engine its likely pronounced "find"
@Jorsten
@Jorsten 10 ай бұрын
My chatGPT 3.5 solved both of them at first try... We need something harder.
@rickhoro
@rickhoro 10 ай бұрын
Does anyone know if this would be reasonable to run on a Linux machine with the following configuration: Intel® Core™ i7-7820X CPU @ 3.60GHz × 16 NVIDIA Corporation GP106 [GeForce GTX 1060 3GB] 64 GB RAM Thanks. I'm new to the whole Python and AI world. Help would be very much apprec!
@engineerprompt
@engineerprompt 10 ай бұрын
I suspect, you will not be able to run it. To just run a 4bit version of 34B model, you need at least 16GB RAM. see here discord.com/channels/1104593592229572640/1134401808425549824/1141459504421146777
@rickhoro
@rickhoro 10 ай бұрын
@@engineerprompt Thanks for your reply. Do you mean 16GB of main memory RAM or GPU RAM? Because I do have 64GB of main memory RAM.
@pocketsfullofdynamite
@pocketsfullofdynamite 10 ай бұрын
@@rickhoro 16GB installed RAM is good to start depending on how much dependencies are required to run on your local machine. Also check how much VRAM is available. These models take huge amount of VRAM considering your display monitor also uses VRAM.
@rickhoro
@rickhoro 10 ай бұрын
@@pocketsfullofdynamite Well, as I said, I have 64GB of DRAM (main memory) but only 3GB of GPU/VRAM. What do you think?
@morsmagne
@morsmagne 10 ай бұрын
I’m skeptical of all those assessments: when it comes to coding what you need is a very large token limit - like Claude 2 with 100,000 tokens. This is because you want the AI to understand all the code that’s written and remember it for the rest of the discussion. It’s no good having a clever AI that’s got the memory of someone living with Alzheimer’s! Because that’s exactly how it feels like trying to do any kind of code work with GPT-4 etc.
@hemeleh8683
@hemeleh8683 10 ай бұрын
Exactly! Which model is your preference? Gpt 4 32k is getting quite expensive on my end, would be happy to hear alternatives😅
@temp911Luke
@temp911Luke 10 ай бұрын
I spoke to someone from WIzardCoder and they say the token limit is same as in CodeLLama so the token window size might actually be potentially 100k.
@morsmagne
@morsmagne 10 ай бұрын
@@hemeleh8683 Personally, I use GPT-Engineer using GPT-3.5-turbo (only because I don't have access to GPT-4 API) then I use Claude 2 (free version) to review the code and make suggestions. However, I'm lucky enough to live in the UK where Claude 2 is free. Therefore, you might need a VPN to access it. You're welecome!
@morsmagne
@morsmagne 10 ай бұрын
And also, I start a new Claude 2 chat for every change I intend to make because even 100,000 tokens doesn't go very far.
SmartGPT: Make ChatGPT Smarter
12:57
Prompt Engineering
Рет қаралды 10 М.
The End Of Programming
24:55
Matthew Berman
Рет қаралды 137 М.
Please be kind🙏
00:34
ISSEI / いっせい
Рет қаралды 160 МЛН
Мы никогда не были так напуганы!
00:15
Аришнев
Рет қаралды 1,7 МЛН
Must-have gadget for every toilet! 🤩 #gadget
00:27
GiGaZoom
Рет қаралды 11 МЛН
All You Need To Know About Running LLMs Locally
10:30
bycloud
Рет қаралды 116 М.
Should You Use Open Source Large Language Models?
6:40
IBM Technology
Рет қаралды 341 М.
I Tried Every AI Coding Assistant
24:50
Conner Ardman
Рет қаралды 713 М.
Falcon 180b 🦅 The Largest Open-Source Model Has Landed!!
12:50
Matthew Berman
Рет қаралды 28 М.
Using Llama Coder As Your AI Assistant
9:18
Matt Williams
Рет қаралды 64 М.
Fine-tuning Large Language Models (LLMs) | w/ Example Code
28:18
Shaw Talebi
Рет қаралды 256 М.
AI coding assistants just leveled up, again…
4:51
Fireship
Рет қаралды 1,1 МЛН
RAG from the Ground Up with Python and Ollama
15:32
Decoder
Рет қаралды 24 М.
META's New Code LLaMA 70b BEATS GPT4 At Coding (Open Source)
9:25
Matthew Berman
Рет қаралды 78 М.
iPhone 12 socket cleaning #fixit
0:30
Tamar DB (mt)
Рет қаралды 50 МЛН
Hisense Official Flagship Store Hisense is the champion What is going on?
0:11
Special Effects Funny 44
Рет қаралды 1,4 МЛН
Iphone or nokia
0:15
rishton vines😇
Рет қаралды 1,8 МЛН
Обзор Sonos Ace - лучше б не выпускали...
16:33
iOS 18 vs Samsung, Xiaomi,Tecno, Android
0:54
AndroHack
Рет қаралды 81 М.
Asus  VivoBook Винда за 8 часов!
1:00
Sergey Delaisy
Рет қаралды 1 МЛН