NEW WizardCoder-34B - THE BEST CODING LLM

  Рет қаралды 24,728

Prompt Engineering

Prompt Engineering

Күн бұрын

Well, less than 48 hours since the release of Codellama, there is a new model WizardCoder-Python-34B that beats all previous models on the HumanEval benchmark. We will also cover the Phind model that was able to be the second best model.
#wizardcoder #codellama #llama2
▬▬▬▬▬▬▬▬▬▬▬▬▬▬ CONNECT ▬▬▬▬▬▬▬▬▬▬▬
☕ Buy me a Coffee: ko-fi.com/promptengineering
|🔴 Support my work on Patreon: Patreon.com/PromptEngineering
🦾 Discord: / discord
▶️️ Subscribe: www.youtube.com/@engineerprom...
📧 Business Contact: engineerprompt@gmail.com
💼Consulting: calendly.com/engineerprompt/c...
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
LINKS:
WizardCoder: github.com/nlpxucan/WizardLM/...
Phind Model: www.phind.com/blog/code-llama...
Try it yourself: 47.103.63.15:50085/
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
All Interesting Videos:
Everything LangChain: • LangChain
Everything LLM: • Large Language Models
Everything Midjourney: • MidJourney Tutorials
AI Image Generation: • AI Image Generation Tu...

Пікірлер: 51
@engineerprompt
@engineerprompt 9 ай бұрын
Want to connect? 💼Consulting: calendly.com/engineerprompt/consulting-call 🦾 Discord: discord.com/invite/t4eYQRUcXB ☕ Buy me a Coffee: ko-fi.com/promptengineering |🔴 Join Patreon: Patreon.com/PromptEngineering
@OwnOpinions
@OwnOpinions 9 ай бұрын
"Training Llama2 with Function Calling Capabilities"... I am daily checking your channel if this video gets uploaded. Thanks for all the great content.
@chaithanyachaganti4305
@chaithanyachaganti4305 9 ай бұрын
Been following your channel lately, very informative and insightful analysis give by you on llms. Keep up good work
@vishnuitsrocking
@vishnuitsrocking 9 ай бұрын
Super excited to see the progress in open-source model! Soon everyone will be a programmer!
@trpultz
@trpultz 9 ай бұрын
Thanks for the great video! Really appreciate the thoroughness (especially mentioning the concerns surrounding possibility of overfitting to the HumanEval test set) and you actually putting the models to the test yourself! Looking forward to seeing where these models are at by the end of the year and looking forward to your future videos on them. 😀
@williammixson2541
@williammixson2541 9 ай бұрын
Outstanding content. Instant sub!
@engineerprompt
@engineerprompt 9 ай бұрын
Thank you 🙏
@vincentjean6756
@vincentjean6756 9 ай бұрын
Yes, amazing. 🎉
@sveindanielsolvenus
@sveindanielsolvenus 9 ай бұрын
Awesome stuff! Where is the paper with the updated GPT-4 humaneval scores? I can't find that.
@Yewbzee
@Yewbzee 8 ай бұрын
Very good point regarding HumanEval overfitting
@paraconscious790
@paraconscious790 9 ай бұрын
great video as always, thanks! can you please make a video for installing this model locally on Mac? thanks! the Vicuna video may not work as the text-generation-webui is changed since then.
@Nick_With_A_Stick
@Nick_With_A_Stick 9 ай бұрын
I agree with that. I don’t have proof but I have a feeling a lot of these models are writing good code due to good data in the original training, but they don’t understand the code they are writing. A more complex benchmark would help diagnose that, and I think benchmarks should be remade and harder specifically since we are starting to get close to 100% for some of them. God just seeing how good the 34b coder does makes me wonder how good the unreleased regular llama 34 could’ve done if trained with orca. Such a shame.
@NakedSageAstrology
@NakedSageAstrology 9 ай бұрын
Wonderful video sir thank you, could you please do an instructional video tutorial on how to install this on a Windows PC?
@Artorias920
@Artorias920 9 ай бұрын
amazing updates! End of year prediction: GPT-4 level open source models running on blockchain-esque publicly distributed systems
@jonsantos6056
@jonsantos6056 9 ай бұрын
what would that mean for us - free?
@Artorias920
@Artorias920 9 ай бұрын
@@jonsantos6056 probably something close to it. Something akin to renting your GPU to the network and earning credits that can be redeemed for inference
@AlexanderBukh
@AlexanderBukh 9 ай бұрын
No need for blockchain, too much overhead and only needed when everyone wants to steal or manipulate data. As for the distributed inference, they already do that, see Petals. Actually, llama.cpp also has means for distributed computational (MPI)
@STDFme
@STDFme 9 ай бұрын
I am looking for a llm supporting C/C++, is it suited, too?
@VR_Wizard
@VR_Wizard 9 ай бұрын
Very interesting. I woulf love to see the code interpreter compared to the open source .odels and to GPT-4 base model. Havent seen these comparisonst. I havent used code unterpreter yet so i am not sure it is usefull for creating code but I thought thats one of its strenghts so should be interesting to see how much better code interpreter is.
@engineerprompt
@engineerprompt 9 ай бұрын
That’s actually a good point. Let me do a comparison
@3lbios
@3lbios 9 ай бұрын
is the non-python version going to be quantized?
@bmoore813
@bmoore813 9 ай бұрын
Do you have any example of how I could run the WizardCoder-Python-34B-V1.0 on my a100? I ask cause that site is so slow and its takes three minutes to get a reply
@theresalwaysanotherway3996
@theresalwaysanotherway3996 9 ай бұрын
interestingly enough, people are taking the 34B coding model and fine-tuning it for other purposes like RP or tool usage now (airoboros-34B and samantha-34B being the first ones). It may be worth testing out the 34Bs ability to solve non coding logic problems, to see if the coding knowledge generalises to better overall logic.
@user-fm5yy2sx2p
@user-fm5yy2sx2p 9 ай бұрын
What is web-ui? is text-generation-webui?
@sunfisoft
@sunfisoft 8 ай бұрын
does it recognize my project files and respond accordingly and is it necessary to have gpu i dont have gpu
@erikjohnson9112
@erikjohnson9112 9 ай бұрын
5:53 Isn't this a fail because it uses .sort()? That would be an advanced function/methods. Only direct looping and conditionals are allowed.
@engineerprompt
@engineerprompt 9 ай бұрын
You are right, that’s a great catch. Should have caught that.
@SlyNine
@SlyNine 6 ай бұрын
33 billion fits on my 4090 34 normally doesn't launch.
@temp911Luke
@temp911Luke 9 ай бұрын
Has anyone tried this in koboldCPP or gpt4all using CPU only in terms of speed ?
@lsp0
@lsp0 9 ай бұрын
Why are none of these papers touching the HumanEval+ benchmark?
@chirwatra
@chirwatra 7 ай бұрын
Do you have a video about how to build or deploy WizardCoder-34B app?
@engineerprompt
@engineerprompt 7 ай бұрын
I don't but you could use streamlit or check out databutton.com. Its a good platform for quickly deploying streamlit apps
@BrandosLounge
@BrandosLounge 9 ай бұрын
anywhere we can try this model online? Don't hvae the resources to run locally
@engineerprompt
@engineerprompt 9 ай бұрын
Yes, there is a link in the video description. Responses are slow though
@jonsantos6056
@jonsantos6056 9 ай бұрын
can i dl this onto local machine?
@vincentjean6756
@vincentjean6756 9 ай бұрын
Yes.
@Jorsten
@Jorsten 9 ай бұрын
My chatGPT 3.5 solved both of them at first try... We need something harder.
@bradgalaxy8419
@bradgalaxy8419 9 ай бұрын
phind, since its a search engine its likely pronounced "find"
@rickhoro
@rickhoro 9 ай бұрын
Does anyone know if this would be reasonable to run on a Linux machine with the following configuration: Intel® Core™ i7-7820X CPU @ 3.60GHz × 16 NVIDIA Corporation GP106 [GeForce GTX 1060 3GB] 64 GB RAM Thanks. I'm new to the whole Python and AI world. Help would be very much apprec!
@engineerprompt
@engineerprompt 9 ай бұрын
I suspect, you will not be able to run it. To just run a 4bit version of 34B model, you need at least 16GB RAM. see here discord.com/channels/1104593592229572640/1134401808425549824/1141459504421146777
@rickhoro
@rickhoro 9 ай бұрын
@@engineerprompt Thanks for your reply. Do you mean 16GB of main memory RAM or GPU RAM? Because I do have 64GB of main memory RAM.
@pocketsfullofdynamite
@pocketsfullofdynamite 9 ай бұрын
@@rickhoro 16GB installed RAM is good to start depending on how much dependencies are required to run on your local machine. Also check how much VRAM is available. These models take huge amount of VRAM considering your display monitor also uses VRAM.
@rickhoro
@rickhoro 9 ай бұрын
@@pocketsfullofdynamite Well, as I said, I have 64GB of DRAM (main memory) but only 3GB of GPU/VRAM. What do you think?
@morsmagne
@morsmagne 9 ай бұрын
I’m skeptical of all those assessments: when it comes to coding what you need is a very large token limit - like Claude 2 with 100,000 tokens. This is because you want the AI to understand all the code that’s written and remember it for the rest of the discussion. It’s no good having a clever AI that’s got the memory of someone living with Alzheimer’s! Because that’s exactly how it feels like trying to do any kind of code work with GPT-4 etc.
@hemeleh8683
@hemeleh8683 9 ай бұрын
Exactly! Which model is your preference? Gpt 4 32k is getting quite expensive on my end, would be happy to hear alternatives😅
@temp911Luke
@temp911Luke 9 ай бұрын
I spoke to someone from WIzardCoder and they say the token limit is same as in CodeLLama so the token window size might actually be potentially 100k.
@morsmagne
@morsmagne 9 ай бұрын
@@hemeleh8683 Personally, I use GPT-Engineer using GPT-3.5-turbo (only because I don't have access to GPT-4 API) then I use Claude 2 (free version) to review the code and make suggestions. However, I'm lucky enough to live in the UK where Claude 2 is free. Therefore, you might need a VPN to access it. You're welecome!
@morsmagne
@morsmagne 9 ай бұрын
And also, I start a new Claude 2 chat for every change I intend to make because even 100,000 tokens doesn't go very far.
SmartGPT: Make ChatGPT Smarter
12:57
Prompt Engineering
Рет қаралды 10 М.
All You Need To Know About Running LLMs Locally
10:30
bycloud
Рет қаралды 113 М.
She ruined my dominos! 😭 Cool train tool helps me #gadget
00:40
Go Gizmo!
Рет қаралды 52 МЛН
Is CODE LLAMA Really Better Than GPT4 For Coding?!
10:21
Matthew Berman
Рет қаралды 110 М.
I Tried Every AI Coding Assistant
24:50
Conner Ardman
Рет қаралды 705 М.
The End Of Programming
24:55
Matthew Berman
Рет қаралды 137 М.
StarCoder: How to use an LLM to code
12:55
AssemblyAI
Рет қаралды 37 М.
Using Llama Coder As Your AI Assistant
9:18
Matt Williams
Рет қаралды 64 М.
Marker: This Open-Source Tool will make your PDFs LLM Ready
14:11
Prompt Engineering
Рет қаралды 34 М.
AI coding assistants just leveled up, again…
4:51
Fireship
Рет қаралды 1,1 МЛН
Should You Use Open Source Large Language Models?
6:40
IBM Technology
Рет қаралды 339 М.
My ChatGPT 4 Workflow & Tips as a Software Engineer
4:52
Marko
Рет қаралды 628 М.
Iphone or nokia
0:15
rishton vines😇
Рет қаралды 1,7 МЛН
Неразрушаемый смартфон
1:00
Status
Рет қаралды 1 МЛН
One To Three USB Convert
0:42
Edit Zone 1.8M views
Рет қаралды 440 М.
Непробиваемый телевизор 🤯
0:23
FATA MORGANA
Рет қаралды 571 М.