LLAMA-2 🦙: EASIET WAY To FINE-TUNE ON YOUR DATA 🙌

  Рет қаралды 162,007

Prompt Engineering

Prompt Engineering

Күн бұрын

In this video, I will show you the easiest way to fine-tune the Llama-2 model on your own data using the auto train-advanced package from HuggingFace.
Steps to follow:
---installation of packages:
!pip install autotrain-advanced
!pip install huggingface_hub
!autotrain setup --update-torch (optional - needed for Google Colab)
---- HuggingFace credentials:
from huggingface_hub import notebook_login
notebook_login()
--- single line command!
!autotrain llm --train --project_name your_project_name --model TinyPixel/Llama-2-7B-bf16-sharded --data_path your_data_set --use_peft --use_int4 --learning_rate 2e-4 --train_batch_size 2 --num_train_epochs 3 --trainer sft --model_max_length 2048 --push_to_hub --repo_id your_repo_id -
🦾 Discord: / discord
☕ Buy me a Coffee: ko-fi.com/promptengineering
|🔴 Patreon: / promptengineering
💼Consulting: calendly.com/engineerprompt/c...
📧 Business Contact: engineerprompt@gmail.com
Become Member: tinyurl.com/y5h28s6h
💻 Pre-configured localGPT VM: bit.ly/localGPT (use Code: PromptEngineering for 50% off).
⏱️ Timestamps
Intro: [00:00]
Auto-train & installation: [00:17]
Fine-tuning - One Liner: [02:00]
Data Set Format: [05:30]
Training settings: [08:26]
LINKS:
autotrain: huggingface.co/autotrain
autotrain GitHub: github.com/huggingface/autotr...
All Interesting Videos:
Everything LangChain: • LangChain
Everything LLM: • Large Language Models
Everything Midjourney: • MidJourney Tutorials
AI Image Generation: • AI Image Generation Tu...
#llama #finetune #llama2 #artificialintelligence #tutorial #stepbystep #llm #largelanguagemodels #largelanguagemodel

Пікірлер: 226
@samcavalera9489
@samcavalera9489 10 ай бұрын
Thanks SO MUCH brother! You are a true hero! Fine tuning is the most important part of OS llms. That's where the value/wealth is hidden. I cannot wait for your following fine-tuning video.🙏🙏
@jersainpasaran1931
@jersainpasaran1931 10 ай бұрын
Thank you very much champion! We are getting to the true spirit of open source, allowing science to be truly scalable for the public and public interests.
@LainRacing
@LainRacing 10 ай бұрын
Very disappointed you didn't show this actually doing anything. How to verify or test if its working. I can run a script and have it do nothing... How do we see it actually worked or test it.
@arjunv7055
@arjunv7055 10 ай бұрын
One of the best video I have come across. I will definitely share this channel with my colleagues and friends who wants to learn more on this topic.
@engineerprompt
@engineerprompt 10 ай бұрын
Thank you!
@OpenAITutor
@OpenAITutor 10 ай бұрын
So great! Thank you for being so clear!!! loving it
@garyhuntress6871
@garyhuntress6871 10 ай бұрын
I was initially skeptical but this was an excellent short tutorial. Thanks!
@engineerprompt
@engineerprompt 10 ай бұрын
Glad it was helpful!
@PickleYard
@PickleYard 10 ай бұрын
Wow, just what I needed. I just put together a Flan Orca style dataset, I cant wait to try in Colab! Thank you for your hard work.
@engineerprompt
@engineerprompt 10 ай бұрын
Nice, good luck
@teleprint-me
@teleprint-me 10 ай бұрын
I was in the hospital because my lung collapsed and I've been having a seriously rough go at it lately (life long issues with fam, etc), so I really appreciate this video. Thanks for all your hard work. Researching these topics and understanding them is no small feat. Keep it up.
@engineerprompt
@engineerprompt 10 ай бұрын
I am really sorry to hear that! Hope you are recovering well. Wishing you a quick recovery. Also really appreciate all your contributions. Stay strong my friend!
@immortalsun
@immortalsun 7 ай бұрын
Hope you get better!
@bardaiart
@bardaiart 10 ай бұрын
Thank you very much! Looking forward to the dataset preparation video :)
@christianmboula8923
@christianmboula8923 2 ай бұрын
Superb tutorial by its clarity, simplicity and to the point...big Thank you! NOTE Bugfix : replace the underscore with corresponding dash to make the autotrain command run on colab
@lallaaichakone2106
@lallaaichakone2106 10 ай бұрын
wooow, after days of seraching for videos. I see everything that i wanted in this video and in simple terms. Great work
@engineerprompt
@engineerprompt 10 ай бұрын
Happy to hear that!
@miriamramstudio3982
@miriamramstudio3982 10 ай бұрын
Thanks for the update. Very interesting.
@adriantang5811
@adriantang5811 10 ай бұрын
Great Sharing again. Many thanks!
@sb98052
@sb98052 8 ай бұрын
Thank you for these very clear videos. Do you have any thoughts or pointers on resources for doing this type of training on code models such as CodeLlama?
@karthigeyan88
@karthigeyan88 10 ай бұрын
Hi, thanks for the video, could you explain in detail how to load the model and create an inference api in the local machine? that would be really helpful. thanks in advance
@MuhammadFhadli
@MuhammadFhadli 10 ай бұрын
hi, have you find a way to do the inference?
@karthigeyan88
@karthigeyan88 10 ай бұрын
@@MuhammadFhadli yeah, we have provisioned a Nvidia 64GB GPU machine and created an inference pipeline with llama.cpp library. Using an GGML model versiom from TheBloke huggingface
@immortalsun
@immortalsun 7 ай бұрын
‘Could you explain in detail […]’ Talking to him like he’s ChatGPT
@bahramboutorabi5971
@bahramboutorabi5971 10 ай бұрын
Great video. Thank you
@photojeremy
@photojeremy 10 ай бұрын
would be great to have a colab notebook for this that included inference on the finished pushed model
@MuhammadFhadli
@MuhammadFhadli 10 ай бұрын
hi, have you find a way to do the inference?
@manujmalik9843
@manujmalik9843 9 ай бұрын
@@MuhammadFhadli did you find it?
@gerardorosiles8918
@gerardorosiles8918 9 ай бұрын
I was thinking that once you push to huggingface you could use something like text generarion webui to play with the model
@serenditymuse
@serenditymuse 10 ай бұрын
The major work looks to be in making your dataset properly. Which is pretty common. Do you have or are you planning another video that is for training models simply by handing it a lot of files of say web content or better still the raw urls and perhaps something like tags and such? In other words how to add to unsupervised learning from a corpus.
@sohailhosseini2266
@sohailhosseini2266 8 ай бұрын
Thanks for sharing!
@ilyaskydyraliev6498
@ilyaskydyraliev6498 8 ай бұрын
Thank you for the video! May I ask, how big of a dataset should I have to see that fine tuning actually worked and model learnt new data?
@learn2know79
@learn2know79 10 ай бұрын
Hi Thanks for the detail explanation. Could you please make another video explaining the RLHF with code implementation.
@bagamanocnon
@bagamanocnon 10 ай бұрын
how can i incorporate my own data into the 'assistant' fine tune? for example, a 100 page document about a company product. do i format it into the something similar to what's in the openassistant dataset and add it to the dataset? or finetuning on own data will be another finetuning step? i.e. after finetuning on the openassistant dataset, i need to run another finetune for my own data? cheers and thanks for all your hardwork to share your knowledge to us!
@dec13666
@dec13666 7 ай бұрын
Nice video! A recurring aspect I have seen amongst these tutorials however, is that they never mention how to use the custom LLM model (i.e., doing some inference with the custom LLM model), or how to obtain metrics about it... Do you have any other video, where you discuss those 2 topics? Thank you!
@AA-rd6nm
@AA-rd6nm 10 ай бұрын
Very deatiled thanks for sharing. I ❤ it.
@engineerprompt
@engineerprompt 10 ай бұрын
You are so welcome!
@jongheebae6269
@jongheebae6269 5 ай бұрын
I have the autotrain error as follows. autotrain [] llm: error: the following arguments are required: - -project-name So I changed '--project-name' instead of '--project_name'. Then faced another error.
@MaralSheikhzadeh
@MaralSheikhzadeh 6 ай бұрын
well explained video. thank you:)
@engineerprompt
@engineerprompt 6 ай бұрын
Thank you
@zhirongchen9861
@zhirongchen9861 10 ай бұрын
Hi, how can I choose a method to finetune the model. For example, if I want to use LoRA to finetune lamma2, how can I do it?
@swauce507
@swauce507 8 ай бұрын
After you finetune the model, how do you use it as a chat interface to query the model and see its results?
@user-hz3oh3xc9t
@user-hz3oh3xc9t 10 ай бұрын
Can you make a video on fine tuning a llm model on a recipe dataset.
@Koyaanisqatsi2000
@Koyaanisqatsi2000 9 ай бұрын
Thank you very much! Where can I view the loss of my training or evaluation data using this method?
@user-me5zg7is3k
@user-me5zg7is3k 7 ай бұрын
Great video thank you! I have a question; I have a prompt, an output from a model, and a desired output, how I can format this data, please?
@anjakuzev592
@anjakuzev592 10 ай бұрын
Please make a video for creating your own dataset and actually using the model
@engineerprompt
@engineerprompt 10 ай бұрын
That is work in progress.
@Yash-mk8tc
@Yash-mk8tc 10 ай бұрын
how to use this trained model? can you please make video on this?
@deepakkrishna837
@deepakkrishna837 7 ай бұрын
Hi Great Video. Thanks a lot for this. QQ: if I am building an information extractor and the max token length of the training data is 2750 and hence I have kept model_max_length as 3000. Do I need to strictly keep the block_size as well to 3000? Please answer!
@vijayendrasdm
@vijayendrasdm 9 ай бұрын
What is the relation between max token size and the model kind of repeats itself ? The one you talk in the things to consider
@krishnareddy9
@krishnareddy9 10 ай бұрын
Thank you for the video, I am looking forward video about how to prepare our own dataset without using huggingface dataset !!
@engineerprompt
@engineerprompt 10 ай бұрын
It's up now, enjoy!
@Techarium
@Techarium Ай бұрын
@@engineerprompt video link please.... And this one-line command throws error on colab: unknown argument, any suggestions pls?
@prakhargurha267
@prakhargurha267 10 ай бұрын
2 questions. Is autotrain-advanced fine tuning is only available as a CLI format, or any other technique i available?Do we need collab pro for llama-2-7b-bf16.Can you suggest some smaller models to try?
@dr.aravindacvnmamit3770
@dr.aravindacvnmamit3770 3 ай бұрын
Hi, the way you are explaining is very positive !!!! One solution am not getting is If I want to train my custom data on regional languages how to proceed can you share your knowledge on this. Which model is best on this and if we pass the Prompt in English will it gets converted to regional language and generates the ouput?
@ShiftKoncepts
@ShiftKoncepts 10 ай бұрын
I am a little confused, so the Llama LLM on gpt4all has to be trained first before usage with local docs?
@VerdonTrigance
@VerdonTrigance 3 ай бұрын
How to train on unstructured data (a book for example) with self-supervized train algorythm and eventually make a chat from it?
@georgekokkinakis7288
@georgekokkinakis7288 10 ай бұрын
I really love your tutorials, they are deeply informative. I was wondering for the following. Unfortunately 😔 all these LLMs are trained in English , but the world has so many other languages. If I follow the fine tuning you described in your video would I be able to fine tune the lama model for a specific dataset which has questions about mathematical definitions and methodologies with their according responses written in Greek? The amound off samples is about 100 questions with answers, I know it is really small but could this give good results for thebspecific dataset? And one last question , do you know any multilingual LLM which supports Greek. Thanks once more and keep up with your excellent ❤ presentations.
@AymanEL-BACHA
@AymanEL-BACHA 9 ай бұрын
hi @georgekokkinakis7288, have you tried training with your 100 sample/questions ? any improvements ?
@georgekokkinakis7288
@georgekokkinakis7288 8 ай бұрын
@@AymanEL-BACHA No I haven't yet
@noraalzamil2660
@noraalzamil2660 8 ай бұрын
Thank you very much 🙏 Can I apply it with TheBlock llama-2-7b ggml?
@ilhemwalker9145
@ilhemwalker9145 3 ай бұрын
hey please i copied the same line but i'm getting error : autotrain [] llm: error: the following arguments are required: --project-name. i don't know what to do
@sharadpatel107
@sharadpatel107 10 ай бұрын
can you please put in a link for a colab notebook for this
@emrahe468
@emrahe468 10 ай бұрын
finished running the autotrain in about 6h. And upload the model to hugginface. so what to do next? How to use this?
@waelmashal7594
@waelmashal7594 10 ай бұрын
Just amazing
@engineerprompt
@engineerprompt 5 ай бұрын
Thank you!
@nitingoswami1959
@nitingoswami1959 10 ай бұрын
Can we train this model on any data or it requires some specific format ? Does every llm requires some specific tabular data or any raw data ?
@PickaxeAI
@PickaxeAI 9 ай бұрын
What GPU should we select to complete this training? Could the T4 handle it?
@oxydol3456
@oxydol3456 Ай бұрын
learnt a lot from the video.Thanks. Is it easy to revert the model to the state before a tuning?
@engineerprompt
@engineerprompt Ай бұрын
Thanks, yes, you are merging the extra "LoRA Adapters" layers to the model. The actual model actually remains unchanged so you can just reuse it for other purposes.
@Noscov
@Noscov 8 ай бұрын
Thanks for the video. I have a further question. At 5:50 your dataset has the columns instruction and input. What is the input-column for?
@immortalsun
@immortalsun 7 ай бұрын
For example a question.
@user-nj7ry9dl3y
@user-nj7ry9dl3y 9 ай бұрын
For fine-tuning of the large language models (llama-2-13b-chat), what should be the format(.text/.json/.csv) and structure (like should be an excel or docs file or prompt and response or instruction and output) of the training dataset? And also how to prepare or organise the tabular dataset for training purpose?
@rainchengcode4fun
@rainchengcode4fun 9 ай бұрын
timdettmers/openassistant-guanaco has introduction about the dataset, it should be a list of json with instruction, response in it.
@GEfromNJ
@GEfromNJ 7 ай бұрын
See this is one the thing that gets completely glossed over in videos like this. If you take a look at timdettmers/openassistant-guanaco, you'll see that it's some nicely formatted data. It doesn't answer the question about how someone would take their own data and get it into this format.
@user-wy3rr4dp2c
@user-wy3rr4dp2c 5 ай бұрын
Hello, I am a beginner in LLM. I generated the model folder locally according to the video operation, but the folder size is only about 130Mb. The base model I use is 7b llama2. Is this normal? Why is the model size reduced so much? How do I get the normal size model? I would be grateful if you could answer it for me
@machineUnlearner
@machineUnlearner 3 ай бұрын
i have a time series data, with 7 to 10 parameters. What should I do ?
@alx8439
@alx8439 10 ай бұрын
Does it use lora or qlora techniques?
@anantkabra6825
@anantkabra6825 7 ай бұрын
Hello I am getting this error can someone please help me out with it: ValueError: Batch does not contain any data (`None`). At the end of all iterable data available before expected stop iteration.
@okopyl
@okopyl 8 ай бұрын
Amazing, but how to do the inference properly with this peft thing?
@mdfarhananis8950
@mdfarhananis8950 10 ай бұрын
Please teach how to create dataset for finetuning
@caiyu538
@caiyu538 5 ай бұрын
How to save the fine tuned model to local disk instead of pushing to hub. Could you show us the model pushed to hub? These video graphs will make it clearer. Great.
@DikHi-fk1ol
@DikHi-fk1ol 7 ай бұрын
Please make another tutorial on how to fine-tune a model on custom dataset rather than using the hugging face ones.
@MarceloLimaXP
@MarceloLimaXP 10 ай бұрын
Thanks guy ;)
@SadeghShahmohammadi
@SadeghShahmohammadi 10 ай бұрын
It took a few hours, everything went well but at the end the model is not in my hf repository! Cannot find it anywhere!
@nufh
@nufh 8 ай бұрын
Other than google colab, what is other platform that we can use? I'm still new, just started to learn about python.
@justabacteria
@justabacteria 10 ай бұрын
Could you explain or make a video on how to use your new fine-tuned model?
@engineerprompt
@engineerprompt 10 ай бұрын
Yes, that's coming very soon
@fangxiaoyuan-fm6vr
@fangxiaoyuan-fm6vr 10 ай бұрын
Could you introduce how to deploy our model to a website? Thanks!
@sravanavvaru4473
@sravanavvaru4473 10 ай бұрын
hey the thing I did not get is on what data is the model getting trained ??
@arjunv7055
@arjunv7055 10 ай бұрын
some of my friends who followed this tutorial mentioned they see an argument issue. I think it is because of the command being broken down into multiple lines. Running the command in multiple lines requires a '\' to be added at the end of every line. Final command should look like this !autotrain llm --train --project_name '' \ --model TinyPixel/Llama-2-7B-bf16-sharded \ --data_path timdettmers/openassistant-guanaco \ --text_column text \ --use_peft \ --use_int4 \ --learning_rate 2e-4 \ --train_batch_size 2 \ --num_train_epochs 3 \ --trainer sft \ --model_max_length 2048 \ --push_to_hub \ --repo_id /'t \ --block_size 2048 > training.log &
@nayyershahzad8051
@nayyershahzad8051 5 ай бұрын
getting following error, kindly help: autotrain [] llm: error: the following arguments are required: --project-name
@bfam7110
@bfam7110 10 ай бұрын
Is there embeddings or RAG with this approach?
@ajlahade2201
@ajlahade2201 9 ай бұрын
can you please make a video on how to push this model to hugging face (like production level with model card) and call that model
@youwang9156
@youwang9156 6 ай бұрын
thank you for ur video, literally save my life, just have one little question about the prompt format, you were using ### human and ### Assistant, so does this format basically depend on the pre-train model prompt format? like Llama-2 chat which has a certain unique format, but some like the Llama 2 base model, if there's no specific mention of that, then we can define our own format for the prompt? do I understand it correctly ? Thank you for your video again !!!!
@engineerprompt
@engineerprompt 6 ай бұрын
Glad you found it helpful. The template depends on whether you are using the base or the chat version. For the base model, you can define your own template as I am doing here because there is no template for it for using it as assistant (base model is actually the next word prediction model). But if you are finetuning a chat version then you will have to use the specific template that was used for finetuning the model. Hope this helps
@adapalarajyalakshmi3728
@adapalarajyalakshmi3728 10 ай бұрын
Thanqu for the video can u explain how to use postgress database dataset
@Dave-nz5jf
@Dave-nz5jf 10 ай бұрын
you would probably need to pull the data in batches, in the right format, and then run this autotrainer on a batch basis. But it's an interesting question - if you have data that's changed (in the database), and you retrain the model, how does the updated data impact the model output.
@MicaleAntonio
@MicaleAntonio 5 ай бұрын
Does auto train do multi-label text classification?
@user-jx3wy6fe4s
@user-jx3wy6fe4s 4 ай бұрын
I am facing issues in the autrain line where its stating argument should be project-name instead of project_name and even if i change that its not taking arguments like data_path, use_peft. can someone help me out?
@Shahawir
@Shahawir 10 ай бұрын
I wonder if it is possible train LLAMA, on data where input are numbers and categorical variables(string), of fixed length, to predict a timer series of fixed size, anyone knows if this possible?
@gamingisnotacrime6711
@gamingisnotacrime6711 8 ай бұрын
I have a custom dataset with 50 rows. For how many epochs should i fine tune thr model? Each line in my dataset is in this format - ###Human: Who is John?### Assistant: John is a famous youtuber (My dataset has only a single column named text and 50 rows which have the data in above format So also are there any issues with my dataset?
@pareak
@pareak 4 ай бұрын
What is the difference between the SFT and the Generic trainer?
@deepjyotibaishya7576
@deepjyotibaishya7576 10 ай бұрын
Colab always stuck and show me complete on 57% when it running on merging It it possible to upload folder to Hugging face and laster on can i Mergin it and make it ai model ??
@Noshiru
@Noshiru Ай бұрын
Hello! The question might be stupid, but how come this is so difficult to learn to the AI our own data ? I mean, when you talk to ChatGPT for example, if you tell it stuff, it will remember (if you use the same chat) what you said and it will be able to answer your questions about it. Why can we just give the AI a documentation for example ?
@okopyl
@okopyl 8 ай бұрын
Now when i generate responses, i get input generated as well. Why? How to avoid that?
@contractorwolf
@contractorwolf 10 ай бұрын
subscribed!
@engineerprompt
@engineerprompt 10 ай бұрын
Thanks :)
@user-zt1ie5ir2p
@user-zt1ie5ir2p 9 ай бұрын
I haven't tried it on colab yet but was wondering, do we need colab pro or colab pro+ for this tutorial?
@engineerprompt
@engineerprompt 9 ай бұрын
For this, you can use the sharded model with free version but for full model you will need pro
@tubesarkilar
@tubesarkilar 7 ай бұрын
can you show a sample of time series data file to feed into Autotrain?
@8eck
@8eck 10 ай бұрын
What if i only want to feed a specific non-instruction data into the model? For example some financial data or some books or some glossary? Can i just keep the ###Output empty, will the model learn from that data? Also, do i need to split that data into train and test parts or it is not required and is optional for pre-trained models?
@curtisho5255
@curtisho5255 10 ай бұрын
i have the exact same question! omg!
@phoenixfire6559
@phoenixfire6559 10 ай бұрын
If you leave the output empty then the model will learn to give you empty responses every time you put that type of data in. The best way to make the data for your finetune is thing about it from reverse. When you put the input in, what do expect the output to be? That's what you should be filling output with.
@8eck
@8eck 10 ай бұрын
@@phoenixfire6559 i'm talking about pre-training like fine-tuning, models in the pre-training phase doesn't get any output examples, they just learn from the data, that's what i'm trying to understand. Is fine-tuning is only about question & answer pairs? How to continue pre-training of the model with frozen base weights. Just like transfer learning.
@curtisho5255
@curtisho5255 10 ай бұрын
@@8eck exactly. he don't get it. We want it to train on pure data, not train on Q&A responses. He must have not played with chatbase.
@robosergTV
@robosergTV 10 ай бұрын
@@curtisho5255 lmao the author of the video knows this. The video is clickbait for farm views (which is money) from noobs, who cant use simple google search.
@sanj3189
@sanj3189 10 ай бұрын
How can i use LLama2 for generating synthetic data
@user-fc5nz9wp2o
@user-fc5nz9wp2o 9 ай бұрын
Is there a link for the google colab notebook?
@meteor1
@meteor1 7 ай бұрын
Can I fine-tune llama-13b-GPTQ using autotrain-advanced ?
@prestonmccauley5467
@prestonmccauley5467 10 ай бұрын
I followed this exactly in collab, but seems that something is wrong with the arguments, Can you share your colab file?
@arjunv7055
@arjunv7055 10 ай бұрын
if you are breaking the command into multiple line please make sure to add \ towards the end so finally the command looks like this !autotrain llm --train --project_name '' \ --model TinyPixel/Llama-2-7B-bf16-sharded \ --data_path timdettmers/openassistant-guanaco \ --text_column text \ --use_peft \ --use_int4 \ --learning_rate 2e-4 \ --train_batch_size 2 \ --num_train_epochs 3 \ --trainer sft \ --model_max_length 2048 \ --push_to_hub \ --repo_id / \ --block_size 2048 > training.log &
@ElNinjaZeros
@ElNinjaZeros 9 ай бұрын
Thanks for sharing, by the way does auto-train need to be paid to be able to use it?
@engineerprompt
@engineerprompt 9 ай бұрын
The cli version is free to use unless they changed something recently
@nexusinfosec
@nexusinfosec 10 ай бұрын
Could you please create a video on the dataset creation?
@VadiyalaRR
@VadiyalaRR 9 ай бұрын
kzbin.info/www/bejne/Y6bMaYyBr5mXY6c hope it helps you
@titangadget
@titangadget 3 ай бұрын
I'm using this one line training code but is giving me error... can you update it?
@bharatkaushik9916
@bharatkaushik9916 8 ай бұрын
Can someone tell how to inference this model ?after pushing it to hub thanks
@rafaelferreiradesouza9972
@rafaelferreiradesouza9972 10 ай бұрын
I'm looking for a way to create a local server, that uses my trained IA for answers like a personal assistent, can anyone tell where can I learn that?
@ScottzPlaylists
@ScottzPlaylists 2 ай бұрын
🤯 Wow Wow Wow ❗
@engineerprompt
@engineerprompt 2 ай бұрын
thanks :)
@Yash-mk8tc
@Yash-mk8tc 10 ай бұрын
can you make a video on hugging face basics
@ajaym4257
@ajaym4257 4 ай бұрын
usage: autotrain [] AutoTrain advanced CLI: error: unrecognized arguments: --use-int4 --learning-rate 2e-4 --num-train-epochs 3 --model-max-length 2048 i'm getting this error
@mavensept
@mavensept 9 ай бұрын
Can we directly use this or should we have llama2 installed locally as a pre-req before using auto train ? pls advise
@engineerprompt
@engineerprompt 9 ай бұрын
You can just use this directly, don’t need to have llama2 installed
@hvbris_
@hvbris_ 10 ай бұрын
Really cool video, saved mea lot of time, how much memory would the GPU need to train a model like llama2-13B - is 12G enough or should I consider getting something beefier? Thanks in advance!
@engineerprompt
@engineerprompt 10 ай бұрын
Atleast 16GB
@carlsagan9808
@carlsagan9808 10 ай бұрын
This command is not working at all. Happening to anyone else? I get the repeated error >> RuntimeError: CUDA error: CUBLAS_STATUS_NOT_SUPPORTED when calling `cublasSgemm( handle, opa, opb, m, n, k, &alpha, a, lda, b, ldb, &beta, c, ldc)`
@user-ns4rg6wu5h
@user-ns4rg6wu5h 8 ай бұрын
I would like to ask, without GPU, can LLAMA-2-13b be fine-tuned on the CPU server, and how should it be modified
@engineerprompt
@engineerprompt 8 ай бұрын
Not really, if you try, it will take forever
@souvickdas5564
@souvickdas5564 9 ай бұрын
Can we fine-tune LLaMA model on MNLI or SNLI dataset? Is it worth doing ? Give me your thought.
@engineerprompt
@engineerprompt 9 ай бұрын
Yes, I think it’s possible. These might already be in the training data.
@jesusic1320
@jesusic1320 9 ай бұрын
A doubt I cannot solve anywhere: running this locally is free right? And after that, can you use it to generate images locally, also for free? I'd like to practice but avoid costs of that practice. I think for example Replicate runs the training online, so you have to pay. But I have a GTX3070 so I think I can do my practice locally for free
@gerardorosiles8918
@gerardorosiles8918 9 ай бұрын
Were you able to fine tune locally with the GTX3070?
LocalGPT & Llama-2: Adding Chat History & Custom Prompt Templates
14:23
Prompt Engineering
Рет қаралды 28 М.
Fine-tuning Large Language Models (LLMs) | w/ Example Code
28:18
Shaw Talebi
Рет қаралды 249 М.
100❤️
00:20
Nonomen ノノメン
Рет қаралды 67 МЛН
WHY IS A CAR MORE EXPENSIVE THAN A GIRL?
00:37
Levsob
Рет қаралды 21 МЛН
2000000❤️⚽️#shorts #thankyou
00:20
あしざるFC
Рет қаралды 14 МЛН
LLAMA-3 🦙: EASIET WAY To FINE-TUNE ON YOUR DATA 🙌
15:17
Prompt Engineering
Рет қаралды 50 М.
Qwen-Agent: Build Autonomous Agents with The Best Open Weight Model
19:34
Prompt Engineering
Рет қаралды 4,2 М.
"okay, but I want Llama 3 for my specific use case" - Here's how
24:20
Google's MESOP: Building Beautiful Web UIs in Pure Python
13:42
Prompt Engineering
Рет қаралды 2,7 М.
QLoRA-How to Fine-tune an LLM on a Single GPU (w/ Python Code)
36:58
NEVER lose dotfiles again with GNU Stow
14:33
typecraft
Рет қаралды 12 М.
Fine-Tune Your Own Tiny-Llama on Custom Dataset
14:32
Prompt Engineering
Рет қаралды 22 М.
I Analyzed My Finance With Local LLMs
17:51
Thu Vu data analytics
Рет қаралды 415 М.
Main filter..
0:15
CikoYt
Рет қаралды 4 МЛН
Карточка Зарядка 📱 ( @ArshSoni )
0:23
EpicShortsRussia
Рет қаралды 784 М.
How To Unlock Your iphone With Your Voice
0:34
요루퐁 yorupong
Рет қаралды 21 МЛН
5 НЕЛЕГАЛЬНЫХ гаджетов, за которые вас посадят
0:59
Кибер Андерсон
Рет қаралды 1,5 МЛН