I like how you explain and give us the time to understand what you are talking about (especially for non-English native speakers ) so I'm so happy to discover your channel thank you so much.
@xbigwigglyx Жыл бұрын
You explained this so well and made it easy to understand. Going to go watch the next videos!
@code4AI Жыл бұрын
Awesome, thank you!
@theunknown2090 Жыл бұрын
Awesome tutorial video, keep up the good work.
@waeldimassi3355 Жыл бұрын
Nice video !! Looking forward for the upcoming ones ! Great job
@code4AI Жыл бұрын
More to come!
@TheUnguidedExplorer2 ай бұрын
Thank you for your video, it's really useful to understand t5 models. I have query, can we use this model for text extraction from user input? How robust are these compared to NER model
@SerenityScriber Жыл бұрын
You are a legend!
@code4AI Жыл бұрын
Thank you!
@anantvignesh9721 Жыл бұрын
This is a great video compiling all the necessary information about Google's family of Open Source LLMs. Not sure what fine-tuning technique you'd use in your upcoming videos but I'd like to see how to fine tune using PEFT from Huggingface.
@code4AI Жыл бұрын
Thanks for your feedback. Next step is classical fine-tuning of T5 and Flan-T5. And then I'll dive in Parameter-Efficient Fine-Tuning, since you'll read everywhere, that recent State-of-the-Art PEFT techniques achieve performance comparable to that of full fine-tuning, so why not let's have a quantitative and qualitative comparison between those two techniques of fine-tuning?
@theunknown2090 Жыл бұрын
@@code4AI 100%
@iamrxn6174 Жыл бұрын
I have a very important question. For the dataset containing only question and answer features. How should I approach? For eg. If the user input question, model must generate answer.
@AmmarAsim-c4s Жыл бұрын
How do I structure my inputs and outputs in my datasets when using huggingface trainer???
@testerpce Жыл бұрын
Can you please make a video on finetuning the Flan T5 with PEFT/Lora ? And possibly show some results ?
@healthtourguide Жыл бұрын
Hello, thank you for your valuable content I currently use gpt4 for my tasks I have a system message that is about 3k tokens( a scoring criteria). A user input that is also about 3k (contains info to be scored) and a assistant output that is 1k I know that flan-T5 has maximum token length about 4096 tokens, Can i do instruced finetuning using the user input and assistant output only, without the system message?
@riser9644 Жыл бұрын
Do we need to do some text cleaning when useing t5
@riser9644 Жыл бұрын
Is t5 better than Roberta for classification
@RedCloudServices Жыл бұрын
A question please how can we fine tune with custom data without formatting it in question-answer?
@code4AI Жыл бұрын
If all of your data have a coherent format, all have the same structure, it doesn't matter how you call them, the neural network will learn the pattern. You just have to be consistent all over your computation. Hint: check the pre-training data set for your transformer to find similar structural patterns and adapt accordingly.
@sunnyy62953 ай бұрын
Getting valueError in flan-t5 finetune
@vinsmokearifka Жыл бұрын
Sir, what is different train model between english vs non english? thank you
@code4AI Жыл бұрын
You need a LLM model, that has been pre-trained on different languages, not just English. Because your tokenizer has to adapt to your (multiple) language(s). But if you have a multi-lingual LLM like BLOOM, that has been pre-trained on 50+ languages, you can fine-tune it like any other language model.