Fine Tune Flux Diffusion Models with Your Photos

  Рет қаралды 5,120

Trelis Research

Trelis Research

Күн бұрын

Пікірлер: 21
@reo4153
@reo4153 Ай бұрын
Excellent video , well explained
@loicbaconnier9150
@loicbaconnier9150 4 ай бұрын
Excellent 😊
@thanartchamnanyantarakij9950
@thanartchamnanyantarakij9950 4 ай бұрын
Great from Thailand
@TrelisResearch
@TrelisResearch 4 ай бұрын
thanks
@loicbaconnier9150
@loicbaconnier9150 4 ай бұрын
Hello from Paris
@TrelisResearch
@TrelisResearch 4 ай бұрын
Salut Loic
@loicbaconnier9150
@loicbaconnier9150 4 ай бұрын
thanks for all
@exzxm3801
@exzxm3801 4 ай бұрын
Great content, your channel is great. In this video, you added [Trigger] to your txt files, do you think it should be the actual token word defined in your script. Else you are training on the word '[trigger]' which isn't the token you will use later to trigger the lora. I found that flux (dev) loras are strong enough without needing a token in prompt anyway. Also mostly people are saying no description needed in the txt files as flux understands the images well enough on its own with clip+t5, unlike previous SD training which neeed the guidance?
@exzxm3801
@exzxm3801 4 ай бұрын
also be great to get your take on fine tuning this hyped reflection type prompt (like how you finetuned for functions previously)
@TrelisResearch
@TrelisResearch 4 ай бұрын
ok, that's quite a nuanced question. So, what the trigger word offers is theoretically a way to have the model use your lora ONLY when that trigger word is present. However, that possibly relies on not overfitting the model... and basically everyone (including me here) overfits the model by doing tons of epochs. If you're going to only use your lora for your own purpose (where you always want the lora's effect), then this is moot. You don't need a trigger and the model will "use" your lora for all prompts.
@TrelisResearch
@TrelisResearch 4 ай бұрын
@@exzxm3801 well there are two vids on function calling you can check out . Regarding reflection, I haven't dug too deeply. I have a video coming out on thursday about DSPy which is basically advanced prompting and the effect is VERY strong on performance. What reflection is doing is specifically putting in a and step. Anthropic already do thiking and my gut feel is that their engineers would have found any obvious tweaks to this kind of approach. Probably does help a lot. Does it help to also close the data gap from llama 70B to claude sonnet 3.5, probably not. #speculation
@exzxm3801
@exzxm3801 4 ай бұрын
@@TrelisResearch i think i agree with your hunch on overfitting is why loras often work without the token in the prompt. But I still think you want to be putting a relevant token in the txt files.
@exzxm3801
@exzxm3801 4 ай бұрын
@@TrelisResearch I guess the curiuous part (if there is one in any of this possibly fakery) is are training structured/ in depth complicated prompts into a finetune showing any improvement over using a simple complicated prompt up front (in the same way finetuning on functions made your function models improve). I do think there is room for improvement on prompt design, the issue in the reflection (aside from the does it work bs) is it is requiring the extra tokens to get to an eventual correct response. I wonder if training in some sort of stop and surmise tag halfway through so it can reviews its own progress might help (would need to be finetuned on specific structured synthetic data I imagine for this)
@ravitejakonda9236
@ravitejakonda9236 4 ай бұрын
Can we merge different trained LoRAs, such as combining a person LoRA model with another person's LoRA model to create a single model? Is it possible to combine both models into one and continue training them as a unified model? Additionally, can we train a model with two different subjects combined into one, ensuring that both subjects are learned together in the training process?
@TrelisResearch
@TrelisResearch 4 ай бұрын
You can train two loras separately and then apply both to the same model to get a combined effect. OR you can take all of the data and just train one lora. To first order, there shouldn't be all that much difference in the two approaches - although training separately gives you more flexibility as to which style to apply.
@AICohort
@AICohort 4 ай бұрын
Can you please create a video on merging multiple Lora and generating frame by frame images with prompting ​@@TrelisResearch
@sharadsisodiya3853
@sharadsisodiya3853 4 ай бұрын
please share code as well
@TrelisResearch
@TrelisResearch 4 ай бұрын
All in the description!
@Username56291
@Username56291 4 ай бұрын
Hi
@TrelisResearch
@TrelisResearch 4 ай бұрын
howdy
Fine tuning Pixtral - Multi-modal Vision and Text Model
55:22
Trelis Research
Рет қаралды 4,5 М.
EASIEST Way to Fine-Tune a LLM and Use It With Ollama
5:18
warpdotdev
Рет қаралды 274 М.
OCCUPIED #shortssprintbrasil
0:37
Natan por Aí
Рет қаралды 131 МЛН
$1 vs $500,000 Plane Ticket!
12:20
MrBeast
Рет қаралды 122 МЛН
ВЛОГ ДИАНА В ТУРЦИИ
1:31:22
Lady Diana VLOG
Рет қаралды 1,2 МЛН
Their Boat Engine Fell Off
0:13
Newsflare
Рет қаралды 15 МЛН
Large Language Models (LLMs) - Everything You NEED To Know
25:20
Matthew Berman
Рет қаралды 166 М.
How Stable Diffusion Works (AI Image Generation)
30:21
Gonkee
Рет қаралды 165 М.
Fine-tuning LLMs with PEFT and LoRA
15:35
Sam Witteveen
Рет қаралды 141 М.
Why Does Diffusion Work Better than Auto-Regression?
20:18
Algorithmic Simplicity
Рет қаралды 448 М.
Flux: all samplers, schedulers, guidance, shift tested!
24:33
Latent Vision
Рет қаралды 37 М.
Multi modal Audio + Text Fine tuning and Inference with Qwen
56:31
Trelis Research
Рет қаралды 1,9 М.
Optimize Your AI - Quantization Explained
12:10
Matt Williams
Рет қаралды 16 М.
OCCUPIED #shortssprintbrasil
0:37
Natan por Aí
Рет қаралды 131 МЛН