For those of you wondering why he has the Network Rank (Dimension) so high (256), I am fairly certain it is because of the thousands of reference images to men he is using in his training. If you arent using that many pictures and are just sticking to your 15-50 reference pictures you are probably fine to leave it on 32-64 unless your training something less human like. This will also cut down on your training time immensely! Also yes, unless you want your training to stop early (around 3 out of the 10) then make sure to change your "Max trained Steps" to 0, (default is 1600). Good video though Thank you!
@Heldn1003 ай бұрын
thanks for that, this was so useful
@maxdeniel4 ай бұрын
Bro, this tutorial is sooo straight forward, and I really appreciate that you were taking the time to do an updated version, AI world is evolving so fast that the tutorials that were made 6 months ago are outdated, the interface for kohya changed a little bit and your tutorial just walk me through the new version step by step... I just clicked Start Training and I'm waiting to finish and star to run and check how the LoRA comes out. Thanks again!
@Heldn1003 ай бұрын
i make a lora and its come out great better than what i want... but 1.7g for one lora is too much we need to know how we can fix this
@weeliano5 ай бұрын
Thank you for producing this video it has helped me tremendously to figure out the training settings. I realized that the LORA that I trained without any regularization images look better than those that have. Been having great fun rendering many iterations of my alter-egos.
@OCGamingz6 ай бұрын
thats probably the best and most useful LoRa guide I've seen so far. thank you very much it helped me alot!
@allyourtechai6 ай бұрын
Thank you!!
@maxdeniel4 ай бұрын
Certainly is the best one. I have seen some other and they skip important steps and focus on other not that important. This guy knows how to impart a master class.
@zei_lia5 ай бұрын
Very good tuto ! Everything worked on my end, I just had to create the “log”, “images” and “models” folders, as it didn't do it automatically. My model works perfectly, thank you! 🙏
@allyourtechai5 ай бұрын
Awesome! Great to hear
@metanulski6 ай бұрын
What I would do different is to use WD14 captioning since it has more details on the picture. Also my setting need only one hour training on my 4060. I have to check the difference. Here is a nice trick: Once the training is done, it saves a setting file in the result folder. So if you need to train another model, you just need to load the setting file again, change the pictures, captions and model name and then hit start. :-)
@metanulski6 ай бұрын
Also, I think it is a good idea to let kohya generate a picture each epoch, so it is easy to pick the correct epoch for focus. Like "Full body picture of a man standing in a Forrest". Low epoch will not look like you, and at some point the Forrest might disappear because of overfitting. Use the highest epoch with the Forrest still in the picture.
@Heldn1003 ай бұрын
how in one hour ? i have 4070 and it took 9 hour for me !
@metanulski3 ай бұрын
@@Heldn100 20 pictures max, 768 max resolution and 2000 to 3000 total steps max :-)
@Heldn1003 ай бұрын
@@metanulski mine is 13 pictures, 1024x1024, but i dont touch max steps or anything i just do like what he do. i will try yours too, thanks
@metanulski3 ай бұрын
@@Heldn100 21:44:24-383786 INFO Valid image folder names found in: D:\SDXL Lora Training\models\img 21:44:24-384788 INFO Folder 40_Test woman: 21 images found 21:44:24-385788 INFO Folder 40_Test woman: 840 steps 21:44:24-385788 INFO Total steps: 840 21:44:24-386789 INFO Train batch size: 1 21:44:24-387791 INFO Gradient accumulation steps: 1 21:44:24-388791 INFO Epoch: 4 21:44:24-389792 INFO Regulatization factor: 1 21:44:24-390792 INFO max_train_steps (840 / 1 / 1 * 4 * 1) = 3360 21:44:24-391794 INFO stop_text_encoder_training = 0 21:44:24-392795 INFO lr_warmup_steps = 336 21:44:24-393795 INFO Saving training config to D:\SDXL Test\Test_20240811-214424.json
@jasonlisenbee37475 ай бұрын
I noticed your Optimizer was on Adafactor by default. Mine wasn't so I changed it. You didn't mention the setting for LR Scheduler, but I see in the video yours is set to Constant. Mine was set to Cosine. I changed it to match yours, but my LoRa's came out goofy and I got 3 instead of 10 somehow. Could that have anything to do with it?
@NateMac0006 ай бұрын
thanks for the tutorial... one thing I did differently was use that BLIP2 for captioning which IMO did a much more detailed caption of the image... at that point I didn't have a prefix so I used ChatGPT to make me a windows bat file, to add the prefix (trigger word) to all the txt files. Great tutorial, thanks again!
@allyourtechai6 ай бұрын
Great tips! Thanks for sharing, I need to test BLIP2 now!
@hleet3 ай бұрын
Thank you it works ! But the final result is not really as I would have expected (I didn't use the extra dataset of man or woman). Anyway I was able to make a lora file with your tutorial, that's the main point :)
@xnauwfallproductions2 ай бұрын
you remind of a very famous actor, very versatile actor who has a wide range of roles, i don't know who his manager/director is but he always ends up tired in his films
@allyourtechai2 ай бұрын
haha, he must have kids. I always end up tired too.
@Art0691p3 ай бұрын
Great video. Clear, no hype and to the point. Thanks.
@allyourtechai3 ай бұрын
Thank you!
@mnedix5 ай бұрын
EDIT: for some reason it started working just fine, I have no idea what I did to it. I think it ok, I have to do more testing with the optimizers; so far, the training is - 30 training pics / 20 reps / 5 Epochs / Rank 32 / Alpha 16 = 3~4hrs Thank you for the tutorial, I really hoped I could create LoRAs. I followed it to the letter and I get a RuntimeError: NaN detected in latents. I'm on a brand new 4070 and the resolution is 512, 512, so I should have enough vram for it.
@Heldn1003 ай бұрын
i do 1024x1024 with 4070 and there was no problem. you need to close games or anything use your vram background programs or wallpaper anmition or anything use so much power. i get really greate lora with it for wonyoung
@jungbtc6 ай бұрын
lol i was just getting confusing with your older tutorial! thanks for the update.
@Fanaz106 ай бұрын
yeahh this seems like civitai uses for training?
@NickPanick5 ай бұрын
I'm completely new to all of this. Will this work using sd3_medium as the pretrained model or should I stick with your template from your Patreon for SDXL base 1.0?
@birbb06 ай бұрын
This video was really good but I was wondering why you had the Network Rank at 256 while you had the Network Alpha at 1, which is a really small value when compared to the network rank? I've seen people use 64,32 a 2:1 ratio or just use the same number for both, I'd love to hear your explanation!
@andresmartin88953 ай бұрын
To avoid the "long process of resize every image" you can do from the terminal "mogrify -resize 1024 *.jpg" and it resizes all files ending with .jpg in the current folder"
@FirstLast-tr3ub3 ай бұрын
This has been really helpful and gotten good results from, thank you
@K-A_Z_A-K_S_URALA15 күн бұрын
I can ask you a question: what is the maximum number of photos you need to upload for training on a real person in full growth? I have 250 pcs of photos of my wife, I trained on a 1.5 model and everything is cool. I wrote 150_gen and top quality and now I'm busy with sdhl and I'm curious whether this amount will be a style or a character or you need to make up to 100 pcs of photos???? thanks
@allyourtechai15 күн бұрын
For a flux model on pixel dojo you only need 10-12 images
@K-A_Z_A-K_S_URALA15 күн бұрын
@@allyourtechai спасибо денег нет!)
@KlausMingo3 ай бұрын
Great guide, but you didn't tell us why you generated 10 tensor files.
@RayMark6 ай бұрын
Thanks for the tutorial! Your last one gave me my best results so was excited to try this. Question: I only got 3 tensor files after 10hrs. They’re all quite big (over 1 gb). Not sure where I went wrong? I have epochs set to 10 like you said. Thanks!
@allyourtechai6 ай бұрын
What type of gpu do you have? The network rank settings at 256 take a long time to train and produce large files, but the quality is also higher. You can lower the rank setting and train faster, smaller files but with lower quality
@RayMark6 ай бұрын
@@allyourtechai RTX 3080 (10gb)... everything seemed to work fine throughout the process and the Loras worked pretty well even though I only got 3 of them to test. Not sure if I had some setting wrong that it would only create 3. Thanks for your help, really appreciate the guides!
@avaloki95776 ай бұрын
@@RayMark I was having a similar problem look for the option in parameters: "Max train steps" it is by default set as: '1600' change it to '0' so it does not limit the max number of steps of the training. it worked for me now it is doing the whole training
@nixonmanuel645927 күн бұрын
Thank you so much for doing this!
@TravelPostcards2 ай бұрын
Thank You! .... I have this Error when I hit Prepare training data: RecursionError: maximum recursion depth exceeded while calling a Python object
@Gamer4Eire4 ай бұрын
The approach should be select images, set resolution to 1024, 1024, add tags, edit tags, refine tags, do it again, repeats x images x epochs / batches = steps. Always always use epochs, these ensure you generate a number of evolving Loras so you can try each one and see what fits.
@Mranshumansinghr6 ай бұрын
Exactly what I was looking for. Thank you.
@lostinoc35283 ай бұрын
What resolution should the training images be? Should they all be the same resolution? Or is a mix, is square, landscape and portrait fine, or preferred? Also when you choose to train for the basic checkpoint, could you instead train for a specific one somehow? Or is it better to train on a basic one?
@elizagarcia87992 ай бұрын
Watch again the video and you can train WHEREVER checkpoint
@lilly23796 ай бұрын
Thank you so much, I'm so happy you updated this! However, I can't seem to find your low VRAM config file. The patreon link only leads to the 3090 one along with the regularization files. I may have missed something (and it's not a big deal) but I thought I'd bring it up just in case. Thanks
@insurancecasino57906 ай бұрын
Wow. I'm thinking of getting an external GPU now. I can do comics with this.
@Hyokkuda12 күн бұрын
I forgot to mention that most people might want to keep "Save precision" at "fp16", because "bf16" is more commonly supported on newer architectures like the NVIDIA A100 (Ampere, specifically for data centers) which most people probably do not even have. An RTX 30XX or an RTX 40XX will not cut it. So if you own a a professional, high-performance GPU like the NVIDIA A100, the RTX 6000 Ada Generation and the NVIDIA RTX A2000. AI generation is definitely much faster with those cards, but the quality is the same. And the price tag on those cards really hurts.
@jasonlisenbee37475 ай бұрын
I think mine stopped. I checked it after maybe an hour and it said it was complete, but I only had 3 finished files, not 10. And they were named Final 1,2, and 3 which is strange. I closed the command window and they're all pretty bad. I've got 16vram and matched the Network Rank to the numbers shown in the video. I'm wondering if that was a mistake. I'm trying it again but have lowered it to 101 and 13 on the Network Alpha and going to bed to see what I come back to in the morning.
@theironneon3 ай бұрын
am i the only one that comparing the result images with his face on the bottom right? :D
@febercantes5 ай бұрын
and to train styles?
@-Belshazzar-4 ай бұрын
thanks for the tutorial, i am wondering though, what has really changed since the last tutorial? if i remember correctly, nothing really, exact same setting, no? except the fact this time you remembered to point out the prepare training data and model selection. which brings me to the question, why do you choose sdxl base and not a better trained checkpoint to start from like juggernaut for example? also, i noticed that even without regularization images i get good results and with an rtx 3090 and 23 highrez images training is about an hour and half long (same settings from your tut) not sure if it’s the lack of reg images, but you said 10 hours!? seems a bit too much i think. anyway, thank you again!
@allyourtechai4 ай бұрын
The settings were the same, but the software changed completely. They moved everything around in the UI, so I end up with 10+ questions a day about where to edit vari pi s settings. Hopefully it remains the way it is for a bit lol
@KingBerryBerry3 ай бұрын
@@allyourtechai What are the more common questions and responses? I follow every step from this today and IT WORKS really good maybe change one thing or two. In a 4090 20 images (differents quality) take me 2 hours, is that normal?
@samu7015Ай бұрын
Should I worry about this: min_bucket_reso and max_bucket_reso are ignored if bucket_no_upscale is set, because bucket reso is defined by image size automatically?
@ollyevans6365 ай бұрын
I’m experiencing an issue with my model training process. When I clicked “train,” the terminal indicated that it was done in less than 30 seconds, but no safetensor files appeared. Do I need to leave it overnight, or did my machine not execute the process correctly because the terminal said it was complete?
@allyourtechai5 ай бұрын
If the terminal said complete without any errors then something went wrong. Hard to say without errors though, so I would go through the settings and folders again
@HiramosCM5 ай бұрын
Same here!
@Jammy1up5 ай бұрын
Same for me, did you ever find a fix? Followed this awesome tutorial to a T
@slann3035 ай бұрын
You most probably ran out of VRAM. Try to not use the regularization-images, this should lower the VRAM usage
@Jammy1up5 ай бұрын
@@slann303 well I had the same issue, 16gb of vram and I did not use regularization images. Pretty sure that's not it
@artyfly3 ай бұрын
Sorry, I missed, why at the end we have 10 trained files? Were is the setting, when did we set this up?? Can we get one file at the end? Thanks)
@satoshidarikotamasara9105 ай бұрын
Can you make a tutor on how to train Lora slider, such as Lora detailer, detail tweaker etc.
@Hyokkuda17 күн бұрын
Gosh, I never had so much trouble in my life installing a single thing. So many things were going wrong. Torch was missing for some reason, the Library was missing, the SSL Certificate or whatever was missing or couldn't be accessed. And nobody had answers to anything. I had to throw all my errors at ChatGPT or Aria AI in order to get the help I need and guess what.. still not running. I even got a similar all-in-one package to install that, could be yours that got leaked, I dunno, but I ran into the same issues regardless.. Either Python's version was too high, or too low or the .whl file for Torch couldn't be identified. Then I had other errors about different versions suddenly conflicting with one another. One was suddenly too high, the other was too low or outdated. And I already had most of the stuff already installed, and did a bunch of reinstallation. I am so sick of looking at that CMD window... lol
@allyourtechai16 күн бұрын
Yep, it’s a massive pain in the ass. That’s why I spent the past year building Pixeldojo.ai
@Heldn1003 ай бұрын
thanks for it i try it and i get great result, but 1.7g for one lora is too much we need to know how we can fix this
@superhachi5 ай бұрын
best tutorial so far!!
@gabrieljuchem5 ай бұрын
Thank you so much for this. I have a RTX 4060 Ti 8GB, do you think it's possible to train a SDXL LoRA with only 8GB of VRAM? If so, what settings would you recommend on Kohya? Thanks!
@anmolagrawal53585 ай бұрын
Same boat, am using OneTrainer
@gavinm575024 күн бұрын
i dont know what im doing wrong. when its done training it only makes config files, TOML, JSON no TENSORS files??? help
@kenshisaan22075 ай бұрын
at 2:55 can you train on other checkpoints like juggernaut?
@allyourtechai5 ай бұрын
Yes, definitely
@kenshisaan22075 ай бұрын
@@allyourtechai im sorry to bother you first thanks for the 1 click setup and reg. i appreciate and subscribe to your patreon at 10:10 you cut the reg file directory is it the new folder it made in final or its the original one with all the picture
@stugryffin361912 күн бұрын
11:12 that word isn't epic.. it's epoch.. pronounced ee-POCK. Otherwise, great vid. You're really suited to this. Good content delivered really well.
@Nikida186 ай бұрын
I'm training LORA in this moment, I hade to delete this scale_parameter=False relative_step=False warmup_init=False because i got error "returned non-zero exit status 1". Anyway, why the GPU usage is 0% around but GPU memory is 100%? I have a RTX4070 for laptop and i set Network Rank (Dimension) 32 ------------- Network Alpha 16. After 30 minutes i'm at 8%, this is what i see: steps: 8%|████▍ | 133/1600
@Nikida186 ай бұрын
Update. I had 12 of my own pictures and 4990 man's regulations images. After 5 hours It's finished creating only 2 tensor files, about 210 MB each one. I tried in fooocus and I didn't get good results. I will repeat training again by changing settings.
@El_Rey_Diamante4 ай бұрын
@allyourtechai can you do this in comfyui instead of kohya?
@allyourtechai4 ай бұрын
I should be able to do a guide on that :)
@daviduartep5 ай бұрын
Thanks for the amazing tutorial! Sadly I couldn't manage to get it working. I am getting "A tensor with all NaNs was produced in Unet." during generation. Loss also appear as NaN during training. It appears to be related to the optimizer, things work when removing relative_step=False, but model quality becomes very poor. A similar training with Prodigy optimizer worked though. I have a 4090. I recall I trained a Lora with these extra optimizer params for adafactor months ago when I had a 3080 and it worked.
@allyourtechai5 ай бұрын
Interesting, I’ll check it out as well
@omegablast20022 ай бұрын
i thought Adafactor was a self adjusting learning rate and you set it to 1 just like prodigy, can someone chime in?
@timeisupchannel3 ай бұрын
Hello, Is there any way to continue training after stop? Thank you!
@thedevilgames82176 ай бұрын
i click on start training but nothing happened
@allyourtechai6 ай бұрын
check your command prompt window for errors.
@indiffident28 күн бұрын
frustrating, keep getting errors RuntimeError: Distributed package doesn't have NCCL built in
@caseyj789456Ай бұрын
Hi. I have to remove all "extra argument" (12:53) otherwise it gives errors and do not start 🤨(Rtx 4070 ti super 16gb).
@webtrabajoscolombia41243 ай бұрын
muchas gracias muy valioso
@gkiss2030Ай бұрын
Is 8 GBs of (GDDR6) VRAM enough for this? Thanks!
@Beauty.and.FashionPhotographer5 ай бұрын
tried in in Pinokio on a mac where this Koya SS can be installed with one button . YET, as is the case with 99% of all ai apps , it does not work. Same settings. terminal gave me some errors after the start button was pushed and after it would stop its processes just a few second later . so its a dead end street and yet another useless phantom ai exercise .
@allyourtechai5 ай бұрын
What errors were in the terminal?
@Beauty.and.FashionPhotographer5 ай бұрын
@@allyourtechai no blip prompts in any of the generated TXT BLIP captions besides the name ,neysalora, (which is yours blove), but no description on any of the 120 images in their respective text files...so blip never really did anything...this is how it starts to go wrong, and where i can detect it being wrong, being myself a total NEWBIE and all, ...here the terminal : The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Users/akos/pinokio/api/kohya_ss.pinokio.git/app/sd-scripts/finetune/make_captions.py", line 21, in import library.train_util as train_util File "/Users/akos/pinokio/api/kohya_ss.pinokio.git/app/sd-scripts/library/train_util.py", line 46, in from diffusers import ( File "", line 1075, in _handle_fromlist File "/Users/akos/pinokio/api/kohya_ss.pinokio.git/app/venv/lib/python3.10/site-packages/diffusers/utils/import_utils.py", line 701, in __getattr__ value = getattr(module, name) File "/Users/akos/pinokio/api/kohya_ss.pinokio.git/app/venv/lib/python3.10/site-packages/diffusers/utils/import_utils.py", line 701, in __getattr__ value = getattr(module, name) File "/Users/akos/pinokio/api/kohya_ss.pinokio.git/app/venv/lib/python3.10/site-packages/diffusers/utils/import_utils.py", line 700, in __getattr__ module = self._get_module(self._class_to_module[name]) File "/Users/akos/pinokio/api/kohya_ss.pinokio.git/app/venv/lib/python3.10/site-packages/diffusers/utils/import_utils.py", line 712, in _get_module raise RuntimeError( RuntimeError: Failed to import diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion because of the following error (look up to see its traceback): Failed to import diffusers.loaders.ip_adapter because of the following error (look up to see its traceback): module 'torch' has no attribute 'compiler' 15:32:53-239016 INFO ...captioning done
@Beauty.and.FashionPhotographer5 ай бұрын
@@allyourtechai i did reply and paste the few lines of terminal here, but i am guessing that youtube deleted my reply here... ? the first issue is that BLIP captioning does not return text files with descriptions of whats in the images,... besides the LORA NAME which i decided on, in my cas "neysalora*,.... so only one word in these text files . terminal does say there was an erros with a loading of a diffusion file. The People at Pinokio , with which i installed this, tried to help , they are super great over there, amazing support,... yet it still does not work . i am going to assume that without image-descriptive words or content in these text files, beside neysalora, its never going to work, ...right?,
@uKnowMister5 ай бұрын
tried everything and if iam starting training i get the following error: 'Parameters: scale_parameter=False relative_step=False warmup_init=False ' is not a valid settings string. what is wrong?
@yourhighnessla5 ай бұрын
delete the mark before the P
@uKnowMister5 ай бұрын
@@yourhighnesslaI will try later again, but I double checked, there wasn’t a mark before the P. Maybe a imaginary one 🤣
@TyreII4 ай бұрын
Such a frustrating process for me. Got it all setup but when I hit the training button it throws an error message. I have reinstalled everything like 12 times. torch.distributed.elastic.multiprocessing.redirects: [WARNING] NOTE: Redirects are currently not supported in Windows or MacOs. [W socket.cpp:663] [c10d] The client socket has failed to connect to
@allyourtechai4 ай бұрын
It is a massive pain in the ass to be honest. I’ve done a half dozen of these guides and it’s still painful to install and use every time. I’m building a model trainer into PixelDojo to simplify this whole training process.
@PretendBreadBoy4 ай бұрын
Yeah it's a pain. It's frustrating when you install everything and just know it's not going to work lol.
@valentinotrinidad4 ай бұрын
Understand nothing in parameters tab, but it Fworks 🤣
@geoffreybirt88994 ай бұрын
Followed exactly and double checked. It is done in like an hour and I only have 1 LORA file...
@geoffreybirt88994 ай бұрын
Found the issue. Set max training steps from 1600 to 0.
@Ishmaam6 ай бұрын
Thank you so much for the useful video, My GPU is Nvidia 3060, can I put Value 256 as Network Rank?
@allyourtechai6 ай бұрын
How much vram does your card have? 256 requires at least 12gb
@Ishmaam6 ай бұрын
@@allyourtechai Thank you, Its a 12GB card.
@prismoZTN5 ай бұрын
I cant find my Safetensor files :(
@JaysterJayster4 ай бұрын
I have SDForge rather than automatic1111, would this work with that?
@allyourtechai4 ай бұрын
It should
@alexalves32932 ай бұрын
How can I train the model using the CPU? I know it's not ideal...
@allyourtechai2 ай бұрын
That would take an insanely long time if even possible. You can use something like pixeldojo.ai
@brunohof29724 ай бұрын
My character with the lora is pretty bad, its like what we got with ai two years ago. I guess I have to play with all available settings before training.
@flow94634 ай бұрын
No one cares
@allyourtechai4 ай бұрын
The prompting afterwards plays a major role as well.
@auzland152Ай бұрын
Fooocus is SDXL right?
@allyourtechaiАй бұрын
Correct
@silentsubz5 ай бұрын
Where are the epochs? Can't seem to find them.
@allyourtechai5 ай бұрын
They should be in your output folder specified earlier before you start the process
@jahinmahbub82374 ай бұрын
after clicking start training, getting an error saying accelerate not found. 01:46:14-810348 WARNING Regularisation images are used... Will double the number of steps required... 01:46:14-811350 INFO Regulatization factor: 2 01:46:14-811350 INFO Total steps: 18300 01:46:14-812351 INFO Train batch size: 10 01:46:14-813352 INFO Gradient accumulation steps: 1 01:46:14-813352 INFO Epoch: 4 01:46:14-814353 INFO Max train steps: 1600 01:46:14-815354 INFO stop_text_encoder_training = 0 01:46:14-815354 INFO lr_warmup_steps = 160 01:46:14-820872 ERROR accelerate not found
@allyourtechai4 ай бұрын
Accelerate is a PyTorch library and must be missing from your system for some reason. You should be able to install it manually though
@jahinmahbub82374 ай бұрын
@@allyourtechai I pip installed it. Still doesn't show up. How can I manually install it? And configure it?
@yoyosfsf90213 ай бұрын
i have rtx 4060 8GB . can i do that ?
@allyourtechai3 ай бұрын
12GB of vram is about the minimum for an XL model. You can probably train a stable diffusion 1.5 model
@greengenesis3 ай бұрын
I always get no data found... -.-
@better_vibes_level_23 ай бұрын
I tried EVERYTHING and I can't get the same results as you did. They must have update it again, 'cause I followed what you've done and I got only 2 final models insted 10 like you did. I'm getting sick of this kohya. I have a rtx 3060 12gb
@allyourtechai3 ай бұрын
They seem to break the workflow every couple weeks. It’s one of the reasons I built PixelDojo.ai . I wanted to be able to control the process and the quality
@vk28a126 ай бұрын
I'm getting "returned non-zero exit status 1" in the error log. Any insights?
@allyourtechai6 ай бұрын
are you loading stable diffusion 1.5 instead of SDXL? github.com/kohya-ss/sd-scripts/issues/1041
@vk28a126 ай бұрын
@@allyourtechai I was pretty sure that I set it to sdxl as in the video, but I'll try the whole thing from scratch and try again. I'll keep an eye on the model! Thanks. EDIT: I retraced the steps and made sure it was sdxl base, doubled checked the path as well. This time I paid more attention to the log and saw: "NotImplementedError: No operator found for `memory_efficient_attention_forward` with inputs:", some stuff about caching, and xformers. I tried to toggle a bunch of things related to caching, with no luck until I switched xformers to sdpa in the advanced section. Now I'm getting further than before, and it appears to be working!
@magneticanimalism74196 ай бұрын
Have you tried deleting the "Optimizer extra arguments" that he pasted as "Parameters: scale_parameter=False relative_step=False warmup_init=False" in his description. This worked for me.
@vk28a126 ай бұрын
@@magneticanimalism7419 I have not tried that, but I will. Thanks!
@vk28a125 ай бұрын
@@magneticanimalism7419 I have tried that. In the end I had to go with onerainer, as koyha just refuses to work. I used essentially the same settings in onetrainer and got some decent results. @allyourtechai Have you considered diving into onetrainer for tutorial purposes? It might be helpful for viewers like me that struggle to get kohya working.
@Fanaz106 ай бұрын
Does anyone know how to make this work on colab?
@Strawberry_ZA5 ай бұрын
Khoya is painful to get working - for w/e reason the optimiser extra arguments your provided were causing errors and preventing khoya from initializing training
@quinn4794 ай бұрын
should my gpu be making brrrr
@allyourtechai4 ай бұрын
Haha, yep
@agrocoding-ia6 ай бұрын
Someone is getting avr_loss = nan?
@daviduartep5 ай бұрын
Yes I am. My loras also produce all nans during generation. I fixed by removing relative_step=False, but generates a Lora with very poor quality.
@uncleMartinWithTheMountainBike6 күн бұрын
Nice tutorial but I just checked the kohya repo and it is made in the most idiotic way possible. If that's how python programming works I have lost all interest in learning it.