Is there a LCM sampler for "hires fix"? The hires fix is the one that takes a lot of time
@matheussoares113822 сағат бұрын
Thanks!!
@tonywhite4476Күн бұрын
So much work for so little return.
@Stephan8087 күн бұрын
where did you got those classification images?
@AI-HowTo7 күн бұрын
generated them from SD1.5 for women with clothes, that are different from the clothes that i am training
@Stephan8087 күн бұрын
@@AI-HowTo Ok! Thx!
@Stephan8087 күн бұрын
@@AI-HowTo Great tutorial! Do you have an tutorial for training poses?
@AI-HowTo6 күн бұрын
you are welcome, no i have not created any, but poses are often not trained, they are guided using Open Pose (dw) Control net, in which you present the skeleton and it gives you the pose perfectly including fingers, such as in kzbin.info/www/bejne/Z2TJmHV4fpKNjrc
@Stephan8086 күн бұрын
@@AI-HowTo thank you! I will definitly try that. Thats one of my topics to learn... but i want to train my ai-influencer for specific poses, where I dont want to do every image with control net and open pose... for a more convenient and efficent workflow. So the question for me is how can i achieve that in a trained loRa... or an own checkpoint...
@Piyush_Sharma--8 күн бұрын
Thanks Dude its solve my problem..
@AI-HowTo8 күн бұрын
Glad to hear that
@nicklynchard75948 күн бұрын
Hi, I have replicated your settings for Lycoris exactly, but keep getting a CUDA error. I have a 16gig 4070 ti super - so, I cannot imagine why it should be overloading as I have read of people running Kohya for Lycoris with much, much less power. Do you have a patreon or some way of tipping? I would be willing to purchase 1 on 1 help.
@AI-HowTo8 күн бұрын
I am sorry, not at the time being... you might want to check this video settings which might be of help kzbin.info/www/bejne/rHKVp2V_qsx-bNk and have more recent update on character training ... it even shows how to train 1024x1024 image sizes without running into CUDA errors ... usuaully this error happens if the image size is too large ... i did training on my 8GB RTX 3070 Laptop card without having this issue...but this worked only for SD15, but for 1024x1024 images i get CUDA error for SDXL ... you just need to recheck you image dimenstions once again to make sure that all are proper.
@lukasdaus172910 күн бұрын
Olivia Casta
@4848kitty13 күн бұрын
Hello! This was an extremely useful video! I'm currently a graduate student in ML and I'm working on training a model for style then using that in img2img. Do you think it's possible to use LoRA with img2img?
@AI-HowTo12 күн бұрын
great to know... yes ofcourse, especially for Video generation, the use of a LoRA can increase the consistency of the video generated... and depending on the denoising level, the output would change to become more affected by the LoRA or less.
@pouyab995216 күн бұрын
OMG THANKS BRO
@matanlevi445718 күн бұрын
Thank you!
@gustavosuarez794521 күн бұрын
Great tutorial! Can this be achieved with Comfyui ?
@AI-HowTo20 күн бұрын
Yes, absolutely, as both use the same Controlnet models, and the same principles, the application methodology changes... what matters is to understand what these models do then discover how they can be used in the target tool such as A1111/ComfyUI
@slookify26 күн бұрын
its always the same scene
@ElDespertar28 күн бұрын
Thank you so much for this super useful tutorial!
@AgileIntentionsАй бұрын
Hello. May I ask about your hardware? I have 4070ti and... my the speed of training is around 6-8 secs per it. I see your speed is around 3 it per second! Very interesting and curiously.
@AI-HowToАй бұрын
ti is more powerful than mine (laptop 3070) so you should get better speed than me, if you are training on same image sizes as me with this speed then xformers options might not be turned on, or the drivers require some update i think not, possibly.
@KiritoxNemesisАй бұрын
thank a lot
@mengwang-io7fwАй бұрын
Paid Sponsorship & Business Consulting,may i get your e-mail?
@AI-HowToАй бұрын
sorry, not doing that at the time being, thanks for offering though.
@banninghammaАй бұрын
I like the mentioning of "real person" when "Olivia Casta" is not a real person (it is a face-swapped character of a real, older model) LOL!
@AI-HowToАй бұрын
:), ya, i got that note from so many peopel, to be honest at that time i didnt know she was an Artifical character herself, I just wanted a sample data set with so many images that looks real, any way, things still apply.
@___x__x_r___xa__x_____f______Ай бұрын
Hi, the settings seem to have changed for adamw8bit at 0.0001. the model seems to overfit. have you noticed a change?
@AI-HowToАй бұрын
have not done any recent trainining, and usually the algorithm is fixed so learning rate changes are unlikely to been used differently, not sure, anyway, if you see things overfit quickly then using smaller number of steps could be better, besides, learning rate smaller than 0.0001 doesnt make much sense i think, so we usually consider increasing it not decreasing it to learn faster for instance...not sure if any recent changes in Kohya have made things different.
@divye.ruhelaАй бұрын
Subbed! Very good tutorial! I know this is an old video, but I had a few queries. Is it harder to create/ train a 'realistic character LoRA' if the original dataset contains AI generated images created on realistic checkpoints, instead of a real person's photos? I guess, what I mean to ask is, can a LoRA created using AI generated datasets achieve such realism? PS. Also, what would be the best checkpoint to create such an AI generated dataset? TIA!
@AI-HowToАй бұрын
yes as long as the training data set are of good quality and do not contain deformations in the eyes or fingers, even slightest deformation could be augmented after training if they repeated. as for the best check point, not sure unfortunately at the time being, previosuly for 1.5 i got the best results with majicmix v4, even for wester characters despite that the checkpoint was asian and for SDXL the Juggernaut XL , not sure now....I think in general the principles of training do not change overtime, so the video is still good to rely on for training.
@abdulrehmanrehan6734Ай бұрын
how to make it appear on windows pc
@AI-HowToАй бұрын
if you disabled and wanted to enable it again, you follow same steps and change the start up type to (Automatic) then click Start from services.
@vascocerqueiraАй бұрын
how do you do this on a mac? .bat is for windows correct?
@AI-HowToАй бұрын
not sure, but i think it is the same file but with .sh extension, and git reset is a command that is independent of the operating system.
@RatScalpАй бұрын
THANK YOU
@shitokenjpnАй бұрын
Little confusing here. You have skipped the part at image selection , as per your video one image from the first frame of dance video been used in imgtoimg. How did the generated image created all the exact same images as per frames of the video in animation?
@AI-HowToАй бұрын
I used one image to test the output only, once satisfied with the results, we go to Batch tab and just filled (input directory with source images, and output directory for the output) and batch will generated all the images based on the prompt/details on my test on the single image
@SithmaАй бұрын
I tried this but It didnt solve this problem I'm having, since I made a clean installation, I can't use lora anymore. I have a long list of errors for each lora i have in my folder
@AI-HowToАй бұрын
sometimes deleting the venv folder inisde A1111 installation could help solver lots of errors, some A1111 are buggier than others.
@dreamzdziner8484Ай бұрын
How could I miss this gem of a video for so long. Thank you so much for this mate💛🤝😍
@AI-HowToАй бұрын
Glad you find it useful, you are welcome.
@apnavu007Ай бұрын
I'm thinking a buy a Laptop 8GB Vram Will I be able to run a Stable Diffusion XL model?
@AI-HowToАй бұрын
Yes it is possible, but it will be slightly slower than you hope for, it can take 20 seconds and more for a 1024x1024 image using Forge UI or using ComfyUI and more ... currently with how AI stuff is heading, if you plan to buy something, you better save and buy 24GB VRAM, it is very expensive, but it is the only option that allows you to run everything such as Animate Diff models without suffocating on memory or suffering slow generation
@apnavu007Ай бұрын
@@AI-HowTo Then I'll just have to buy a PC good choice.
@AI-HowToАй бұрын
yes, PC is alot more practical, cheaper, and more powerful, avoid laptops unless you extremely need to move around often, even lower VRMAM RTX 3060 for PC is alot more powerful than its Laptop counterpart and has more VRAM.
@nicolaseraso162Ай бұрын
Hey bro, do you know how to install insightface in Automatic1111 (I use PaperSpace) in order to use the option of Face ID in IP Adapter?
@AI-HowToАй бұрын
not sure, for me it worked without any problems, just downloaded the IP Adapter Face id models into the Controlnet models folder, and the Face ID loras into the LoRA folder, and made sure Controlnet was upto date, and it automatically downloaded necessary extra models related to insightface such as buffalo_l , not sure, why some have troubles with this while others dont.
@sahilchowdhari5298Ай бұрын
my webui keeps showing timer on checkpoint and all lora/embeddinsgs/etc and tries to load all again on every start is there a fix for that?
@AI-HowToАй бұрын
I have not seen this issue, before, but it could be due to low memory in your PC which forces the A1111 to load/unload models on multiple stages, not sure, I suggest to try github.com/lllyasviel/stable-diffusion-webui-forge which has same UI as A1111, but has better memory management and might run faster and automatically detect best memory settings that could run on your PC.
@sahilchowdhari5298Ай бұрын
@@AI-HowTo thanks for reply it was working fine for months and broke out of nowhere will wait a week before fresh installing.
@AI-HowToАй бұрын
i see, you can try using Git pull (in the command line after set COMMANDLINE_ARGS= line as well, maybe that helps bring up any updates and fix the problem, if auto update is not enable in your installation
@solutionxpress23782 ай бұрын
Te amo, miles de turoriales en español y este fue el que me ayudo
@AI-HowToАй бұрын
great to know, thanks
@gboediman2 ай бұрын
Thanks SO MUCH - you saved my time!!
@Hshjshshjsj727272 ай бұрын
For better results we should use what? Splinter and what else did you say
@AI-HowTo2 ай бұрын
I think, for perfect consistency, Blender (3D Software) is the right tool, stable diffusion is not the one for perfect consistency of objects/faces/clothes, while one can still achieve good resutls with it.
@Hshjshshjsj727272 ай бұрын
Thank you. Do you have video on that you can add link yet? Also, would it allow me to create photorealistic portraits like for social media, as that is my goal.
@AI-HowTo2 ай бұрын
sorry, Blender is a 3D Free software tool, it is for games and realistic 3D stuff and not AI based.... for social media stuff such as creating realistic pictures, then Stable diffusion is really good to create photo realistic images, you can check other videos in this channel hopefully you find something useful, the only drawback for stable diffusion i think is just the Computing power (good graphics card) which makes it difficult for most people to really dive into it and it's capabilities.
@AI-HowTo2 ай бұрын
kzbin.info/www/bejne/rHKVp2V_qsx-bNk this video shows how one can create a LoRA model that can be used for instance for creating anything related to a specific photo realistic character ... two LoRAs could also be combined to create a new Character with features from both using specific ratio of both ... or one can use the same LoRA with 0.7 weight to get a different variation of the same character... there are no limits on what one can do .... if you were new to Stable diffusion then kzbin.info/www/bejne/iKXNdaatl7h4e7s could be a good starting point ... and you may want to check other creatos content such as www.youtube.com/@sebastiankamph Sebastian has lots of content about creating stuff or www.youtube.com/@OlivioSarikas Olivio start is also fun to watch ... my content contains just limited number of videos and mostly about techniques such as LoRA creation, training, or provides tools and techniques to create certain stuff that other creators may not have covered it properly or in depth.
@Hshjshshjsj727272 ай бұрын
@@AI-HowTooh excellent, thank you very much 😊
@monkeysit78262 ай бұрын
I have questions on photos pick up. For example, I want to create a real character with all different face angles, including face close-up and face with upper body. Because it seems like that if the training images include too much face close-up while only few face with upper body, the images generated afterwards with upper body will have failed face whereas close-up photos generation will be fine. So is the ratio, or portion of different kinds of image important to prevent overfitting of one type and increase photos diversity? In general, how many photos per each types and how many training steps per that type of image would give good flexibility of that type as well as good quality. to make it more understandable, let’s say I just need to create a good close-up 45 degree face and 90 degree side face for my whole project, how many photos and training step should I use in general.
@AI-HowTo2 ай бұрын
There is no rule... everything is purely experimental, even the creators of Stable diffusion do not know. I think that close up and upper body shots should have the same number to have a balance in training. we would usually have few full body shots such as 10% and most shots are protraits/close up/upper body You might be interested in watching this newer video kzbin.info/www/bejne/rHKVp2V_qsx-bNk which uses smaller number of images... as far as i have seen, full body shots are difficult to reproduce with high quality, this is why we often use After detailer to repaint the face (With prompt that has face with the LoRA inside) some people train only close up in a separate lora and upper body in another lora to get better results as for full body, since they will never function as perectly as one wants, they are best to have small ratio such as less than 10% (For realistic models in which details matter). experimenting is the key eventually, some models may work from first experiemnt, others might take 10s till you get something good, even use different regularization images could afffect the output greatly.
@monkeysit78262 ай бұрын
@@AI-HowTo Thanks you. It's helpful.
@dragongaiden19922 ай бұрын
Friend, you can do it with XL since it is very difficult to guide yourself if you use SD 1.5, basically it is doing everything differently from your video and I find many errors and deformed images
@AI-HowTo2 ай бұрын
true, XL is certainly better, but I still dont use it unfortunately on my 8GB video card.
@530214172 ай бұрын
I get this problem where Reactor wil skip any image/frame of a video that doesn't has a face in it. So when I put it back into video(output) the audio will be out of synch and video will be all choppy. is there anyway to solve this problem?
@AI-HowTo2 ай бұрын
even if the image doesnt have a face, Reactor will output the image (based on my batch image tests), not sure why this is not working out for you, you should do some tests for 5 images for instance to see where the problem happens
@Lell198620102 ай бұрын
Is there any possibility to use batch with openpose, making each image with a different seed?
@chiptaylor11242 ай бұрын
I really appreciate you for making this video. It solved my issue. Thank you so much.
@AI-HowToАй бұрын
not sure, but i think its the same file on mac but named webui-user.sh and you edit it in the same way and put the command there, if you go throught he .sh file it might give you some guidance
@rbdesignguy2 ай бұрын
Why not just crop in photoshop and save yourself a step?
@AI-HowTo2 ай бұрын
I think I did that at some point
@damned75832 ай бұрын
where do I download the ip_adapter_clip_sd15 processor?
@AI-HowTo2 ай бұрын
I think it is (ip-adapter_sd15.bin) ... all 1.5 models are in huggingface.co/h94/IP-Adapter/tree/main/models
@damned75832 ай бұрын
@@AI-HowTo I work with Google Colab, could you tell me which folder to place this file in?
@AI-HowTo2 ай бұрын
I think it should be the same as local installation folder --- which is the Controlnet model's folder, on my local installation that is stable-diffusion-webui\extensions\sd-webui-controlnet\models ... but i think A1111 also looks inside stable-diffusion-webui\models\ControlNet folder as well
@kakaoxi11 күн бұрын
@@AI-HowTo I downloaded that but i don 't have the preprocessor ip_adapter_clip_sd15
@AI-HowTo10 күн бұрын
it is fine, the developer might have removed the clip version or renamed it, you can use any other sd15 version and try them out.
@Damage232 ай бұрын
IT DIDENT WORK
@beanbean99262 ай бұрын
THANK YOU SO MUCH I'M AT MY KNEES KISSIN UR FEET FRL
@mothishraj44632 ай бұрын
Hey, I have two questions, 1) How did you get the image output for each epochs ? I'm getting only the tensor data 2) Can I train a color and a pattern (Leapord pattern Fabric) and use it on any garment ? (By eliminating anything related to leopord or animal pattern ?
@AI-HowTo2 ай бұрын
1) from sample images config section, as in kzbin.info/www/bejne/rXu7ZZV4qamlb7s we choose for instance 1 for epochs to generate 1 image each epoch, and we write in sample prompts the prompt that we want to dispaly, it must be written as shown in the sample text completed including the image size to dispaly 2) yes, as in kzbin.info/www/bejne/iIWVm51jatlrrpo training guide which is a style, this helps train styles/patterns rather than objects -- and yes we eliminate anything related to the pattern in the image description for the training images (leopord or animal pattern ) and keep everything else in the image description
@moulichand98522 ай бұрын
is there is any script availabe without using web ui?
@AI-HowTo2 ай бұрын
the web ui is built on top of python scripts, so everything in stable diffusion image generation or training is based on scripts, so they can be automated, but i have not used that unfortunately so i dont have enough expertise to guide you on that
@ReinhardMozes2 ай бұрын
Since when DreamBooth appears there???? I can't understand this :(((
@AI-HowTo2 ай бұрын
this video is from few months ago, back then, dreambooth training was easier on A1111 and i think it appeared on the GUI by default... if that is what you are asking about.
@Eustas53 ай бұрын
thank you so much bro
@AI-HowTo2 ай бұрын
You're welcome!
@--91993 ай бұрын
THANK YOU!!
@AI-HowTo2 ай бұрын
You're welcome!
@ricardoc94363 ай бұрын
Sorry but I don’t follow you. I think is a good video, but I don’t understand you very well…
@rustyMetal993 ай бұрын
i got rtx 3060 12gb vram but my reactor face swapping takes too much time, at least 1 hour for like 300 frames or so, please HELP!
@AI-HowTo3 ай бұрын
not sure, this could indicate a problem with onnxruntime installation possibly, or that you are using frames with large dimensions which results in longer processing time, check the frame dimension first, I have not faced this issue before this is why i cannot advise you more that on the matter, i didnt even had any troubles with onnxruntime installatino before... but when it is slow, then there is an issue.
@rustyMetal993 ай бұрын
@@AI-HowTo i really don't know and can't find a solution cuz i've tryed multiple command args but none of them reduced my generation time, imagine 9 seconds took 25 minutes (220 frames) my specs are : rtx 3060 12gb / i7 3th gen / 24gb ram / normal 1TB HDD. my command args : --xformers --no-half-vae
@AI-HowTo3 ай бұрын
I think this is a good speed for applying ReActor through A1111, 25minutes for 220 frames ... sometimes the size of the image used in ReActor also may affect the output... but 25 minutes it means it is using the GPU and I dont it gets faster than this using A1111.... anything related to AI or face swap can take a while to produce a decent result and creating a long video could take hours or days depending on the quality and length of the clip.
@AI-HowTo3 ай бұрын
my video card is RTX 3070GB Laptop , it can swap and generate using ReActor at speed of 15 images per minutes, the source video input image sizes are 720x1280, and i used 512x512 face input image in this test.
@rustyMetal993 ай бұрын
@@AI-HowTo thank you for your response and assitance, i think it's okay to be patient to get a 2 minutes video in arround 3 or 4 hours. better thank nothing. i was just asking myself how other famous apps give me results within few minutes even if the video is long, maybe they have multiple computers and they split the input for each computer and re-link the parts later as one video. MAYBE
@HPCAT883 ай бұрын
thanks. now let's scam some simps on OF
@Ziko6753 ай бұрын
I am confused about one thing. During captioning do we need to caption all the keywords that we want to train in a model or do we have to remove those tags if we want them to train in a model. Which one of them is correct?
@AI-HowTo3 ай бұрын
yes, anything that we want to be part of our LoRA must not be captioned... for example, if our subject always have green eyes, then we better not include that in the captions... if she has blonde hair, then we better not include that in the captions either ... we caption only the things that changes, for example: her clothes, the background...etc.... it is a bit confusing... but this is the best way to caption things and improve accuracy for the LoRA model.
@Ziko6753 ай бұрын
@@AI-HowTo hmm that’s interesting. So you saying that anything which is not consistent among images should be not be captioned/tagged. For example if I am training some background images as a style whose theme is neon light cyberpunk, so I should not caption cyberpunk or neon light as it will be consistent among the images but I could caption a group of people for example or tall building because they will not be always there
@AI-HowTo3 ай бұрын
exactly...we can also add a trigger word that refers to your style (which absorbs all things related to your style common features in all the images) for example which may help increase the chance of bring that style out when prompting too.
@Ziko6753 ай бұрын
@@AI-HowTo I think I understand some of it even if it sounds somewhat confusing. Thanks mate😀
@AI-HowTo3 ай бұрын
you are welcome.
@ManDogAndCows3 ай бұрын
i want to run this of a server i have in it a gt1030 only 2Gb wil it work? also has 64Gb ram and 2x 10 core CPU.. render time is no issue for me since the server works while i do something else i just want to utilise my server for something other than storage.. also a quadro p2000 fits in my server im thinking about upgradeing it has 5 GB
@AI-HowTo3 ай бұрын
it will be impractical to run on 2GB, with A1111 this might not work propertly, but Forge have better automatic memory management github.com/lllyasviel/stable-diffusion-webui-forge this repositry provides same things as A1111 with same UI but has better memory management and can run on 2GB for SD 1.5.... it might run SDXL too but it will use the CPU then which will be slow.
@ManDogAndCows3 ай бұрын
@@AI-HowTo yes slow these days is unusable.. the gt1030 was the dumbest purchase i have ever made 2gb idk. if it is the drivers but i cant get it to render or transcode anything..i found a quadro p2000 for cheap so i will run with that thank you for fast response
@AI-HowTo3 ай бұрын
You are welcome, these days, RTX graphics cards are game changers, they are the way to go for AI/Gaming/3D, they are expensive, but they seem like the only option to save time and be able to stay uptodate with the technolgoy...best of luck.