SDXL ComfyUI Stability Workflow - What I use internally at Stability for my AI Art

  Рет қаралды 79,793

Scott Detweiler

Scott Detweiler

10 ай бұрын

Since we have released stable diffusion SDXL to the world, I might as well show you how to get the most from the models as this is the same workflow I use on a daily basis at stability.ai. In this video I show you some of the basics on how to get the model from the models to generate your best AI artwork from our models. You will need some of the custom nodes over at civit, but you can choose the package that works best for you, as they are all pretty similar.
We will start with a basic workflow and then complicate it with a refinement pass, but then we will add in another special twist I am sure you will enjoy. #stablediffusion #sdxl #comfyui
Grab some of the custom nodes from civit.ai: civitai.com/tag/comfyui
Grab the SDXL model from here (OFFICIAL): (bonus LoRA also here)
huggingface.co/stabilityai/st...
The refiner is also available here (OFFICIAL):
huggingface.co/stabilityai/st...
Additional VAE (only needed if you plan to not use the built-in version)
huggingface.co/stabilityai/sd...

Пікірлер: 298
@TedWillingham
@TedWillingham 10 ай бұрын
I would love if you could go over some of those settings in advanced detail - like "oh, I fiddle with more conditioning steps when I want to X", etc. There are so many superstitious people out there giving bunk advice that your level-headed breakdown would be super valuable!
@sedetweiler
@sedetweiler 10 ай бұрын
Great idea! I will have to ponder where to start! :-)
@dxnxz53
@dxnxz53 23 күн бұрын
it blew my mind that you can load an entire workflow from the image! thanks for the great content.
@iiiCorrosiveiii
@iiiCorrosiveiii 10 ай бұрын
thank you so much, i've become really proficient with A1111 and moving to comfyui was a big switch, so your help with how the workflows work in comfyui has made it just as easy as using A111 for me.
@lakislambrianides7619
@lakislambrianides7619 10 ай бұрын
This is a great video congrats. Very informative very thorough and you left no doubts. Can't wait for the next step!
@sedetweiler
@sedetweiler 10 ай бұрын
More to come!
@henryabela2672
@henryabela2672 10 ай бұрын
Thank You so much! videos like these are a blessing and help people to get into it more professionally.
@sedetweiler
@sedetweiler 10 ай бұрын
You're so welcome!
@me.shackvfx5911
@me.shackvfx5911 10 ай бұрын
I've grown to understand and enjoy comfy UI more that the one i was using before thanks to your videos.I really appreciate you and the effort you put into making these tutorials. One of these days you can show us how to train sdxl 1 or it's lora with our faces . Thanks :)
@sedetweiler
@sedetweiler 10 ай бұрын
Great to hear! Training will be coming soon! Cheers!
@angryDAnerd
@angryDAnerd 10 ай бұрын
Excellent tutorial, thanks! I got SDXL up and running with the refiner. If you have the time I'd like to see you make a video explaining how Stable Diffusion works and explain exactly what the program is doing as it sends the data through the nodes in Comfy so I can have a greater conceptual understanding of what is happening. Believe me I could watch hours of technical stuff lol.
@Pfaeff
@Pfaeff 10 ай бұрын
Why are there width and height values in the CLIPTextEncoderSDXL and what is the difference between width and target_width and why is one of them 4096?
@courtneyb6154
@courtneyb6154 9 ай бұрын
Great questions and hopefully Scott can take the time to explain. Building out the workload is a great first step, but not knowing what everything does so that you can fine tune it is lame.
@shallowandpedantic2320
@shallowandpedantic2320 10 ай бұрын
Thanks. If you're looking for recommendations, a video focused on comparing upscalers and incorporating upscaling into this kind of workflow might help people. Seems like a nice next step. Appreciate what you've shared so far.
@rsunghun
@rsunghun 10 ай бұрын
I was waiting for it. These are very difficult for ordinary people to figure out how to use it. Thank you for the video!
@sedetweiler
@sedetweiler 10 ай бұрын
Glad it was helpful!
@ImAlecPonce
@ImAlecPonce 10 ай бұрын
Thanks!!! these boxes are actually starting to make sense
@sedetweiler
@sedetweiler 10 ай бұрын
Woot!
@imperfectmammal2566
@imperfectmammal2566 10 ай бұрын
Thank you so much! Even though I couldn’t understand much, it helped me get started with comfy.
@sedetweiler
@sedetweiler 10 ай бұрын
You’re welcome 😊 Just keep working with it and it will start to click into place.
@wschieber
@wschieber 10 ай бұрын
Thanks, Scott. I was really looking for something like this to get started with SDXL in ComfyUI.
@sedetweiler
@sedetweiler 10 ай бұрын
Glad it was helpful!
@lionhearto6238
@lionhearto6238 10 ай бұрын
all the info i was looking for. great video. thank you
@sedetweiler
@sedetweiler 10 ай бұрын
Glad to hear it!
@Ziixu
@Ziixu 9 ай бұрын
Great tutorial! This is my first time using ComfyUI and this video helped me a lot, tyvm!
@sedetweiler
@sedetweiler 9 ай бұрын
You are most welcome!
@PieterLaroy
@PieterLaroy 10 ай бұрын
My first steps into ComfyUI, and it's the kind of thing I really like 🙂
@sedetweiler
@sedetweiler 10 ай бұрын
Glad to hear it!
@Feelix420
@Feelix420 10 ай бұрын
I'm mind blown! never thought of using comfyui ever but seems like I'm sold over this video. very nice sir and thank you for sharing your knowledge
@sedetweiler
@sedetweiler 10 ай бұрын
Glad you liked it! It will also teach you a lot more about how things work, which I always feel is a good idea.
@Feelix420
@Feelix420 10 ай бұрын
@@sedetweiler absolutely! downloading it now hehe
@gbresaleking
@gbresaleking 10 ай бұрын
Thanks for this ! I was using comfy and using refiner at the beginning and it was coming up with really wacky pictures, now I can use it correctly thx
@14MTH3M00N
@14MTH3M00N 5 ай бұрын
Love your disgust for the negative prompts lists haha. relatable stuff
@sedetweiler
@sedetweiler 5 ай бұрын
(((((((((extra arms!))))))))) :-)
@JRis44
@JRis44 10 ай бұрын
This is getting fun! Cant wait to work on img to img tomorrow after work!
@sedetweiler
@sedetweiler 10 ай бұрын
Have fun!
@CMak3r
@CMak3r 10 ай бұрын
Prompt switching can be realized with additional KSampler that will render first steps with completely different prompt. For example you may want to create triangle composition, or a symmetrical image, and it can be done at early steps of a generation. Good for abstract art. And also I like that in ComfyUI it's seed can be fixed while base model and refiner will be generating on different seeds
@zacharykrevitt7560
@zacharykrevitt7560 7 ай бұрын
good idea! just tried this out and it worked in an ineresteing way. Essentially prompting an init image
@russellmm
@russellmm 10 ай бұрын
quickly becoming my goto channel. keep up the great work
@sedetweiler
@sedetweiler 10 ай бұрын
Thank you!
5 ай бұрын
Thanks Scott, as a beginner your videos are great. Very well explained and "easy to learn than a bunsh of others.
@sedetweiler
@sedetweiler 5 ай бұрын
Glad to help
@eliasdelreal3863
@eliasdelreal3863 10 ай бұрын
You are da man! Thank You so much for this tutorial!!!
@zerosequels
@zerosequels 8 ай бұрын
Thanks Scott, your tutorials are great
@sedetweiler
@sedetweiler 8 ай бұрын
Glad you like them!
@archielundy3131
@archielundy3131 4 ай бұрын
A million thanks for these. As finicky and frustrating as the program is for beginners, your calm expertise is just what's needed.
@sedetweiler
@sedetweiler 4 ай бұрын
Thank you!
@DarnSylon
@DarnSylon 9 ай бұрын
I kept to the two sampler idea but introduced separate prompts for the first sampler. It creates an interesting variation. Especially if you put different colors in each positive prompt. For example, make a white bridge and a black animal. I was hoping to create a background and a foreground. But the twist turned out pretty cool anyway.
@Yggdrasil777
@Yggdrasil777 10 ай бұрын
I have been in love with ComfyUI since I found it (coming from Unreal Blueprints, very familiar system). I am currently working out some torch issues with my current system, but I generate whenever I can. It is great to see you building out the workflow and explaining the nodes that you use and why. Very informative and THANKS for the tip with the shift-click to copy nodes AND connections. NICE!
@sedetweiler
@sedetweiler 10 ай бұрын
Great to hear! I am really happy with the nodes, but I hope they really update to things like docking, etc. Cheers!
@digitalbear3831
@digitalbear3831 10 ай бұрын
Same here since I come from Houdini, just love the node spagetti
@sedetweiler
@sedetweiler 10 ай бұрын
Yusss! I also used Houdini as well as Substance Designer and I am hoping to get into nested nodes here as well. Cheers!
@conquestofninjacats
@conquestofninjacats 10 ай бұрын
Love your videos :D just started using ComfyUI with SDXL. Having a lot of fun so far!
@sedetweiler
@sedetweiler 10 ай бұрын
Glad you enjoy it!
@nicolasmarques7460
@nicolasmarques7460 7 ай бұрын
Amazing tutorial, thanks for sharing!
@sedetweiler
@sedetweiler 7 ай бұрын
Glad it was helpful!
@potusuk
@potusuk 6 ай бұрын
yay finally got it working :) learning fast thank you
@sedetweiler
@sedetweiler 6 ай бұрын
Great!
@clonosaurios
@clonosaurios 7 ай бұрын
Thank you for your video! I learnt that comfyui is awesome :)
@sedetweiler
@sedetweiler 7 ай бұрын
You're so welcome!
@tomaseriksson5430
@tomaseriksson5430 8 ай бұрын
thanks for this! looks so crazy to a beginner but I followed and it's a great place to start generating images
@sedetweiler
@sedetweiler 8 ай бұрын
It really is! You have officially leveled up, as this specific type of workflow shows you how things work, not how to use a specific UI. Cheers!
@ColbstaD
@ColbstaD 6 ай бұрын
Thank you for this! I've created my own custom workflow based on this one with lots of inputs --> primitives to change stuff quickly.
@sedetweiler
@sedetweiler 6 ай бұрын
Fantastic!
@TomMaiaroto
@TomMaiaroto 10 ай бұрын
I'm new to ComfyUI all and really love your videos. Thanks! Maybe this is obvious to folks, but one thing I recently learned was the ability to condition after one KSampler ran so you can continue to refine your final image. It ended up being an alternative (or another tool in the toolbelt) to inpainting. I wasn't just refining, I was adding to or dramatically changing the final image - all without losing the "base" starting point that was all "locked down" in that the seed was fixed, the cfg and steps didn't change, etc. So it was a very non-destructive compositional workflow. If I wanted to add an object to the image, I could do that through a second prompt that was applied to a second KSampler. I could also introduce new LoRAs later on in those steps. I'm going to continue to experiment with this strategy and go through this more than once. So instead of a long prompt followed by a smaller corrective one, do more of a build up of prompts. Start simple and continue to add on to it so that elements within the image can be independently adjusted, removed, or re-arranged. Again, a more compositional approach during image generation to hopefully reduce the amount of work in post (or a series of very similar images that can be worked together in post processing). This could get a bit messy too, but maybe not if they are arranged left to right in a linear fashion building up the scene.
@sedetweiler
@sedetweiler 10 ай бұрын
That's great! It is a lot of fun adding into the pipeline. It's what we do internally as well when testing models and playing with new ideas. Cheers!
@lukeovermind
@lukeovermind 10 ай бұрын
fantastic! I am looking at some advance workflows, however with no real explanations how they work. I want to use it but I dont know what some of the nodes and flows do! However I found alot of value from your vids and at this stage I am happy to just play and learn comfy and put of creating art projects/ideas with SDXL for the time being. That 3rd Sampler is neat! I tried to see if you can use latent upscale method in your previous video with SDXL base and refiner, didnt work but that is the beauty of comfy! You get to try stuff
@sedetweiler
@sedetweiler 10 ай бұрын
I also think it is a pretty great way to learn how all of this works together. It really is limitless!
@paullangdon1852
@paullangdon1852 10 ай бұрын
Great video, very usefull. I am struggling to do a workflow from load image and the put through an ultimate upscale node.
@benkilowhiskey
@benkilowhiskey 19 күн бұрын
Great watch, thank you!
@MonzonMedia
@MonzonMedia 10 ай бұрын
Appreciate this Scott, you helped me fill in the blanks! I was wondering how the primitive nodes were used, it was driving me nuts! Hahaha! I was able to add an extra step to add an upscale process and it works very well! Looking forward to more.
@MonzonMedia
@MonzonMedia 10 ай бұрын
By the way, is there a way to create an image gallery somehow? Sort of like how invoke Ai is set up?
@sedetweiler
@sedetweiler 10 ай бұрын
I just use Bridge, since I already have an Adobe sub and it's better than most gallery apps. Comfy really isn't good for that type of thing at this point.
@novantha1
@novantha1 10 ай бұрын
Huh. I wonder what would happen if you had dedicated models for a variety of tasks (hands, eyes, hair, reflections, contrast, and so on) and fed a few steps from each of them in a daisy chain until you got to the first "true" sampler... Truly the possibilities are endless; thanks for the food for thought and the hard work!
@sedetweiler
@sedetweiler 10 ай бұрын
That's a great idea, and we do have those as loras. It's fun to combine them to help get what you want.
@tripleheadedmonkey6613
@tripleheadedmonkey6613 10 ай бұрын
That is an interesting idea. The multitude of experts approach is proving to be the more effective of what we have developed recently. Not too mention that you could also combine this with prompt blending syntax to ensure that each part of the processing is focusing entirely on one subject in the prompt while still maintaining an overall mixed composition. If for simplification purposes you set up 5 samplers, each with an equal number of steps, 4 for the limbs and 1 for the head/torso. Then you set up a prompt blending which focuses 20% of the processing on each limb etc. it may even have better results.
@tripleheadedmonkey6613
@tripleheadedmonkey6613 10 ай бұрын
And yeah using LORA chains would mean that we could have a separate model output for each limb, while maintaining the same initial model. Allowing for less resources used at the same time compared to multiple dedicated models.
@tripleheadedmonkey6613
@tripleheadedmonkey6613 10 ай бұрын
I think I'm going to play around with this now actually xD Minus the dedicated limb lora of course.
@Adreitz7
@Adreitz7 10 ай бұрын
Thanks for this look at the setup that Stability uses internally. I'm not so familiar with Comfy, but I've been using and enjoying SDXL through Invoke, which has a similar Nodes capability. I have a few questions and comments: 1. What are the Original and Target W/H actually doing for the CLIP conditioning nodes and what is the logic to setting those values? I played around with it, testing various combinations, and the only thing I could confidently say is that setting Original W/H smaller than 1024 causes the image to become blurry. I couldn't see any specific benefit to any other value, as I tried 1024, 4096, and 40960 for Original and between 64 and 40960 for Target -- setting different values made the image different, but not obviously better or worse. I settled on just setting them the same as the output image dimensions. 2. Why are there two prompt inputs for the base text encoder node when you provide the same input to both? Invoke calls one input the prompt and the other the style. What effects are caused by, e.g. separating your prompts into a prompt and a style and sending them independently to the two inputs, switching the inputs (so prompt goes to the "style" input and vice versa), setting them both the same, or leaving one or the other blank? I've found that if I prompt the base model for a roller coaster in the first input, I get a roller coaster. But if I prompt "roller coaster" for the first input and "photograph" for the second, I get anything BUT a roller coaster -- ruined buildings, abstract paintings, etc. 3. Connected with #2, Invoke's refiner conditioning node only includes a "style" input, but I've found that only giving it a style prompt can cause the refiner to do weird things (like making architecture look like it's made of tent fabric). 4. You've indicated that initializing the noise with the refiner is an interesting idea, which it is, but have you seen any consequences other than just making the images different? Does it provide any actual benefit? 5. I've experimented with higher resolution SDXL generations. I'm on a Mac and there are some apparent generation bugs with Invoke on MPS (about 1856 square and above it becomes debilitating). But I've noticed that my scenes at higher resolution (photographic sci-fi style architecture) tend to become wide angle and taken from a high vantage point, almost as if the resolution setting is correlated with the position and zoom of the virtual camera. Has Stability done any experiments at higher resolutions than 1024x1024? 6. Is there a benefit or danger to sending the same noise seed to both the base and refiner?
@bobbyboe
@bobbyboe 6 ай бұрын
Good questions... I also would like to know the answer. Did you understand the concept of why there is a field of dimensions in a node that is supposed to provide only text?
@larryross9380
@larryross9380 Ай бұрын
Perhaps things have changed since this was published nine months ago, because this workflow just gave me dark, abstract images. But I learned a lot about how to build out a workflow! Thanks!5
@hleet
@hleet 10 ай бұрын
WOW ! that's a super tutorial of ComfyUI there ! Thanks. I never know that there was this new addition of clipnode for SDXL ! The only drawback that I find in ComfyUI is the way it manage the workflows. I mean when you want to change your original workflow, you need to save a local file, and if you want to do something else (like inpainting) you have to redo ALL your workflow and save it to a file to recall your workflow and switch by loading one workflow or another depending on what you want to do. Definitly not fond of this way of managing workflows. They could have done some kind of "favorite" workflow. Like 5 or more "workflow ready" that you could custom afterwards and save your "favorite custom workflow" and switch whenever you like. it would skyrocket the use and adoption of comfyui !
@sedetweiler
@sedetweiler 10 ай бұрын
I just drop the json you get from using "save" into the interface and it loads. But,*do agree that would be nice.
@hleet
@hleet 10 ай бұрын
@@sedetweiler ooh ! Nice another tip ! Drag and drop the json just works too ! I might be able to explore more versatile stuff with comfyui now :)
@badmojjo
@badmojjo 10 ай бұрын
I am definitely going to search for a good upscale workflow on your channel.
@sedetweiler
@sedetweiler 10 ай бұрын
It's coming this weekend!
@reekster30
@reekster30 9 ай бұрын
wow - great tutorial dude. I've only recently got into comfy and wondered why all the controlNETs were failing last week :D All new ones install thanks to your videos and loving all the sdxl videos... fun times ahead (but I really need a pc gaming rig for speed) haha Out of interest - what kind of set up for a pc would you recommend for quicker generation/processing? massive 128gb RAM and like a RTX4090? :D thanks for your videos - amazing
@Aaabii
@Aaabii 10 ай бұрын
Thank you very much. ı prefer comfyUI over A1111 and you are my go to channel for my purposes.
@sedetweiler
@sedetweiler 10 ай бұрын
Happy to hear that!
@matthewharrison3813
@matthewharrison3813 10 ай бұрын
Thanks for the great video. Could you please talk more on the clip encoder width and height and target width and height? What do they do and is there any documentation? Why are you using a different value for the target than the base?
@4richis
@4richis 7 ай бұрын
I would love to see and answer to this as well
@DarnSylon
@DarnSylon 9 ай бұрын
When you added the third or 'pre-sampler', why did you not pass the noise information as you had done with the first of the two samplers? I messed with that setting on the first two and didn't notice much of a change. Thank you for the videos and instructions. They are extremely helpful. And you suggest not to add things like extra fingers to the negative prompt. What is your method of not getting extra fingers or limbs, etc?
@MikevomMars
@MikevomMars 10 ай бұрын
Using SDXL with a 2060 SUPER 8GB + ComfyUI and it works great 👍Turned out that Comfy is 15x(!!!) faster than A1111 for the same tasks using SDXL! Also I never got any errors with ComfyUI while A1111 always gives me "NaN tensor" errors when working with SDXL. For SDXL, Comfy is a MUST! So I`m looking forward for more tutorials on ComfyUI.
@sedetweiler
@sedetweiler 10 ай бұрын
Great to hear!
@technoprincess95
@technoprincess95 10 ай бұрын
I am on the opposite; I am only be able to use Colab. When using A111, it's very fast and smooth, allowing me to work with videos and other tasks, including upscaling to 8k,comfuyi just pictures. However, with Comfuyi on Colab, I encounter issues such as disconnects and running out of ram after just 5 images, im using workflow of olivio sakiras, using base ckpt sd1.0 + refiner sd 1.0 . overheard refiner consumes lots of Ram
@MikevomMars
@MikevomMars 10 ай бұрын
@@technoprincess95 I`d double check your workflow used then. Also did you try to run ComfyUI locally? You don`t need a decent GPU to do so. Even an old 8GB NVIDIA will do it.
@A.polon.i.a
@A.polon.i.a 2 ай бұрын
Great video Scott, I wonder could you explain how to change the image size? What do I have to alter to produce an image of 832 x 1216 for example? Or point me to a future video that explains it, as I'm only on ep.2 Thanks💖
@demoran
@demoran 10 ай бұрын
So I was following the guidance here, and found that UniPC and the 2M variants will barf on you when the refiner steps are higher than the base steps. I tried with the 12/20 pair you've demoed here, and got an image with nasty vertical streaks in it. It was fine at 20/20, but barfed again at 20/50.
@San4itos
@San4itos 7 ай бұрын
You set Base KSampler to return the leftover noise but there is no leftover noise because it does all of its steps. Then Refiner adds its own noise and process it further. You may see it in Base preview. I guess if you turn off the leftover noise from Base, result is gonna be the same. What you need to do to pass the leftover noise to the Refiner is to use for example 20 steps but end on step 12. Then disable add noise function on Refiner KSampler.
@florentraffray1073
@florentraffray1073 2 ай бұрын
Thanks for these tutorials, great to have an in depth dive into the UI. I'm a little confused about the start/end steps and steps in the KSampler. In your second sampler in the chain of them, if you start at step 3 and do 12 steps, wouldn't that leave you at step 15 for your starting point in the next one?
@sedetweiler
@sedetweiler 2 ай бұрын
There are some advantages to skipping steps in some cases. It all has to do with the residual noise.
@MisterKerstov
@MisterKerstov 10 ай бұрын
Thank for this really concise and helpful tutorial. Just one thought, you did not enable the "return with leftover noise" for the "initial conditioning" node. Wouldn't it make sense to do so?
@sedetweiler
@sedetweiler 10 ай бұрын
It actually returns so much that things go sideways. Give it a try. I have not found that to work well.
@renderartist
@renderartist 10 ай бұрын
Thank you! 👍🏼
@sedetweiler
@sedetweiler 10 ай бұрын
You're welcome!
@JohnSundayBigChin
@JohnSundayBigChin 8 ай бұрын
Hi Scott, im rewatching the whole series again, you have done a good job. I have a question in this particular episode with the Sampler...why do you have the possibility of using the denoise within the KSampler but not with the advanced KSampler? Do they work differently?
@sedetweiler
@sedetweiler 8 ай бұрын
it was to simplify things. when you start at a later step with the advanced sampler, you are "skipping" some of the pieces you do not want to denoise, so it is the same thing but harder to explain.
@JohnSundayBigChin
@JohnSundayBigChin 8 ай бұрын
When you make Img2Img in one of the videos I saw that you used the common Ksampler because you needed the denoiser. Now everything is much clearer to me, thank you very much for answering.
@vanstonecypher
@vanstonecypher 10 ай бұрын
Thank you so much for the tutorial. It really helped with some basic knowledge that was not obvious as a new user of ComfyUI (double-clicking to get a list of nodes for example.) A couple of question though. Why are the width and height of the CLIPTextEncodeSDXL nodes are set to 4096? What does this mean since the output is still 1024?
@sedetweiler
@sedetweiler 10 ай бұрын
It is the resolution CLIP was conditioned at prior to scaling. I tend to use it and prefer the result.
@ysy69
@ysy69 7 ай бұрын
ComfyUI is truly about fine tuning the way one approaches the creation of an image using AI
@sedetweiler
@sedetweiler 7 ай бұрын
I agree!
@courtneyb6154
@courtneyb6154 9 ай бұрын
Excellent video Scott. If you could do some of us a favor and go into detail about what everything is and how it works within the cliptextencode nodes then that would be of tremendous value. I have scoured the net and am only able to find limited info about the options and nothing i have found has explained how or why they work. Building out the workflow is a great first step but not knowing how to fine tune is lame 😂Thanks!!!!
@sedetweiler
@sedetweiler 9 ай бұрын
Sure thing!
@ysy69
@ysy69 7 ай бұрын
Very helpful. Have you experimented with learned with using multiple Ksamplers? Are you still keeping at least 3 at different steps today as part of your workflow?
@sedetweiler
@sedetweiler 7 ай бұрын
I use 2 most of the time.
@ysy69
@ysy69 10 ай бұрын
Hi Scott, really appreciate your giving us the most recent update on SDXL. Do you know how to fine tuned a model using SDXL 1.0 and Dreambooth? Is this something you can create a tutorial video for us?
@sedetweiler
@sedetweiler 10 ай бұрын
That is coming soon. It is going to be easier to train, results wise, but still getting methodology together.
@ysy69
@ysy69 10 ай бұрын
@@sedetweiler 🙏🙏looking forward to… do you know if the new dataset should be set at minimum at 1024 by 1024?
@GamingDaveUK
@GamingDaveUK 10 ай бұрын
Very useful video and great to see videos that are not automatic 1111 lol
@Smashachu
@Smashachu 7 ай бұрын
Hmm i'm messing around with rendering the first 2-3 steps as something that i know SDXL is trained very well in so for example a brown horse racing for a positive prompt on the first 3 frames, then using a negative prompt for the Brown, with the new color being purple with a (purple horse:1.3). It's been working very well especially for harder to generate things, it's like it's erasing the colors and redrawing it now that there's a rough shape. I'd love to see how it will workout in combination with controlnet to maintain consistency in textures and shapes.
@sedetweiler
@sedetweiler 7 ай бұрын
That method can also help with LoRA images that are not as strong as you prefer. It's a great workflow. 🥂
@Aksuloid
@Aksuloid 9 ай бұрын
Would it be possible for you to make a video explaining the ideas behind negative prompting? You've mentioned it twice in the ComfyUI playlist now, and I am quite interested to know why and what one would actually put in there.
@sedetweiler
@sedetweiler 9 ай бұрын
It was really only an issue with SD 1.5, and not so much with SDXL. Consider for a moment that the model has no clue how many fingers a person has, so using a negative prompt of ((((((extra fingers))))) is just silly. People making it higher priority with all the "(" is like yelling English to someone that doesn't speak the language, cause that helps. The model was not trained on what "extra fingers" are, and that is just one example.
@Aksuloid
@Aksuloid 9 ай бұрын
@@sedetweiler That makes alot of sense, thank you. From what I've gathered then, its going to be more useful for removing unwanted objects/themes/concepts from an image (for example), or steering the model away from unwanted styles, instead of fixing small anatomical mistakes, etc.
@case0h
@case0h 10 ай бұрын
thank you!
@sedetweiler
@sedetweiler 10 ай бұрын
You're welcome!
@TomSweeney-ov8qs
@TomSweeney-ov8qs 9 ай бұрын
Do you have any videos (or recommendations for other videos) that go in depth on debunking the negative prompt urban legends you mention?
@sedetweiler
@sedetweiler 9 ай бұрын
No, but I should make one. It's just terrible what people pass on as the perfect negative. Do they think the model was trained on "bad anatomy" and "extra fingers?"
@user-kn6vx2xz4e
@user-kn6vx2xz4e 9 ай бұрын
very good
@sedetweiler
@sedetweiler 8 ай бұрын
Thanks
@skylightikab443
@skylightikab443 7 ай бұрын
Thanks for no nativ english speaker this was a good tutorial. It was very helpful! :)
@sedetweiler
@sedetweiler 7 ай бұрын
Glad you enjoyed it!
@deafponi
@deafponi 4 ай бұрын
Hi there Scott, thank you for the excellent tut. I must admit though, my robots did not look anything close to how refined yours came out. I wonder if I missed anything somewhere...
@tripleheadedmonkey6613
@tripleheadedmonkey6613 10 ай бұрын
One question I had. Is there any reason why you recommend using the VAE from the refiner, when there is only 1 version of the VAE (barring custom fixes for FP16) publicly available? If I choose to merge the fixed FP16 base VAE with the refiner, am I getting the same experience as you are (besides fp16-fp32 differences) ?
@appolonius4108
@appolonius4108 10 ай бұрын
Another great video. I learned a lot doing this, but could you build off of this one and show how to add a lora to this setup? thanks :D
@sedetweiler
@sedetweiler 10 ай бұрын
Sure thing!
@appolonius4108
@appolonius4108 10 ай бұрын
@@sedetweiler oh! I forgot, maybe also, if you could, do an upscaler node tutorial? These would be very helpful.. I think that would be all basic stuff. These tutorials are well done and easy to follow. They really help me understand all this stuff lol.
@sedetweiler
@sedetweiler 10 ай бұрын
@@appolonius4108 I did schedule a video for tomorrow that will cover basic img2img. That is the start of the basic steps. I will get an upscale video done super soon.
@appolonius4108
@appolonius4108 10 ай бұрын
@@sedetweiler that is awesome news, I'll be on the look out for those! Thanks, you rock!
@adam4adam890
@adam4adam890 10 ай бұрын
For some reason I always get an "overcooked" or flat images from refiner, in some cases it takes good result from the base and just removes the depth from image and turn it into drawings. Regardless very helpful information, TY
@sedetweiler
@sedetweiler 10 ай бұрын
It is a balancing act. You might want to minimize the steps in base to around 12 and also be sure to send over the remaining noise.
@karenreddy
@karenreddy 10 ай бұрын
Even the simpler setup is convoluted. I've worked with shader graphs, so it's alright, but I can see how this has a bad learning curve for many. I just don't see the big gain in using this setup for this utility quite yet. Stability should allow it to be "baked" into a simple GUI, so you can create a front end with different graphs, then not mess with it much, unless you want to add more pieces to the front end. Saving this front end would allow it to be shared with beginners and make it easy to get into, the complexity would be hidden until they're ready to explore. What is the advantage to conditioning with the refiner first?
@kabochsaygo6976
@kabochsaygo6976 8 ай бұрын
Is it possible for you to do a tutorial showing the ComfyUI ->Models folder structure, and what goes into each of them? I manually installed the manager with no issues. But other things such as diffusers, embeddings, clip_vision, etc. are unknown to me. And a lot of things on huggingface can't be found within the manager. Thanks. PS: Just getting started with SDXL and using Comfy. So going through your videos one at a time.
@davewills6121
@davewills6121 9 ай бұрын
Excellent tutorial, learned a whole lot in a short time. Why is it that the Refined images are indeed sharper, they seem to loose some of the more acute details. EXAMPLE: undersea shot without Refiner shows a murky underwater world with subtle light refraction and a sense of DOF, refiner seems to strip that away, leaving a sterile shot with little atmosphere?. Kudos
@sedetweiler
@sedetweiler 9 ай бұрын
There is some balancing going on for sure!
@davewills6121
@davewills6121 9 ай бұрын
@@sedetweiler Im just an eager noob getting his toes wet. PS: Can i trouble you for one question: If i pick one image from say my ''history'', can i build from that, run further batches based off of that one image, so that i can fine tune my results?
@parsley8188
@parsley8188 10 ай бұрын
Nice
@sedetweiler
@sedetweiler 10 ай бұрын
Thanks
@spiralofhope
@spiralofhope 7 ай бұрын
I was able to follow the tutorial well. I'm a bit confused at the three separate seeds. I can adjust the first (the conditioner/initializer) and get changed results, do I care about the others? In a previous video you said it wouldn't matter much for that context. Is that also true here?
@dreaminspirer
@dreaminspirer 10 ай бұрын
thanks so much for the video. I'm having BASE Steps and TOTAL Steps Primitives. So I'm trying to use a Primitive node to feed the PRERUN steps to 1st Refiner (let's call it PRERUN KSampler) but i bumped into a problem. - Feeding "steps" into PRERUN Ksampler is fine but I can not feed this "steps" INT to "start at step" for the BASE KSampler . they're both INT, but perhaps ComfyUI considers "steps" and "start/end at step" are different types. 😒 - The other way around is feeding "end at step" for PRERUN and feeding this value to "start at step" for BASE and feed all KSampler with same "steps" value. But for some reason, the PRERUN Ksampler needs to be fed with exact amount of steps otherwise the result is nothing but NOISE. 😒 please help , thanks again.
@sedetweiler
@sedetweiler 10 ай бұрын
I have also noted that, and I think it is a bug. That should work just fine. I got around it by using a math node, since that was the end goal anyway.
@dreaminspirer
@dreaminspirer 10 ай бұрын
@@sedetweiler thats exactly what i found. Derfuu VAR nodes and MATH nodes did the trick without any problem. Having said that, i found PRERUN step should not be more than 3 or it's all crap :) Thanks again and pls keep sharing with us the quirky tricks to play with Comfyui
@ethanhorizon
@ethanhorizon 29 күн бұрын
Thanks for the tutorial! Is the "noise seed" in Ksampler Advanced same as "seed" in Ksampler? You set noise seed as 4, what's the meaning of the number? What if I left it as zero?
@martdiamond6109
@martdiamond6109 10 ай бұрын
Do you also use an upscaler with SDXL? All the Comfyui examples I've seen never include it, so I'm just wondering how that would look in this workflow?
@sedetweiler
@sedetweiler 10 ай бұрын
Ya, you can use any upscaler and use them repeatedly. It's way more flexible than AUTO1111. I will do a video on this super soon. Cheers!
@ianwilliams7740
@ianwilliams7740 8 ай бұрын
on that third sampler you added you kept the return with leftover noise to disable.. does that mean you use up all the noise in those 3 early steps? what's the thought in not setting that to enable??
@97DarkSkull
@97DarkSkull 10 ай бұрын
How would you process sequences of images (I'm going or similar) with Confyui?
@lechatsportif124
@lechatsportif124 5 ай бұрын
Couple of things, isn't it recommended for the refiner to actually be started at 80% of total steps? Also, is conditioning via the refiner really a thing or did you just kind of mess around with it? You didn't select pass on noise, so I'm not sure what that means. Thank you for the tutorials, they are great!
@97DarkSkull
@97DarkSkull 10 ай бұрын
How would you precess sequences of images (I'm going or similar) with Comfyui?
@Zizos
@Zizos 9 ай бұрын
I just did download the official base and refiner but it seems I've got the VAE version from somewhere else in the past. What's the difference? I get that the VAE is built-in to the model. Does this mean you get to delete the VAE Decode node or some other node? Can you just keep the VAE version and follow your workflow with no difference in results or at least no negative results in quality? As in the last step you showed you can 1st generate a blank latent and then into the base and refiner... Seems like you can do all sorts of tricks like that to experiment with the resulting image. I wonder if it makes sense. If I get it right, it seems that the latent creates a base noise ignoring the models so that you can just get something a bit out of the box (model). Is that right? Thank you for the tutorial. I have lot's of stuff to learn.
@bigbo1764
@bigbo1764 9 ай бұрын
I’m curious, how would I implement a lora in this setup? I tried inserting 2 lora nodes after the checkpoint nodes and connecting them like I would in SD 1.5, but it seems to not be registering the existence of my Lora and just skipping over it. My checkpoints are connected to the Lora nodes only, except for the VAE, which is used for the decoding, what am I doing wrong and how exactly do I fix this?
@dkf-nl1703
@dkf-nl1703 9 ай бұрын
@14:20, doesn't the second sampler go up to step 15? And as a result, shouldn't the third sampler start at 15? And thanks for a great video!
@sedetweiler
@sedetweiler 9 ай бұрын
They are exclusive, the step start is correct.
@Darkwing8707
@Darkwing8707 10 ай бұрын
Why did you choose 4096 for the height and width in the conditioners?
@digitalbear3831
@digitalbear3831 10 ай бұрын
I'd like to know that one too
@AdamDesrosiers
@AdamDesrosiers 10 ай бұрын
also would like to know what these conditioners numbers do. And somehow, I've been happier with outputs when I set those number to 2048. But why? I don't know what those are doing.
@sedetweiler
@sedetweiler 10 ай бұрын
The refiner was initially conditioned at that size prior to scaling, so we tend to use that size.
@jonnyfat
@jonnyfat 10 ай бұрын
@@sedetweiler Thanks for this tutorial - great reference. Great to have tutorials on this by someone who knows what they're talking about :-) I picked up on the size thing too - so it's 4096 for the base and 1024 for the refiner? Thanks!
@petec737
@petec737 3 ай бұрын
@@sedetweiler "we tend to use that size" isn't really an answer. The only reason you'd have those numbers different is if you want to CROP a portion of the image..so in your case it's like wanting to crop out a 4096x4096 OUT OF a 1024x1024 image; which obviously is not how math works :)
@gloorbit5471
@gloorbit5471 6 ай бұрын
Being that this video is now four months old can I assume that your checkpoint is now named differently? The one I have that was downloaded when I installed Searge's script yesterday has the vae included in the filename like so: sd_xl_base_1.0_0.9vae.safetensors and a refined one named respectively.
@sedetweiler
@sedetweiler 6 ай бұрын
Sure, feel free to rename them. I do because they all have generic names and need to be changed to keep things sane.
@TissaUnderscore
@TissaUnderscore 2 ай бұрын
Should i use a refiner for a custom model? for example if i use juggernaut xl?
@jaredbeiswenger3766
@jaredbeiswenger3766 10 ай бұрын
I'm curious what's happening with your 2nd refiner when it starts at step 12 while the base model is also running to step 15. Are the 2 models alternating steps (acting simultaneously) or do they still run discretely? I'm curious if the starting step is logically useful or if it's straight voodoo magic.
@m4dbutt3r
@m4dbutt3r 9 ай бұрын
Yes I was just going to comment that the math does not seem to add up in that 3 sampler version at the end (first refiner: start 0, steps 3; base model: start 3, steps 12; 2nd refiner: start 12 [??? why not 15??] steps 20). I tried it at both 12 and 15 and actually liked the 12 better, but that may have been a coincidence and in fact it doesn't really matter. Very curious what is actually happening if you "mess up" these numbers. If I really mess with them, most of the time it comes out black and white (In one iteration I forgot to change the numbers when I copied over the second refiner to make the first, and I got beautiful, but black and white, versions of my images!!). Voodoo magic indeed.
@jaredbeiswenger3766
@jaredbeiswenger3766 9 ай бұрын
@@m4dbutt3r appreciate this. Would be nice to know if this power can be harnessed for good
@alexlindgren1
@alexlindgren1 7 ай бұрын
I'm aware that SD don't take account of spatial relationships, but I want to be able to replace for example a sofa in an existing image with an image of another sofa, but not sure on how to take on that challenge with SD, do you have any suggestions where to start? I don't want to manually mask each image, but I want the AI to recognize what part of the image is a sofa and mask it for me, I should just provide the image of the sofa and the "base image" of the livingroom.
@GuitarWithMe100
@GuitarWithMe100 10 ай бұрын
Im still confused on what the ClipTextEncodeSDXL does? and how does the value 4096 affect it?
@sedetweiler
@sedetweiler 10 ай бұрын
That was the initial conditioning prior to scaling, so we just prefer that for the refiner.
@kenjix7316
@kenjix7316 9 ай бұрын
why exactly dont we change the return leftover noise for the first sampler (the 2nd refiner one added at the end)?
@imperfectmammal2566
@imperfectmammal2566 10 ай бұрын
Can you tell me how to use the Loras offset that came with sdxl in comfy
@sedetweiler
@sedetweiler 10 ай бұрын
Yes, I will post a video on that and it is SUPER easy to do! Cheers!
@RobertWildling
@RobertWildling 5 ай бұрын
Hmmm... at around 14:15, when you add the first refiner with the 3 steps, shouldn't the last refiner's "start_at_step" be changed to 15?
@scottmahony4742
@scottmahony4742 4 ай бұрын
models, the refiner, etc. Where can I find definitions for all these variables?
@technoprincess95
@technoprincess95 10 ай бұрын
Would you mind sharing this workflow through a gdrive ❤
@lioncrud9096
@lioncrud9096 10 ай бұрын
any tips on adding an upscaler?
@sedetweiler
@sedetweiler 10 ай бұрын
Upscaler video is out today! Woot!
@lioncrud9096
@lioncrud9096 10 ай бұрын
thanks!!@@sedetweiler
@gameplayfirst6548
@gameplayfirst6548 10 ай бұрын
How is there any noise left during handover to the refiner, if you don't use the "end_at_step" parameter? Don't you get images without any noise from the base sampler if you don't limit the end in any way? Your base preview image confirms that you don't have any noise left after the base, which doesn't match the workflow described in the SD-XL documentation. And why do you overlap steps? For example you do 12 steps in base, but start at step 12 in refiner, instead of starting at step 13.
Cat story: from hate to love! 😻 #cat #cute #kitten
00:40
Stocat
Рет қаралды 16 МЛН
How many pencils can hold me up?
00:40
A4
Рет қаралды 19 МЛН
ComfyUI - Hands are finally FIXED!  This solution works with all models!
12:17
ComfyUI Multi Person Masking With IPADAPTER Workflow
12:31
Grafting Rayman
Рет қаралды 6 М.
ComfyUI: Yolo World, Inpainting, Outpainting (Workflow Tutorial)
37:46
Why You're Prompting Wrong, Do This (Per Leonardo AI)
18:31
metricsmule
Рет қаралды 24 М.
Animations with IPAdapter and ComfyUI
16:07
Latent Vision
Рет қаралды 31 М.
Magnific AI Upscaler Free Alternatives! Krea and Comfy UI Workflows
10:21
From Stills to Motion - AI Image Interpolation in ComfyUI!
11:32
Nerdy Rodent
Рет қаралды 29 М.
ЖАПАНДАҒЫ ЖАЛҒЫЗ ҮЙ / ФИЛЬМ / 02.06.2024
39:03
Жетісу телеарнасы
Рет қаралды 148 М.
Spiderman And Hulk Human real - Good Friend and Bad Friend#spiderman
1:00
Dad takes baby too unreliable# husband and wife humor funny video
0:16
Country Humor TV
Рет қаралды 2,4 МЛН