Don't make these 7 mistakes in Stable diffusion.

  Рет қаралды 89,255

Sebastian Kamph

Sebastian Kamph

Күн бұрын

Top 7 mistakes people make in Stable diffusion. Check out the bonus content! Which joke did you like best?
Chat with me in our community discord: / discord
Support me on Patreon to get access to unique perks!
/ sebastiankamph
Ultimate Stable diffusion guide
• Stable diffusion tutor...
The Rise of AI Art: A Creative Revolution
• The Rise of AI Art - A...
Ultimate Animation guide in Stable diffusion
• Stable diffusion anima...
Dreambooth tutorial for Stable diffusion
• Dreambooth tutorial fo...
5 tricks you're not using
• Top 5 Stable diffusion...
Avoid these 7 mistakes
• Don't make these 7 mis...
How to fix live render preview:
• Stable diffusion gui m...
CHAPTERS
0:00 Introduction
0:32 Mistake 1 - Prompting
1:28 Mistake 2 - Denoising strength
2:20 Mistake 3 - Time
3:12 Mistake 4 - Copying settings
3:46 Mistake 5 - Not copying settings
4:16 Mistake 6 - Messing with the resolution
5:15 Mistake 7 - Restore faces
5:48 Bonus
7:12 Closing words & final joke
7:44 Discord community

Пікірлер: 234
@sebastiankamph
@sebastiankamph Жыл бұрын
The FREE Prompt styles I use here: www.patreon.com/posts/sebs-hilis-79649068
@voyowee
@voyowee Жыл бұрын
Don't ignore negative prompts.
@Doughy_in_the_Middle
@Doughy_in_the_Middle Жыл бұрын
On the Denoising strength, when I'm trying to find just the right generation of an image, I like to do an X/Y prompt grid with CFG and Denoising. Denoising goes from 0 to 1 for no change to craziness, and CFG goes from 1 to 30 for wild creativity to "do exactly as I typed!". So, I do something like CFG "30, 25, 20, 15, 10, 5, 0" and Denoising "0, 0.15, 0.35, 0.5, 0.65, 0.85, 1". That's a 7x7 grid, but I've gone larger too. When the grid completes, you get the upper right as almost the exact original image, and that guy in the lower right is coming straight out of a Lovecraft novel.
@edtp79
@edtp79 Жыл бұрын
If you have a very specific pose you want, use it as a base in img-to-img, even if the style is completely different. I sometimes even do some rough editing in PS beforehand. Then work with a denoise of 0,2 to 0,4 and you're golden.
@sebastiankamph
@sebastiankamph Жыл бұрын
Great comment Eduardo! 🌟
@beetheimmortal
@beetheimmortal Жыл бұрын
This is what I always do. I really like using historical leaders, like that one portrait of Charles XIV John or even Victoria Louise of Prussia, because their poses are really excellent.
@ZeroAngel
@ZeroAngel Жыл бұрын
Use a depth mask to maintain the composition/pose while changing everything else
@eyoo369
@eyoo369 Жыл бұрын
It's amazing how fast this technology moves. 3 months ago it was indeed quite a problem. Then depth masks came which slightly helped but still not the perfect solution. Now with the new ControlNET extension you have perfect poses every single time.
@thebrokenglasskids5196
@thebrokenglasskids5196 9 ай бұрын
I think the biggest mistake people make when new to AI art is not realizing what an iterative process it is to get a good final result. Don't obsess too much trying to generate a perfect result initially in txt2img(it will never happen). txt2img is really only for dialing in your prompt and getting the composition of your image correct so you can get a good base seed to work from. Refining details, increasing quality, getting your perfect result is going to happen from working that base seed in img2img and inpainting. If you master inpainting, there is nothing you can't change or control in your image. I cannot emphasize that enough - Learn Inpainting! For example, wanna change the eyes or facial expression? Mask what you want to change in inpaint(use Only Masked, not Whole Picture when inpainting the face), then keep the quality portions of your prompt, but remove anything not related to what you want to change(you don't need things like "extra limbs" in your negative prompt for changing eyes, etc), and then replace those things with new prompt items that relate specifically to what you want to change in inpaint(for eyes you could add things like "highly detailed eyes, looking left, rich hazel color, subtle eyeliner" to the BEGINNING of your prompt). Render batches with the cfg unchanged, but the steps maxed out to 150. Boom, now you not only have amazing quality eyes, but can choose from the batches which ones you want exactly. Obviously there is a lot more iterative workflow steps to getting that perfect image(like using your X/Y/Z plot tool in txt2img to dial in the optimal steps and cfg for your prompt), but this is how you need to think when rendering AI images. It's an i t e r a t i v e p r o c e s s! TL/DR: Txt2Img is for dialing in prompting, cfg and steps to get a good base seed to work from. Img2Img and Inpainting is for refining and changing the details and increasing image quality. AI art is an iterative process - don't fall into the trap of playing endless rounds of Russian Roulette to get a good result! Get a good base seed to work from and then work it and refine it. 😎
@NeroZenith
@NeroZenith 6 ай бұрын
this was really well explained, thanks so much
@thebrokenglasskids5196
@thebrokenglasskids5196 6 ай бұрын
@@NeroZenith Thanks for the thanks! Glad someone got something out of it. 😁
@shambhangal438
@shambhangal438 Жыл бұрын
Here's a good one - If you want something that is proving difficult with prompting, photo-bash it first then use the prompt and the photo-bash together via img2img, rather than prompt only and txt2img. The photo-bash doesn't have to look pretty (no need for color corrected, seamless joins etc) because img2img will fix those, assuming you give it a high enough denoising strength (I start at 0.75).
@jackslaughter3233
@jackslaughter3233 10 ай бұрын
That's actually a REALLY great idea! Thanks! 😁
@tomaszwota1465
@tomaszwota1465 8 ай бұрын
photo-bash?
@zvit
@zvit Жыл бұрын
The most annoying thing is, when denoiser is too high, it still draws a cat even when you put "cat" in the negative prompts.
@sebastiankamph
@sebastiankamph Жыл бұрын
It gives the AI more creative freedom, for sure 😅
@jackslaughter3233
@jackslaughter3233 10 ай бұрын
If you set that denoiser to 9+ you're in the wild west! Hold on to your hat, god knows what'll happen 😂
@spearcy
@spearcy Жыл бұрын
Another common mistake is to close the main window when you meant to close the folder of currently rendering images. I’ve never made that mistake, of course, but this could help others. 💁‍♂️
@sebastiankamph
@sebastiankamph Жыл бұрын
I've also heard of this mistake, that only other people do 🌟
@mattlegge8538
@mattlegge8538 Жыл бұрын
Pin the tab, then there is no close button
@pranshumittal3374
@pranshumittal3374 9 ай бұрын
I would like to know how to properly close sd like I close the cmd window or what?
@spearcy
@spearcy 9 ай бұрын
@@pranshumittal3374 I always close the main window and then the cmd window.
@PriestessOfDada
@PriestessOfDada Жыл бұрын
Thought: You can get some good stuff at 512, but you're still going to have trouble with eyes and facial details. 570-640 is usually where I work. And most of the time, it's not too much for my six year old graphics card. Also restore faces only works on realistic stuff. If you're making cartoony stuff, they'll break your image as easily as fixing it. That's why you want to work at a slightly higher resolution
@thebrokenglasskids5196
@thebrokenglasskids5196 9 ай бұрын
I find that being good at inpainting is far more important for getting good facial details, such as eyes, rather than upping the initial resolution when generating base images. Staying at 512 for initial batch generation allows more batches and images faster. Pick the one you want to work from as a base and send it to inpaint. Mask what you want to change(make sure to always select Only Masked instead of Whole Image when inpainting faces or the results will not be as good). The key here is to not use your original image prompt, but rather change the prompt to exactly whatever it is you want to change in the face. So for eyes I would simplify the original prompt to keep anything in it that is related to quality, but removing anything that has nothing to do with eyes and then adding to the beginning something like "highly detailed eyes, looking left, vivid hazel color, subtle eyeliner" to the beginning of the prompt. Same for the negative prompt - simplify it by keeping terms that relate to details and removing anything that has nothing to do with eyes(extra limbs, poorly drawn hands, poorly drawn feet, etc). This is a lot more effective way of changing any facial feature, facial expression, etc. as you get much more control over what is changed, how it is changed, and the quality of the changes through using specific prompts for the masked area and denoising strength. Plus you can generate the changes in batches and select the very best one. Specifically speaking to eyes, you can get consistent amazing quality eyes using this method by doing all of the above, not changing the cfg, but maxing out the steps to 150 when you generate your inpainted batches. Works regardless of style once you master this method. 😎
@PriestessOfDada
@PriestessOfDada 9 ай бұрын
@@thebrokenglasskids5196 Yeah, that original comment is pretty ancient. We've come a long way since last year. Also try adetailer. Take care
@JackedOnMtnDew
@JackedOnMtnDew Жыл бұрын
I can't tell you how happy your bonus tip made me. I was literally just wondering today if this was possible and then happened to catch your video. Turns out it was enabled by default and I can pull up all my best images again. Thank you!
@badradish2116
@badradish2116 Жыл бұрын
funny you should mention a poker playing friend with a prosthetic arm, i have one as well. his luck just hasnt been good lately. first he loses his arm, then last time he played, when the pot was $10000, his draw was so bad he had to replace his entire hand.
@ZeroAngel
@ZeroAngel Жыл бұрын
I actually took a second with the prosthetic joke because I was expecting a "bad hand" joke and got the difficulty dealing joke 😅🙃
@MugiwaraRuffy
@MugiwaraRuffy 10 ай бұрын
Henlooo! Watched a couple of your videos on Stable Diffusion setup, basics, etc. I've got a free weekend coming ahead, and I want to get into and fiddle around with it a little bit. So, a "Thank You" in advance for your tutorials.
@SwiteFilms
@SwiteFilms Жыл бұрын
What a banger Sebastian! Great video input and also big step to more professional videos! Love the style :)!
@sebastiankamph
@sebastiankamph Жыл бұрын
Thank you, glad you liked it and happy to hear feedback on the new style! 😊🌟
@blackvx
@blackvx Жыл бұрын
Thank you. The best tip is to starts at 512x512 and then upscale your favorite images.
@sebastiankamph
@sebastiankamph Жыл бұрын
Happy to hear you found something valuable to you! 🌟
@gordonbrinkmann
@gordonbrinkmann Жыл бұрын
My main problem is that upscaling with any kind of upscaling tool doesn't really get you that much detail as upscaling in SD with img2img - but then I'm very limited in size due to my bad GPU.
@TheIllusionRecords
@TheIllusionRecords Жыл бұрын
Very true about the prompt differences between SD and MJ. SD has the potential to give you much more of what you want, but you've got to learn how to speak to the machine ;)
@sebastiankamph
@sebastiankamph Жыл бұрын
Exactly how I feel! 🌟
@ShawnFumo
@ShawnFumo Жыл бұрын
I've heard interesting discussions around this too, in relation to things like ChatGPT. Whether someone is impressed or not often comes down to if someone intuitively knows how to ask in a way that the AI understands.
@OriBengal
@OriBengal Жыл бұрын
The Dad jokes were the best part :) ---- Btw, Face restore often blurs out the image a little.. makes it too perfect. I tell people to be ok with a crappy face, then go in and inpaint it using "only masked" -- and this will build you out a high-res face, with all the skin texture, etc....
@bearlyphased
@bearlyphased Жыл бұрын
I feel like I have mastered stable diffusion. I have done it all and my images are far superior to most that I see. I understand how to engineer every part of the picture at every spot in the prompt and there is a hierarchy even without weights I spend a lot of time going through lots of seed configurations, cfg scales, clip steps, and denoising to sometimes get a closer result to my prompt and sometimes to let AI surprise me. I like to do a large batch and sometimes single out the ones that I really like and then run them back in hi-res
@upsxace
@upsxace 10 ай бұрын
😂😂😂 👏
@MBeccaro
@MBeccaro Жыл бұрын
9. Forgetting to return back to the actual checkpoint after changing to an inpainting checkpoint for that finishing touch, and wondering why all the next 50 images you generate look nothing like that crisp one you made earlier
@joshlin50ja
@joshlin50ja Жыл бұрын
Nice tunes, good advice condensed into a short video. Subbed.
@DrNothing23
@DrNothing23 Жыл бұрын
Hey, Sebastian! I just installed 1111 last week and am now diving into SD. Mainly for Deforum Animating, but definitely generating and tweaking images, too. I'd say the biggest noob mistake I've made so far is not watching this and your deep dive SD videos yet... :) Thanks for all your work and info sharing!!
@sebastiankamph
@sebastiankamph Жыл бұрын
Thank you kindly for the nice words! Welcome to the world of SD 🌟🤩
@SJ76
@SJ76 Жыл бұрын
Excellent tips! Thanks.
@bird_s_eye
@bird_s_eye Жыл бұрын
Thanks for the video Sebastian, definitely gonna improve my renders. btw, do you have any idea how to make a lyric video for a song? I had trouble displaying words so idk if there's like something specific I have to do with my prompt or something like that. Thank you
@sebastiankamph
@sebastiankamph Жыл бұрын
You're most welcome! Are you referring to SD actually creating the text for you? If so, no, it's terrible at text. I'd slap those on afterwards in a video editing software.
@alexsanders8881
@alexsanders8881 Жыл бұрын
Friend, first of all congratulations for the video, by the way, is it possible to train a model and during that training pause to continue on another day? Thanks
@swagmum4203
@swagmum4203 Жыл бұрын
When upscaling in imagetoimage, lower the denoising strenght, otherwise your picture will consist of multiple tiles
@lyricsdepicted5628
@lyricsdepicted5628 Жыл бұрын
Always wondered what codeformer is when booting automatic1111 :). Have you tried "asthetics gradients" (auto1111 plugin) where you can slab LAION beautifiers over your creations? I didn't see much of a change in my tests. It's a lot of more of these over-sensitive knobs, but promising from the idea and claim. Also: those different gradient files are crazy-small.
@sebastiankamph
@sebastiankamph Жыл бұрын
I haven't tried that particular plugin/extension. I'll be sure to check it out, thanks for the tip! 🌟
@wernerblahota6055
@wernerblahota6055 6 ай бұрын
Great video input and also big step to more professional videos!
@general_fail4887
@general_fail4887 8 ай бұрын
DO NOT use "restore faces" if you want to generate anime/ cartoon style images. Otherwise, SD tries to humanize the faces, which then ends in alien-like results
@LoriLinstruth
@LoriLinstruth Жыл бұрын
Loving these videos (and hearing your lovely Swedish accent makes me homesick for Sweden...I lived there for nearly 20 years). Any tips about how to avoid heads being cropped off? Or how to build out an image if you like it but half the head is missing? So far that is by far my biggest annoyance.
@sebastiankamph
@sebastiankamph Жыл бұрын
Thank you for the kind words! Now I don't understand why you would ever leave Sweden, but I do admit, it's kinda cold here right now. Heads being cropped off is a regular annoyance for everyone. I would recommend working more with img2img. Find a composition that you like in txt2img, and then move on to img2img. You could also outpaint, but that's more finicky tbh. Seeing as you've practiced drawing I'd even recommend starting from your own sketches and use those in img2img. That way you'll also have the heads like you want them.
@sub-jec-tiv
@sub-jec-tiv Жыл бұрын
Wish i could live there (i have family there but 3 generations back). Starting to feel a little crazy living in this country 😢😅😢 Have had some luck with strong negative prompts like “cut off head” etc.
@LoriLinstruth
@LoriLinstruth Жыл бұрын
@@sebastiankamph I had to leave for luuuuuuuv...but I get SO homesick for Sweden, even after 15 years. Oh well. Thanks for the tips! Seriously, cut off stuff is the bane of my existence. I've had mixed success with outpainting, so I'll give img2img a try and see how it goes. LOL that you know about my drawing experiment! I have to say, it was a year well spent. Thanks again for your response...you rock!
@sebastiankamph
@sebastiankamph Жыл бұрын
@@LoriLinstruth You can also check out high res fix for expanding. Like start with 512x512 and then a higher resolution to get more stuff in there. Oh, I sometimes have a quick look at commenters if they have content on their own KZbin. I'm curious by nature.
@ShawnFumo
@ShawnFumo Жыл бұрын
Not sure if you kept with SD or not, but make sure you check out openpose with ControlNet if you haven't. Should prevent heads from ever being chopped off again.
@tylerelyt
@tylerelyt Жыл бұрын
Agree with all of your points besides "stick with a low resolution." While 512x512 does generally result in a consistently cohesive and attractive image, the "Highres. fix" works incredibly well with 1024x1024. Waaaayy sharper and more detailed than 512x512, and looks even better than a 4x upscaled 512x512 image. With the "Highres. fix" I leave "Denoising strength" at 0.7, and set the "Firstpass width/height" to 512. Takes around ~8GB VRAM for 1024x1024. It's by far my favorite feature in the WebUI.
@NathanLorenzana
@NathanLorenzana Жыл бұрын
Where is the Hires.fix? Is it a setting or do you download it? I just started using WebUI, thanks!
@KadayiPolokov
@KadayiPolokov Жыл бұрын
@@NathanLorenzana It's a checkbox on TXTIMG in the same area as Restore Faces
@sebastiankamph
@sebastiankamph Жыл бұрын
You're right, highres.fix has been improving a lot. I think it depends if you have the hardware and time to use it.
@KillFrenzy96
@KillFrenzy96 Жыл бұрын
You can also upscale a 512x512 image to 1024x1024 with img2img using the same prompt. Use around 0.2 to 0.3 denoising. This is pretty much what high res fix does. This does alter the image a bit, but adds in the fine detail where upscalers fail. This is useful because you can generate 512x512 way faster than 1024x1024. Once you find an image with the right layout, 'upscale' it using img2img. After that, you can use the regular upscale to make a 4096x4096 image.
@NathanLorenzana
@NathanLorenzana Жыл бұрын
@@KadayiPolokov Got it! Thanks!!
@vladislavvolkoff
@vladislavvolkoff Жыл бұрын
Thank you for your content! What is the music at 4:06?
@LelioAngelus
@LelioAngelus Жыл бұрын
I just like and suscribe with the first dad joke. Thanks for the content. It's really usefull
@LexChan
@LexChan Ай бұрын
i like your jokes sometime.. give me a chill out start.. good work sebastian
@justinwhite2725
@justinwhite2725 Жыл бұрын
Biggest mistake is garbage words. If a model isn't trained with a word, that word sounds up tokenizing to effectively random noise that will warp your image in unpredictable ways.
@hotdog8008
@hotdog8008 Жыл бұрын
Web Designer! That was too good to be a dad joke! My biggest mistake? Forgetting to reset my seed to random. That was an infuriating 10 minutes...
@sebastiankamph
@sebastiankamph Жыл бұрын
Hah, but at least you learned something 😊
@hotdog8008
@hotdog8008 Жыл бұрын
@@sebastiankamph oh you bet I did. Keep up the great work Sebastian ;)
@arthurjeremypearson
@arthurjeremypearson Жыл бұрын
Specifically taking my sketches and finishing the art with flat color and inked
@GameDevNerd
@GameDevNerd Жыл бұрын
I'm gonna go ahead and clone the repo and take this thing for a spin. How do ya think my RTX-3080Ti will do with the algorithm? And is it worth paying for cloud compute and hosting to run my instances remotely, or is running it locally a decent and productive enough experience? I'm a software engineer, so I don't mind tinkering around or building extra utilities and components to automate things and make it more pleasant to work with ... Also, I noticed that it seems like there are not many good choices out there other than automatic1111's web-ui front-end. Why a browser? It seems to me that it would be a better experience with the headless back-end and a native front-end UI with better features on desktop would be a vastly superior user experience, and I could write some additional tools and utilties to make it easier to work with. If anyone has any ideas/suggestions for a better front-end UI, improved features and user experience, etc, please let me know ... I'm strongly considering making a free / open source project for an improved toolset and front-end layer, to address many of the things that seem rough around the edges upon my first impressions. Any integrations, shortcuts, automations, etc people think would be valuable would be great to hear about too. I really wanna have some fun with this soon as do some things to make it better, faster, more accessible, etc. I have some ideas for performance optimizations and possibly improving the model/algorithm itself, but I will need to spend some time studying the existing ones and experimenting.
@sebastiankamph
@sebastiankamph Жыл бұрын
Your 3080ti will run perfectly fine. I run on a 3080 myself. In regard to the browser question, it's where we are atm. Model has only been open source a few months. Anyone building a superior ui will surely take market shares.
@GameDevNerd
@GameDevNerd Жыл бұрын
@@sebastiankamph I think we can both agree that a _4090_ would run better and is necessary and justified ... the children may have to tighten their belts a bit for Christmas ... LOL, kidding, but I already planned on getting one, now I have a great idea for what to do with the 3080Ti afterward also :D I hit the ground doing some heavy research, found _very few_ projects implementing better UI and management tools for this other than a couple work-in-progress things ... outside of that, there's basically just a slew of websites that want to charge you a fee to use it and dole out little bits and pieces of functionality at insane costs ... that's definitely something that needs to change, I know _plenty_ other ways as a developer to make money that doesn't involve extortion or creating artificial roadblocks to create monetized gateways and pay-walls, lol. There are also a _lot_ of unexplored optimizations and enhancements that could be made to this, which I think is largely due to web developers who don't really know what they're doing just reading the setup instructions, putting it on some servers with cloud compute rental and building a simple UI and authentication + monetization layer around it. That's neither how this tech is supposed to be used _nor_ the best way to employ it. I'm not a super AI-expert, my expertise is more in graphics and engineering (mostly 3D real-time) and native engines, applications, games, etc. It seems like people are using _very_ compute-intensive (i.e., _expensive_ and inefficient) methods of upscaling, denoising, filtering, etc ... they probably should have checked with graphics engineers first, haha. I'm getting a lot of great ideas and "opportunist vibes" from this thing ... I've still got more research and tinkering to do though, and need to talk to some "hardcore" users about what could be done differently to make this better ...
@sub-jec-tiv
@sub-jec-tiv Жыл бұрын
@@GameDevNerd Be careful, 4090 may melt down and set your house on fire 😂
@GameDevNerd
@GameDevNerd Жыл бұрын
@brainswashedthisways nonsense, you just need a cold-fusion or zero-point quantum reactor to power it, which can be obtained from any local intergalactic computer store for a low price of 2,300 Blemkflarkz. Insulated wires, cables and plugs? Ha, so primtive, so inefficient ... if you use Earth technologies to power a 4090 you've made a critical error ...
@FlorianMaeder
@FlorianMaeder Жыл бұрын
What's with those video clips you show during the video (e.g. 2:33)? Were they generated with Stable Diffusion?
@infographie
@infographie Жыл бұрын
Excellent
@emmettbrown6418
@emmettbrown6418 Жыл бұрын
Spider joke, and my big mistake (started Yesterday) is not putting in the negative prompts. I got a lot of stuff I didn't want that way.
@hhvg7057
@hhvg7057 Жыл бұрын
he is looking up as if he is thinking and then reads it from the screen, haha I luv this guy
@sebastiankamph
@sebastiankamph Жыл бұрын
He's on to us! Run!
@THENEONGRID
@THENEONGRID Жыл бұрын
whats your best tip to avoid double heads apart from keeping the resolution low? negative prompts that work for you? I used two persons, two heads and such, what are yours?
@sebastiankamph
@sebastiankamph Жыл бұрын
I don't mess around too much with long negative prompts, just use some basic words. But I iterate a lot in img2img when I have a base composition that I want with the right number of heads. You can also check the highres-fix box when you're working outside of 512x512.
@unit1871
@unit1871 Жыл бұрын
Generally, using "highrez fix" with square resolution works best.
@THENEONGRID
@THENEONGRID Жыл бұрын
​@@unit1871 cool thank you
@THENEONGRID
@THENEONGRID Жыл бұрын
@@sebastiankamph thank you!
@767corp
@767corp Жыл бұрын
Get negative prompt embedding ,slap it in there and your golden, for me saved me tons of time messing with negatives for different models and image types I wanted to achieve. Search for bad prompts version 2 it's super easy to find and use.
@tzgaming207
@tzgaming207 Ай бұрын
I always seem to get horrible body distortions when I've included some detail in the prompt that the model doesn't appear to have 'awareness' of.. I could have a prompt that's making some decent images (still very new to this), but then I change an adjective the model doesn't seem to know, & it either alters the image in a way that it can avoid having to incorporate that detail (slacker!), or it's limbs, limbs, limbs 😬 btw, I'm physically incapable of telling dad jokes... I have no children 🥁😄
@atdfilms360
@atdfilms360 Жыл бұрын
Thank you the helpful video. Appreciate it! Anyone know where "codeforner" setting is please? Do you have to dl it as a separate module? Thank you very much
@sebastiankamph
@sebastiankamph Жыл бұрын
It's included in automatic1111. Check settings tab under face restoration.
@atdfilms360
@atdfilms360 Жыл бұрын
@@sebastiankamph Thank you very much! Do you find codeformer better than GFPGAN?
@enmodevrai2vrai91
@enmodevrai2vrai91 11 ай бұрын
5:48 I love this music, who's the artist? Thanks for the helpful video bro
@darth_sidious_sheev_palpatine
@darth_sidious_sheev_palpatine Жыл бұрын
Play around with models, merge them, test them. Easy way to add alot of flare
@besocialable482
@besocialable482 Жыл бұрын
Getting used to Sebastian's dad jokes. Funny stuffs!
@sebastiankamph
@sebastiankamph Жыл бұрын
🥰
@swannschilling474
@swannschilling474 Жыл бұрын
Very nice!! 🤩 But use Stable Diffusion images for your B-roll! 😁
@sebastiankamph
@sebastiankamph Жыл бұрын
I tend to use a mix, but best would be for everything to be AI, even videos! :)
@sub-jec-tiv
@sub-jec-tiv Жыл бұрын
I like his B-roll 🎉 he has nice taste in stock clips 😂
@JudgeDreadz
@JudgeDreadz Жыл бұрын
My fav joke was taking out the spider! Thank you :) Sub'd!
@teambellavsteamalice
@teambellavsteamalice Жыл бұрын
Now I'm thinking of a process with ChatGPT generating dad jokes, matching this with images using stable diffusion and put the joke text in text balloons... 🤣
@dimitopalidis2732
@dimitopalidis2732 Ай бұрын
could u give me tipps for making: good eyes, teeth, nails, fingers etc.
@user-vr2ey8wk5h
@user-vr2ey8wk5h Жыл бұрын
Thanks!
@sebastiankamph
@sebastiankamph Жыл бұрын
You're very welcome 😊
@ZeroTimelineX
@ZeroTimelineX Жыл бұрын
Ok noob question here. Do you know how prevent elements from combining? For example, if you want a cat and a dog, not a cat mixed with a dog. Thanks.
@Android-tc3me
@Android-tc3me 11 ай бұрын
4:53 wait... Can you explain to me what I have to do in order to upscale the resolution? Pls (🙏🏻)
@fakejaat1704
@fakejaat1704 Жыл бұрын
Whichs apps you use to make your thumbnail?
@cyberhard
@cyberhard Жыл бұрын
Second dad joke was the best.
@frankschannel2642
@frankschannel2642 9 ай бұрын
"Nice guy, he's a web designer." ROFL!!!!
@sebastiankamph
@sebastiankamph 9 ай бұрын
😂
@CaptainSkeletor
@CaptainSkeletor Жыл бұрын
how do you actually paste the copied generation data?
@cuteguy-2001
@cuteguy-2001 11 ай бұрын
i have bigger problem cant generate image and also even using prompt doesnt work im using rog flow x13 1060 gpu and other solution
@FredRosa
@FredRosa Жыл бұрын
I had to think twice about that web designer (lack of sleep perhaps) XD win!
@sebastiankamph
@sebastiankamph Жыл бұрын
😅
@rufinolarson635
@rufinolarson635 Жыл бұрын
Using models like Waifu Diffusion, Anywhere v3, etc. which are trained on Danbooru tags has made sure that I never prompt like a human whenever making waifus. But that first mistake is definitely something I should keep in mind using SD 1.5 and other models, that's for sure. The rest were also definitely something I can learn from. Thanks for all the useful advice!
@hmmmmmm_3429
@hmmmmmm_3429 Жыл бұрын
anything v3
@sebastiankamph
@sebastiankamph Жыл бұрын
Glad you liked it, thanks for the support! 🌟
@yesyoucan4619
@yesyoucan4619 Жыл бұрын
Imagine Iiving in a world where you ргіḓəḟսꞁꞁꭚ talk about your іпԀῦІƺҽлт мᾲᴤtսꭇβάϮtіѳח habits on KZbin, while ргꬲէӟոժіпᶃ to be a סתгмἇ׀ person without any ꝓѕꭚꞇԩσΙоϧίсᶐӏ enwarpments. It sure is a new generation of Internet ꞇꞕᴉլԀꭋἔꞃ we have in our woꭇId today. And it's also ẜꝺꭇьіԀɗɘὴ to spеаk սp against іt, because that's ԁιςᴐᵲίպḯꭒӑϯσꝵꭚ, so it has to become normal now.
@hmmmmmm_3429
@hmmmmmm_3429 Жыл бұрын
@@yesyoucan4619 writing like that is crime who taught you to write like that EW!!
@KasumiRINA
@KasumiRINA Жыл бұрын
​@@yesyoucan4619 Imagine being so insecure you have to attack strangers online for having fun with drawing pretty girls, something humanity did since before antiquity, and being so paranoid from your persecution complex you mask your words like that.
@hutukaTB
@hutukaTB Жыл бұрын
Is restore faces useful for anime?
@anilax
@anilax Жыл бұрын
How do you have soo low subs You deserve way more
@sebastiankamph
@sebastiankamph Жыл бұрын
Thank you for the kind words, it warms my heart! 😊 You're a real 💎
@simonavittoria586
@simonavittoria586 Жыл бұрын
THX for this good informations. I have a question to the automatic update SD111 with the command git pull in the webui.batch. It always ran, but now i have an error message if execute : error: The following untracked working tree files would be overwritten by merge: javascript/generationParams.js modules/script_loading py. Please move or remove them before you merge. Anybody now what does this means and what should do? THX
@sebastiankamph
@sebastiankamph Жыл бұрын
I think you can remove that git pull in the file for now. I did too, sometimes updates come too fast and things break. So might as well leave as is and update when necessary. And if you keep getting that error in the future and need to update, do a fresh install. And thanks for the kind words 😊
@simonavittoria586
@simonavittoria586 Жыл бұрын
@@sebastiankamph Thank you and you are welcome. I have learned a lot about stable diffusion from your good videos and tutorials. Keep up the good work 😊
@bojeppe6828
@bojeppe6828 Жыл бұрын
you had me at the dad jokes❤
@sebastiankamph
@sebastiankamph Жыл бұрын
Glad I could be of service! 😘🌟
@moldofvalgard738
@moldofvalgard738 Жыл бұрын
best tip for upscaling? please :)
@sebastiankamph
@sebastiankamph Жыл бұрын
SD does upscaling pretty good. Check my ultimate guide for some upscaling comparisons.
@LouisGedo
@LouisGedo Жыл бұрын
5:10 😆 🤣 😂 And multiple legs and toes on certain animals too!
@sebastiankamph
@sebastiankamph Жыл бұрын
So many weird images, right!? 😅
@Shibby27ify
@Shibby27ify 7 ай бұрын
How the hell do you make actions? Like people doing things, pulling, walking, holding... Most of the time is just makes people standing there. If I want someone picking something up or throwing something, there are just random hands or arms in a strange place. It just seems to make people standing there but not doing things!
@sebastiankamph
@sebastiankamph 6 ай бұрын
ControlNet extension is your friend here
@Mocorn
@Mocorn Жыл бұрын
You guys seem to know how all of this works so let me ask. Sometimes I see people taking a picture of themselves into img2img and remake it in the style of Arcane or something similar. This requires the arcane.ckpt file which means you don't have the style transfer setting that you get with the 1.5 inpaint.ckpt which means that your resulting picture is in the style of Arcane but the composition is not close to your start image at all. Corridor crew did this in a recent movie, how? When I try I need the 1.5 inpaint.ckpt to get a matching image but the arcane.ckpt file for the actual style which changes the comp drastically.. what am I missing here?
@sebastiankamph
@sebastiankamph Жыл бұрын
You can merge models, but you can also adjust the denoising strength in img2img. I'd try merging a trained Dreambooth model on you together with arcane.
@ShawnFumo
@ShawnFumo Жыл бұрын
Also, I think there is an arcane LoRA now, so should be able to use that on top of an inpainting capable model, if I understand how it works correctly.
@peterpui7219
@peterpui7219 Жыл бұрын
8. Don't keep sitting on your desk for Stable Diffusion. You need better inspiration from outside.Your output results are determine by your emotional intelligence and have better understanding of life and people around you. Don't lost in keeping in touch with people, pets,arts, nature and music if you're seeking better AI creativity.
@EriIaz
@EriIaz Жыл бұрын
Or creativity in general, I'd say. Observing the best is good for inclining your work, but having a meaningful life helps you a lot best when it comes to finding your own direction
@russelnormandia2876
@russelnormandia2876 Жыл бұрын
your spider joke secured the subscribe
@sebastiankamph
@sebastiankamph Жыл бұрын
Hah, glad to hear it :)
@jakobhovman
@jakobhovman Жыл бұрын
How do you salute a sailor, that uses AI...? AI AI Capten...! Oh...Video is great, Thank You...! Looking forward, to get my own SD and rewatch you. Learning in Playground now...I just can't use their SD inpaint proper...Is It Me...? PS: I think I will settle for the spider...most in tune.
@sebastiankamph
@sebastiankamph Жыл бұрын
Hah, great with another one! Thank you for the kind words. And I absolutely think you should jump on SD. Looking forward to seeing the pixel light stuff 🌟
@Refused56
@Refused56 Жыл бұрын
Come for SD, stay for the jokes😁
@sebastiankamph
@sebastiankamph Жыл бұрын
Glad you liked them! 😊🌟
@dresdenk
@dresdenk Жыл бұрын
I'm not sure how much it matters, but when I specify "irises" instead of "eyes", the eyes maybe look slightly better?
@sebastiankamph
@sebastiankamph Жыл бұрын
Cool tip!
@MoS910
@MoS910 9 ай бұрын
I guess short Midjourney prompt looks lazy. More details more specifications, I would use same prompt fro SD in Midjourney because Midjourney prompts like let AI decide instead of the user... like wombo dream, I don't like short prompts!
@Nagrom
@Nagrom Жыл бұрын
you forgot to use the "no cross eyes" prompt
@TheAiConqueror
@TheAiConqueror Жыл бұрын
Thank you 4 the dad jokes 😁✌️
@sebastiankamph
@sebastiankamph Жыл бұрын
You're very welcome! And a huge thank you for the support, your DNA is made of diamonds, you absolute gem you! 💎💎💎
@reitakuhemphill
@reitakuhemphill Жыл бұрын
It was a mistake to compliment him on the dad gags.
@sebastiankamph
@sebastiankamph Жыл бұрын
I've ventured down a path I cannot stray from now.
@ThePixelkd
@ThePixelkd Жыл бұрын
Definitely the spider one
@sebastiankamph
@sebastiankamph Жыл бұрын
Glad you liked it! 🌟😊
@paulmcallen
@paulmcallen Жыл бұрын
i fundametally disagree. i generate the straight image from txt2img and upscale. very few times i had to img2im to fix a crop. i use img2img to clean up archival photos. running the same image 20 times is a waste of time and you will probably end up ruining the things you liked about it in the first place. 1000x1200 txt2img resolution means you dont have to do much to it
@Steamrick
@Steamrick Жыл бұрын
So yeah... I forgot about restore faces. Dang.
@Ich.kack.mir.in.dieHos
@Ich.kack.mir.in.dieHos Жыл бұрын
i dont get what is automatic 111 ?? is this another name for stable diff??
@sebastiankamph
@sebastiankamph Жыл бұрын
Automatic1111 is currently the #1 Stable diffusion graphical user interface.
@qaesarx
@qaesarx Жыл бұрын
Those faaaaaaaaaans 🌀🌀
@digidope
@digidope Жыл бұрын
#1 Mistake: Thinkin you NEED to use 512x512 resolution. Nope. SD has zero problems with say 1024x2048 or my current most used resolution 384x2048. That works just fine with 1.5 and 2.0.
@tomaszwota1465
@tomaszwota1465 8 ай бұрын
I wouldn't say "zero" problems. There certainly are issues as different resolutions and ratios will influence composition. For example, make the image too tall and you might find your characters growing another waist from their thighs, things like that. But yeah, no need to limit yourself to 512x512, just need to up that batch number.
@digidope
@digidope 8 ай бұрын
@@tomaszwota1465 Thinkin that my comment is now 8mo old and how little SD has improved in that time :D You are correct that it makes weird human anatomy. SDXL has the same problem. Midjourney 5.2 does that rarely if ever. But i rarely do humans for real work. For landscapes SD did great job with big resolutions like 2024x1024.
@max477
@max477 Жыл бұрын
What is your opinion playground AI
@sebastiankamph
@sebastiankamph Жыл бұрын
It doesn't look different from every other online Stable diffusion ui out there.
@LouisGedo
@LouisGedo Жыл бұрын
I have had pretty bad results with many of the batches of images I've had SD render at and above 768 X 768.
@devnull_
@devnull_ Жыл бұрын
Then use the native resolution, 512x512
@LouisGedo
@LouisGedo Жыл бұрын
@@devnull_ Yes, of course. The larger rez is experimentation. Some of it renders nice results.....but generally that's a relatively small percentage
@ShawnFumo
@ShawnFumo Жыл бұрын
Though important for people reading to know SD 2.1 is natively at 768x768, so is very appropriate with a model/mix based on that (like Illuminati Diffusion)
@VovanEkb
@VovanEkb Жыл бұрын
do not try to turn on the restoration of faces, when generating art, or any non-photo realistic effect, try to do without it until the last, this jackdaw (yes, I know about the possibility of setting it up) greatly affects the overall style, and often completely destroys it. I'm saying that you should try to achieve this without her, and you will get a much cooler and more consistent result at the output!
@KasumiRINA
@KasumiRINA Жыл бұрын
Just make it save both before and after pic, and you can choose.
@kishirisu1268
@kishirisu1268 9 ай бұрын
You dont understand how ML learning model works. “Cat in anhat” and “painting” is two DIFFERENT tokens, and model will create two different objects. If you dont belive me, go to Hugginface and test how Clip model (which used for tokenisation in SD) is reacting on your prompts. The correct promt often long, like “picture of cat in a hat”. Same with people. Not “woman”, “dress”, “beautiful”, it is wrong prompt. The correct one “a beautiful woman wearing dress”. “Wearing” is indication that dress and woman is one intitty.
@schmutly
@schmutly Жыл бұрын
last joke the funniest 😆🤣😂
@sebastiankamph
@sebastiankamph Жыл бұрын
🌟🌟🌟
@divinesimmortus5655
@divinesimmortus5655 Жыл бұрын
Anyone got a spare graphics card haha with my radeon r9 290x it takes about 20 min for a 512x512 with 10 sampling steps xD
@RHYTE
@RHYTE Жыл бұрын
Why is it that, when u use a specific seed and generate a batch of maybe 10 images from it, they all look different. Shouldn't they all look the same?
@androsforever500
@androsforever500 Жыл бұрын
I think when you generate a batch of 10 it automatically uses a different variation seed for every one of those images
@RHYTE
@RHYTE Жыл бұрын
@@androsforever500 oh yeah that makes sense
@Raven_RVN
@Raven_RVN Жыл бұрын
If you are using the automatic1111 web ui then it actually changes the seed. If you enter Seed 1, and generate 4 images, then Image 1 will have Seed 1, Image 2 Seed 2, etc. It adds always adds 1. You can check this yourself. You can click on, say, image 7 and then use the button to retrieve the seed and can then compare this to the seed of image 6 and 8, it should always be 1 apart from each other
@pirobot668beta
@pirobot668beta 9 ай бұрын
If you save txt2img as JPG, the comment field of the file will have all your settings.
@ajudicator
@ajudicator Жыл бұрын
He’s a web crawler Would have been a better end to the punchline Thanks for the videos
@mattlegge8538
@mattlegge8538 Жыл бұрын
web developer
@Venoxium
@Venoxium Жыл бұрын
Is it just me or is IMG-IMG literally black magic? I can't figure it out. I've tried using it SO many times and can never get any good results. Currently using Anything V3
@sebastiankamph
@sebastiankamph Жыл бұрын
Sure is, and works great. Check some basics on img2img here: kzbin.info/www/bejne/ennEfWhshZuZa68
@brianclark7719
@brianclark7719 Жыл бұрын
Has anyone found a solution to AI lackings in multiple actors in a given prompt? For example - Dog sitting on top of a camel which is licking a cat that is chewing on a mouse? I know that the AI is somewhat in it's infancy but it bewilders me why it cannot see that level of description and interpret the final result when it draws the image. Anyone? I know I may be expecting too much, but using the reality that it can produce a dog sitting on top of a camel wearing a sombrero in the desert at night, why couldn't it?
@sebastiankamph
@sebastiankamph Жыл бұрын
Check my latest videos on ControlNet. That fixes that. Wasn't possible previously.
@ShawnFumo
@ShawnFumo Жыл бұрын
That level of spatial awareness is still definitely tricky for the AIs in general. MidJourney V5 may or may not be a bit better if they manage to get a better text engine into it (they said they weren't sure on the time), and perhaps DeepFloyd though that isn't totally clear (and who knows when that actually comes out). Like Sebastian said, ControlNet can help, especially the posing and new MultiDiffusion region stuff.
@lpanzieri
@lpanzieri Жыл бұрын
The jokes are so terrible they end up making me laugh :D
@sebastiankamph
@sebastiankamph Жыл бұрын
Great, eh? 😊
@captainkaput7281
@captainkaput7281 Жыл бұрын
Please include (for absolute beginners) in the title. Thanks.
@rogerc7960
@rogerc7960 Жыл бұрын
Use a thesaurus
@SaintMatthieuSimard
@SaintMatthieuSimard Жыл бұрын
OMG! 🤣A 🕸Web designer 🕷
@WizardF3-pc2li
@WizardF3-pc2li Жыл бұрын
Great video. But using "Restore Faces" is a trap. Believe me, do not use it. It changes so much of the face, turning a great model into meh most of the time. Just make sure you write a good prompt, using a good model and you will never have to use Restore Faces.
@sebastiankamph
@sebastiankamph Жыл бұрын
You're right! But this was good info back when this video was released. Nowadays I inpaint to improve faces.
@ameet21
@ameet21 9 ай бұрын
0:57 I feel now more embarrassed after watching this 😂 a month ago I had an argument with davinci ai where I told them why they put a lot of nsfw restrictions on normal words like crazy, beautiful. But now I completely understand, Ai don't understand or care feelings but technical breefing. There I have one issue that where in negative prompt I add a double head, double body, double figure, double hand, disfigured body but why even that i often get 2 head and double or triple bodies instead adding these words in negative prompt I get more often weird results like in 50 attempt only 2-3 or 4-5 output which come properly. So please solve this issue 🙏 Oh wait 5:06 so changing aspect ration from square to others causes this double head double body issue? Ok bro I will render with square and after upscale I will crop that art in Photoshop. Now i realised why the heck even I bloody add specifically in prompt mid-shot portrait and I use to get headshot close-up
Revealing my Workflow to Perfect AI Images.
13:31
Sebastian Kamph
Рет қаралды 304 М.
【獨生子的日常】让小奶猫也体验一把鬼打墙#小奶喵 #铲屎官的乐趣
00:12
“獨生子的日常”YouTube官方頻道
Рет қаралды 79 МЛН
McDonald’s MCNUGGET PURSE?! #shorts
00:11
Lauren Godwin
Рет қаралды 28 МЛН
Barriga de grávida aconchegante? 🤔💡
00:10
Polar em português
Рет қаралды 10 МЛН
Create consistent characters with Stable diffusion!!
26:41
Not4Talent
Рет қаралды 182 М.
How to create VIRAL Motion Graphic Animations IN CANVA! 💯
13:03
Grow with Alex
Рет қаралды 77 М.
Stable Diffusion - Prompt 101 #ai
30:05
Not That Complicated
Рет қаралды 20 М.
Intro to LoRA Models: What, Where, and How with Stable Diffusion
21:01
Laura Carnevali
Рет қаралды 176 М.
Stable Diffusion for Flawless Portraits
13:24
Vladimir Chopine [GeekatPlay]
Рет қаралды 228 М.
Forget MidJourney, This Is How I Make AI Art Now
8:23
Maximize
Рет қаралды 478 М.
How to use Stable Diffusion. Automatic1111 Tutorial
27:10
Sebastian Kamph
Рет қаралды 278 М.
How to change clothes with AI.
14:45
Sebastian Kamph
Рет қаралды 74 М.
【獨生子的日常】让小奶猫也体验一把鬼打墙#小奶喵 #铲屎官的乐趣
00:12
“獨生子的日常”YouTube官方頻道
Рет қаралды 79 МЛН