Don't make these 7 mistakes in Stable diffusion.

Поділитися
Вставка
  • Опубліковано 6 лип 2024
  • Top 7 mistakes people make in Stable diffusion. Check out the bonus content! Which joke did you like best?
    Chat with me in our community discord: / discord
    Support me on Patreon to get access to unique perks!
    / sebastiankamph
    Ultimate Stable diffusion guide
    • Stable diffusion tutor...
    The Rise of AI Art: A Creative Revolution
    • The Rise of AI Art - A...
    Ultimate Animation guide in Stable diffusion
    • Stable diffusion anima...
    Dreambooth tutorial for Stable diffusion
    • Dreambooth tutorial fo...
    5 tricks you're not using
    • Top 5 Stable diffusion...
    Avoid these 7 mistakes
    • Don't make these 7 mis...
    How to fix live render preview:
    • Stable diffusion gui m...
    CHAPTERS
    0:00 Introduction
    0:32 Mistake 1 - Prompting
    1:28 Mistake 2 - Denoising strength
    2:20 Mistake 3 - Time
    3:12 Mistake 4 - Copying settings
    3:46 Mistake 5 - Not copying settings
    4:16 Mistake 6 - Messing with the resolution
    5:15 Mistake 7 - Restore faces
    5:48 Bonus
    7:12 Closing words & final joke
    7:44 Discord community

КОМЕНТАРІ • 234

  • @sebastiankamph
    @sebastiankamph  Рік тому +3

    The FREE Prompt styles I use here:
    www.patreon.com/posts/sebs-hilis-79649068

  • @voyowee
    @voyowee Рік тому +32

    Don't ignore negative prompts.

  • @Doughy_in_the_Middle
    @Doughy_in_the_Middle Рік тому +18

    On the Denoising strength, when I'm trying to find just the right generation of an image, I like to do an X/Y prompt grid with CFG and Denoising. Denoising goes from 0 to 1 for no change to craziness, and CFG goes from 1 to 30 for wild creativity to "do exactly as I typed!".
    So, I do something like CFG "30, 25, 20, 15, 10, 5, 0" and Denoising "0, 0.15, 0.35, 0.5, 0.65, 0.85, 1". That's a 7x7 grid, but I've gone larger too. When the grid completes, you get the upper right as almost the exact original image, and that guy in the lower right is coming straight out of a Lovecraft novel.

  • @zvit
    @zvit Рік тому +22

    The most annoying thing is, when denoiser is too high, it still draws a cat even when you put "cat" in the negative prompts.

    • @sebastiankamph
      @sebastiankamph  Рік тому +8

      It gives the AI more creative freedom, for sure 😅

    • @jackslaughter3233
      @jackslaughter3233 Рік тому +1

      If you set that denoiser to 9+ you're in the wild west! Hold on to your hat, god knows what'll happen 😂

  • @NotAnAlchemist_Ed
    @NotAnAlchemist_Ed Рік тому +53

    If you have a very specific pose you want, use it as a base in img-to-img, even if the style is completely different. I sometimes even do some rough editing in PS beforehand. Then work with a denoise of 0,2 to 0,4 and you're golden.

    • @sebastiankamph
      @sebastiankamph  Рік тому +3

      Great comment Eduardo! 🌟

    • @beetheimmortal
      @beetheimmortal Рік тому +1

      This is what I always do. I really like using historical leaders, like that one portrait of Charles XIV John or even Victoria Louise of Prussia, because their poses are really excellent.

    • @ZeroAngel
      @ZeroAngel Рік тому +1

      Use a depth mask to maintain the composition/pose while changing everything else

    • @eyoo369
      @eyoo369 Рік тому +6

      It's amazing how fast this technology moves. 3 months ago it was indeed quite a problem. Then depth masks came which slightly helped but still not the perfect solution. Now with the new ControlNET extension you have perfect poses every single time.

  • @spearcy
    @spearcy Рік тому +22

    Another common mistake is to close the main window when you meant to close the folder of currently rendering images. I’ve never made that mistake, of course, but this could help others. 💁‍♂️

    • @sebastiankamph
      @sebastiankamph  Рік тому +7

      I've also heard of this mistake, that only other people do 🌟

    • @mattlegge8538
      @mattlegge8538 Рік тому +3

      Pin the tab, then there is no close button

    • @pranshumittal3374
      @pranshumittal3374 11 місяців тому

      I would like to know how to properly close sd like I close the cmd window or what?

    • @spearcy
      @spearcy 11 місяців тому

      @@pranshumittal3374 I always close the main window and then the cmd window.

  • @shambhangal438
    @shambhangal438 Рік тому +19

    Here's a good one - If you want something that is proving difficult with prompting, photo-bash it first then use the prompt and the photo-bash together via img2img, rather than prompt only and txt2img.
    The photo-bash doesn't have to look pretty (no need for color corrected, seamless joins etc) because img2img will fix those, assuming you give it a high enough denoising strength (I start at 0.75).

  • @thebrokenglasskids5196
    @thebrokenglasskids5196 11 місяців тому +10

    I think the biggest mistake people make when new to AI art is not realizing what an iterative process it is to get a good final result.
    Don't obsess too much trying to generate a perfect result initially in txt2img(it will never happen). txt2img is really only for dialing in your prompt and getting the composition of your image correct so you can get a good base seed to work from.
    Refining details, increasing quality, getting your perfect result is going to happen from working that base seed in img2img and inpainting. If you master inpainting, there is nothing you can't change or control in your image. I cannot emphasize that enough - Learn Inpainting!
    For example, wanna change the eyes or facial expression? Mask what you want to change in inpaint(use Only Masked, not Whole Picture when inpainting the face), then keep the quality portions of your prompt, but remove anything not related to what you want to change(you don't need things like "extra limbs" in your negative prompt for changing eyes, etc), and then replace those things with new prompt items that relate specifically to what you want to change in inpaint(for eyes you could add things like "highly detailed eyes, looking left, rich hazel color, subtle eyeliner" to the BEGINNING of your prompt). Render batches with the cfg unchanged, but the steps maxed out to 150. Boom, now you not only have amazing quality eyes, but can choose from the batches which ones you want exactly.
    Obviously there is a lot more iterative workflow steps to getting that perfect image(like using your X/Y/Z plot tool in txt2img to dial in the optimal steps and cfg for your prompt), but this is how you need to think when rendering AI images. It's an i t e r a t i v e p r o c e s s!
    TL/DR: Txt2Img is for dialing in prompting, cfg and steps to get a good base seed to work from. Img2Img and Inpainting is for refining and changing the details and increasing image quality. AI art is an iterative process - don't fall into the trap of playing endless rounds of Russian Roulette to get a good result! Get a good base seed to work from and then work it and refine it. 😎

    • @NeroZenith
      @NeroZenith 8 місяців тому +2

      this was really well explained, thanks so much

    • @thebrokenglasskids5196
      @thebrokenglasskids5196 8 місяців тому +2

      @@NeroZenith Thanks for the thanks! Glad someone got something out of it. 😁

  • @PriestessOfDada
    @PriestessOfDada Рік тому +20

    Thought: You can get some good stuff at 512, but you're still going to have trouble with eyes and facial details. 570-640 is usually where I work. And most of the time, it's not too much for my six year old graphics card. Also restore faces only works on realistic stuff. If you're making cartoony stuff, they'll break your image as easily as fixing it. That's why you want to work at a slightly higher resolution

    • @thebrokenglasskids5196
      @thebrokenglasskids5196 11 місяців тому +1

      I find that being good at inpainting is far more important for getting good facial details, such as eyes, rather than upping the initial resolution when generating base images.
      Staying at 512 for initial batch generation allows more batches and images faster. Pick the one you want to work from as a base and send it to inpaint.
      Mask what you want to change(make sure to always select Only Masked instead of Whole Image when inpainting faces or the results will not be as good). The key here is to not use your original image prompt, but rather change the prompt to exactly whatever it is you want to change in the face. So for eyes I would simplify the original prompt to keep anything in it that is related to quality, but removing anything that has nothing to do with eyes and then adding to the beginning something like "highly detailed eyes, looking left, vivid hazel color, subtle eyeliner" to the beginning of the prompt. Same for the negative prompt - simplify it by keeping terms that relate to details and removing anything that has nothing to do with eyes(extra limbs, poorly drawn hands, poorly drawn feet, etc).
      This is a lot more effective way of changing any facial feature, facial expression, etc. as you get much more control over what is changed, how it is changed, and the quality of the changes through using specific prompts for the masked area and denoising strength. Plus you can generate the changes in batches and select the very best one.
      Specifically speaking to eyes, you can get consistent amazing quality eyes using this method by doing all of the above, not changing the cfg, but maxing out the steps to 150 when you generate your inpainted batches. Works regardless of style once you master this method. 😎

    • @PriestessOfDada
      @PriestessOfDada 11 місяців тому

      @@thebrokenglasskids5196 Yeah, that original comment is pretty ancient. We've come a long way since last year. Also try adetailer. Take care

  • @JackedOnMtnDew
    @JackedOnMtnDew Рік тому

    I can't tell you how happy your bonus tip made me. I was literally just wondering today if this was possible and then happened to catch your video. Turns out it was enabled by default and I can pull up all my best images again. Thank you!

  • @badradish2116
    @badradish2116 Рік тому +2

    funny you should mention a poker playing friend with a prosthetic arm, i have one as well.
    his luck just hasnt been good lately. first he loses his arm, then last time he played, when the pot was $10000, his draw was so bad he had to replace his entire hand.

    • @ZeroAngel
      @ZeroAngel Рік тому

      I actually took a second with the prosthetic joke because I was expecting a "bad hand" joke and got the difficulty dealing joke 😅🙃

  • @SwiteFilms
    @SwiteFilms Рік тому +1

    What a banger Sebastian! Great video input and also big step to more professional videos! Love the style :)!

    • @sebastiankamph
      @sebastiankamph  Рік тому

      Thank you, glad you liked it and happy to hear feedback on the new style! 😊🌟

  • @blackvx
    @blackvx Рік тому +4

    Thank you. The best tip is to starts at 512x512 and then upscale your favorite images.

    • @sebastiankamph
      @sebastiankamph  Рік тому

      Happy to hear you found something valuable to you! 🌟

    • @gordonbrinkmann
      @gordonbrinkmann Рік тому

      My main problem is that upscaling with any kind of upscaling tool doesn't really get you that much detail as upscaling in SD with img2img - but then I'm very limited in size due to my bad GPU.

  • @bird_s_eye
    @bird_s_eye Рік тому

    Thanks for the video Sebastian, definitely gonna improve my renders.
    btw, do you have any idea how to make a lyric video for a song? I had trouble displaying words so idk if there's like something specific I have to do with my prompt or something like that.
    Thank you

    • @sebastiankamph
      @sebastiankamph  Рік тому +1

      You're most welcome! Are you referring to SD actually creating the text for you? If so, no, it's terrible at text. I'd slap those on afterwards in a video editing software.

  • @MugiwaraRuffy
    @MugiwaraRuffy Рік тому

    Henlooo!
    Watched a couple of your videos on Stable Diffusion setup, basics, etc. I've got a free weekend coming ahead, and I want to get into and fiddle around with it a little bit. So, a "Thank You" in advance for your tutorials.

  • @TheIllusionRecords
    @TheIllusionRecords Рік тому +3

    Very true about the prompt differences between SD and MJ. SD has the potential to give you much more of what you want, but you've got to learn how to speak to the machine ;)

    • @sebastiankamph
      @sebastiankamph  Рік тому

      Exactly how I feel! 🌟

    • @ShawnFumo
      @ShawnFumo Рік тому

      I've heard interesting discussions around this too, in relation to things like ChatGPT. Whether someone is impressed or not often comes down to if someone intuitively knows how to ask in a way that the AI understands.

  • @DrNothing23
    @DrNothing23 Рік тому +4

    Hey, Sebastian! I just installed 1111 last week and am now diving into SD. Mainly for Deforum Animating, but definitely generating and tweaking images, too.
    I'd say the biggest noob mistake I've made so far is not watching this and your deep dive SD videos yet...
    :) Thanks for all your work and info sharing!!

    • @sebastiankamph
      @sebastiankamph  Рік тому +2

      Thank you kindly for the nice words! Welcome to the world of SD 🌟🤩

  • @MBeccaro
    @MBeccaro Рік тому +2

    9. Forgetting to return back to the actual checkpoint after changing to an inpainting checkpoint for that finishing touch, and wondering why all the next 50 images you generate look nothing like that crisp one you made earlier

  • @lyricsdepicted5628
    @lyricsdepicted5628 Рік тому +3

    Always wondered what codeformer is when booting automatic1111 :). Have you tried "asthetics gradients" (auto1111 plugin) where you can slab LAION beautifiers over your creations? I didn't see much of a change in my tests. It's a lot of more of these over-sensitive knobs, but promising from the idea and claim. Also: those different gradient files are crazy-small.

    • @sebastiankamph
      @sebastiankamph  Рік тому

      I haven't tried that particular plugin/extension. I'll be sure to check it out, thanks for the tip! 🌟

  • @OriBengal
    @OriBengal Рік тому +2

    The Dad jokes were the best part :) ---- Btw, Face restore often blurs out the image a little.. makes it too perfect. I tell people to be ok with a crappy face, then go in and inpaint it using "only masked" -- and this will build you out a high-res face, with all the skin texture, etc....

  • @THENEONGRID
    @THENEONGRID Рік тому

    whats your best tip to avoid double heads apart from keeping the resolution low? negative prompts that work for you?
    I used two persons, two heads and such, what are yours?

    • @sebastiankamph
      @sebastiankamph  Рік тому +1

      I don't mess around too much with long negative prompts, just use some basic words. But I iterate a lot in img2img when I have a base composition that I want with the right number of heads. You can also check the highres-fix box when you're working outside of 512x512.

    • @unit1871
      @unit1871 Рік тому +1

      Generally, using "highrez fix" with square resolution works best.

    • @THENEONGRID
      @THENEONGRID Рік тому

      ​@@unit1871 cool thank you

    • @THENEONGRID
      @THENEONGRID Рік тому

      @@sebastiankamph thank you!

    • @767corp
      @767corp Рік тому +1

      Get negative prompt embedding ,slap it in there and your golden, for me saved me tons of time messing with negatives for different models and image types I wanted to achieve.
      Search for bad prompts version 2 it's super easy to find and use.

  • @bearlyphased
    @bearlyphased Рік тому +2

    I feel like I have mastered stable diffusion. I have done it all and my images are far superior to most that I see. I understand how to engineer every part of the picture at every spot in the prompt and there is a hierarchy even without weights
    I spend a lot of time going through lots of seed configurations, cfg scales, clip steps, and denoising to sometimes get a closer result to my prompt and sometimes to let AI surprise me. I like to do a large batch and sometimes single out the ones that I really like and then run them back in hi-res

    • @upsxace
      @upsxace Рік тому +1

      😂😂😂 👏

  • @joshlin50ja
    @joshlin50ja Рік тому

    Nice tunes, good advice condensed into a short video. Subbed.

  • @alexsanders8881
    @alexsanders8881 Рік тому

    Friend, first of all congratulations for the video, by the way, is it possible to train a model and during that training pause to continue on another day?
    Thanks

  • @FlorianMaeder
    @FlorianMaeder Рік тому

    What's with those video clips you show during the video (e.g. 2:33)? Were they generated with Stable Diffusion?

  • @SJ76
    @SJ76 Рік тому

    Excellent tips! Thanks.

  • @swagmum4203
    @swagmum4203 Рік тому

    When upscaling in imagetoimage, lower the denoising strenght, otherwise your picture will consist of multiple tiles

  • @LoriLinstruth
    @LoriLinstruth Рік тому

    Loving these videos (and hearing your lovely Swedish accent makes me homesick for Sweden...I lived there for nearly 20 years). Any tips about how to avoid heads being cropped off? Or how to build out an image if you like it but half the head is missing? So far that is by far my biggest annoyance.

    • @sebastiankamph
      @sebastiankamph  Рік тому +1

      Thank you for the kind words! Now I don't understand why you would ever leave Sweden, but I do admit, it's kinda cold here right now. Heads being cropped off is a regular annoyance for everyone. I would recommend working more with img2img. Find a composition that you like in txt2img, and then move on to img2img. You could also outpaint, but that's more finicky tbh. Seeing as you've practiced drawing I'd even recommend starting from your own sketches and use those in img2img. That way you'll also have the heads like you want them.

    • @sub-jec-tiv
      @sub-jec-tiv Рік тому

      Wish i could live there (i have family there but 3 generations back). Starting to feel a little crazy living in this country 😢😅😢
      Have had some luck with strong negative prompts like “cut off head” etc.

    • @LoriLinstruth
      @LoriLinstruth Рік тому

      @@sebastiankamph I had to leave for luuuuuuuv...but I get SO homesick for Sweden, even after 15 years. Oh well. Thanks for the tips! Seriously, cut off stuff is the bane of my existence. I've had mixed success with outpainting, so I'll give img2img a try and see how it goes. LOL that you know about my drawing experiment! I have to say, it was a year well spent. Thanks again for your response...you rock!

    • @sebastiankamph
      @sebastiankamph  Рік тому

      @@LoriLinstruth You can also check out high res fix for expanding. Like start with 512x512 and then a higher resolution to get more stuff in there. Oh, I sometimes have a quick look at commenters if they have content on their own UA-cam. I'm curious by nature.

    • @ShawnFumo
      @ShawnFumo Рік тому

      Not sure if you kept with SD or not, but make sure you check out openpose with ControlNet if you haven't. Should prevent heads from ever being chopped off again.

  • @atdfilms360
    @atdfilms360 Рік тому

    Thank you the helpful video. Appreciate it! Anyone know where "codeforner" setting is please? Do you have to dl it as a separate module? Thank you very much

    • @sebastiankamph
      @sebastiankamph  Рік тому +1

      It's included in automatic1111. Check settings tab under face restoration.

    • @atdfilms360
      @atdfilms360 Рік тому

      @@sebastiankamph Thank you very much! Do you find codeformer better than GFPGAN?

  • @justinwhite2725
    @justinwhite2725 Рік тому +1

    Biggest mistake is garbage words. If a model isn't trained with a word, that word sounds up tokenizing to effectively random noise that will warp your image in unpredictable ways.

  • @Android-tc3me
    @Android-tc3me Рік тому

    4:53 wait... Can you explain to me what I have to do in order to upscale the resolution? Pls (🙏🏻)

  • @simonavittoria586
    @simonavittoria586 Рік тому

    THX for this good informations. I have a question to the automatic update SD111 with the command git pull in the webui.batch. It always ran, but now i have an error message if execute : error: The following untracked working tree files would be overwritten by merge: javascript/generationParams.js modules/script_loading py. Please move or remove them before you merge.
    Anybody now what does this means and what should do? THX

    • @sebastiankamph
      @sebastiankamph  Рік тому +1

      I think you can remove that git pull in the file for now. I did too, sometimes updates come too fast and things break. So might as well leave as is and update when necessary. And if you keep getting that error in the future and need to update, do a fresh install. And thanks for the kind words 😊

    • @simonavittoria586
      @simonavittoria586 Рік тому

      @@sebastiankamph Thank you and you are welcome. I have learned a lot about stable diffusion from your good videos and tutorials. Keep up the good work 😊

  • @CaptainSkeletor
    @CaptainSkeletor Рік тому

    how do you actually paste the copied generation data?

  • @ZeroTimelineX
    @ZeroTimelineX Рік тому +1

    Ok noob question here. Do you know how prevent elements from combining? For example, if you want a cat and a dog, not a cat mixed with a dog. Thanks.

  • @cuteguy-2001
    @cuteguy-2001 Рік тому

    i have bigger problem cant generate image and also even using prompt doesnt work im using rog flow x13 1060 gpu and other solution

  • @LelioAngelus
    @LelioAngelus Рік тому

    I just like and suscribe with the first dad joke. Thanks for the content. It's really usefull

  • @GameDevNerd
    @GameDevNerd Рік тому +2

    I'm gonna go ahead and clone the repo and take this thing for a spin. How do ya think my RTX-3080Ti will do with the algorithm? And is it worth paying for cloud compute and hosting to run my instances remotely, or is running it locally a decent and productive enough experience? I'm a software engineer, so I don't mind tinkering around or building extra utilities and components to automate things and make it more pleasant to work with ...
    Also, I noticed that it seems like there are not many good choices out there other than automatic1111's web-ui front-end. Why a browser? It seems to me that it would be a better experience with the headless back-end and a native front-end UI with better features on desktop would be a vastly superior user experience, and I could write some additional tools and utilties to make it easier to work with. If anyone has any ideas/suggestions for a better front-end UI, improved features and user experience, etc, please let me know ... I'm strongly considering making a free / open source project for an improved toolset and front-end layer, to address many of the things that seem rough around the edges upon my first impressions. Any integrations, shortcuts, automations, etc people think would be valuable would be great to hear about too. I really wanna have some fun with this soon as do some things to make it better, faster, more accessible, etc. I have some ideas for performance optimizations and possibly improving the model/algorithm itself, but I will need to spend some time studying the existing ones and experimenting.

    • @sebastiankamph
      @sebastiankamph  Рік тому +1

      Your 3080ti will run perfectly fine. I run on a 3080 myself. In regard to the browser question, it's where we are atm. Model has only been open source a few months. Anyone building a superior ui will surely take market shares.

    • @GameDevNerd
      @GameDevNerd Рік тому

      @@sebastiankamph I think we can both agree that a _4090_ would run better and is necessary and justified ... the children may have to tighten their belts a bit for Christmas ... LOL, kidding, but I already planned on getting one, now I have a great idea for what to do with the 3080Ti afterward also :D
      I hit the ground doing some heavy research, found _very few_ projects implementing better UI and management tools for this other than a couple work-in-progress things ... outside of that, there's basically just a slew of websites that want to charge you a fee to use it and dole out little bits and pieces of functionality at insane costs ... that's definitely something that needs to change, I know _plenty_ other ways as a developer to make money that doesn't involve extortion or creating artificial roadblocks to create monetized gateways and pay-walls, lol. There are also a _lot_ of unexplored optimizations and enhancements that could be made to this, which I think is largely due to web developers who don't really know what they're doing just reading the setup instructions, putting it on some servers with cloud compute rental and building a simple UI and authentication + monetization layer around it. That's neither how this tech is supposed to be used _nor_ the best way to employ it. I'm not a super AI-expert, my expertise is more in graphics and engineering (mostly 3D real-time) and native engines, applications, games, etc. It seems like people are using _very_ compute-intensive (i.e., _expensive_ and inefficient) methods of upscaling, denoising, filtering, etc ... they probably should have checked with graphics engineers first, haha. I'm getting a lot of great ideas and "opportunist vibes" from this thing ... I've still got more research and tinkering to do though, and need to talk to some "hardcore" users about what could be done differently to make this better ...

    • @sub-jec-tiv
      @sub-jec-tiv Рік тому

      @@GameDevNerd Be careful, 4090 may melt down and set your house on fire 😂

    • @GameDevNerd
      @GameDevNerd Рік тому +1

      @brainswashedthisways nonsense, you just need a cold-fusion or zero-point quantum reactor to power it, which can be obtained from any local intergalactic computer store for a low price of 2,300 Blemkflarkz. Insulated wires, cables and plugs? Ha, so primtive, so inefficient ... if you use Earth technologies to power a 4090 you've made a critical error ...

  • @besocialable482
    @besocialable482 Рік тому

    Getting used to Sebastian's dad jokes. Funny stuffs!

  • @Mocorn
    @Mocorn Рік тому

    You guys seem to know how all of this works so let me ask. Sometimes I see people taking a picture of themselves into img2img and remake it in the style of Arcane or something similar. This requires the arcane.ckpt file which means you don't have the style transfer setting that you get with the 1.5 inpaint.ckpt which means that your resulting picture is in the style of Arcane but the composition is not close to your start image at all.
    Corridor crew did this in a recent movie, how? When I try I need the 1.5 inpaint.ckpt to get a matching image but the arcane.ckpt file for the actual style which changes the comp drastically.. what am I missing here?

    • @sebastiankamph
      @sebastiankamph  Рік тому

      You can merge models, but you can also adjust the denoising strength in img2img. I'd try merging a trained Dreambooth model on you together with arcane.

    • @ShawnFumo
      @ShawnFumo Рік тому

      Also, I think there is an arcane LoRA now, so should be able to use that on top of an inpainting capable model, if I understand how it works correctly.

  • @fakejaat1704
    @fakejaat1704 Рік тому

    Whichs apps you use to make your thumbnail?

  • @emmettbrown6418
    @emmettbrown6418 Рік тому

    Spider joke, and my big mistake (started Yesterday) is not putting in the negative prompts. I got a lot of stuff I didn't want that way.

  • @hhvg7057
    @hhvg7057 Рік тому

    he is looking up as if he is thinking and then reads it from the screen, haha I luv this guy

  • @darth_sidious_sheev_palpatine

    Play around with models, merge them, test them. Easy way to add alot of flare

  • @dimitopalidis2732
    @dimitopalidis2732 3 місяці тому

    could u give me tipps for making: good eyes, teeth, nails, fingers etc.

  • @dresdenk
    @dresdenk Рік тому +1

    I'm not sure how much it matters, but when I specify "irises" instead of "eyes", the eyes maybe look slightly better?

  • @RHYTE
    @RHYTE Рік тому

    Why is it that, when u use a specific seed and generate a batch of maybe 10 images from it, they all look different. Shouldn't they all look the same?

    • @androsforever500
      @androsforever500 Рік тому +3

      I think when you generate a batch of 10 it automatically uses a different variation seed for every one of those images

    • @RHYTE
      @RHYTE Рік тому

      @@androsforever500 oh yeah that makes sense

    • @Raven_RVN
      @Raven_RVN Рік тому +4

      If you are using the automatic1111 web ui then it actually changes the seed. If you enter Seed 1, and generate 4 images, then Image 1 will have Seed 1, Image 2 Seed 2, etc. It adds always adds 1.
      You can check this yourself. You can click on, say, image 7 and then use the button to retrieve the seed and can then compare this to the seed of image 6 and 8, it should always be 1 apart from each other

  • @vladislavvolkoff
    @vladislavvolkoff Рік тому

    Thank you for your content!
    What is the music at 4:06?

  • @teambellavsteamalice
    @teambellavsteamalice Рік тому

    Now I'm thinking of a process with ChatGPT generating dad jokes, matching this with images using stable diffusion and put the joke text in text balloons... 🤣

  • @hutukaTB
    @hutukaTB Рік тому

    Is restore faces useful for anime?

  • @JudgeDreadz
    @JudgeDreadz Рік тому

    My fav joke was taking out the spider! Thank you :) Sub'd!

  • @hotdog8008
    @hotdog8008 Рік тому +1

    Web Designer! That was too good to be a dad joke!
    My biggest mistake? Forgetting to reset my seed to random. That was an infuriating 10 minutes...

    • @sebastiankamph
      @sebastiankamph  Рік тому +2

      Hah, but at least you learned something 😊

    • @hotdog8008
      @hotdog8008 Рік тому

      @@sebastiankamph oh you bet I did.
      Keep up the great work Sebastian ;)

  • @LexChan
    @LexChan 3 місяці тому

    i like your jokes sometime.. give me a chill out start.. good work sebastian

  • @swannschilling474
    @swannschilling474 Рік тому +1

    Very nice!! 🤩 But use Stable Diffusion images for your B-roll! 😁

    • @sebastiankamph
      @sebastiankamph  Рік тому +1

      I tend to use a mix, but best would be for everything to be AI, even videos! :)

    • @sub-jec-tiv
      @sub-jec-tiv Рік тому +1

      I like his B-roll 🎉 he has nice taste in stock clips 😂

  • @wernerblahota6055
    @wernerblahota6055 9 місяців тому

    Great video input and also big step to more professional videos!

  • @tzgaming207
    @tzgaming207 4 місяці тому

    I always seem to get horrible body distortions when I've included some detail in the prompt that the model doesn't appear to have 'awareness' of.. I could have a prompt that's making some decent images (still very new to this), but then I change an adjective the model doesn't seem to know, & it either alters the image in a way that it can avoid having to incorporate that detail (slacker!), or it's limbs, limbs, limbs 😬 btw, I'm physically incapable of telling dad jokes... I have no children 🥁😄

  • @tylerelyt
    @tylerelyt Рік тому +2

    Agree with all of your points besides "stick with a low resolution." While 512x512 does generally result in a consistently cohesive and attractive image, the "Highres. fix" works incredibly well with 1024x1024. Waaaayy sharper and more detailed than 512x512, and looks even better than a 4x upscaled 512x512 image.
    With the "Highres. fix" I leave "Denoising strength" at 0.7, and set the "Firstpass width/height" to 512. Takes around ~8GB VRAM for 1024x1024. It's by far my favorite feature in the WebUI.

    • @NathanLorenzana
      @NathanLorenzana Рік тому +1

      Where is the Hires.fix? Is it a setting or do you download it? I just started using WebUI, thanks!

    • @KadayiPolokov
      @KadayiPolokov Рік тому +2

      @@NathanLorenzana It's a checkbox on TXTIMG in the same area as Restore Faces

    • @sebastiankamph
      @sebastiankamph  Рік тому +2

      You're right, highres.fix has been improving a lot. I think it depends if you have the hardware and time to use it.

    • @KillFrenzy96
      @KillFrenzy96 Рік тому +2

      You can also upscale a 512x512 image to 1024x1024 with img2img using the same prompt. Use around 0.2 to 0.3 denoising. This is pretty much what high res fix does. This does alter the image a bit, but adds in the fine detail where upscalers fail.
      This is useful because you can generate 512x512 way faster than 1024x1024. Once you find an image with the right layout, 'upscale' it using img2img.
      After that, you can use the regular upscale to make a 4096x4096 image.

    • @NathanLorenzana
      @NathanLorenzana Рік тому

      @@KadayiPolokov Got it! Thanks!!

  • @infographie
    @infographie Рік тому

    Excellent

  • @arthurjeremypearson
    @arthurjeremypearson Рік тому

    Specifically taking my sketches and finishing the art with flat color and inked

  • @enmodevrai2vrai91
    @enmodevrai2vrai91 Рік тому

    5:48 I love this music, who's the artist? Thanks for the helpful video bro

  • @FredRosa
    @FredRosa Рік тому

    I had to think twice about that web designer (lack of sleep perhaps) XD win!

  • @Venoxium
    @Venoxium Рік тому

    Is it just me or is IMG-IMG literally black magic? I can't figure it out. I've tried using it SO many times and can never get any good results. Currently using Anything V3

    • @sebastiankamph
      @sebastiankamph  Рік тому

      Sure is, and works great. Check some basics on img2img here: ua-cam.com/video/DHaL56P6f5M/v-deo.html

  • @Ich.kack.mir.in.dieHos
    @Ich.kack.mir.in.dieHos Рік тому

    i dont get what is automatic 111 ?? is this another name for stable diff??

    • @sebastiankamph
      @sebastiankamph  Рік тому

      Automatic1111 is currently the #1 Stable diffusion graphical user interface.

  • @frankschannel2642
    @frankschannel2642 11 місяців тому

    "Nice guy, he's a web designer." ROFL!!!!

  • @user-vr2ey8wk5h
    @user-vr2ey8wk5h Рік тому

    Thanks!

  • @moldofvalgard738
    @moldofvalgard738 Рік тому

    best tip for upscaling? please :)

    • @sebastiankamph
      @sebastiankamph  Рік тому

      SD does upscaling pretty good. Check my ultimate guide for some upscaling comparisons.

  • @bojeppe6828
    @bojeppe6828 Рік тому

    you had me at the dad jokes❤

  • @anilax
    @anilax Рік тому

    How do you have soo low subs
    You deserve way more

    • @sebastiankamph
      @sebastiankamph  Рік тому

      Thank you for the kind words, it warms my heart! 😊 You're a real 💎

  • @LouisGedo
    @LouisGedo Рік тому

    5:10
    😆 🤣 😂
    And multiple legs and toes on certain animals too!

  • @digidope
    @digidope Рік тому +1

    #1 Mistake: Thinkin you NEED to use 512x512 resolution. Nope. SD has zero problems with say 1024x2048 or my current most used resolution 384x2048. That works just fine with 1.5 and 2.0.

    • @tomaszwota1465
      @tomaszwota1465 10 місяців тому

      I wouldn't say "zero" problems. There certainly are issues as different resolutions and ratios will influence composition. For example, make the image too tall and you might find your characters growing another waist from their thighs, things like that.
      But yeah, no need to limit yourself to 512x512, just need to up that batch number.

    • @digidope
      @digidope 10 місяців тому

      @@tomaszwota1465 Thinkin that my comment is now 8mo old and how little SD has improved in that time :D You are correct that it makes weird human anatomy. SDXL has the same problem. Midjourney 5.2 does that rarely if ever. But i rarely do humans for real work. For landscapes SD did great job with big resolutions like 2024x1024.

  • @LouisGedo
    @LouisGedo Рік тому

    I have had pretty bad results with many of the batches of images I've had SD render at and above 768 X 768.

    • @devnull_
      @devnull_ Рік тому

      Then use the native resolution, 512x512

    • @LouisGedo
      @LouisGedo Рік тому

      @@devnull_
      Yes, of course. The larger rez is experimentation. Some of it renders nice results.....but generally that's a relatively small percentage

    • @ShawnFumo
      @ShawnFumo Рік тому

      Though important for people reading to know SD 2.1 is natively at 768x768, so is very appropriate with a model/mix based on that (like Illuminati Diffusion)

  • @TheAiConqueror
    @TheAiConqueror Рік тому +1

    Thank you 4 the dad jokes 😁✌️

    • @sebastiankamph
      @sebastiankamph  Рік тому

      You're very welcome! And a huge thank you for the support, your DNA is made of diamonds, you absolute gem you! 💎💎💎

  • @russelnormandia2876
    @russelnormandia2876 Рік тому

    your spider joke secured the subscribe

  • @paulmcallen
    @paulmcallen Рік тому

    i fundametally disagree. i generate the straight image from txt2img and upscale. very few times i had to img2im to fix a crop. i use img2img to clean up archival photos. running the same image 20 times is a waste of time and you will probably end up ruining the things you liked about it in the first place. 1000x1200 txt2img resolution means you dont have to do much to it

  • @Refused56
    @Refused56 Рік тому +1

    Come for SD, stay for the jokes😁

  • @rufinolarson635
    @rufinolarson635 Рік тому +2

    Using models like Waifu Diffusion, Anywhere v3, etc. which are trained on Danbooru tags has made sure that I never prompt like a human whenever making waifus. But that first mistake is definitely something I should keep in mind using SD 1.5 and other models, that's for sure. The rest were also definitely something I can learn from. Thanks for all the useful advice!

    • @hmmmmmm_3429
      @hmmmmmm_3429 Рік тому

      anything v3

    • @sebastiankamph
      @sebastiankamph  Рік тому

      Glad you liked it, thanks for the support! 🌟

    • @yesyoucan4619
      @yesyoucan4619 Рік тому +1

      Imagine Iiving in a world where you ргіḓəḟսꞁꞁꭚ talk about your іпԀῦІƺҽлт мᾲᴤtսꭇβάϮtіѳח habits on UA-cam, while ргꬲէӟոժіпᶃ to be a סתгмἇ׀ person without any ꝓѕꭚꞇԩσΙоϧίсᶐӏ enwarpments. It sure is a new generation of Internet ꞇꞕᴉլԀꭋἔꞃ we have in our woꭇId today. And it's also ẜꝺꭇьіԀɗɘὴ to spеаk սp against іt, because that's ԁιςᴐᵲίպḯꭒӑϯσꝵꭚ, so it has to become normal now.

    • @hmmmmmm_3429
      @hmmmmmm_3429 Рік тому +4

      @@yesyoucan4619 writing like that is crime who taught you to write like that EW!!

    • @KasumiRINA
      @KasumiRINA Рік тому +1

      ​@@yesyoucan4619 Imagine being so insecure you have to attack strangers online for having fun with drawing pretty girls, something humanity did since before antiquity, and being so paranoid from your persecution complex you mask your words like that.

  • @Shibby27ify
    @Shibby27ify 9 місяців тому

    How the hell do you make actions? Like people doing things, pulling, walking, holding... Most of the time is just makes people standing there. If I want someone picking something up or throwing something, there are just random hands or arms in a strange place.
    It just seems to make people standing there but not doing things!

    • @sebastiankamph
      @sebastiankamph  9 місяців тому

      ControlNet extension is your friend here

  • @max477
    @max477 Рік тому

    What is your opinion playground AI

    • @sebastiankamph
      @sebastiankamph  Рік тому

      It doesn't look different from every other online Stable diffusion ui out there.

  • @cyberhard
    @cyberhard Рік тому

    Second dad joke was the best.

  • @Nagrom
    @Nagrom Рік тому

    you forgot to use the "no cross eyes" prompt

  • @brianclark7719
    @brianclark7719 Рік тому

    Has anyone found a solution to AI lackings in multiple actors in a given prompt? For example - Dog sitting on top of a camel which is licking a cat that is chewing on a mouse? I know that the AI is somewhat in it's infancy but it bewilders me why it cannot see that level of description and interpret the final result when it draws the image. Anyone? I know I may be expecting too much, but using the reality that it can produce a dog sitting on top of a camel wearing a sombrero in the desert at night, why couldn't it?

    • @sebastiankamph
      @sebastiankamph  Рік тому +1

      Check my latest videos on ControlNet. That fixes that. Wasn't possible previously.

    • @ShawnFumo
      @ShawnFumo Рік тому

      That level of spatial awareness is still definitely tricky for the AIs in general. MidJourney V5 may or may not be a bit better if they manage to get a better text engine into it (they said they weren't sure on the time), and perhaps DeepFloyd though that isn't totally clear (and who knows when that actually comes out). Like Sebastian said, ControlNet can help, especially the posing and new MultiDiffusion region stuff.

  • @jakobhovman
    @jakobhovman Рік тому

    How do you salute a sailor, that uses AI...?
    AI AI Capten...!
    Oh...Video is great, Thank You...! Looking forward, to get my own SD and rewatch you.
    Learning in Playground now...I just can't use their SD inpaint proper...Is It Me...?
    PS: I think I will settle for the spider...most in tune.

    • @sebastiankamph
      @sebastiankamph  Рік тому

      Hah, great with another one! Thank you for the kind words. And I absolutely think you should jump on SD. Looking forward to seeing the pixel light stuff 🌟

  • @general_fail4887
    @general_fail4887 10 місяців тому

    DO NOT use "restore faces" if you want to generate anime/ cartoon style images.
    Otherwise, SD tries to humanize the faces, which then ends in alien-like results

  • @divinesimmortus5655
    @divinesimmortus5655 Рік тому +1

    Anyone got a spare graphics card haha with my radeon r9 290x it takes about 20 min for a 512x512 with 10 sampling steps xD

  • @VovanEkb
    @VovanEkb Рік тому

    do not try to turn on the restoration of faces, when generating art, or any non-photo realistic effect, try to do without it until the last, this jackdaw (yes, I know about the possibility of setting it up) greatly affects the overall style, and often completely destroys it. I'm saying that you should try to achieve this without her, and you will get a much cooler and more consistent result at the output!

    • @KasumiRINA
      @KasumiRINA Рік тому

      Just make it save both before and after pic, and you can choose.

  • @MoS910
    @MoS910 Рік тому

    I guess short Midjourney prompt looks lazy.
    More details more specifications, I would use same prompt fro SD in Midjourney because Midjourney prompts like let AI decide instead of the user... like wombo dream, I don't like short prompts!

  • @Steamrick
    @Steamrick Рік тому

    So yeah... I forgot about restore faces. Dang.

  • @lpanzieri
    @lpanzieri Рік тому

    The jokes are so terrible they end up making me laugh :D

  • @reitakuhemphill
    @reitakuhemphill Рік тому +1

    It was a mistake to compliment him on the dad gags.

    • @sebastiankamph
      @sebastiankamph  Рік тому

      I've ventured down a path I cannot stray from now.

  • @pirobot668beta
    @pirobot668beta 11 місяців тому

    If you save txt2img as JPG, the comment field of the file will have all your settings.

  • @ThePixelkd
    @ThePixelkd Рік тому +1

    Definitely the spider one

  • @captainkaput7281
    @captainkaput7281 Рік тому

    Please include (for absolute beginners) in the title. Thanks.

  • @ajudicator
    @ajudicator Рік тому

    He’s a web crawler
    Would have been a better end to the punchline
    Thanks for the videos

  • @qaesarx
    @qaesarx Рік тому

    Those faaaaaaaaaans 🌀🌀

  • @SaintMatthieuSimard
    @SaintMatthieuSimard Рік тому

    OMG! 🤣A 🕸Web designer 🕷

  • @schmutly
    @schmutly Рік тому

    last joke the funniest 😆🤣😂

  • @rogerc7960
    @rogerc7960 Рік тому

    Use a thesaurus

  • @Kolesoprodut
    @Kolesoprodut 6 місяців тому

    04:06

  • @rafaelestevam
    @rafaelestevam Рік тому

    I get better results in 512 x 448, then hires

  • @123nicanor
    @123nicanor Рік тому

    How bout trying the grab a pen and try real drawing?

  • @Grimleal-ws1mk
    @Grimleal-ws1mk Рік тому

    I fatfingered my mouse and clicked into the middle of the video to hear "anime waifu".
    I'm scared