You have one of the absolute BEST channels for intermediate/advanced SD users, with real-world workflows. Your explanations, tips, speed, and dialog ist absolute top! PLEASE continue what you're doing here, and I would even try using "Advanced, Real-World Project Workflow" or something like that in your titles. Definitely NOT for beginners, and it should keep the complaints down. We NEED more of you and your tutorials.
Very good tutorial. I came to the same conclusion that pixel upscale is much better than latent and this confirms it. I haven't bothered trying an additional tile controlnet on top, even though that's what I used to do with auto1111. Now that you've mentioned it, I'll have to give it a try. I think this format is great, progressing from left to right, and your demonstration is very concise and easy to follow. I think this is a good way to cover more intermediate topics, and if someone is just learning the basics they can review step by step and do further research if they don't understand something. At least that's my opinion. Thanks for posting!
yeah you need very high denoise for latent-upscale. if fidelity to the original image is not important latent upscale can be still used, possibly together with SDXL in the second pass
You're definitely a master of ComfyUI. Your videos are great-both educational and fun to watch. I love that you always provide your workflows without any restrictions. Keep up the excellent work!
These are by far the best videos about this topic by far. Many other influencers are just users, they don't understand the concepts behind this. I learned so much. Keep the videos coming.
I feel like my understanding of ComfyUI and StableDiffusion have progressed so rapidly since discovering your videos. I really appreciate the way you don’t just show a workflow, but provide an explanation and show alternatives and really make it clear why you’re making the choices you make. It’s the difference between giving a man a fish versus teaching him how to fish.
I love your tutorials I learn so much from your work!!! I know it takes a lot of time and effort to make these. I really appreciate you sharing your knowledge with us. Keep them coming, I would watch anything you would likely want to make and share. 🙌
Your format is great! Trial and error seams to be the name of the game and you don't cherry pick. This makes your videos incredibly helpful. And this specific use case with pixel art is very interesting! I'm following the space carefully and personally I'm looking for use cases that I can implement and make a business from. But in my experience consistency is the hardest part. Say a client loves the white haired sorcerer - can we deliver more around that character or are we stuck? Idk, but I have a feeling with your videos I'm going to be able to unlock a new super power that 99% of the world is sleeping on. Thank you 🙏 so helpful, so informative 🤙
Thanks for taking the time to write. What I don't like of many youtube tutorials is that they concentrate on the latest/coolest technology ignoring to consolidate the basics. So I'm trying to do fill that gap. For example prompt engineering is very much ignored but very important. Regarding "variations". ControlNet and IPAdapter together generally grant a pretty stable image generation.
@latentvision yes, I have the same experience with everyone wanting to be first. But ironic you are first, cutting edge and an actual contributer to the space. Much respect!
I like what you're doing here, and that you make the workflow files available. I have no complaints about these videos, and I'm very glad that you're sharing how you do these things. I've learned more from your videos than most others simply because you spend that extra time to show how and why you wire things the way you do.
This was awesome and very insiteful, I am still getting to grips with Comfy. Your videos are great, thanks. Also, thank you so much for the free workflow :)
Thank you very much, I really appreciate your work! You are really creating big stuff for the community by coding this gem and even making videos about the usage. This is fantastic and I learned tons from your videos. I really like, that you show more advanced techniques, most other channels do all the basic stuff like explaining how to install a new node in each and every video. Again, thank you so much!
This is amazing! I've been trying to just use controlnet with line art and color grid models to redo pixel art but this is so much more accurate. Can't wait to try
your knowledge is amazing and great videos...I am always surprised at how endless comfyui appears to be...There is so much in some ways it is purely intimidating and I spend more time thinking where do I start....lol....its almost a university course to fully understand comfyui...lol......amazing :-)
Thank you for the clear explanations. It's very useful to know why you're doing things a certain way. Little things like "the order doesn't matter" on condition concat nodes are very valuable. Edit: So in a next video you actually mentioned that it does matter. The from has lower influence.
Great videos. The only constructive critique is slow down a little and show where you’re connecting things to more closely. I’m glad I found your channel, seems geared for people who have an 20:42 intermediate understanding of comfy, which is severely lacking on UA-cam
thanks for your feedback! It's always hard to find the right pace as I don't want the videos to be too long (you know watching a workflow is not really that much fun). I tried to increase the resolution, that should help a little. (be sure to select 1440p and enlarge the video). Most of the times there's only one place where the connection could go, but I'll try to zoom in when it's ambiguous. Also check the video description for the workflows!
@@latentvision you Bet! I was able to follow along fine, you made a great tutorial. Seeing things laid out did give me a better insight on how upscaling works and I don’t know why I didn’t think to do other passes. These are educational videos and people who watch are here to learn and I bet they rewind and watch certain parts over and over. All the other channels out there spend so much time talking about how to install comfy and python, what feels like, on every damn video.
Love your tutorials! 🔥Always learn something new! But I had to learn the basics before I went in on your excellent videos. I had them saved for when I was ready for them. 📚
also, really hoping to see a workflow like this form into something that could be used for an animation. from motion jpg, feeding each image into the pipeline and seeing what a 10 second clip would look like. =)
Very nice. Mixing the good old Amiga 500 / Atari ST times with the modern AI tech. Really impressed me and made me nostalgic. :D One issue I have: After you were chosing the Speedball character, I am always hearing "ICE CREEEAM, ICE CREEEAM". That's annoying. :D
Inspiring! Liked and subbed. This is my first of your videos..and it sounds like you’re testing workflow tutorials (like this)..well, It’s just the sort of thing I was looking for. ,{^_^}”
incredible wooow, I have a silly question???????????? I was not able to use the workflow after downloading, since tries to load and it doesnt support text versions, changed to .png still not. would you please help me????
I notice that ComfUI with IMG 2 IMG and Inpainting, it seems to degrade the entire photo each time it is used. So if you feed the output back as the input and keep editing, it will get unusable. Automatic1111 doesn't have this issue, unless it is the model or settings i'm using idk.
Great stuff as always. One question, what controlnet would you use for SDXL? I've achieved great results by using the tile controller, but it doesn't work when using any SDXL checkpoint or controlnet, and as there is no tile controlnet for SDXL, as of now, the results are no so good. I've tried different controlnet types, but unfortunately haven't been able to reproduce the same results as with the tile controlnet. Anyway, I really hope they will release it for SDXL as it really makes a difference in this type of image reproduction.
Another great video, thanks. One question - why do you need to change the seed from the 1st stage to the 2nd stages? I had been thinking for upscaling you'd want the same seed?
@@EH21UTB SD is a game of chances. It's not about finding the best workflow but finding the workflow that have the highest chances to get what you want. Over time you discover the best practices that in your specific scenario give you the highest rate of best results.
Complimenti, gran bel video. Ottimo format, mi sembra di livello intermedio: nè troppo basic, nè troppo complesso. Durata giusta. Un paio di appunti al workflow: ci sono dei nodi utilizzabili per ricreare il prompt a partire dall'immagine caricata: mai provati? (simili al CLIP interrogator di A1111). Secondo, visto che parli di upscaling, come mai non hai inserito un workflow in cui utilizzi Ultimate SD Upscaler? Tra i vari metodi, di solito è questo con cui ottengo i migliori risultati. Comunque ripeto, ottimo video e spero continuerai a pubblicare video di livello intermedio-esperto. Grazie mille
i text-to-image possono essere utili in alcune circostanze, ma tendono a essere un po' generici. Ultimate SD Upscale implementa molti dei metodi che ho illustrato nel filmato con in aggiunta la possibilità di dividere l'immagine in tiles (e seam-fix), ma non fa molto di diverso da quello illustrato qui.
Amazing work, thanks so much! FYI IP Adapter does great at adding skin detail, just give it a cropped example of some close up skin, even works with batch images/face model!.. 🪄✨
Hi, I don't imagine doing this video "in reverse" is something you'd be interested in, but I've found it very difficult finding resources for generating pixel art that sticks to specific styles. Do you think it's worth trying to do it with a variant of this workflow? For example, say you'd like to use your realistic human end result from this video as the reference image for a sprite head. Do you think that workflow would make sense? Sorry for all the questions, I just find that there's a significant amount of paid tools related to pixel art in the SD sphere, but very little actual information on how to generate it without specifically trained models that lead to inflexibility.
that would be really interesting. Pixel art is very difficult to do without a trained model and even so the result is generally either not great or over trained. SDXL is pretty good at pixel art if you don't need "pixel perfect" resolution. I'll make some tests to see if there's enough material for a video. Thanks for the suggestion.
I think youtube ate my comment due to having links on it, but I just wanted to add that in my research I found two good tools that help us lower the "entry bar" from the base model by not needing pixel perfection. These are ComfyUI-PixelArt-Detector by dimtoneff (palette control and pixel grid) and the 1x PixelSharpen upscaler. Anyways, I hope there's enough to make something interesting out of it! I think there's enough tools to make a good transition from illustration to pixel art happen, but putting it together with control tools feels completely out of my depth personally.
Tile ControlNet is not available on SDXL unfortunately. There's a tile model for XL but it's only for anime. I haven't checked if they released a generic model... things move so fast. Apart from that when doing videos SD1.5 is much faster so it's a little easier. Anyway you can totally use SDXL as upscale model in the second pass and actually the result is generally very good.
You have one of the absolute BEST channels for intermediate/advanced SD users, with real-world workflows. Your explanations, tips, speed, and dialog ist absolute top! PLEASE continue what you're doing here, and I would even try using "Advanced, Real-World Project Workflow" or something like that in your titles. Definitely NOT for beginners, and it should keep the complaints down. We NEED more of you and your tutorials.
I really appreciate the motivational comment, thanks. It works 😄
i am just totally shocked by the quality of your content, absolutely astonishing
Very good tutorial. I came to the same conclusion that pixel upscale is much better than latent and this confirms it.
I haven't bothered trying an additional tile controlnet on top, even though that's what I used to do with auto1111. Now that you've mentioned it, I'll have to give it a try.
I think this format is great, progressing from left to right, and your demonstration is very concise and easy to follow.
I think this is a good way to cover more intermediate topics, and if someone is just learning the basics they can review step by step and do further research if they don't understand something. At least that's my opinion.
Thanks for posting!
yeah you need very high denoise for latent-upscale. if fidelity to the original image is not important latent upscale can be still used, possibly together with SDXL in the second pass
You're definitely a master of ComfyUI. Your videos are great-both educational and fun to watch. I love that you always provide your workflows without any restrictions. Keep up the excellent work!
6:00 Wow I knew the VAE was hungry for pixels but *gosh!* , also 18:36 "Zura Janai, Katsura Da!"
These are by far the best videos about this topic by far. Many other influencers are just users, they don't understand the concepts behind this. I learned so much. Keep the videos coming.
I feel like my understanding of ComfyUI and StableDiffusion have progressed so rapidly since discovering your videos. I really appreciate the way you don’t just show a workflow, but provide an explanation and show alternatives and really make it clear why you’re making the choices you make. It’s the difference between giving a man a fish versus teaching him how to fish.
I love your tutorials I learn so much from your work!!! I know it takes a lot of time and effort to make these. I really appreciate you sharing your knowledge with us. Keep them coming, I would watch anything you would likely want to make and share. 🙌
This is one of the few great channels that explains how it works!
Your format is great! Trial and error seams to be the name of the game and you don't cherry pick. This makes your videos incredibly helpful. And this specific use case with pixel art is very interesting! I'm following the space carefully and personally I'm looking for use cases that I can implement and make a business from. But in my experience consistency is the hardest part. Say a client loves the white haired sorcerer - can we deliver more around that character or are we stuck? Idk, but I have a feeling with your videos I'm going to be able to unlock a new super power that 99% of the world is sleeping on. Thank you 🙏 so helpful, so informative 🤙
Thanks for taking the time to write. What I don't like of many youtube tutorials is that they concentrate on the latest/coolest technology ignoring to consolidate the basics. So I'm trying to do fill that gap. For example prompt engineering is very much ignored but very important.
Regarding "variations". ControlNet and IPAdapter together generally grant a pretty stable image generation.
@latentvision yes, I have the same experience with everyone wanting to be first. But ironic you are first, cutting edge and an actual contributer to the space. Much respect!
OH MY GOD!
Again, a tons of helpfull info in your videos. A whole pipeline here)
Thank you very much for sharing your techniques.
You are really good at explaining. I'm learning so much with you.
For sure MagnificAI was born from this tutorial. Magnificent information.
dude you dont need to change the style of the tutorials, this is really amazing toptier content
Great video! Your explanations are so clear!
And btw.. The format is great..!
Cool process, The wig in negative is a nice trick!
I like what you're doing here, and that you make the workflow files available. I have no complaints about these videos, and I'm very glad that you're sharing how you do these things. I've learned more from your videos than most others simply because you spend that extra time to show how and why you wire things the way you do.
The best person to teach us everything about Ai image generation. I've dazzled of how effortless you make the process seem. Sir, you have my respect.
This was an amazing tutorial. Thank you for the in-depth descriptions. I had no idea that latent upscalars were so efficient.
your mastery of these concepts is simply mind numbing. love this.
This was awesome and very insiteful, I am still getting to grips with Comfy. Your videos are great, thanks. Also, thank you so much for the free workflow :)
Amiga Forever!!! This is excellent. Thank you!
Excellent. Thank you very much for the technique and the explanation!
Thank you very much, I really appreciate your work! You are really creating big stuff for the community by coding this gem and even making videos about the usage. This is fantastic and I learned tons from your videos. I really like, that you show more advanced techniques, most other channels do all the basic stuff like explaining how to install a new node in each and every video.
Again, thank you so much!
I've learnt so much from watching through your various videos, you're very knowledgeable. Have a happy holidays if you celebrate them :)
This gave me a fun afternoon recreating the golden axe characters. Awsome tutorial
Loved the video, but loved also the AMIGA
Amiga Strikes back!
Your content is among the best hands down. I like the deep, straight to the point approach. Maybe just raise the mic volume a little bit?
This was a master class. God bless you.
This is amazing! I've been trying to just use controlnet with line art and color grid models to redo pixel art but this is so much more accurate. Can't wait to try
I mean line art / canny / soft edge usually one of those.
Ciao Matteo, ti seguo dagli Stati Uniti. Ho appena scoperto questo software e i tuoi video. Ottimo lavoro!
your knowledge is amazing and great videos...I am always surprised at how endless comfyui appears to be...There is so much in some ways it is purely intimidating and I spend more time thinking where do I start....lol....its almost a university course to fully understand comfyui...lol......amazing :-)
Oohhh... the diffusion MAESTRO is here with an amazing new video!! My week is fulfilled. Many infinite thanks amigo!! ❤️🇲🇽❤️
Thank you for the clear explanations. It's very useful to know why you're doing things a certain way. Little things like "the order doesn't matter" on condition concat nodes are very valuable. Edit: So in a next video you actually mentioned that it does matter. The from has lower influence.
Dude. You are doing God's job here. Love your tutorials.
This is great. I will book mark this and use it many times.
Terrific video, thanks! The detailed walk through is great. You have a new subscriber.
Great videos. The only constructive critique is slow down a little and show where you’re connecting things to more closely.
I’m glad I found your channel, seems geared for people who have an 20:42 intermediate understanding of comfy, which is severely lacking on UA-cam
thanks for your feedback! It's always hard to find the right pace as I don't want the videos to be too long (you know watching a workflow is not really that much fun). I tried to increase the resolution, that should help a little. (be sure to select 1440p and enlarge the video).
Most of the times there's only one place where the connection could go, but I'll try to zoom in when it's ambiguous. Also check the video description for the workflows!
@@latentvision you Bet!
I was able to follow along fine, you made a great tutorial. Seeing things laid out did give me a better insight on how upscaling works and I don’t know why I didn’t think to do other passes.
These are educational videos and people who watch are here to learn and I bet they rewind and watch certain parts over and over.
All the other channels out there spend so much time talking about how to install comfy and python, what feels like, on every damn video.
Best comfyui videos out there
wow, I didn't realize how many different ways there were to upscale
Love your tutorials! 🔥Always learn something new! But I had to learn the basics before I went in on your excellent videos. I had them saved for when I was ready for them. 📚
also, really hoping to see a workflow like this form into something that could be used for an animation. from motion jpg, feeding each image into the pipeline and seeing what a 10 second clip would look like. =)
Stay tuned! I have great news about animations!
muahahaha! exciting times. =) @@latentvision
Love your work! Thanks! Learn soo much every time! Cheeers
Wonderful as always...
Non male il tuo canale Matteo. P.S. Ero un fan di Speedball 2. Iscritto!
your videos are always so great!
Great work. I like the style of your video
Grazie! More such tutorials please!
What about something to improve dallE images? And to have an input face?
If you want an easy 100k views, do this to each of the Stardew Valley pixel portraits. I’d love to see that video!
best teacher ever 😊
Very nice. Mixing the good old Amiga 500 / Atari ST times with the modern AI tech. Really impressed me and made me nostalgic. :D One issue I have: After you were chosing the Speedball character, I am always hearing "ICE CREEEAM, ICE CREEEAM". That's annoying. :D
don't worry you are not alone I had "Ice cream!" in my head during the whole video 😄
THE GOAT
great, thanks a lot!
SPEEDBALL 2! My favorite Amiga game!
Inspiring! Liked and subbed.
This is my first of your videos..and it sounds like you’re testing workflow tutorials (like this)..well, It’s just the sort of thing I was looking for.
,{^_^}”
Another great tutorial
great work please keep going 💪
it‘s not easy but well explain. thank you
incredible wooow, I have a silly question????????????
I was not able to use the workflow after downloading, since tries to load and it doesnt support text versions, changed to .png still not. would you please help me????
I notice that ComfUI with IMG 2 IMG and Inpainting, it seems to degrade the entire photo each time it is used.
So if you feed the output back as the input and keep editing, it will get unusable.
Automatic1111 doesn't have this issue, unless it is the model or settings i'm using idk.
set the latest mask together with the inpainting. that is what a1111 does
Love you videos!
Can you do a video on how to do this process in reverse? And control for pixelisation size?
that would be pretty complicated without a trained model... but it would very interesting indeed
Great stuff as always. One question, what controlnet would you use for SDXL? I've achieved great results by using the tile controller, but it doesn't work when using any SDXL checkpoint or controlnet, and as there is no tile controlnet for SDXL, as of now, the results are no so good. I've tried different controlnet types, but unfortunately haven't been able to reproduce the same results as with the tile controlnet. Anyway, I really hope they will release it for SDXL as it really makes a difference in this type of image reproduction.
Another great video, thanks. One question - why do you need to change the seed from the 1st stage to the 2nd stages? I had been thinking for upscaling you'd want the same seed?
keeping the same seed you risk to burn the image (ie: you impress the same image twice). That is especially true for img2img (without upscaling)
@@latentvision Thank you for the reply and information. I guess then it's not such a good idea to use the Global Seed? It's such a convenient node.
@@EH21UTB SD is a game of chances. It's not about finding the best workflow but finding the workflow that have the highest chances to get what you want. Over time you discover the best practices that in your specific scenario give you the highest rate of best results.
Complimenti, gran bel video. Ottimo format, mi sembra di livello intermedio: nè troppo basic, nè troppo complesso. Durata giusta. Un paio di appunti al workflow: ci sono dei nodi utilizzabili per ricreare il prompt a partire dall'immagine caricata: mai provati? (simili al CLIP interrogator di A1111). Secondo, visto che parli di upscaling, come mai non hai inserito un workflow in cui utilizzi Ultimate SD Upscaler? Tra i vari metodi, di solito è questo con cui ottengo i migliori risultati.
Comunque ripeto, ottimo video e spero continuerai a pubblicare video di livello intermedio-esperto. Grazie mille
i text-to-image possono essere utili in alcune circostanze, ma tendono a essere un po' generici. Ultimate SD Upscale implementa molti dei metodi che ho illustrato nel filmato con in aggiunta la possibilità di dividere l'immagine in tiles (e seam-fix), ma non fa molto di diverso da quello illustrato qui.
do we have tile controlnet for sdxl that may work with comfy?
I was on Atari, I couldn't wait to test it on Teenage Queen... uh I mean on Xenon 2!!! 😅
That's amazing
Is big difference with model Realistic Vision v51_v51VAE like you used and new Realistic Vision V6.0? It is possible to download v51_v51VAE?
Amazing work, thanks so much! FYI IP Adapter does great at adding skin detail, just give it a cropped example of some close up skin, even works with batch images/face model!.. 🪄✨
that's a great tip! thanks!
This was a great idea :) fun stuff.
I didn't know VAE was lossy. Would it be a better idea generally to use a checkpoint with a baked-in VAE?
Amaaazing but Im not abble to use the workflow, would you please help me??
Amazing tutorial, but I would prefer a1111, comfyui for me is too complicated
Really cool - this was Stabel diffusion?
yes
Hi, thanks for tutorial. Could you share what hardware do you use so that it so quick? Thank you.
All the renderings are sped up in post production otherwise it would be a terribly boring video :) I have a 4090 anyway
Hi, I don't imagine doing this video "in reverse" is something you'd be interested in, but I've found it very difficult finding resources for generating pixel art that sticks to specific styles. Do you think it's worth trying to do it with a variant of this workflow?
For example, say you'd like to use your realistic human end result from this video as the reference image for a sprite head. Do you think that workflow would make sense?
Sorry for all the questions, I just find that there's a significant amount of paid tools related to pixel art in the SD sphere, but very little actual information on how to generate it without specifically trained models that lead to inflexibility.
that would be really interesting. Pixel art is very difficult to do without a trained model and even so the result is generally either not great or over trained.
SDXL is pretty good at pixel art if you don't need "pixel perfect" resolution. I'll make some tests to see if there's enough material for a video. Thanks for the suggestion.
I think youtube ate my comment due to having links on it, but I just wanted to add that in my research I found two good tools that help us lower the "entry bar" from the base model by not needing pixel perfection. These are ComfyUI-PixelArt-Detector by dimtoneff (palette control and pixel grid) and the 1x PixelSharpen upscaler.
Anyways, I hope there's enough to make something interesting out of it! I think there's enough tools to make a good transition from illustration to pixel art happen, but putting it together with control tools feels completely out of my depth personally.
Great idea! How it will look in game, with animations?
that would be an interesting experiment to do :)
Where do you get the software?
very cool! How would you do the same with SDXL?
SDXL doesn't have tile controlnet, there are sone alternative techniques but they are not as effective.
why still use sd1.5 inplace of sdxl?
Tile ControlNet is not available on SDXL unfortunately. There's a tile model for XL but it's only for anime. I haven't checked if they released a generic model... things move so fast. Apart from that when doing videos SD1.5 is much faster so it's a little easier.
Anyway you can totally use SDXL as upscale model in the second pass and actually the result is generally very good.
What about the other way around - converting a real world face into a pixel character? 😊
there are some checkpoints that are pretty well trained at pixel art...
Not working with directml :/
Can confirm that this one works: ComfyUI_IPAdapter_plus/tree/c28a04466b17d760a345aea41d6a593c0a312c95
the link of the workflow dosnt work
he has green hair not blue 😅
let's agree on turquoise 😄
now we have the knowledge, but not the gpu 😂