As an artist I initially did not like the AI stuff.. But as I am working on remaking a game I did way back in high school, this is a massive time saver. I'd take my designs and run it through the AI and get different variations that have allowed me to quickly finish a character's final redesign. This is quite amazing, it also saves money as I am using this very technique to have 3D models done of the main character, that I will later have printed for a statue. I will also be using it to have the figures done of the characters to another project.
it's heart warming to see that some people find the right use behind the black spots of this revolution. i've started learning blender for a bit for chara modelling. I was painfuly missing orginal 2D reference. then i saw Royal Skies video and was sold Instantly... however i did not touch blender since. time and such you know :/
Thanks for making these videos! I started playing with stable diffusion a couple of days ago and binged all your videos. SD is honestly too fun, I sat up to like 4 am yesterday inpainting instead of in-bedding😅 I'm really impressed that your videos are so concise without being hard to understand. Not to mention funny! Everything looked really daunting at first but I just want to learn more, and you make that a lot easier, and a lot more entertaining. So thanks!
For those who are lazy (like me) here's the text at around 11:08 close up of a man,{{1$$__cameraView__}}, {{1$$orientation__}}, {{1$$__expression__}} full body shot of a man, dynamic pose, {{1$$cameraView__}}, {{1$$__orientation__}} upper body shot of a man, {{1$$__orientation__}} Change Man to Woman if you want a woman
If you struggle with getting it to make consistent faces, I highly recommend making the face in Vroid Studio and Photoshopping them in. It's also necessary when making a lora to get close ups of the faces, and also for details on the clothes that you will want to be able to inpaint later in case the program struggles.
Good Idea. I know Vroid and it have gotten pretty good at making anime oc. You can use any videogame character marker to get close to the style you look for.
I trained a checkpoint on a 3D model when Dreambooth first came out last year and it turned out fairly well in that I could change backgrounds and poses. I tried again the other day on a Lora and it was terrible. I was left scratching my head until I saw your video and you explained that all of the auto captioning (which did not exist back then) was likely throwing it off. Thanks so much for the tip, can't wait to try it again! Exciting stuff!
ty!!! We are working on a fairly in depth LORA training guide so If you keep running into a wall I hope that video helps when it comes out! (And then there is the discrod for help ofc XD)
@@lefourbe5596 I would share the one I made last year but alas it is for work. Was pretty easy in dreambooth, I wager if I just filter the excess of the auto captioning I should be OK.
Great accumulation of info, workflow and the topic of consistency is so important.++ Consider breaking your next videos into 10 minute segments ( which means more videos for you!) so they're more digestible for us. Separate them into bite-sized skills all under one related thread or collection of videos. Yours is the only tutorial I've had to slow the playback down for to fully hear what you're saying because your visuals are also moving very quickly. You can fill up the extra time by both being sure you're explaining the next step and not just showing what to do but WHY you're doing it; plus you can give some alternative examples regarding what you're demonstrating. Great work but it feels like I'm "rewinding" more than I'm playing!
Thanks for the feedback! I dont know about making them 10 min each, but we plan on making a firt video covering the basics and how to understand them. And then another video with more advanced info. Tyty!!
Thank You so much. I do anime edits and copyright is a hige issue right now and switching over to ai has been easier but people love the stable diffusion more then just pictures 😅 ive been looking for a way to get more consistent characters too ❤ you get a sub my brother!
I have gave you a thumbs down on the first video I watched and I was about to walk away and I don’t know what happened but I was just really interested about the topic so I listened to it and then gave you the thumb up and then subscribed you are now one of my favourites UA-camr on the topic I appreciate your genuine interest and your dedication🎉🎉🎉🎉
@@Not4Talent_AI I don’t know how I felt but it is so rare that I give thumbs down normally I do mean it and given that you are in my top list make me feel like… I remember that it was genuine then and now I don’t know why it was like that… but knowing that I would have never heard you saying *Popimpokin* many times in the other video makes me feel very happy that I changed my mind
I will kindly argue about some of your points: 1. When you caption the character you should describe the outfit and any accessories as well. Trust me, you will have an easier time if you want to prompt your character in any others outfits which is not the default one or to undress it. 2. Increasing the max resolution of the training dataset to 768x768 does actually make a difference on the overall quality of the images. I would drop the number of batch size to 2 (this wil be comfortable for most people without getting the Cuda memory error) and set the image resolution to 768x768. Lora training already takes little time so don't sacrifice image quality for training speed. 3. If you're a beginner in Lora training don't bother with regularization images, you're overcomplicating yourself (I know in your video you said it's optional, just wanted to make a mention about this)
Yes! I agree with everything. For better results having 768 or even 1024 is the best option. But more time consuming. If you want the perfect training then thats perfect. In this case we were testing so I thin 512 by 512 is the fastest testing option. Also true what you say about tagging. Even though I wasnt looking for sn ourfit change as that is prettt much the character. The face and hairstyle are very standard😂😂 And for epochs also true, gpu is pretty much everything there hahaha
That exactly what i told N4T when WE made the vid 😁. 1. You are right and it's a choice we made. Describing everything will make the Lora harder to prompt and make comparison between diffents designs difficult cause of bleeding concepts. There is more in the video than i have planned. It would have been in part 2 for many of these details to me. Trust me he is well aware. My best Lora is a one that need a long freak prompt to get every part of it's complex design. (My profile pic) Facial mark, Horn, asymetrical, heterochromia, high collar, black sclera, ponytail, gauntelet, belt, coffin pouch... For second point, i disagree cause GTX and 20** RTX cards exist and are slow with less VRAM. I get it as i have a 3090 but even then i prefer to be able to tweak LR, models and network dim first before. Especialy for N4T was short on Time and run a GTX 1080. So yeh 512 training first with your 1024 dataset.
Super cool video and explanation. I have made a character of some sort just yesterday, so training my own lora would be really helpfull, I guess. I'll try it next week and see how it works out.
You seem to know so much, thanks for sharing! Frankly, for a beginner the majority of the videos are hard to understand though. You present so many different features and techniques in one video that it gets overwhelming. Would appreciate it if you also make some videos in the future for noobs like me, with a slower pacing and less concepts but more in-depth explanations 😅❤
Tyty!! Hahaha I ve been tolda that, true that the channel is more aimed towards people with more experience. I have a vid for most concepts touched in this one, but they are still just as fast😂 I might go back and update some vids on the basics eventually. In a more calmed and beginerfriendly way hahah Ty for the feedback!!
Great idea! But a character sheet in one drawing style will make the Lora learn that style. That's why it is suitable to use the first result to create new images with more variety and retrain again as suggested.
true, that's a pretty good idea. Instead of having the lora already trained in a style and retrain from there. Create different styles directly, right? (not sure if that is what you mean. but sounds possible and nice)
You should try the same process but start with a 3d model character you can pose and use as the Lora training data, I feel that would best process for a small studio, pay an artist to make a custom 3d character then use that as the Lora training base
I think that could be interesting yeah. Mainly seeing how good can the LORA get the character. (Lefourbe tried something like that, with a very hard character, and he is getting pretty nice results). so should be possible I think- Things to note in the "future testing" notebook for sure! tyty!
Yeah i'm doing that mostly ! I'm trying to make some example for the next video. The 3D character i have are not so well made and i was hopping to improve their visuals with SD without sacrificing their design.
Wow, very detailed and beautifully produced video. My guess is that there is no fully online service that would allow a creator to do the things you describe here, but please correct me if I am wrong.
thank you so much!! As far as FREE online services for this, I wouldt say it is possible atm, since google colabs are RIP. But there are a few payed options out there. Like thinkdiffusion for Automatic1111 (you probably could find a free solution to this as well, even though I cant tell you one atm cuz they usually either close up or end up changing to a payed model over time). And for the training Dreamlook.AI is a very good option. (I have been sponsored by them in the past. But I still think they are an awesome option) If you have any question or doubt, you can ask on discord, there is a lot of people that might be using online tools there as well.
Hi!! It is a way to try for AI to get better colors. I explain it a lil in this video: ua-cam.com/video/wso_O2vk2dw/v-deo.html and yes, it is necessary to use capital letters :3
@@Not4Talent_AI thanks for the instant reply! Does it work in ANY other platforms like midjourney? Also thanks for the provided link, definitely will check it!
20:50 I saw one guy provided various results experimenting with Network Rank and Network Alpha. So the best was 128 to 1. I also experimented with my characters and also found that the best is 128 to 1. But my characters were photorealistic, for anime maybe your parameters better.
Hey man, great job. Are you still using a1111 and lora training or have you switched to comfyui? Another question I have is, when you were solving all this stuff, did you have any prior knowledge? Like software knowledge or graphic design. Especially when I look at comfyui workflows, it seems impossible to reinstall it.
Hi! Tyty Im still using a1111, even though I dont train many lora. But thats bc I have no need for that atm. All the knowladge I have is from studing animation in uni. Comfy ui I havent gotten to it but I have worked with nodal stuff before like nuke, maya, ue, blender... so it isnt as intimidating to me. I dont know about installing all that stuff tho, havent done it yet 😂
It is okay to use for commercial use. Just check if the model you are using allows it or not I'm trying to find a way to make the same thing for non-humanoids. I think maybe you could use midjourney, or batch generate without controlnet and pray xD (ty for watching, btw :3 )
Thanks for this video . although I doubt able to prepare and fix the dataset like yours, Now I`m bit understand the basic concept - idea and workflow needed And about the discord invitation , yeah ... when Lex offering you to hangout in his crib you should come and ready to feel amazed
Thanks!! It will depend on your pc speed and luck when generating the character. Also the complexity of it. So i cant really giva good estimate for this. For me it took like 2 hour to prepare an ok dataset once I knew what I was doing. And then training took a bit more. With my current pc it would be a total of maybe 2 hours total if the character isnt super hard. If you dont really care super ultra much bout the character having a lot of precission. Then you could do this in 30 min + wtv it takes for the training
Hello there ! It's Le_Fourbe ! Damn why YT trow me an error when i comment ? I'm so late ! Anyway i'm here so i can answers some of your question that will probably gonna cover in part 2 :)
hello friend, a friend and I are starting to develop lora character models but we don't fully understand what regularization images are, we have a question about whether regularization images are random images for the model to convert them into your character or if they are images that the model has created and that have gone wrong in order to correct the model, we also borrowed a better computer during this weekend (since our pc are "potatoes", our best graphics and with more vram is an RX 590 the which doesn't even support cuda xd) so we want to take advantage of the fact that they lent it to us to train better models, currently we train a model with 80 images, 3000 steps and 1 epoch and we train another model with 460 images, 4600 steps and 1 epoch, what kind of training do we get? would it give better results? in neither of the 2 we used regularization images because we did not understand that point :c
If you have more than 60-70 images dont even bother with regularization! I exain it a little more here tho: ua-cam.com/video/xXNr9mrdV7s/v-deo.html I have no idea what training you will get, since it highly depends on the concept trained and the complexity of it. Also depends on the quality, learning rate, resolution, captioning... there is a hige list of things that go into a good training other that steps! You can get into the discord if anything, there the community will most likely help. Amma be almost gone for a few days so cant help much myself tho
yes but no. if you want to avoid a generic look you should leave the model express itself. negative embedding will drop creativity and randomness greatly (especially negative embeddings). model quality, style lora and details tweaker tools can make good picture by themself. negative are very useful but they suck at randomness. for my part i avoid hem on recent model or i use them later in the generation using *[ lowres, blur, artifacts:0.3 ]*, with that the original composition is mostly unafected in the first frame and so is the resulting generation
I watched many times, but could not follow this video, Could you please make video on "Character Consistency SOLVED in Stable Diffusion" since i could not find it on youtube?
Sorry for that? My idea of "character consistency solved" is pretty much this video😂 im trying to find other ways with 3D and stuff. But for now this is all I was able to get. If you want, you could tell me what is it the video doesnt explain clearly, so I can try to improve on it when, eventually, I make an update on the method. Again, srry. And thanks!
You need to download them from the official hugging face page! (I think it was in hugging face. Cant really check rn but I have a video on it where all the info should be)
hi! Ended up not using it, but it's a LORA. You basically download it and, adding it to de prompt, it will make the image a turn around. Sorry for the late response!!!
Don't feel sad : you Can generate great reference sheet on midjourney. And way more easily. Try out "character expression sheet, reference art, 9 head --ar 1:1" Then train it on stable diffusion 😂
Interesting. I'm a skilled artist that spends way too much time making details most people will never even notice. I have been thinking about trying to get into training an AI to draw things in my art style. 🤔
If you decide to do so, I'd be super interested in how you feel about the results. I think it is a pretty interesting idea that is talked about in the AI space but never actually seen anyone do it and comment on it. Hope to see you in the discord sharing your process! (if tou dont mind ofc xD. If you do try it we would love to help finding the best training result). Thanks btw!
you are basically the audience i'm trying to reach and save. i'm often seen as the greedy ennemy in the field. but i know that the work SD do is from artist. i'm measely an manager of an custom assistant. artists should use image AI as they aready have their own art style to train on. making their work with fidelity and efficiency. from this point on, self made animation is not far away. full story would be illustrated correctly by the hand of the master behind it. true artists have the most power over diffusion model and will put our bests attempt to shame once they edit their generated images.
@@lefourbe5596 It sounds like it could be fun and I do have a lot of ideas for stuff that would take forever to draw. But not sure my PC can handle it. 😅
I am artist too, learning Ai for a long time. I am also ethnographer and have to recreate costumes and armoire. For 20 plus years I have collected a lot information to train Ai to help me with reconstruction of costumes just by telling the name of an item and the historical period.
Very Cool Vid, It's a shame channels like yours are always shoved in the closets so people who really would be interested in it will never find it. I found it on a lark. Been hogging Stable Diffusion/Ai Channels for more than 7 months. Stable Diffusion has reignited my love for Digital art... (cranked up my Roland SP300 again).... deep diving into everything SD.. the 4090 is violently fanning the room 24/7.... soon to be accompanied by either a ADA Titan... A6000...orrrrrrr... an H100🤑🤑🤑. LORAs are my fave next to TI... (you can adjust TIs also with (xxx:0.6)s or encapsulating whole prompt and lowering weight using 'Model Keyword' extension.)
@@lefourbe5596 YesYess, I bought 4090 primarily to run my new PIMAX Crystal.... but havent spent 1 second with it.... on Stable Diffusion/Kohyass 24/7.... trying to figure the best way of captioning, numbers for different parameters... When I spend more time on a (100+PIX)...LORA it comes out crappier... I make a LORA with 12 pix.... it's perfect... UGGHHHH
I use all the same settings and i get super bad results. Its not following the poses from open pose and if I put control weight to 2 (max) then its following the lines but its creating really bad fractured results.
Too much control will give you broken result. Too Big of a Map will lead to subject shifting. Too rich of a prompt will lower creativity. Too few of a resolution will give blurry mess. Draft your first work. Use it for img2img with utlimate SD upscale with Tile of 512*512. optionnaly use the Tile controlnet to better guide the image at higher denoising.
@@lefourbe5596 @Not4Talent_AI Thank you. I got some better resulst now by amping up the resolution and making my own poses. Its not nearly as good as in your video but I think I can go from there :) Pretty amazing to see what is possible. And a bit scarry.
is there a all in 1 program or site that does all for you? i know midjourney isnt capable to accurately replicate same characters.. im checking on civitai and nightcafe anything better than that?
Hmmm dreamlook.ai has training and generation. But idk if there are many other options ngl Edit: Now that I think about it, if you see other videos on this same topic. There are sites that took what I did here, or similar, and then automated it. So if thats what you are looking for there should be videos that point to that side. I dont remember it tho
do you have suggestions for creaating a character sheet based on my own character? I have one illustration in my own style and now want to make a character sheet based on that reference.
I think it is now possible, maybe with help of sites like: huggingface.co/spaces/sudo-ai/zero123plus-demo-space Also IP adapters can help. I know that @lefourbe5596 made a dataset from just 1 image, but no idea how atm hahahhaha
Someone got something like a spreetsheet of a warrior slashing with SD? The greatest problema I face was (other IA's) was that something like a slash, poses, frames and sequencies in the action was completyly unknow, so my question is if someone can generate consistent spritesheet of actions like slash, smash dash and others with SD.
When I select the OpenPose control type and select the preprocessor as "none" my model also appears as "None." I feel like this is causing my outputs to end up not looking like a character sheet despite using the provided OpenPose references. Is there anyway I can fix this?
UPDATE: It's been a good couple weeks since I've tried this "tutorial." Back when I installed controlnet I didn't realize the models themselves. That was my main issue right there. I'm going to use this comment as a log for my progress (if I decide to continue)
I take it as you dont currently have the "none" issure, right? Just in case, you need to download the models. Once you have them properly placed in your stable diffusion, models, controlnet folder. You'll be able to select any model you want. You can do this by clicking on the drop down menu. If you dont find it you can just click on the "open pose" buttont. That will automatically add the openpose preprocessor and model. You cna just take out the preprocessor and it should work fine. @@acetum_
tokens at the start of the prompt have more weight. BREAK acts like a second "start of the prompt" to reset the token weights after it. Can help with prompt comprehension
Hello, it's the first time I use control net, and I'm trying to follow your tutorial, but I can't get controlnet to give me any results, it's as if it doesn't exist, it doesn't affect the final image.
probably you are using a preprocessor when it should be at "none". If that's not the case, please contact me via email or discord with a screen capture of what you have as settings!
@@Not4Talent_AI I re installed stable difusion and then I downloaded the open pose file from github, its working perfectly now, i don´t understand what went wrong before, but ill finish ur tutorial now, thank you for your reply, your videos are amazing.
You have to use *Higher Resolutions* to get better generations from the get go. I've personally found if I generate at anything lower than 1024, the Ai can't produce enough detail to make a complete and coherent character/image, sometimes you might get something good, but the lower res you generate at, the worse the actual Design/Art quality.
I am curious- would this work for a character that you already have the reference image for. Basicallly I generated a character and am pretty happy with it but I'm trying to work out how to generate that same character in different poses while keeping hair, face and clothing consistent. If I try to change the pose it either changes how the character looks or the pose doesn't change even with controlnet and openpose
Yep, thats the usual problem. We havent found much of a solution tbh. There are a few options. Wont b perfect but its your best shot. 1- train a lora with that one image /separating it flping it editing it etc..(to have as many variations of that same image as possible and make the training a little more flexible). 2- generating a lot of images describing your character and cherrypickinh the ones that look most like it. Then training a lora with those 3 a mix of the 2 options
@@SorcerWizard-f8f actually i'm doing that thing ... i got decent results but it have to be forced with openpose. the less you have the more frozen your generation gets. move an arm and it goes to hell FAST ! fliping manually saved the day in the dataset as the character can somehow face two direction and have both arms down. i've yet to generate a correct lower body to feed. my V1 lora have 9 of the same image, my V3 Lora have 22 of carefully selected and cleaned images. however you will fight your SD model. i have a anime girl that is BLACK and SD anime models are usually racist... can't really get more than taned skin color. my solution is to merge your favorite model with a good general/digital model. in my case AnyAnimeMix with Dreamshaper bring back the dark skin tone a bit along some finer details that AnyAnimeMix lacks.
Im running on an AMD Radeon rx580. I've heard/read SD will not support training on AMD videocards. Is this accurate? I figured I'd ask before I dive into this.
So, I don't quite understand. Are you using an image you have chosen initially or are you generating the character using the character sheet as a first step?
in the video it appears you are using Counterfeit V3.0 for a SD model. Is that what you recommend using still or is there a better model that also works with OpenPose? I've been getting fairly poor results using SD 1.5
C@@Not4Talent_AI would you mind recommending one that works well with the process you describe here? I've discovered that not every model (XL for one) adheres to OpenPose in ControlNet.
Ok it finaly discovered some models that work with your OP character sheets (some models ignore them). I fould ComicBabes to work perfectly for my needs. I can't thank you enough for your video... very helpful!
Hey I am working on a project which is to create a comic book. First the user would define the character and then narrate the whole story. Can you tell if for my case it would be a good idea to train a lora using the character description and then use it while generating images of narration part. AND How much time it takes to train a character LORA given I am working on AWS EC2 instance with 16 GB GPU access? Also i want to automate all the steps in code itself(without manually doing). Can you tell if it is possible. THANKS
I think it is a possible idea, some websites have already started doing similar stuff so it also must be possible to automate. hardest part would be the correct upscaling and cleanup. (making sure that the generated character makes sense before starting the training) then, for a 16GB gpu, a lora of something around 20 images, should take 15-20 min? maybe? I'm not sure tbh, has a lot of "ifs" involved. It would take a while to do and figure out how to solve some of the possible issues you might encounter along the way, but I do think it is possible to do. Would do some manual testing before investing a lot of time into it tho
I get it. Even though for me it would be kind of opposite. I feel like AI opens the door for people that like telling stories. It opens the door for them to tell THEIR OWN sotries. And not just working for a corporation that only knows how to take an existing IP and re-animate it. As someone with a degree on 3D animation. Thats ultimately my goal. That people with the skills are able to show it and make masterpieces that would have been buried in a drawer in any other time period. I completely understand the though of "sure, that sounds cool. But if anyone can do that then Im wont make a living off of it". But I disagree for two reasons. 1- I really dont think it will ever get to a place where AI is able to make a full animation for someone without manual labor. /maybe it does, who knows do/ . I think it will just help artists speed up certain processes that where tedious and repetitive. 2- there is so much stuff in animation that will make yours stand out from the others. And only someone with actual animating knowladge will be able to choose. Where do yo strech and squash and for how long. How do you manage impact frames. What souns will you exagerate. Which of the 178476266273 possible camera angles are you picking for this scenes? How many cuts? What is the pacinf on these cuts? Do we exagerate prespective to give emphassis to the main character, or do we show the crowds reaction to it? Is it a narrated story or shall my composition and easter eggs tell a more complex and concealed story? I really dont think someone without the actual experience and skills cant choose all of this. Even if they do, it isnt a "correct/incorrect" thing. You still have the chance of making your style unic in this. And live off what you (I suppose) enjoy. Telling stories. Not drawing frames or moving riggs for 10, 60 second clips and getting payed 2 bucks a frame
havent really used animatediff yet. So idk. If animate diff allows the use of LORA, then I don't see why not. I'd test with a pre-made lora first. Just download a random one from civit AI and try if it works. So that you don't go through the pain of doing all of this for nothing hahahha
hello, I don't have any knowledge in this kind of stuff but I would like to know what kind to software did you use to create all of this? I literally don't know anything but I am interested to learn
Hi!! Im using stable diffusion, with the Automatic1111 ui. It is a local version that you can use for free in your Pc as long as you have a decent pc. I have a video on how to install it if you are interested.ua-cam.com/video/bzTTUrGv77M/v-deo.htmlsi=v920P2TIfBUWicS7 I also use some extensions like controlnet. But dont rush it, there is a lot of stuff and its easy to get overwhelmed 😂
I am "relatively" new to this, and everything was going very well, but halfway through the video I got lost, there were many fields that I don't know, but basically it is loading a good amount of images of the character and then creating a LORA, I think I can try
I love that this video doesn't gloss over the fact that a lot of touch up is necessary.
I always try to encourage the use of external tools and skills if possible hahha tyty!
As an artist I initially did not like the AI stuff.. But as I am working on remaking a game I did way back in high school, this is a massive time saver. I'd take my designs and run it through the AI and get different variations that have allowed me to quickly finish a character's final redesign. This is quite amazing, it also saves money as I am using this very technique to have 3D models done of the main character, that I will later have printed for a statue. I will also be using it to have the figures done of the characters to another project.
super great to hear!!! really curious on the 3D model aspect tbh (as a 3D modeler xD)
I think the people who would benefit the most from these AI tools are real artists like yourself.
@@Not4Talent_AI That is awesome! Thanks for this video, definitively glad I subscribed too!
it's heart warming to see that some people find the right use behind the black spots of this revolution.
i've started learning blender for a bit for chara modelling. I was painfuly missing orginal 2D reference.
then i saw Royal Skies video and was sold Instantly... however i did not touch blender since. time and such you know :/
@@lefourbe5596 time is a b***
Thanks for making these videos! I started playing with stable diffusion a couple of days ago and binged all your videos. SD is honestly too fun, I sat up to like 4 am yesterday inpainting instead of in-bedding😅
I'm really impressed that your videos are so concise without being hard to understand. Not to mention funny! Everything looked really daunting at first but I just want to learn more, and you make that a lot easier, and a lot more entertaining. So thanks!
thank you for the kind comment!! Glad you are enjoying it :3
bro, there is channels with millions of subscribers and i can't learn as much out of those channels. this channel is a GOLD. liked and subbed.
Thank you so much!!!!
Yep that is a dense video.
I just digged into stable diffusion and your infos are an absolute life saver!!
thank you so much!!!
Thank you for taking the time to share your knowledge!
thank you for watching and the posstive comment!
Super interesting. Thanks for your hard work, I'm exhausted just watching 😂
hahahahaha tyty!!
What I love about videos like this is *someone* is going to make a tool that simplifies all of these steps, maybe AI to generate new poses too?
That would b super cool!
For those who are lazy (like me) here's the text at around 11:08
close up of a man,{{1$$__cameraView__}}, {{1$$orientation__}}, {{1$$__expression__}}
full body shot of a man, dynamic pose, {{1$$cameraView__}}, {{1$$__orientation__}}
upper body shot of a man, {{1$$__orientation__}}
Change Man to Woman if you want a woman
once i finish with my client i think this video deserve a good remaster
Let me know if thats the case then! Hahahah
If you struggle with getting it to make consistent faces, I highly recommend making the face in Vroid Studio and Photoshopping them in. It's also necessary when making a lora to get close ups of the faces, and also for details on the clothes that you will want to be able to inpaint later in case the program struggles.
Thanks!!
Good Idea.
I know Vroid and it have gotten pretty good at making anime oc.
You can use any videogame character marker to get close to the style you look for.
This video tutorial and magic numbers are legends.
Lol, 'freedom signs', this guy is a comedian😂
I trained a checkpoint on a 3D model when Dreambooth first came out last year and it turned out fairly well in that I could change backgrounds and poses. I tried again the other day on a Lora and it was terrible. I was left scratching my head until I saw your video and you explained that all of the auto captioning (which did not exist back then) was likely throwing it off. Thanks so much for the tip, can't wait to try it again! Exciting stuff!
ty!!! We are working on a fairly in depth LORA training guide so If you keep running into a wall I hope that video helps when it comes out! (And then there is the discrod for help ofc XD)
you're just like me then. my avatar here is made up from a 3D video game model i made. lucky you i've not given up trying and example will follow soon
@@lefourbe5596 I would share the one I made last year but alas it is for work. Was pretty easy in dreambooth, I wager if I just filter the excess of the auto captioning I should be OK.
thank you man that u remembered to make this. big love from me
tyty!!! Hope it helps :3
Good stuff bro! Keep it up!
yooo sup royal! Thank you so much!
(fun fact I'm having to learn blender for the next vid XD)
🥳 i'm sure you could make a 3 min version of this !
there is much cleaning to be done and part to divide
Wow, pretty detailed explanation and i am a LoRA expert, subscribed!
Thanks!!¡ hope it helped
"Look What They Need To Mimic A Fraction Of Our Power"
hahhahahhahaha so true xD
AI people planning 3 weeks of projects to get a character.
Artists: *draws*
It is what it is tho, still fun
@@Not4Talent_AI it must be so hard to type words im sry 🥺🥺🥺
@@wallacewells6969don't know if you watched the video but I'm assuming you didint XD
I'm agreeing with you my man
@@Not4Talent_AI no i didnt watch ur dumb vid learn to draw instead of stealing art pls
Great accumulation of info, workflow and the topic of consistency is so important.++
Consider breaking your next videos into 10 minute segments ( which means more videos for you!) so they're more digestible for us. Separate them into bite-sized skills all under one related thread or collection of videos.
Yours is the only tutorial I've had to slow the playback down for to fully hear what you're saying because your visuals are also moving very quickly. You can fill up the extra time by both being sure you're explaining the next step and not just showing what to do but WHY you're doing it; plus you can give some alternative examples regarding what you're demonstrating.
Great work but it feels like I'm "rewinding" more than I'm playing!
Thanks for the feedback! I dont know about making them 10 min each, but we plan on making a firt video covering the basics and how to understand them. And then another video with more advanced info.
Tyty!!
He got that Ghosthunters scanner!
Wtf😂😂 where?
At 1:32. That’s exactly what the spirits be looking like!
@@NezD hhahhahahhahahaha ok I see now xD
your videos are just too good!!!!
thank you so much!! hope they help :3
Very good in-depth video on Lora Training!
Thanks!! an even more indepth one incoming soon enough xD
Thank You so much. I do anime edits and copyright is a hige issue right now and switching over to ai has been easier but people love the stable diffusion more then just pictures 😅 ive been looking for a way to get more consistent characters too ❤ you get a sub my brother!
Thank you so much, hope it helps!!!
Epic video, huge value.
Tyty! Hope it helps
I have gave you a thumbs down on the first video I watched and I was about to walk away and I don’t know what happened but I was just really interested about the topic so I listened to it and then gave you the thumb up and then subscribed you are now one of my favourites UA-camr on the topic I appreciate your genuine interest and your dedication🎉🎉🎉🎉
super glad to hear that, thank you!!!
@@Not4Talent_AI I don’t know how I felt but it is so rare that I give thumbs down normally I do mean it and given that you are in my top list make me feel like… I remember that it was genuine then and now I don’t know why it was like that… but knowing that I would have never heard you saying *Popimpokin* many times in the other video makes me feel very happy that I changed my mind
hahahahahaha popimpokin changed everything @@Luxcium
lost it at the freedom signs 💀💀💀
I will kindly argue about some of your points:
1. When you caption the character you should describe the outfit and any accessories as well. Trust me, you will have an easier time if you want to prompt your character in any others outfits which is not the default one or to undress it.
2. Increasing the max resolution of the training dataset to 768x768 does actually make a difference on the overall quality of the images. I would drop the number of batch size to 2 (this wil be comfortable for most people without getting the Cuda memory error) and set the image resolution to 768x768. Lora training already takes little time so don't sacrifice image quality for training speed.
3. If you're a beginner in Lora training don't bother with regularization images, you're overcomplicating yourself (I know in your video you said it's optional, just wanted to make a mention about this)
Yes! I agree with everything. For better results having 768 or even 1024 is the best option. But more time consuming. If you want the perfect training then thats perfect.
In this case we were testing so I thin 512 by 512 is the fastest testing option.
Also true what you say about tagging. Even though I wasnt looking for sn ourfit change as that is prettt much the character. The face and hairstyle are very standard😂😂
And for epochs also true, gpu is pretty much everything there hahaha
That exactly what i told N4T when WE made the vid 😁. 1. You are right and it's a choice we made. Describing everything will make the Lora harder to prompt and make comparison between diffents designs difficult cause of bleeding concepts.
There is more in the video than i have planned. It would have been in part 2 for many of these details to me.
Trust me he is well aware. My best Lora is a one that need a long freak prompt to get every part of it's complex design. (My profile pic)
Facial mark, Horn, asymetrical, heterochromia, high collar, black sclera, ponytail, gauntelet, belt, coffin pouch...
For second point, i disagree cause GTX and 20** RTX cards exist and are slow with less VRAM.
I get it as i have a 3090 but even then i prefer to be able to tweak LR, models and network dim first before. Especialy for N4T was short on Time and run a GTX 1080.
So yeh 512 training first with your 1024 dataset.
Super cool video and explanation. I have made a character of some sort just yesterday, so training my own lora would be really helpfull, I guess. I'll try it next week and see how it works out.
thanks!!! If you need help we'll be happy to give you a hand on the discord :3 Hope the video helped hahahaha
You seem to know so much, thanks for sharing! Frankly, for a beginner the majority of the videos are hard to understand though. You present so many different features and techniques in one video that it gets overwhelming. Would appreciate it if you also make some videos in the future for noobs like me, with a slower pacing and less concepts but more in-depth explanations 😅❤
Tyty!!
Hahaha I ve been tolda that, true that the channel is more aimed towards people with more experience.
I have a vid for most concepts touched in this one, but they are still just as fast😂
I might go back and update some vids on the basics eventually. In a more calmed and beginerfriendly way hahah
Ty for the feedback!!
Thank you so much for your resources. you are a legend bro!, have also joined your discord channel.
thank you for watching! hope it helps!
Que dhora!!!
Muito obrigado pelo vídeo!!!
hahaha tyty!
HERE I AM! Ready for another GREAT video! =D
hahhahahha tyty! Hope it lives up to the expectations XD
um classico do youtube @guilhermegamer
Great video! Fun ideas and it was great to follow along!
Tyty!!!
Thanks for the tutorial!
np! hope it helps!!
感谢您制作这些视频!
thank you for watching!!
Great idea! But a character sheet in one drawing style will make the Lora learn that style. That's why it is suitable to use the first result to create new images with more variety and retrain again as suggested.
true, that's a pretty good idea. Instead of having the lora already trained in a style and retrain from there. Create different styles directly, right? (not sure if that is what you mean. but sounds possible and nice)
Fantastic video! Subscribed! ❤
Thankss!!!
This workflow is so fun when you're fluid with your Photoshop skills. In my case, I open up another tutorial how to edit in Photoshop :D
Hahahah yess!!
so much information
Yeah, I had to hold back too xD
Omg Abraham Lincoln was my first thought when I clicked this video 😂
hahahahhahahhahahhahaa
You should try the same process but start with a 3d model character you can pose and use as the Lora training data, I feel that would best process for a small studio, pay an artist to make a custom 3d character then use that as the Lora training base
I think that could be interesting yeah. Mainly seeing how good can the LORA get the character. (Lefourbe tried something like that, with a very hard character, and he is getting pretty nice results). so should be possible I think-
Things to note in the "future testing" notebook for sure! tyty!
Yeah i'm doing that mostly !
I'm trying to make some example for the next video.
The 3D character i have are not so well made and i was hopping to improve their visuals with SD without sacrificing their design.
im about to give it a shot
Hope it works well, gl!!
@@Not4Talent_AI gettign better. trying to get a model sheet so i can model a character
Wtf. Literally was looking for this exact thing to get a consistent character yesterday. You're a legend.
hahahha hope it helps!!
Your phone is watching you
So good thank you for the new knowledge.
thank you!
This video is a goldmine
Tyy! Hope it helps :3
Thanks for your help!
ty for watching!
Wow, very detailed and beautifully produced video. My guess is that there is no fully online service that would allow a creator to do the things you describe here, but please correct me if I am wrong.
thank you so much!!
As far as FREE online services for this, I wouldt say it is possible atm, since google colabs are RIP.
But there are a few payed options out there.
Like thinkdiffusion for Automatic1111 (you probably could find a free solution to this as well, even though I cant tell you one atm cuz they usually either close up or end up changing to a payed model over time).
And for the training Dreamlook.AI is a very good option. (I have been sponsored by them in the past. But I still think they are an awesome option)
If you have any question or doubt, you can ask on discord, there is a lot of people that might be using online tools there as well.
This is great, thank you!
Hope it helps!! :3
I feel like you'd get more mileage actually picking up a pencil... but this works too.
100%. If you can create the cajaracter ny drawing it then go for it. Thats always the bes way ahhahah
Hi... I am totally new in AI stuffs. Can you tell me what does it mean to write "BREAK" in the prompt? Also, is it necessary to write it in capitals?
Hi!! It is a way to try for AI to get better colors. I explain it a lil in this video: ua-cam.com/video/wso_O2vk2dw/v-deo.html
and yes, it is necessary to use capital letters :3
@@Not4Talent_AI thanks for the instant reply! Does it work in ANY other platforms like midjourney?
Also thanks for the provided link, definitely will check it!
@@muhammadzazulirizki1000 No problem!! I dont think it does, its specific to Automatic1111 I think
20:50 I saw one guy provided various results experimenting with Network Rank and Network Alpha. So the best was 128 to 1. I also experimented with my characters and also found that the best is 128 to 1. But my characters were photorealistic, for anime maybe your parameters better.
From what I've been seeing this days, there is a lot of different opinions about that, tbh just said what worked for me. But could be what you say too
Hey man, great job. Are you still using a1111 and lora training or have you switched to comfyui?
Another question I have is, when you were solving all this stuff, did you have any prior knowledge? Like software knowledge or graphic design. Especially when I look at comfyui workflows, it seems impossible to reinstall it.
Hi! Tyty
Im still using a1111, even though I dont train many lora. But thats bc I have no need for that atm.
All the knowladge I have is from studing animation in uni. Comfy ui I havent gotten to it but I have worked with nodal stuff before like nuke, maya, ue, blender... so it isnt as intimidating to me.
I dont know about installing all that stuff tho, havent done it yet 😂
brush looking stuff-- cackle cackle cackle!🎉🍾
Thank you for the video.
Is it really okay to use it for comersial use?
How if the character is robot or machine. Would you make a video about it?
It is okay to use for commercial use. Just check if the model you are using allows it or not
I'm trying to find a way to make the same thing for non-humanoids. I think maybe you could use midjourney, or batch generate without controlnet and pray xD
(ty for watching, btw :3 )
ㅇ호우 대박적!!
Thanks!!! Hope it helps
Thanks for this video . although I doubt able to prepare and fix the dataset like yours, Now I`m bit understand the basic concept - idea and workflow needed And about the discord invitation , yeah ... when Lex offering you to hangout in his crib you should come and ready to feel amazed
ty!! I hope it gets easier with time and testing. In the case it does I'll do a new video. atm there is quite a bit of manual labor involved xD
Thank you so much! 👍👍👍
hope it helps! ty for watching!
Great video as always!
tyty!! hope it helps :3
Great video!!!!! Only one question: about how much time does it takes to do all this process with details and refinement?
Thanks!!
It will depend on your pc speed and luck when generating the character. Also the complexity of it.
So i cant really giva good estimate for this.
For me it took like 2 hour to prepare an ok dataset once I knew what I was doing. And then training took a bit more.
With my current pc it would be a total of maybe 2 hours total if the character isnt super hard.
If you dont really care super ultra much bout the character having a lot of precission. Then you could do this in 30 min + wtv it takes for the training
This will be soooo useful for making webcomics 😮
I hope so!!! Demiurge enthusiasts xD
Manga in general
Hello there ! It's Le_Fourbe !
Damn why YT trow me an error when i comment ? I'm so late !
Anyway i'm here so i can answers some of your question that will probably gonna cover in part 2 :)
Wtf no idea tbh 😂 this one worked😂
hello friend, a friend and I are starting to develop lora character models but we don't fully understand what regularization images are, we have a question about whether regularization images are random images for the model to convert them into your character or if they are images that the model has created and that have gone wrong in order to correct the model, we also borrowed a better computer during this weekend (since our pc are "potatoes", our best graphics and with more vram is an RX 590 the which doesn't even support cuda xd) so we want to take advantage of the fact that they lent it to us to train better models, currently we train a model with 80 images, 3000 steps and 1 epoch and we train another model with 460 images, 4600 steps and 1 epoch, what kind of training do we get? would it give better results? in neither of the 2 we used regularization images because we did not understand that point :c
our original language is not english so maybe that's why we missed the point
If you have more than 60-70 images dont even bother with regularization!
I exain it a little more here tho: ua-cam.com/video/xXNr9mrdV7s/v-deo.html
I have no idea what training you will get, since it highly depends on the concept trained and the complexity of it. Also depends on the quality, learning rate, resolution, captioning... there is a hige list of things that go into a good training other that steps!
You can get into the discord if anything, there the community will most likely help.
Amma be almost gone for a few days so cant help much myself tho
I am finding out it's not the prompts, it is the negative prompts that make good pictures.
they can make and break the image, yep
yes but no. if you want to avoid a generic look you should leave the model express itself. negative embedding will drop creativity and randomness greatly (especially negative embeddings).
model quality, style lora and details tweaker tools can make good picture by themself.
negative are very useful but they suck at randomness.
for my part i avoid hem on recent model or i use them later in the generation using *[ lowres, blur, artifacts:0.3 ]*, with that the original composition is mostly unafected in the first frame and so is the resulting generation
Lol. Micheal Jackson but shows Santa Claus...was that an Eminem reference?
hahahahaha tbh I dont know if Eminem had a something like that in a song. I just like messing up characters names xd
23:53 now imagine a neural network that takes raw dataset of an existing character and goes through this cycle on it's own 🤤
that'd be so nice hahahhahaha Hope it happens eventually
I watched many times, but could not follow this video, Could you please make video on "Character Consistency SOLVED in Stable Diffusion" since i could not find it on youtube?
Sorry for that? My idea of "character consistency solved" is pretty much this video😂 im trying to find other ways with 3D and stuff. But for now this is all I was able to get.
If you want, you could tell me what is it the video doesnt explain clearly, so I can try to improve on it when, eventually, I make an update on the method.
Again, srry. And thanks!
This looks amazing! I hope I figure this out on how to do it xD for now I gotta figure out why I've no models to be found in my control net portion ha
You need to download them from the official hugging face page! (I think it was in hugging face. Cant really check rn but I have a video on it where all the info should be)
awesome tyty! I'll dig it up and see how it's done
Crazy Man
Hi! I didn't get so much how CharacterTurner works as I'm no totally familiar with the program. Can you explain that more in detail? Thank you!
hi! Ended up not using it, but it's a LORA. You basically download it and, adding it to de prompt, it will make the image a turn around. Sorry for the late response!!!
wish midjourney could do this... damn invested it the wrong one lol
Don't feel sad : you Can generate great reference sheet on midjourney. And way more easily.
Try out "character expression sheet, reference art, 9 head --ar 1:1"
Then train it on stable diffusion 😂
great video
Thanks!!!
Interesting. I'm a skilled artist that spends way too much time making details most people will never even notice. I have been thinking about trying to get into training an AI to draw things in my art style. 🤔
If you decide to do so, I'd be super interested in how you feel about the results. I think it is a pretty interesting idea that is talked about in the AI space but never actually seen anyone do it and comment on it.
Hope to see you in the discord sharing your process! (if tou dont mind ofc xD. If you do try it we would love to help finding the best training result).
Thanks btw!
you are basically the audience i'm trying to reach and save.
i'm often seen as the greedy ennemy in the field. but i know that the work SD do is from artist. i'm measely an manager of an custom assistant.
artists should use image AI as they aready have their own art style to train on. making their work with fidelity and efficiency.
from this point on, self made animation is not far away. full story would be illustrated correctly by the hand of the master behind it. true artists have the most power over diffusion model and will put our bests attempt to shame once they edit their generated images.
@@lefourbe5596 It sounds like it could be fun and I do have a lot of ideas for stuff that would take forever to draw. But not sure my PC can handle it. 😅
I am artist too, learning Ai for a long time. I am also ethnographer and have to recreate costumes and armoire. For 20 plus years I have collected a lot information to train Ai to help me with reconstruction of costumes just by telling the name of an item and the historical period.
I would like to be onboard of your server. This was one of the most comprehensive tutorials, but it is hard to repeat and get the same result.
When making the character sheet in the beginning would it be ok to make them completely naked then add clothing when you go ahead to use them?
Wouldnt recomend that but it is possible. The problem is that the character will be naked most of the time if you train it like that
16:25 I might be blind, but it seems the link is not in the description.
My bad, added it now!
ua-cam.com/video/xXNr9mrdV7s/v-deo.html
Very Cool Vid, It's a shame channels like yours are always shoved in the closets so people who really would be interested in it will never find it. I found it on a lark. Been hogging Stable Diffusion/Ai Channels for more than 7 months. Stable Diffusion has reignited my love for Digital art... (cranked up my Roland SP300 again).... deep diving into everything SD.. the 4090 is violently fanning the room 24/7.... soon to be accompanied by either a ADA Titan... A6000...orrrrrrr... an H100🤑🤑🤑. LORAs are my fave next to TI... (you can adjust TIs also with (xxx:0.6)s or encapsulating whole prompt and lowering weight using 'Model Keyword' extension.)
Ohhhh.... I am '@SkitzoBunito'
Thank you fr!!! It did for me too tbh. And thats a nice setup, got me jealous hahaaha
are you a clone of me or something 😂?(9 month on my side) i too hope for the ADA Titan :D. not worth upgrading my 3090 yet.
@@lefourbe5596 YesYess, I bought 4090 primarily to run my new PIMAX Crystal.... but havent spent 1 second with it.... on Stable Diffusion/Kohyass 24/7.... trying to figure the best way of captioning, numbers for different parameters... When I spend more time on a (100+PIX)...LORA it comes out crappier... I make a LORA with 12 pix.... it's perfect... UGGHHHH
I use all the same settings and i get super bad results. Its not following the poses from open pose and if I put control weight to 2 (max) then its following the lines but its creating really bad fractured results.
weird, have you tried changing checkpoints? Maybe going with a higher Resolution from the start instead of highres fix?
Too much control will give you broken result.
Too Big of a Map will lead to subject shifting.
Too rich of a prompt will lower creativity.
Too few of a resolution will give blurry mess.
Draft your first work. Use it for img2img with utlimate SD upscale with Tile of 512*512.
optionnaly use the Tile controlnet to better guide the image at higher denoising.
@@lefourbe5596 @Not4Talent_AI Thank you. I got some better resulst now by amping up the resolution and making my own poses. Its not nearly as good as in your video but I think I can go from there :)
Pretty amazing to see what is possible. And a bit scarry.
is there a all in 1 program or site that does all for you? i know midjourney isnt capable to accurately replicate same characters.. im checking on civitai and nightcafe anything better than that?
Hmmm dreamlook.ai has training and generation. But idk if there are many other options ngl
Edit:
Now that I think about it, if you see other videos on this same topic. There are sites that took what I did here, or similar, and then automated it. So if thats what you are looking for there should be videos that point to that side. I dont remember it tho
do you have suggestions for creaating a character sheet based on my own character? I have one illustration in my own style and now want to make a character sheet based on that reference.
I think it is now possible, maybe with help of sites like: huggingface.co/spaces/sudo-ai/zero123plus-demo-space
Also IP adapters can help.
I know that @lefourbe5596 made a dataset from just 1 image, but no idea how atm hahahhaha
Two freedom signs lol
Someone got something like a spreetsheet of a warrior slashing with SD? The greatest problema I face was (other IA's) was that something like a slash, poses, frames and sequencies in the action was completyly unknow, so my question is if someone can generate consistent spritesheet of actions like slash, smash dash and others with SD.
Atm I have no idea, but you pose an interesting issue that might be very worth to look into. I'll note it down and see what I can do!
would you say using regularization images is worth the extra effort?
only if you dont have anough images for a good dataset
When I select the OpenPose control type and select the preprocessor as "none" my model also appears as "None." I feel like this is causing my outputs to end up not looking like a character sheet despite using the provided OpenPose references. Is there anyway I can fix this?
UPDATE: It's been a good couple weeks since I've tried this "tutorial." Back when I installed controlnet I didn't realize the models themselves. That was my main issue right there. I'm going to use this comment as a log for my progress (if I decide to continue)
I take it as you dont currently have the "none" issure, right?
Just in case, you need to download the models. Once you have them properly placed in your stable diffusion, models, controlnet folder. You'll be able to select any model you want. You can do this by clicking on the drop down menu. If you dont find it you can just click on the "open pose" buttont. That will automatically add the openpose preprocessor and model. You cna just take out the preprocessor and it should work fine. @@acetum_
Thanks for the video! What does BREAK do?
tokens at the start of the prompt have more weight. BREAK acts like a second "start of the prompt" to reset the token weights after it.
Can help with prompt comprehension
@@Not4Talent_AI thanks!
Hello, it's the first time I use control net, and I'm trying to follow your tutorial, but I can't get controlnet to give me any results, it's as if it doesn't exist, it doesn't affect the final image.
probably you are using a preprocessor when it should be at "none". If that's not the case, please contact me via email or discord with a screen capture of what you have as settings!
@@Not4Talent_AI I re installed stable difusion and then I downloaded the open pose file from github, its working perfectly now, i don´t understand what went wrong before, but ill finish ur tutorial now, thank you for your reply, your videos are amazing.
glad to hear that! and thank you so much! @@Gamecore-cdmx
Awesome!!! Sadly, I will need a new computer. :((
Oh... hope you find a nice price for it!
You have to use *Higher Resolutions* to get better generations from the get go. I've personally found if I generate at anything lower than 1024, the Ai can't produce enough detail to make a complete and coherent character/image, sometimes you might get something good, but the lower res you generate at, the worse the actual Design/Art quality.
yeah, 100% true. character sheets with 512 are unintelligible xD
Thanks so much for the video. I'm wondering is there a way to make each scene move and transition like a short film?
With AI I'm not sure tbh. Dont really get what you mean either without a visual refrence, srry.
Thanks for watching btw!!
I am curious- would this work for a character that you already have the reference image for. Basicallly I generated a character and am pretty happy with it but I'm trying to work out how to generate that same character in different poses while keeping hair, face and clothing consistent. If I try to change the pose it either changes how the character looks or the pose doesn't change even with controlnet and openpose
Yep, thats the usual problem. We havent found much of a solution tbh. There are a few options. Wont b perfect but its your best shot.
1- train a lora with that one image /separating it flping it editing it etc..(to have as many variations of that same image as possible and make the training a little more flexible).
2- generating a lot of images describing your character and cherrypickinh the ones that look most like it. Then training a lora with those
3 a mix of the 2 options
@@Not4Talent_AI I see thanks for the advice - I will give it a look over the week
@@SorcerWizard-f8f actually i'm doing that thing ... i got decent results but it have to be forced with openpose.
the less you have the more frozen your generation gets. move an arm and it goes to hell FAST !
fliping manually saved the day in the dataset as the character can somehow face two direction and have both arms down. i've yet to generate a correct lower body to feed. my V1 lora have 9 of the same image, my V3 Lora have 22 of carefully selected and cleaned images.
however you will fight your SD model. i have a anime girl that is BLACK and SD anime models are usually racist... can't really get more than taned skin color.
my solution is to merge your favorite model with a good general/digital model. in my case AnyAnimeMix with Dreamshaper bring back the dark skin tone a bit along some finer details that AnyAnimeMix lacks.
Im running on an AMD Radeon rx580. I've heard/read SD will not support training on AMD videocards. Is this accurate? I figured I'd ask before I dive into this.
Its not impossible to use with AMD but it is way harder and I think it works worse too
So, I don't quite understand.
Are you using an image you have chosen initially or are you generating the character using the character sheet as a first step?
Generating the character with the sheet. Then cleaning up that character and training AI with it
Can you tell me the modal that makes consistency the same character
there are a few, I used Ivory_v2 and counterfeit
the model of your choice actually. the one with the style that you prefer.
En donde hiciste las poses, con caras y mano? He estado buscando y no encuentro 😢
app.posemy.art
Muchas gracias
in the video it appears you are using Counterfeit V3.0 for a SD model. Is that what you recommend using still or is there a better model that also works with OpenPose? I've been getting fairly poor results using SD 1.5
I'd never use the base 1.5 model. Counterfeit is nice, but you have a lot of other options on civit AI
C@@Not4Talent_AI would you mind recommending one that works well with the process you describe here? I've discovered that not every model (XL for one) adheres to OpenPose in ControlNet.
Ok it finaly discovered some models that work with your OP character sheets (some models ignore them). I fould ComicBabes to work perfectly for my needs. I can't thank you enough for your video... very helpful!
glad to hear that! Thanks!! @@thinkinginmotion
Hey I am working on a project which is to create a comic book. First the user would define the character and then narrate the whole story. Can you tell if for my case it would be a good idea to train a lora using the character description and then use it while generating images of narration part. AND How much time it takes to train a character LORA given I am working on AWS EC2 instance with 16 GB GPU access? Also i want to automate all the steps in code itself(without manually doing). Can you tell if it is possible. THANKS
I think it is a possible idea, some websites have already started doing similar stuff so it also must be possible to automate.
hardest part would be the correct upscaling and cleanup. (making sure that the generated character makes sense before starting the training)
then, for a 16GB gpu, a lora of something around 20 images, should take 15-20 min? maybe? I'm not sure tbh, has a lot of "ifs" involved.
It would take a while to do and figure out how to solve some of the possible issues you might encounter along the way, but I do think it is possible to do.
Would do some manual testing before investing a lot of time into it tho
Thanks for your reply. Really appreciate your content!!@@Not4Talent_AI
np! thank you for watching it! @@nitingoyal1495
this is really demotivating me to persue my dream of being a illustrator/animator
I get it. Even though for me it would be kind of opposite.
I feel like AI opens the door for people that like telling stories. It opens the door for them to tell THEIR OWN sotries. And not just working for a corporation that only knows how to take an existing IP and re-animate it.
As someone with a degree on 3D animation. Thats ultimately my goal. That people with the skills are able to show it and make masterpieces that would have been buried in a drawer in any other time period.
I completely understand the though of "sure, that sounds cool. But if anyone can do that then Im wont make a living off of it".
But I disagree for two reasons.
1- I really dont think it will ever get to a place where AI is able to make a full animation for someone without manual labor. /maybe it does, who knows do/ . I think it will just help artists speed up certain processes that where tedious and repetitive.
2- there is so much stuff in animation that will make yours stand out from the others. And only someone with actual animating knowladge will be able to choose.
Where do yo strech and squash and for how long. How do you manage impact frames. What souns will you exagerate. Which of the 178476266273 possible camera angles are you picking for this scenes? How many cuts? What is the pacinf on these cuts? Do we exagerate prespective to give emphassis to the main character, or do we show the crowds reaction to it? Is it a narrated story or shall my composition and easter eggs tell a more complex and concealed story?
I really dont think someone without the actual experience and skills cant choose all of this. Even if they do, it isnt a "correct/incorrect" thing. You still have the chance of making your style unic in this. And live off what you (I suppose) enjoy. Telling stories. Not drawing frames or moving riggs for 10, 60 second clips and getting payed 2 bucks a frame
Does this technique allow to be incorporate to animatediff? So can we have consistent characters in an animation?
havent really used animatediff yet. So idk.
If animate diff allows the use of LORA, then I don't see why not.
I'd test with a pre-made lora first. Just download a random one from civit AI and try if it works.
So that you don't go through the pain of doing all of this for nothing hahahha
Thanks, that's a good idea!
Can you do this with real people too? That is, copy a person's entire body and face into SD?
Yes, you can! With permission from the person ofc, but lora training works with pretty much anything
hello, I don't have any knowledge in this kind of stuff but I would like to know what kind to software did you use to create all of this? I literally don't know anything but I am interested to learn
Hi!! Im using stable diffusion, with the Automatic1111 ui. It is a local version that you can use for free in your Pc as long as you have a decent pc.
I have a video on how to install it if you are interested.ua-cam.com/video/bzTTUrGv77M/v-deo.htmlsi=v920P2TIfBUWicS7
I also use some extensions like controlnet. But dont rush it, there is a lot of stuff and its easy to get overwhelmed 😂
@@Not4Talent_AI OMG an actual person replying 😆. I was expecting other people to tell me what it is but thank you very much.
hahahahha I try to respond to everyone I can, thank you for the kind comment! @@Yuugere
I am "relatively" new to this, and everything was going very well, but halfway through the video I got lost, there were many fields that I don't know, but basically it is loading a good amount of images of the character and then creating a LORA, I think I can try
Hahhahaha sorry! Yep, thats basically the idea. And using a photoediting software to create some variations in the images
@@Not4Talent_AI don't worry,English is not my primary language,that could explain it,but I'm working with all that I understad,thanks bro
@@kabirgomez7967 thank you for watching! hope it helps to some extent :3
Please tell consistent images of stable diffusion on phone as majority are using phone instead of computer
Im not sure if I will, but its a nice idea if people really want it.
I should experiment tho, cuz I have no experience using it with the phone
@@Not4Talent_AI Thanks