Update: Check this video How to Install Forge UI & FLUX Models: The Ultimate Guide ua-cam.com/video/BFSDsMz_uE0/v-deo.html Here are some useful resources for Stable Diffusion: Download Stable Diffusion Webui Forge from: github.com/lllyasviel/stable-diffusion-webui-forge Download Juggernaut XL version 9 from: civitai.com/models/133005/juggernaut-xl?modelVersionId=348913 More info on FreeU: github.com/ChenyangSi/FreeU Download more ControlNet SDXL models huggingface.co/lllyasviel/sd_control_collection/tree/main Extensions used github.com/ahgsql/StyleSelectorXL and github.com/thomasasfk/sd-webui-aspect-ratio-helper If you have any questions you can post them in Pixaroma Community Group facebook.com/groups/pixaromacrafts/ or Pixaroma Discord Server discord.gg/a8ZM7Qtsqq
I have been looking for a tutorial like this for months. You have a real talent for this tutorial style and I HIGHLY encourage you to keep making these videos. Information is packed and logically flowing from one point to the next. Subscribed!
HOLY COW! I've been using A1111 (and now Forge) for a year, so by now I know most of these "hacks", but I wish a video so clear and so thorough existed when I was starting my journey. I even picked up a new nugget here and there. Bravo! Subscribed. And Saved.
Does it crash more or less often? I've been using A1111 for a while now, but it feels like it's been crashing more and more. Especially with SDXL models.
Since i switched to forge it didn't crash at all, only when i used control net ceashed if i didn't use image size divisible with 64 @@edouarddubois9402
The width and height of the image, sometimes i got that error when the size was not divisible with 64, but mostly when i used some extensions@@edouarddubois9402
Watched 10 videos about SD Forge install, but your is the best, Quick to the point, super useful tips for beginners, you even say what to do if it crashes.
It should be noted, for those who stumbled upon this like I did without knowing any better, that this method only works for nVidia graphics cards. WebUI uses CUDA, which is a proprietary API specifically for nVidia...meaning if you don't have their drivers, you can't natively run Web UI. Luckily there are forks that exist that do work for AMD Radeon cards, but you'll have to jump through a few more hoops than what is shown here in order to install, and it probably won't run quite as fast as it does on nVidia cards.
Thanks for your video! I installed Forge yesterday (no stranger to A1111 here), but thought I'd check out a video or 2. For preferred defaults, I've been editing ui-config.json. Didn't realise there was a more straightforward method via settings haha! Dark mode is so much easier on the eye. I set it at browser level so that all pages appear dark, then display settings > high contrast in Windows (7, 8, 10, 11) will give dark mode OS-wide. Thanks for the heads up on SDXL styles and how its extra prompting works. 2 more extensions I use that might help is CivitAI Browser+ which integrates CivitAI into A1111/Forge, and ADetailer (After Detailer) which is an automatic in-paint utility that tidies up facial features and I find this better than GFPGAN and CodeFormer. Your 7-second image generation near the start of your video took 17 seconds for me with the same settings, coming from an RTX 3070 here. You covering file naming was very helpful too as I wanted to add the denoising value to the file names, and use a suffix instead of file numbers :-)
Very helpful, i will be watching all the videos in this playlist, thanks! BTW what do you use for your voice It's great.( If it's not a trade secret that is)
I didnt play with that function yet, it always seems to be complicated to do trainings, I tried also on A1111 but I dont get always good results, needs good settings, good images, captions, too many things involved it seems. And now I saw an anoucement that forge is not going to be updated anymore, like is used more for tests or something.
great info. Also a quick tip, below the image there is a button to upscale using hires fix, just a quicker way to do it. 09:25 I think that option is new with Forge, wasn't in A1111
First of all, I must say thank you. I started with your videos with the latest one, about Flux, and I am stuck here. Forge UI is fantastic! My only question is if I can find log files about prompts? It would be great to keep them.
Well each prompt and settings are saved in the png you generated, so if you drag the png you like into the png info tab you can see the prompt and setting. For more complex probably you need a script or an extension, on a quick search maybe an extension like this could do something similar, didn't test but maybe gives you some ideas github.com/ThereforeGames/unprompted
@@CsokaErno I use XnView MP as my default image viewer. It has meta-info tab on the right, no need to import images anywhere, you can just copy, alt-tab and paste prompt+properties into your browser. Besides that, it's a really handy piece of software compared to vanilla windows image viewers.
i wonder if any of the stable diffusion UI makers (forge, automatic, comfyui etc) has considered a method for capturing 'recommended model settings' like you point out at 3:29 - as going out and hunting down a model's recommended settings is a work slow-down; perhaps be able to configure a 'model or ksampler template' that can be a quick preset based on the model.. would be kinda cool to have the option to be able to on checkpoint load to trigger the preset (but again should be an optional thing, not everyone would want that in all cases). if this already exists someone let us know
There is a preset saving extension so you can just save settings and give it a name similar to the model you use to know for what it is, but many extensions have bugs since with updates
You can read more about here, i didn't play with them in forge only with canny control net mostly, also keep in mind the version you are using there are different forks of forge now, the main one is used for beta testing and many things might not work! github.com/lllyasviel/stable-diffusion-webui-forge/discussions/178
great video! thanks very much. quick question, do you use tts for narration? If so, it's incredible, may I ask which one? I've been trying to find something decent for my videos. Cheers :)
the real tutorial we need is how to run ForgeUI using colab, there are currently no colab notebooks provided for it.... some of us have colab pro subscriptions and want to run this on the big boy GPUs
I haven't used colab for months because I upgraded my PC, i just share tutorials on how I use it and the knowledge I got so far, sorry i can't help more
I'm new to this Stable Diffusion GUI. Experienced people can you please answer is this Forge WebUI is better than Fooocus MRE? If Yes, then in what parameters is it better? Thanks!
Little question: did anyone have a problem with LyCORIS models on Forge? I'm using Forge through Stability Matrix, and no matter how I load -- from my computer or Matrix's models loader -- it just doesn't show up in Lora's tab. And when I load it in Lora's folder, it doesn't work correctly.
This is a great tutorial, but for me using it on Ubuntu makes feel a bit sad, cause most of extensions simply doesn't work, or doesn't want to install. Maybe because of GPU (rtx2060,6gb), but when I had Windows system on same machine, it had more extensions preinstalled and used. Like I don't have Free U and Control Net SDXL. As well as I remember, it worked more better than now. Did they made some new updates, which made it work worse? (Last time used 3 months ago)
They stopped updated for the official version, there are some versions that still around but not sure how much update they get., you can try the last stable version or change to dev2 ua-cam.com/video/RZJJ_ZrHOc0/v-deo.html. I have on older pc same gpu, but i wasnt able to run with control net, it crashed. ComfuUI works ok, but didnt try yet complex stuff
not sure if all those works, but did you installed them from extension? Go to extension tab, click on available, click on Load From button, that will load all, search for an extension, for example tried ratio helper in search and installed just well when i clicked install, and restarted forge.
Hello, really great tutorial. I have a question I want to use the 4xVALAR upscaler but have no idea where to put it. Could you please answer if you have an idea in what folder exactly it should be.
Go to your webui\models folder and there create a folder named ESRGAN, so you will have webui\models\ESRGAN path, and in that ESRGAN you put that upscaler model. That worked for me, hope it works for you.
I have NVIDIA graphic card but i get this error when run forge: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check. how can I solve this?
so in the last month there has been a lot of updates to the forge ui, today I am working on an update video with what is new. Go to this discussion page to see what they changed and in comments some people had the same problem like yours and it seems has something to do with forge, on the bottom of the page you can see comments and click to load more comments, and with ctrl+f you can search for "torch is" to find comments that have those words github.com/lllyasviel/stable-diffusion-webui-forge/discussions/981 also check this page in the comments how some used different settings in arguments - github.com/AUTOMATIC1111/stable-diffusion-webui/issues/1742
Just updated to latest Forge version , the one that can work with Flux, but using only Sdxl on my 8gb card : every time I do inpainting or img2img the result has lower saturation than the original, it's me or what? Assigning a VAE do not solve 😢
There are a lot of bugs on new version so it will take a while for all to get fixed, this has a similar problem github.com/lllyasviel/stable-diffusion-webui-forge/issues/1189 and if you look at the list of open issues are like 600 github.com/lllyasviel/stable-diffusion-webui-forge/issues
My built-in controlnet's IP-Adapter is missing its models, and thus, doesn't work. Any ideas? I wanted to install them manually, but the library is different, and so are the files.
Forge still has some problems with control net, check this discussion maybe it helps github.com/lllyasviel/stable-diffusion-webui-forge/discussions/178#discussioncomment-8572388
Yeah I am using a voice from voiceair.ai , my voice is ok but my spoken English is not so good and the accent is too strong, I am better with writing, that why that voice can produce a clear English and anyone can understand me, and sounds good for an AI voice.
Should be next to run.bat and environment.bat a file called update.bat i have it there since installation, your should have it to. Just careful with updates to have a good stable version, check this video ua-cam.com/video/RZJJ_ZrHOc0/v-deo.htmlsi=rF-9wCmzResJiW3L
Thankyou for this tutorial! ❤Do I need the automatic 1111 stable diffusion installed to be able to install forge? I have the oldest version of automatic 1111 installed and I hvnt used it or upgraded it as I cudnt keep up with the every new update and other troubleshooting issues as I hv zero knowledge of programming language 😢
I explain that in this video starting with minute 3 43 ua-cam.com/video/q5MgWzZdq9s/v-deo.htmlsi=WOTQNXWSwiwWBkU6&t=223 in your forge folder you have webui folder and there is a webui-user.bat file where you can put the a1111 paths, just have to switch backslash with forward slash
please i got this error when i started run.bat C:\Users\moor\stable-diffusion-webui-forge>python launch.py Traceback (most recent call last): File "C:\Users\moor\stable-diffusion-webui-forge\launch.py", line 1, in from modules import launch_utils ImportError: cannot import name 'launch_utils' from 'modules' (C:\Users\moor\AppData\Roaming\Python\Python310\site-packages\modules.py)
Not sure what is with error, but what you can try is to create a folde on another drive and try again fresh maybe can make it run, I don't know coding but looks like can not important a file, maybe something didn't download or is a bug, that why I say to try a fresh install in another folder
@@pixaromathank you, hope that some one can answer. I don't have the requested performance machine to do local installation, so that will give me a great help. why I'm asking? it's just for the seamless pattern setting that exists in the models presented. this capability isn't offered right now in fooocus witch is easily accessible with colab.
From what I know it search for Nvidia driver so no it doesnt work, many fail to load on AMD, i saw one commenting somewhere that got it to run on amd one UI called SD.Next by Vladmandic
Flux nodels can do that most of the time without fixing, so if your video can handle flux you should try. I have video on flux for both forge and comfyui, just forge is still work on progress, works with flux but other things still don't work yet, they are changing the interface and need time to fix all the things
@@pixaroma Yes, you are right, but in this case I talked about only SD. Flux can handle hands and eyes pretty well but when two hand are close or touch each other, it makes fault. Anyway, I suppose it will be fixed very soon like everything else in AI realm :)
@@CsokaErno with sdxl I got ok results if i used control net, there are some more nodes that can be used, I will see if I get enough info to get an episode about that in the future
@pixaroma Owww... I understand, I haven't seen the other videos yet, i dint have time, I came here to see the correct folder to paste the file. Do you have any tips for me? I have a Ryzen 5600g with 32GB RAM and a 3060 with 12GB. What is the best SD for me to install here?
Maybe look here gist.github.com/ShMcK/d14d90abea1437fdc9cfe8ecda864b06 aws.amazon.com/blogs/machine-learning/use-stable-diffusion-xl-with-amazon-sagemaker-jumpstart-in-amazon-sagemaker-studio/ as I don't use aws I can't hellp
When I put my model in control net "control_v11p_sd15_openpose.pth" and when I try to generate the image, I have a error message "TypeError: 'NoneType' object is not iterable. My setup is OpenPOSE and processor Openpose_full, can you help me please
i see you are using v1.5, do you get the same on sdxl? I got that error recently but not on control net but when i used an extension at different image size. It works at 1024x1024px? I got that error on different sizes, but worked at 1024x1024px. I dont used v1.5 anymore since sdxl appeared.
so i tried with sdxl control net and i get the same error if i use certain sizes, for example it works if is 1024x1024px, or 1024x576px, but i get that error if i use 1200x672, or 912x512 or 1024x816
@@pixaroma thank you for your answer, I tried all the sizes 1024, 512 etc. it does not work, I have this problem that when I want to use controlnet otherwise to generate an image no problem but when I use controlnet impossible to generate an picture
Either your video card is not good enogh or forge dont recognize it, I am a designer not a coder, but you can try this in webui-user.bat add the following arguments to see if it works, it need at least 6gb of vram and prefers nvidia cards but try it anyway: set COMMANDLINE_ARGS=--skip-torch-cuda-test --precision full --no-half
I see someone already posted taht on the bugs area, you can watch that to see if it gets any response if nothing else works, github.com/lllyasviel/stable-diffusion-webui-forge/issues
Mine is off also, can be activated with some command in the bat, I tried but made my generation slower not faster so i left it deactivated, it appears as suggestions on the cmd window when you start and there is also a command I think, i don't remember now, just i know activated was slower for me
You solved the problem I was actually worried about the output speed. So I don't have to worry about CUDA but the internet connection, it should be sometimes fast sometimes slow which affects the output speed. Thanks for the tutorial above.
thanks for the controlnet section I got stumped on where to put the models. Edit: can anybody help me with LORA models, I paste them in the LORA folder but FORGE doesnt seem to detect them
i just tested now with a file and seems to be recognized, webui\models\Lora so the folder is Lora, and after you past it there you go to lora in the interface and refresh the page or just restart the stable diffusion so it can see it
What if you use both 1.5 and XL checkpoints? Do you have to keep manually changing the FreeU settings everytime? Also with Hypertile, should the tile size be half of the generated dimensions longest side? what if I am usijng x2 hires fix, should it be half the size of the hires fix? What does the hypertile swap size and depth settings do? and is there a way like in SDNext to have hypertile set to automatic mode?
I don't use those settings too be much to give you more details, i know that for control net you have to manually keep changing models, 1.5 is different from sdxl so it needs different models and settings
@@pixaroma yes that's to be expected for control net as 1.5 and xl use different model. That's easily done. What's not easily done is having to remember the precise settings for freeU for 1.5 and xl. Can't believe there is no preset option to load these on the fly. I also am disappointed that there is no auto mode for hypertile tile size. Damn, all these different UIs should come together and make one ultimate ui
Forge has some basic prompt from image but is not so accurate, in img2img tab under generate it has a paperclip icon, first time will download a model but after that should work faster, and is giving a basic description of the image you uploaded in the img2img.
Try to put VAE at automatic or none, also try different models and samplers.to.see if that fix it, try random seeds, also try updating the forge, see if any of those can fix it, is hard to tell what can be the cause you have to keep trying different things to see what cause it
hi my installed Sd forge doesn't have the update.bat file. Is there anyway to update SD forge without the file? Maybe by adding arguments to look for an update?
So, in the arguments section where you put the dark theme I can add: --pin-shared-memory --cuda-malloc --cuda-stream For optimization, right? Thanks for the video!
Yes, i tried all those as forge suggested but didn't make it faster on my rtx4090 but slower, maybe it does better for you but for me with bo argument was faster
is in the webui-user.bat look for set COMMANDLINE_ARGS= and there after equal you add it, like I added the dark theme you add more, set COMMANDLINE_ARGS=--theme dark --cuda-stream
1. Is there any performance drop if I don't install it on C drive? 2. My C drive is SSD and D drive is HDD. Can I still install on D drive? Will I face any performance difference or issues?
I am not sure, can you join my Facebook group and you can you show me maybe some screenshots or post there to take a look, do you get any errors or what it looks like
You can download more models from civitai website and depends in model some need different vae settings or models, thr juggernaut x i use it has embedded vae and my vae setting is set to auto. Sorry i am not at the computer this weekend to check exactly but if you do exactly like in the video it should work
@@pixaroma thank you so much. I think that I have to go deeper in learning because with same prompts I have spectacular results with midJ and disaster with Forge 🤷♂️
I use forges deforum tab to create animations. I would like to know how to create the animations within a boundary. I projection map so I would like to know how to keep the animations within the map of my house. Would you know how to accomplish this? I have a png map file that I created but unsure what to use it with. TIA
Sorry, I didn't play yet with deforum, so i can't help there yet, I like to create HQ images and the video and animation isn't quite there yet, i am waiting for an improvement before i jump in to it
@@pixaroma I have 4gb of vram as im generating with SD realisticVisionV51 works perfect As you suggested i have installed the Juggernaut_X_RunDiffusion_Hyper but still the same issue
The one you mention seems to be based on sd 1.5 so the sd 1.5 are smaller and you can run it because usually has like 2gb, but when you run an sdxl base model that is 5-6 gb your 4gb memory might not be able to handle that size, so either it takes a lot of time to load like a few minutes or it crashes. so maybe just work with sd v1.5 model for a while or if you can find an sdxl model that is smaller maybe
Using RUN I got the following error: "Torch is not able to use GPU: add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check" Does this mean my GPU will not work for this? I have a GTX 1660 Ti with 6GB GDDR6. Not a new card, but it's getting me by. No good?
I am not sure, try to add what they recommend, add --skip-torch-cuda-test to COMMANDLINE_ARGS, so copy --skip-torch-cuda-test and then open the webui-user.bat with notepad and paste it there after the equal sign, where it says set COMMANDLINE_ARGS=, should be set COMMANDLINE_ARGS= --skip-torch-cuda-test and see if that works. For me it worked on 6gb vram but was on rtx4090, is still a new UI so it might still have bugs.
I received same error and tried suggested command. Restarted run.bat and got new error saying no NVIDIA driver on your system. Makes sense because I have AMD video card. Simply put this is only for NVIDIA users.
I just installed a fresh copy today and I didn't run on any problems, so maybe try again, not sure what is the problem, if you have Nvidia usually works ok
@@pixaroma i solved it , actually i was trying to install some extension , but maybe some error happened , , later on i was panicked , and then i thought carefully , and deleted some extensions , that resolved it
I don't think so, is a different interface based on ,a1111 but is made by different users, so is not an update is a different UI. Also seems that works only with Nvidia card
@@pixaroma Thanks! yeah i do have a Nvidia card and seems like everyone using this one. Mine, web ui looks really outdated and have many things missing.. Looks like i have to install all over again? should i uninstall the SD web ui?
That is strange, you can try report the issue on their page github.com/lllyasviel/stable-diffusion-webui-forge/issues maybe is a bug and so how it used the video card to much and make it crash, but that should not happen. I only tested on 2 computers, and i didnt had a problem on rtx2060 or rtx4090. Try an update in a few days maybe is get fixed. I wish I can help you more, but I am a designers not a programmer, I just share how to use new technology to get images with Ai, when is come to actual code, and bugs if it didnt happened to me not sure how to tell you a fix :(
I tested on 6gb and worked, only control net give me some crashed, but other things worked for me, faster then automatic that took ages. It worth a try, and if you dont like it how it work you can just delete the folder with all files. But i believe you can use most of the functions, if automatic1111 works this should work even faster
I runned forge on 6gb rtx 2060 so it should work, other solution will be comfy ui since forge will not be updated anymore, and once they update automatic 1111 that will be also a good solution
Mostly in optimization on how it handles memory, so it generates images faster then a1111 and have some extra things, but it stopped being updated officially so i switched now to comfy UI
I wanted to ask, I have models and loras in my fooocus folder. is it possible to copy and paste them to the appropriate forge folders? Or do I have to redownload them from civitai? ORRR...(is there a better way to have these models linked from my fooocus to my forge folder that could help on saving space on my cpu)...... I rather not redownload them or copy the file over so i dont fill up my harddrive.
I am adding a few hours a tutorials for forge, and how to link the automatic folder to forge, probably you can do something similar with the fooocus paths. I dont have focus to test it but it worked with the path of automatic1111. Also if you have it once you just copy them in the right location no need to redownload them.
thank you for the reply! I wonder if it will work linking fooocus folder to forge..i guess it is worth a try. ANy idea when your vid will be dropping for that so I can keep an eye out for it?@@pixaroma
I never did it but someone commented with this on reddit Use command prompt in SD directory and type git revert or git reset --hard . You can find the previous version hashes using git log, or there is a list somewhere on the github. So for forge probably you have to go to forge and then to web ui folder, then in the address bar type cmd and press enter, then you can see all those commit with a string of numbers, then I am not sure the next part either you use git revert and put that comit number or something like that
Can anyone tell me which config file(s) need to be changed on Linux(Ubuntu) so I can point to an already existing venv folder, and model directories? I want to run this alongside A1111, which is pretty much working fine except that when I uncomment the line in the config file to point to my venv folder it still creates a new venv folder in it's install directory instead of using my existing venv folder. Same with models, I can't seem to find anywhere to change where forge looks for it's model files?
To have all the models from the Automatic1111 in forge you do like this on Windows (I dont know how is for linux and others but maybe is something similar): Edit or open with notepad the webui-user.bat look for where it says @REM set A1111_HOME=Your A1111 checkout dir and from there down remove all the @REM you can find, that will uncomment those lines first line should look like this set A1111_HOME=D:/StableDiffusion/sdwebui/webui for me that is the path of Automatic1111 D:/StableDiffusion/sdwebui/webui notice the slash, you need to change it, because when you copy and paste your path will look like this D:\StableDiffusion\sdwebui\webui\ and you need to change the slash to look like this D:/StableDiffusion/sdwebui/webui so you should have now in that bat something like this: set A1111_HOME=D:/StableDiffusion/sdwebui/webui set VENV_DIR=%A1111_HOME%/venv set COMMANDLINE_ARGS=%COMMANDLINE_ARGS% ^ --ckpt-dir %A1111_HOME%/models/Stable-diffusion ^ --hypernetwork-dir %A1111_HOME%/models/hypernetworks ^ --embeddings-dir %A1111_HOME%/embeddings ^ --lora-dir %A1111_HOME%/models/Lora call webui.bat then save it and you can run the bat and you should be able to see the models from Automatic in Forge
@@anthonyirwin6698do you have an Nvidia card, someone said that is looking for Nvidia driver, so that might be one cause, i have windows and Nvidia and tested on 2 computers and did work, but i can not check on other systems, you can maybe post on their page on issue tab, maybe someone can help
Update: Check this video How to Install Forge UI & FLUX Models: The Ultimate Guide
ua-cam.com/video/BFSDsMz_uE0/v-deo.html
Here are some useful resources for Stable Diffusion:
Download Stable Diffusion Webui Forge from: github.com/lllyasviel/stable-diffusion-webui-forge
Download Juggernaut XL version 9 from: civitai.com/models/133005/juggernaut-xl?modelVersionId=348913
More info on FreeU:
github.com/ChenyangSi/FreeU
Download more ControlNet SDXL models huggingface.co/lllyasviel/sd_control_collection/tree/main
Extensions used github.com/ahgsql/StyleSelectorXL and github.com/thomasasfk/sd-webui-aspect-ratio-helper
If you have any questions you can post them in Pixaroma Community Group facebook.com/groups/pixaromacrafts/
or Pixaroma Discord Server discord.gg/a8ZM7Qtsqq
Will this work on Mac M1?
Sorry I don't have a mac to test it, but didn't see something that says supports Mac, so probably not yet
@@pixaroma OK tnx
You should probably put these links in the video description. It's way more likely to be noticed.
when i open run.bat it says couldnt install pip, can you please help?
I have been looking for a tutorial like this for months. You have a real talent for this tutorial style and I HIGHLY encourage you to keep making these videos. Information is packed and logically flowing from one point to the next. Subscribed!
Thank you for showing more than just the installation like customizing the ui and settings :-)
Best video on SD Forge on youtube. great contribution to the community!
HOLY COW! I've been using A1111 (and now Forge) for a year, so by now I know most of these "hacks", but I wish a video so clear and so thorough existed when I was starting my journey. I even picked up a new nugget here and there. Bravo! Subscribed. And Saved.
Thank you ☺️
I like A1111 but I have found more performance in forge, I must say that you are very good at explaining, excellent video!
Thank you, yeah for me forge is faster and didn't crash like a1111 :)
Does it crash more or less often? I've been using A1111 for a while now, but it feels like it's been crashing more and more. Especially with SDXL models.
Since i switched to forge it didn't crash at all, only when i used control net ceashed if i didn't use image size divisible with 64 @@edouarddubois9402
@@pixaroma When you say image size, you mean the actual resolution?
The width and height of the image, sometimes i got that error when the size was not divisible with 64, but mostly when i used some extensions@@edouarddubois9402
Finally I found the best tutorial channel on YT. Thanks a lot!!!
This was exactly what I needed to get started. This is perfection and I can't thank you enough for your work. Bless you man.
Watched 10 videos about SD Forge install, but your is the best, Quick to the point, super useful tips for beginners, you even say what to do if it crashes.
Thank you, glad it helped ☺️
Do you know minimum hardware requirements?
Nvidia video card, 6gb of vram tested, possible work with 4gb of vram but not sure
@@pixaroma It works with 4gb :)
most helpful video on AI to ever exist give this guy an award please, very helpful saved me hours thankyou :)
Great walkthrough! Just switched over from EasyDiffusion and Forge is a massive improvement in regard to generation speed.
Duuuude!! So detailed, thank you!! Not hush hush, like well done on your style of explaining. Amazing
Thank you very much! Loved how clear you brought everything across! Definitely am gonna hang around here :)
It should be noted, for those who stumbled upon this like I did without knowing any better, that this method only works for nVidia graphics cards. WebUI uses CUDA, which is a proprietary API specifically for nVidia...meaning if you don't have their drivers, you can't natively run Web UI.
Luckily there are forks that exist that do work for AMD Radeon cards, but you'll have to jump through a few more hoops than what is shown here in order to install, and it probably won't run quite as fast as it does on nVidia cards.
Thank you so much! Now i can make clear and sharp pictures with the settings you've provided, once again, tysm!
Thanks for your video! I installed Forge yesterday (no stranger to A1111 here), but thought I'd check out a video or 2. For preferred defaults, I've been editing ui-config.json. Didn't realise there was a more straightforward method via settings haha! Dark mode is so much easier on the eye. I set it at browser level so that all pages appear dark, then display settings > high contrast in Windows (7, 8, 10, 11) will give dark mode OS-wide. Thanks for the heads up on SDXL styles and how its extra prompting works. 2 more extensions I use that might help is CivitAI Browser+ which integrates CivitAI into A1111/Forge, and ADetailer (After Detailer) which is an automatic in-paint utility that tidies up facial features and I find this better than GFPGAN and CodeFormer. Your 7-second image generation near the start of your video took 17 seconds for me with the same settings, coming from an RTX 3070 here. You covering file naming was very helpful too as I wanted to add the denoising value to the file names, and use a suffix instead of file numbers :-)
glad it helped :) I use rtx4090 so that why the generation was faster. Referring to styles check the latest videos i have one with 260 art styles :)
@@pixaroma Thanks. I have found what you meant + subscribed
holy shit!!! didnt know where the controlnet files where suppose to go was tryin to use on forge. vid help alot thanks!!!!
Incredible and informative! Well done. Thank you so much for the video.
Very helpful, i will be watching all the videos in this playlist, thanks! BTW what do you use for your voice It's great.( If it's not a trade secret that is)
The voice is from VoiceAir ai, they have it from eleven labs from what i know , i got a lifetime deal a while back
Very interesting. Good to know there is also another interface
Nice video and tricks bro, thanks!
...20:06 🤣😂😅
✨👌🙂🤗🙂👍✨
Wow Thanks for putting in the time to make this!! Is there any guide on using the Train tab (embedding, hypernetwork, train)?
I didnt play with that function yet, it always seems to be complicated to do trainings, I tried also on A1111 but I dont get always good results, needs good settings, good images, captions, too many things involved it seems. And now I saw an anoucement that forge is not going to be updated anymore, like is used more for tests or something.
This was VERY helpful, thanks"
great info. Also a quick tip, below the image there is a button to upscale using hires fix, just a quicker way to do it. 09:25 I think that option is new with Forge, wasn't in A1111
Thank you, yeah I didnt notice that :) good tip
Thanks for this tutorial!! I noticed you're able to generate very quickly. Can I ask for your PC specs? Thank you!
I speed up video sometimes, it takes like 5 sec for a 1024px image. I have rtx4090 with 24gb of vram, more vram the faster is the generation
First of all, I must say thank you. I started with your videos with the latest one, about Flux, and I am stuck here. Forge UI is fantastic! My only question is if I can find log files about prompts? It would be great to keep them.
Well each prompt and settings are saved in the png you generated, so if you drag the png you like into the png info tab you can see the prompt and setting. For more complex probably you need a script or an extension, on a quick search maybe an extension like this could do something similar, didn't test but maybe gives you some ideas github.com/ThereforeGames/unprompted
@@pixaroma Genial, thank you!
@@CsokaErno I use XnView MP as my default image viewer. It has meta-info tab on the right, no need to import images anywhere, you can just copy, alt-tab and paste prompt+properties into your browser. Besides that, it's a really handy piece of software compared to vanilla windows image viewers.
Best tutorial ever
This was excellent!!!!!
You should definitely meet your goal with this video!!
Thank you ☺️
@@pixaroma do you use an app to get the time stamps?
Like the chapters on the UA-cam? I use tubebuddy
i wonder if any of the stable diffusion UI makers (forge, automatic, comfyui etc) has considered a method for capturing 'recommended model settings' like you point out at 3:29 - as going out and hunting down a model's recommended settings is a work slow-down; perhaps be able to configure a 'model or ksampler template' that can be a quick preset based on the model.. would be kinda cool to have the option to be able to on checkpoint load to trigger the preset (but again should be an optional thing, not everyone would want that in all cases). if this already exists someone let us know
There is a preset saving extension so you can just save settings and give it a name similar to the model you use to know for what it is, but many extensions have bugs since with updates
Check the extension in this video to see if still works ua-cam.com/video/89YRfqArm-Y/v-deo.htmlsi=1va366VyvAt6s1f8
@@pixaroma You rock! thanks for those informative replies! will check that out. -- updated, yep the config preset still appears to work!
excellent guide! subscribed!
Thank You! Brilliant👍
Great tips, thanks!
Great job keep up the good work
Great tutorial, thanks
Thank you ☺️
quick question what's with ipadapters, I cannot acces any preprocesors there are only 3 encoders available, am I missing something?
You can read more about here, i didn't play with them in forge only with canny control net mostly, also keep in mind the version you are using there are different forks of forge now, the main one is used for beta testing and many things might not work! github.com/lllyasviel/stable-diffusion-webui-forge/discussions/178
great video! thanks very much. quick question, do you use tts for narration? If so, it's incredible, may I ask which one? I've been trying to find something decent for my videos. Cheers :)
Is called voiceair they have the voices from elevenlabs
thanks alot! I'll have a look at it@@pixaroma
Thank You So Much - Bro 😍
You are welcome ☺️
the real tutorial we need is how to run ForgeUI using colab, there are currently no colab notebooks provided for it.... some of us have colab pro subscriptions and want to run this on the big boy GPUs
I haven't used colab for months because I upgraded my PC, i just share tutorials on how I use it and the knowledge I got so far, sorry i can't help more
I'm new to this Stable Diffusion GUI. Experienced people can you please answer is this Forge WebUI is better than Fooocus MRE? If Yes, then in what parameters is it better? Thanks!
You can have both installed and play around, just put it on a different folder. It has more options and extension then fooocus from what i know
thanks for the tutorial. i have many models in stable diffusion. can i use them in forge ui?
Yes you can use it just like in other interfaces if is in the right folder or your settings are changed so it can take it from the folder you put them
Little question: did anyone have a problem with LyCORIS models on Forge? I'm using Forge through Stability Matrix, and no matter how I load -- from my computer or Matrix's models loader -- it just doesn't show up in Lora's tab. And when I load it in Lora's folder, it doesn't work correctly.
I didnt use lycoris yet, but I hope someone can help you with an answer . Is it possible that forge it still has some bugs
This is a great tutorial, but for me using it on Ubuntu makes feel a bit sad, cause most of extensions simply doesn't work, or doesn't want to install. Maybe because of GPU (rtx2060,6gb), but when I had Windows system on same machine, it had more extensions preinstalled and used. Like I don't have Free U and Control Net SDXL. As well as I remember, it worked more better than now. Did they made some new updates, which made it work worse? (Last time used 3 months ago)
They stopped updated for the official version, there are some versions that still around but not sure how much update they get., you can try the last stable version or change to dev2 ua-cam.com/video/RZJJ_ZrHOc0/v-deo.html. I have on older pc same gpu, but i wasnt able to run with control net, it crashed. ComfuUI works ok, but didnt try yet complex stuff
@@pixaroma Thank you!
The tiling bug has been also fixed. I just checked on a generated image
Awesome, thanks for info ☺️
How to use A1111 extensions? I'm trying but it auto turns em off. And integrated extensions are 💩. Help
not sure if all those works, but did you installed them from extension? Go to extension tab, click on available, click on Load From button, that will load all, search for an extension, for example tried ratio helper in search and installed just well when i clicked install, and restarted forge.
is it possible to setup all of this on AWS? could you please make a video? I am looking for an sketch to image model
I dont use aws but saw online that is possible, stable-diffusion-art.com/aws-ec2/ check this article maybe it helps
Hello, really great tutorial. I have a question I want to use the 4xVALAR upscaler but have no idea where to put it. Could you please answer if you have an idea in what folder exactly it should be.
Go to your webui\models folder and there create a folder named ESRGAN, so you will have webui\models\ESRGAN path, and in that ESRGAN you put that upscaler model. That worked for me, hope it works for you.
Nice! Thank you. But unfortunately ControlNET is dead in this built.
For me works if the image size, width and height is divisible with 64
I have NVIDIA graphic card but i get this error when run forge: Torch is not able to use GPU; add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check. how can I solve this?
so in the last month there has been a lot of updates to the forge ui, today I am working on an update video with what is new. Go to this discussion page to see what they changed and in comments some people had the same problem like yours and it seems has something to do with forge, on the bottom of the page you can see comments and click to load more comments, and with ctrl+f you can search for "torch is" to find comments that have those words github.com/lllyasviel/stable-diffusion-webui-forge/discussions/981 also check this page in the comments how some used different settings in arguments - github.com/AUTOMATIC1111/stable-diffusion-webui/issues/1742
Just updated to latest Forge version , the one that can work with Flux, but using only Sdxl on my 8gb card : every time I do inpainting or img2img the result has lower saturation than the original, it's me or what? Assigning a VAE do not solve 😢
There are a lot of bugs on new version so it will take a while for all to get fixed, this has a similar problem github.com/lllyasviel/stable-diffusion-webui-forge/issues/1189 and if you look at the list of open issues are like 600 github.com/lllyasviel/stable-diffusion-webui-forge/issues
@@pixaroma thank you for your answer!
My built-in controlnet's IP-Adapter is missing its models, and thus, doesn't work. Any ideas? I wanted to install them manually, but the library is different, and so are the files.
Forge still has some problems with control net, check this discussion maybe it helps github.com/lllyasviel/stable-diffusion-webui-forge/discussions/178#discussioncomment-8572388
I have attempted many times but am unable to get this to run with AMD R9 580 looking for the CPU ver
I think is looking for a Nvidia driver, you can try comfy UI or automatic 1111 but depends on the video card vram
bro you awesome, thanks
Niice, thanks for sharing
8:37 yes it same cute cat we see everywhere. its picture on model. so far nothing havent AI
:)
Out of curiosity are you using an AI voice tool or is this your real voice? Some crisp amazing voice if its all you :D
Yeah I am using a voice from voiceair.ai , my voice is ok but my spoken English is not so good and the accent is too strong, I am better with writing, that why that voice can produce a clear English and anyone can understand me, and sounds good for an AI voice.
i dont have a update.bat file. where do i find it? i can run forge just fine but been trying to find out how to update.
Should be next to run.bat and environment.bat a file called update.bat i have it there since installation, your should have it to. Just careful with updates to have a good stable version, check this video ua-cam.com/video/RZJJ_ZrHOc0/v-deo.htmlsi=rF-9wCmzResJiW3L
Thankyou for this tutorial! ❤Do I need the automatic 1111 stable diffusion installed to be able to install forge? I have the oldest version of automatic 1111 installed and I hvnt used it or upgraded it as I cudnt keep up with the every new update and other troubleshooting issues as I hv zero knowledge of programming language 😢
You don't need to have it installed it for forge to work, is different UI similar to Automatic, you just install it in different folder
What if I just want it to use my A1111 folders for checkpoints and VAE/embeddings/loras and all that jazz?
I explain that in this video starting with minute 3 43 ua-cam.com/video/q5MgWzZdq9s/v-deo.htmlsi=WOTQNXWSwiwWBkU6&t=223 in your forge folder you have webui folder and there is a webui-user.bat file where you can put the a1111 paths, just have to switch backslash with forward slash
@@pixaroma Thanks a bunch for the response, priceless advice! That slash part was pretty important lol.
It don't mess with the normal AUTO installation, it use his own ENV?
Yeah you can have both it doesn't mess with a1111 or others
please i got this error when i started run.bat
C:\Users\moor\stable-diffusion-webui-forge>python launch.py
Traceback (most recent call last):
File "C:\Users\moor\stable-diffusion-webui-forge\launch.py", line 1, in
from modules import launch_utils
ImportError: cannot import name 'launch_utils' from 'modules' (C:\Users\moor\AppData\Roaming\Python\Python310\site-packages\modules.py)
Not sure what is with error, but what you can try is to create a folde on another drive and try again fresh maybe can make it run, I don't know coding but looks like can not important a file, maybe something didn't download or is a bug, that why I say to try a fresh install in another folder
I do not see the SDXL styles that you show in the video at time 18:12 - how do I enable that.
I explain it in this video, is a file i created that you can download and put it in the right folder ua-cam.com/video/UyBnkojQdtU/v-deo.html
@@pixaroma got it
Thank you so much. Can it be installed in collab as well as fooocus?
I am not sure, maybe someone else can answer that
@@pixaromathank you, hope that some one can answer. I don't have the requested performance machine to do local installation, so that will give me a great help. why I'm asking? it's just for the seamless pattern setting that exists in the models presented. this capability isn't offered right now in fooocus witch is easily accessible with colab.
Thanks for the great video. One question here; does Forge works with AMD GPU ? Or it only run with Nvidia GPU ?
From what I know it search for Nvidia driver so no it doesnt work, many fail to load on AMD, i saw one commenting somewhere that got it to run on amd one UI called SD.Next by Vladmandic
@@pixaromathanks for the reply. i will check it out. much appreciated
One day, could you make a video about how to fix bad eyes and hands? Or you have it but I can't find?
Flux nodels can do that most of the time without fixing, so if your video can handle flux you should try. I have video on flux for both forge and comfyui, just forge is still work on progress, works with flux but other things still don't work yet, they are changing the interface and need time to fix all the things
@@pixaroma Yes, you are right, but in this case I talked about only SD. Flux can handle hands and eyes pretty well but when two hand are close or touch each other, it makes fault. Anyway, I suppose it will be fixed very soon like everything else in AI realm :)
@@CsokaErno with sdxl I got ok results if i used control net, there are some more nodes that can be used, I will see if I get enough info to get an episode about that in the future
@@pixaroma Thank you.
The controlnet doesn't work with me... the openpose the SD is ignoring the pose and... did something random and nothing to see
Not all control net models works, i use mostly canny you can see in my other videos, the rest of them are not working how it should.
@pixaroma Owww... I understand, I haven't seen the other videos yet, i dint have time, I came here to see the correct folder to paste the file. Do you have any tips for me? I have a Ryzen 5600g with 32GB RAM and a 3060 with 12GB. What is the best SD for me to install here?
@@alterinfo9448 I suggest swarm ui or comfyui, I am adding in a hour a tutorial on how to install both including stable diffusion 3
Could you please tell if this can also be run on AWS sagemaker?
Maybe look here gist.github.com/ShMcK/d14d90abea1437fdc9cfe8ecda864b06 aws.amazon.com/blogs/machine-learning/use-stable-diffusion-xl-with-amazon-sagemaker-jumpstart-in-amazon-sagemaker-studio/ as I don't use aws I can't hellp
When I put my model in control net "control_v11p_sd15_openpose.pth" and when I try to generate the image, I have a error message "TypeError: 'NoneType' object is not iterable. My setup is OpenPOSE and processor Openpose_full, can you help me please
i see you are using v1.5, do you get the same on sdxl? I got that error recently but not on control net but when i used an extension at different image size. It works at 1024x1024px? I got that error on different sizes, but worked at 1024x1024px. I dont used v1.5 anymore since sdxl appeared.
so i tried with sdxl control net and i get the same error if i use certain sizes, for example it works if is 1024x1024px, or 1024x576px, but i get that error if i use 1200x672, or 912x512 or 1024x816
Seems that the image size need to be divisible by 64 to work
@@pixaroma
thank you for your answer, I tried all the sizes 1024, 512 etc. it does not work, I have this problem that when I want to use controlnet otherwise to generate an image no problem but when I use controlnet impossible to generate an picture
@@REUBEUCOP75 maybe you can report it on their page, at issues github.com/lllyasviel/stable-diffusion-webui-forge/issues
I did run.bat, it showed runtimeerror: Torch is not able to use GPU, what happened?
Either your video card is not good enogh or forge dont recognize it, I am a designer not a coder, but you can try this in webui-user.bat add the following arguments to see if it works, it need at least 6gb of vram and prefers nvidia cards but try it anyway:
set COMMANDLINE_ARGS=--skip-torch-cuda-test --precision full --no-half
I see someone already posted taht on the bugs area, you can watch that to see if it gets any response if nothing else works, github.com/lllyasviel/stable-diffusion-webui-forge/issues
How big size is the file? 60GB without model? And each model is around 3GB or 30GB?
I am not sure about the total size, but usually a sdxl model is 6gb and a 1.5 model is around 2gb.
Thanks!
Great stuff. Is it possible to run it on AMD?
From what i know it needs Nvidia, so not at the moment
Greetings, it shows CUDA stream activated: False in cdm does it affect this Stable Difussion? If I have to also activate CUDA how do I do that?
Mine is off also, can be activated with some command in the bat, I tried but made my generation slower not faster so i left it deactivated, it appears as suggestions on the cmd window when you start and there is also a command I think, i don't remember now, just i know activated was slower for me
You solved the problem I was actually worried about the output speed. So I don't have to worry about CUDA but the internet connection, it should be sometimes fast sometimes slow which affects the output speed. Thanks for the tutorial above.
thanks for the controlnet section I got stumped on where to put the models. Edit: can anybody help me with LORA models, I paste them in the LORA folder but FORGE doesnt seem to detect them
i just tested now with a file and seems to be recognized, webui\models\Lora so the folder is Lora, and after you past it there you go to lora in the interface and refresh the page or just restart the stable diffusion so it can see it
There is no backwards compatibility with dynamic thresholding when using png info
there are still some bugs and extensions that doesnt work yet, probably will be fixed in time..
What if you use both 1.5 and XL checkpoints? Do you have to keep manually changing the FreeU settings everytime? Also with Hypertile, should the tile size be half of the generated dimensions longest side? what if I am usijng x2 hires fix, should it be half the size of the hires fix? What does the hypertile swap size and depth settings do? and is there a way like in SDNext to have hypertile set to automatic mode?
I don't use those settings too be much to give you more details, i know that for control net you have to manually keep changing models, 1.5 is different from sdxl so it needs different models and settings
@@pixaroma yes that's to be expected for control net as 1.5 and xl use different model. That's easily done. What's not easily done is having to remember the precise settings for freeU for 1.5 and xl. Can't believe there is no preset option to load these on the fly. I also am disappointed that there is no auto mode for hypertile tile size. Damn, all these different UIs should come together and make one ultimate ui
Brother, Do you know how yo generate prompt from image for free as some websites are charge
Forge has some basic prompt from image but is not so accurate, in img2img tab under generate it has a paperclip icon, first time will download a model but after that should work faster, and is giving a basic description of the image you uploaded in the img2img.
For some reason when I'm using img2img it spits out something that looks demorphed and blotchy or at a worse quality. Any tips?
Try to put VAE at automatic or none, also try different models and samplers.to.see if that fix it, try random seeds, also try updating the forge, see if any of those can fix it, is hard to tell what can be the cause you have to keep trying different things to see what cause it
@@pixaroma Thanks, I'll try that. It works decent with Automatic1111.
How do I share existing models on A1111 with forge?
You can edit the forge bat file i explained in this video ua-cam.com/video/q5MgWzZdq9s/v-deo.htmlsi=VQDUDjPvi256KCps
hi my installed Sd forge doesn't have the update.bat file. Is there anyway to update SD forge without the file? Maybe by adding arguments to look for an update?
It should have it there next to run. In the folder you extracted not in the webui folder
Where should we put lora files? There is no lora folder in models
You should have a folder for lora, look at this video how i download and where i put them ua-cam.com/video/q5MgWzZdq9s/v-deo.htmlsi=nKX2enJ7KPEAoGIF
So, in the arguments section where you put the dark theme I can add:
--pin-shared-memory
--cuda-malloc
--cuda-stream
For optimization, right?
Thanks for the video!
Yes, i tried all those as forge suggested but didn't make it faster on my rtx4090 but slower, maybe it does better for you but for me with bo argument was faster
@@pixaroma gonna try it on my 4080.
It's not working. Is it really on web-ui.bat file that I should put the arguments?
is in the webui-user.bat look for set COMMANDLINE_ARGS= and there after equal you add it, like I added the dark theme you add more, set COMMANDLINE_ARGS=--theme dark --cuda-stream
I have Forge installed via Pinokio but it doesn't look anything like this. Can I install this standalone version on a drive other than my C: drive?
You can install on any drive, i have on D mine, you just create a folder where you want and put the installation files there and run it
@@pixaroma Cheers! I'll give it a go, as no tutorials seem to match the Pinokio version
1. Is there any performance drop if I don't install it on C drive?
2. My C drive is SSD and D drive is HDD. Can I still install on D drive? Will I face any performance difference or issues?
Thank you very much! I tried it. It can generate and download images, but the window that shows the generated image does not work. What should I do?
I am not sure, can you join my Facebook group and you can you show me maybe some screenshots or post there to take a look, do you get any errors or what it looks like
Go to settings and paths for saving. Set save file paths (output dirs) to full paths, like C:\pathtoyourimagedir
It worked! Thank you very much!!@@Dark_Lobster
I'm trying it now but the results are slightly blurry (default settings). Any advice?
Not sure, try VAE automatic, also try different models, i use now juggernaut x version 10
@@pixaroma thanks but I have only one model... You mean in the SD checkpoint?
You can download more models from civitai website and depends in model some need different vae settings or models, thr juggernaut x i use it has embedded vae and my vae setting is set to auto. Sorry i am not at the computer this weekend to check exactly but if you do exactly like in the video it should work
@@pixaroma thank you so much. I think that I have to go deeper in learning because with same prompts I have spectacular results with midJ and disaster with Forge 🤷♂️
can you add image to video with this?
The latest version doesn't have the svd video tab only the older version. Maybe they will add it back in new updates
I use forges deforum tab to create animations. I would like to know how to create the animations within a boundary. I projection map so I would like to know how to keep the animations within the map of my house. Would you know how to accomplish this? I have a png map file that I created but unsure what to use it with.
TIA
Sorry, I didn't play yet with deforum, so i can't help there yet, I like to create HQ images and the video and animation isn't quite there yet, i am waiting for an improvement before i jump in to it
which GPU do you recommend for the lower budget?
Any Nvidia rtx that you can afford, more VRAM the better. Minimum 6-8 GB of vram, but if you can get more you will generate faster
I have problem with launching Web UI (1:42). "Found no NVIDIA drivers on your system..." Am i able to run it on rx 580 8gb?
I think it only works with Nvidia Video Card for now, that is why it says it didn't find a driver
as im trying to generate a art with juggernaut model UI showing is in queue any solution?
Do you have enough vram on your video card? Maybe try the hyper version of juggernaut that doesn't need so many steps to generate
@@pixaroma I have 4gb of vram as im generating with SD realisticVisionV51 works perfect As you suggested i have installed the Juggernaut_X_RunDiffusion_Hyper but still the same issue
The one you mention seems to be based on sd 1.5 so the sd 1.5 are smaller and you can run it because usually has like 2gb, but when you run an sdxl base model that is 5-6 gb your 4gb memory might not be able to handle that size, so either it takes a lot of time to load like a few minutes or it crashes. so maybe just work with sd v1.5 model for a while or if you can find an sdxl model that is smaller maybe
@@pixaromathank you for the assistance 🤗
Using RUN I got the following error: "Torch is not able to use GPU: add --skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check"
Does this mean my GPU will not work for this? I have a GTX 1660 Ti with 6GB GDDR6. Not a new card, but it's getting me by. No good?
I am not sure, try to add what they recommend, add --skip-torch-cuda-test to COMMANDLINE_ARGS, so copy --skip-torch-cuda-test and then open the webui-user.bat with notepad and paste it there after the equal sign, where it says set COMMANDLINE_ARGS=, should be set COMMANDLINE_ARGS= --skip-torch-cuda-test and see if that works. For me it worked on 6gb vram but was on rtx4090, is still a new UI so it might still have bugs.
I received same error and tried suggested command. Restarted run.bat and got new error saying no NVIDIA driver on your system. Makes sense because I have AMD video card. Simply put this is only for NVIDIA users.
my forge ui is stuck
[ aut-sd-paint-ext ] Auto-update disabled
I just installed a fresh copy today and I didn't run on any problems, so maybe try again, not sure what is the problem, if you have Nvidia usually works ok
@@pixaroma i solved it , actually i was trying to install some extension , but maybe some error happened , , later on i was panicked , and then i thought carefully , and deleted some extensions , that resolved it
Hello, i am using the SD web ui and anyway i can update it to forge ui without doing everything all over again?
I don't think so, is a different interface based on ,a1111 but is made by different users, so is not an update is a different UI. Also seems that works only with Nvidia card
@@pixaroma Thanks! yeah i do have a Nvidia card and seems like everyone using this one. Mine, web ui looks really outdated and have many things missing.. Looks like i have to install all over again? should i uninstall the SD web ui?
No, you can have in different folder forge ui, just depend on the space on your hard drivez I have both installed in different folder
@@pixaroma Oh great to hear, thank you for the info and the tutorial, i really appreciate.
All this stuff and for some absurd reason they removed the "refresh" button for all tab files.
It is not perfect, I hope they can update it soon since it has been a while since the last update
I installed a couple days ago, and when I try to render images, it shuts down my pc, which hasn't happened with MRE or A1111
this is on a 3060
That is strange, you can try report the issue on their page github.com/lllyasviel/stable-diffusion-webui-forge/issues maybe is a bug and so how it used the video card to much and make it crash, but that should not happen. I only tested on 2 computers, and i didnt had a problem on rtx2060 or rtx4090. Try an update in a few days maybe is get fixed. I wish I can help you more, but I am a designers not a programmer, I just share how to use new technology to get images with Ai, when is come to actual code, and bugs if it didnt happened to me not sure how to tell you a fix :(
Thanks again for the video! I have a question. Will this model work on my video card if it is only 8 GB? and if not, what options can you recommend?
I tested on 6gb and worked, only control net give me some crashed, but other things worked for me, faster then automatic that took ages. It worth a try, and if you dont like it how it work you can just delete the folder with all files. But i believe you can use most of the functions, if automatic1111 works this should work even faster
Wow! Thank you so much, I'll experiment. @@pixaroma
you only had to watch the video for 30 seconds and your question was answered holy shit
@@schinie3777 That question is not for this video!
I runned forge on 6gb rtx 2060 so it should work, other solution will be comfy ui since forge will not be updated anymore, and once they update automatic 1111 that will be also a good solution
21:34 yup allmost like dog you wanted was not found cat instead this candy xD. it AI it should do that to that dog image you choosed. why is not do
because in the prompt was a cat, so it used the lines from the dog to create a cat, if the prompt was a dog, I would have gotten a dog
What is the difference between Auto111 and Forge?
Mostly in optimization on how it handles memory, so it generates images faster then a1111 and have some extra things, but it stopped being updated officially so i switched now to comfy UI
@@pixaroma damn u.u any tip for someone with a 3060 who wants to use pony whitout waiting 30 minutes for an image?
I wanted to ask, I have models and loras in my fooocus folder. is it possible to copy and paste them to the appropriate forge folders? Or do I have to redownload them from civitai? ORRR...(is there a better way to have these models linked from my fooocus to my forge folder that could help on saving space on my cpu)...... I rather not redownload them or copy the file over so i dont fill up my harddrive.
I am adding a few hours a tutorials for forge, and how to link the automatic folder to forge, probably you can do something similar with the fooocus paths. I dont have focus to test it but it worked with the path of automatic1111. Also if you have it once you just copy them in the right location no need to redownload them.
thank you for the reply! I wonder if it will work linking fooocus folder to forge..i guess it is worth a try. ANy idea when your vid will be dropping for that so I can keep an eye out for it?@@pixaroma
I just added it like an hour ago, look for the thumbnail with under the hood in the title with an engine
Hello thanks a lot , i want to downgrade to Forge UI 1.7 but i cant find any information
I never did it but someone commented with this on reddit Use command prompt in SD directory and type git revert or git reset --hard . You can find the previous version hashes using git log, or there is a list somewhere on the github. So for forge probably you have to go to forge and then to web ui folder, then in the address bar type cmd and press enter, then you can see all those commit with a string of numbers, then I am not sure the next part either you use git revert and put that comit number or something like that
@@pixaroma thanks ill give it a try !
Can anyone tell me which config file(s) need to be changed on Linux(Ubuntu) so I can point to an already existing venv folder, and model directories?
I want to run this alongside A1111, which is pretty much working fine except that when I uncomment the line in the config file to point to my venv folder it still creates a new venv folder in it's install directory instead of using my existing venv folder.
Same with models, I can't seem to find anywhere to change where forge looks for it's model files?
To have all the models from the Automatic1111 in forge you do like this on Windows (I dont know how is for linux and others but maybe is something similar):
Edit or open with notepad the webui-user.bat
look for where it says
@REM set A1111_HOME=Your A1111 checkout dir
and from there down remove all the @REM you can find, that will uncomment those lines
first line should look like this
set A1111_HOME=D:/StableDiffusion/sdwebui/webui
for me that is the path of Automatic1111 D:/StableDiffusion/sdwebui/webui notice the slash, you need to change it, because when you copy and paste your path will look like this D:\StableDiffusion\sdwebui\webui\ and you need to change the slash to look like this D:/StableDiffusion/sdwebui/webui
so you should have now in that bat something like this:
set A1111_HOME=D:/StableDiffusion/sdwebui/webui
set VENV_DIR=%A1111_HOME%/venv
set COMMANDLINE_ARGS=%COMMANDLINE_ARGS% ^
--ckpt-dir %A1111_HOME%/models/Stable-diffusion ^
--hypernetwork-dir %A1111_HOME%/models/hypernetworks ^
--embeddings-dir %A1111_HOME%/embeddings ^
--lora-dir %A1111_HOME%/models/Lora
call webui.bat
then save it and you can run the bat and you should be able to see the models from Automatic in Forge
@@pixaroma I'm trying to run Forge standalone so to speak, but even now it still won't work right.
@@anthonyirwin6698do you have an Nvidia card, someone said that is looking for Nvidia driver, so that might be one cause, i have windows and Nvidia and tested on 2 computers and did work, but i can not check on other systems, you can maybe post on their page on issue tab, maybe someone can help