
- 20
- 255 696
Render Realm
Austria
Приєднався 11 лип 2022
Let's dive into StableDiffusion!
Stable Diffusion explained (in less than 10 minutes)
Curious about how Generative AI models like Stable Diffusion work? Join me for a short whiteboard animation where we will explore the basics. In just under 10 minutes, we'll discuss how Deep Learning and Neural Networks are employed to train these models, and how they generate images. Whether you're new to the field or seeking a deeper understanding of these models in a non-mathematical way, this video provides an accessible overview of Stable Diffusion's fascinating mechanisms. Hope you enjoy!
You can also download the whiteboard animation for free from Gumroad:
goblenderrender.gumroad.com/l/orbfz
Donations welcome, but not required.
#stablediffusion #generativeai #neuralnetworks #whiteboardanimation
You can also download the whiteboard animation for free from Gumroad:
goblenderrender.gumroad.com/l/orbfz
Donations welcome, but not required.
#stablediffusion #generativeai #neuralnetworks #whiteboardanimation
Переглядів: 19 006
Відео
Your personal Avatar from a single photo (2D/3D/Images/Animations/LipSync)
Переглядів 1,7 тис.7 місяців тому
Create a consistent Avatar in ComfyUI from a single photo with just 4 Rendering Steps, using SDXL-Lightning. This tutorial will guide you through the whole process of creating high-quality avatars of yourself or any other person, using several IP-Adapters, LoRAs and Controlnets and a pretty fast 4-step SDXL-Lightning model. We will then turn this avatar into a fully animated 3D-Model using Avat...
Turn MoCaps into 3D-Controlnets into ComfyUI Images/Videos
Переглядів 3,1 тис.9 місяців тому
This tutorial explains how to grab Motion-Captures from your Webcam or any video file, using the free BlendARMocap add-on for Blender by cgtinker with Google Mediapipe, and turn them into 3D-animated Controlnets, using the OpenPoseBone tool by toyxyz. No further equipment needed. In the second part of the tutorial, we will go through the process of creating images and animations with these Cont...
Animated ControlNets using the OpenPoseBone tool and ComfyUI
Переглядів 9 тис.10 місяців тому
This tutorial is featuring the great OpenPoseBone tool by toyxyz, available for free on Gumroad. With this tool you can create a great variety of ControlNet poses and animations in Blender, which you then can feed directly into ControlNets in ComfyUI or Automatic1111, without preprocessing any other Video or image footage, while keeping the scene stable and consistent. This video explains the w...
Reviewing & Rating 50 SDXL models
Переглядів 8 тис.Рік тому
I've tested and rated 50 different Stable Diffusion SDXL models in a structured way, using the GoogleResearch PartiPrompts approach, by assigning 107 prompts for each tested model to a Classification and Challenge Matrix, which resulted in rendering more than 5.000 images. I believe that this method - though rather time-consuming - can give a more concise view on the strengths and weaknesses of...
Using SDXL Controlnet with Automatic1111
Переглядів 13 тис.Рік тому
A quick tutorial about installing and using the new Controlnets for Stable Diffusion SDXL in Automatic1111 (yes, it's finally here!). Giving you a short introduction about the new SDXL models for Automatic1111, as well as the new Controlnets. The combination of SDXL and Controlnet allows you to produce stunning, high quality images of people in different poses, while keeping the original image ...
Stable Diffusion - Mac vs RTX4090 vs RTX3060 vs Google Colab - how they perform.
Переглядів 30 тис.Рік тому
A comparison of running Stable Diffusion Automatic1111 on - a Macbook Pro M1 Max, 10 CPU / 32 GPU cores, 32 GB Unified Memory - a PC with a Ryzen 9 and an NVIDIA RTX 4090, 24 GB VRAM, 64 GB RAM - a PC with a Ryzen 5 and an NVIDIA RTX 3060, 12 GB VRAM, 32 GB RAM - Google Colab on a NVIDIA Tesla T4 , 16 GB VRAM, 12 GB RAM Going through a variety of real-life test cases on each system, comparing t...
Convert any mesh into Lego (or other) bricks in Blender. Free Blender file available for download.
Переглядів 2,5 тис.Рік тому
This quick tutorial shows how to convert any mesh into Lego Bricks (or any other kind of bricks and meshes), while keeping the original textures intact. A free Blender file, containing the required Geometry Nodes, is available for download. Blender version 3.xx or above is required. It takes a single mesh with a single material in order to work properly, but the tutorial shows you how to conver...
Clone of ANY voice and create a lip-synched animation with Tortoise-TTS and StableDiffusion
Переглядів 8 тис.Рік тому
In this quick tutorial I will show you how to clone ANY voice and create a flicker-free, lip-synched animation of that person, using only free AI-tools (Tortoise-TTS, Stable Diffusion / Automatic1111/ SadTalker). Some useful links: Download Audacity: www.audacityteam.org/download/ Google Colab link Tortoise-TTS: colab.research.google.com/drive/1N8lXviiwfmrS9vYzuFdQnLHlddXMMMvU#scrollTo=vF1d-My9...
Picture Yourself with Stable Diffusion
Переглядів 32 тис.Рік тому
A quick tutorial about creating your own Stable Diffusion model in four easy steps, using just a few photos of yourself and training them with Dreambooth. All you need to do is running the free fast-Dreambooth notebook by theLastBen, using Google Colab, to create your Custom model and then import it into a local Stable Diffusion app, like Automatic1111. To watch some more great music videos cre...
RunwayML Gen 1 vs Automatic1111 - Video creation with StableDiffusion
Переглядів 4,4 тис.Рік тому
In this video we take a deeper dive into creating videos with StableDiffusion, comparing RunwayML Gen 1 with Automatic1111. First we will create a short animation in Blender, then feed it first into RunwayML Gen1 and later into Automatic1111, Batch img2img / ControlNet and explain the whole workflow of creating a video animation in each system. We also will show, how to reduce flickering and te...
Creating a morphing animation with Stable Diffusion, using any video you like
Переглядів 13 тис.Рік тому
This quick tutorial describes how to create a morphing animation out of an existing video, using StableDiffusion Automatic1111, Deforum Video-input and ControlNet. Deforum doesn't feature prompt interpolation from a video input out of the box, but with a little bit of math it can be easily achieved. If you have any questions, don't hesitate asking! I will also use this method for a new music-vi...
Creating an uber-realistic video animation from an avatar with Stable Diffusion
Переглядів 74 тис.Рік тому
This tutorial will guide you through the process of creating an avatar with ReadyPlayerMe, animating it in Mixamo, building a 3d scene around it in Blender and feeding this scene into StableDiffusion Automatic1111 to create a video animation, using an uber-realistic custom model with the Deforum and ControlNet extensions. To watch some more great music videos created with StableDiffusion, Unrea...
Quick Tutorial - Running Stable Diffusion on a Mac
Переглядів 6 тис.Рік тому
A quick and easy guide how to install Stable Diffusion Automatic1111 on a Mac with Apple Silicon. To watch some great music videos created with StableDiffusion, Unreal Engine and Blender, visit our UA-cam music channel: www.youtube.com/@-vero- Chapters: 00:00 Basic infos 01:53 Installing the required apps and packages 03:20 Downloading the required models for Stable Diffusion 05:48 Optimizing t...
Creating an audio-reactive music video & dancing animation with StableDiffusion
Переглядів 2,3 тис.Рік тому
This quick tutorial covers the whole process of creating an audio-reactive music video, also including a Mixamo dancing animation, with SD-Automatic1111 Deforum, batch img2img and Controlnet. The dancing animation has been prepared with Mixamo and Blender. You can watch the whole video here: ua-cam.com/video/7Db-h01a-zs/v-deo.html To watch some more great music videos created with StableDiffusi...
Quick Tutorial - Mixamo to Stable Diffusion
Переглядів 15 тис.Рік тому
Quick Tutorial - Mixamo to Stable Diffusion
Quick Tutorial - creating a 360° VR video with Stable Diffusion Automatic1111 img2img / Deforum
Переглядів 3,3 тис.Рік тому
Quick Tutorial - creating a 360° VR video with Stable Diffusion Automatic1111 img2img / Deforum
Quick Tutorial - creating a music video with Stable Diffusion Automatic1111, batch img2img
Переглядів 7 тис.Рік тому
Quick Tutorial - creating a music video with Stable Diffusion Automatic1111, batch img2img
Creating Social Media spheres with Blender - how to project images on a sphere
Переглядів 4,2 тис.2 роки тому
Creating Social Media spheres with Blender - how to project images on a sphere
How to create a music video in Blender - complete tutorial
Переглядів 2562 роки тому
How to create a music video in Blender - complete tutorial
Absolutely fantastic! You have given me plenty of good help. However i have a batch of pre existing a.i images i made of an underground abandoned tubeway. I would like to use these and create VR content. I assume i can use these to convert them into 360 ?
I've been using Stable Diffusion for about a month now, and I'm blown away by how amazing the software is-especially considering it's free! The quality is incredible, and with all the available extensions, it becomes even more fantastic. I’ve even canceled and deleted my Photoshop subscription because I no longer need it.
Wow~ No wonder "Stable Diffusion" is so Good to use! Thanks for this in-depth video. I think you explained it so well. I can't help watching it twice!!
This does not work anymore.
Great work
Taylor Amy Jackson Laura Harris Ruth
You put RealVisXL V2.0 into C? Really? You underestimate this model, I guess. Cause I now use RealVis V5, and it's fantastic - nothing to compare with.
A1111 now works with mac with the half precision parameter
Lewis John Williams Nancy Perez Shirley
Nice! Just a nit-pick about the neural network part at around 3:35; you are correct that each neuron has input weights (w1 & w2) plus the bias - these are the trainable parameters (including the bias, it is not a constant) for this neuron. The output calculation b+(w1*w2)=1.1 is unfortunately not correct - it would reduce the whole thing into a single matrix multiplication and not be capable of adapting to the input data in the manner we desire when using NN:s. The calculation is a two step process. In step one, you multiply the weights with the value/output of the previous connected neuron. In your case if the neuron connected through w1 has an output value of - let's say 1.5 - the input value from that neuron is 1.5*0.2. So you add all the inputs (multiplied with their corresponding weights) up and add the bias as step one. Secondly you take this summed up value and put it through a "transfer function" that can be pretty much anything you'd like as long as it is not linear to obtain the neuron output value. But, as you say, it is these weights (including the bias value) that are adjusted by tiny fractions in each training pass.
excelent. no video I found explains how stable diffusion works this simple and concise
I am pretty sure you understood nothing.
wow Thank you very much
Amazing, I hope to see more videos like this. Very much appreciated
well done man great presentation i came here to find a video for a friend im glad i clicked this one :D
Amazing!
Brilliant
My experience with my PCs: Win10, Fooocus 2.5.0 (SDXL 1.0), 1024x1024 image, base "realistic" preset, no other models or LORAs. Latest GPU drivers installed. Ryzen 5600X, 32GB DDR4-3200, 8GB RX7600 with PRO 24.Q2 drivers, Fooocus AMD version: avg 261 seconds per image Same PC with 24.7.1 gaming driver: 318 seconds per image Same PC with ComfyUI ZLUDA, PRO drivers, same model and settings: 340 sec per image Ryzen 5600G, 16GB DDR4-3200, 8GB RTX3050, standard Fooocus CUDA version: avg 33 seconds per image Newegg has a 12GB 2060 (refurb) for $170 so I have it on the way for additional testing this week.
The mac’s speed can be improved by running macOS Sequoia and by converting the Stable Diffusion model from Safetensors (or any other type) to CoreML so it can run on the Nerual Engine instead of the GPU or CPU. These can vastly improve the performance.
rtx 4090 vs rtx 3060, with more than 5 times more expensive. we would only get 3 times of performance.
In other benchmarks rtx 4090 should be have price $1500 instead.
Great tutorial!! I have one question: how to make the talking avatar more realistic by adding some eye blinking?
Please tell me about the difference between learning objects (people etc.) and learning concepts (painting style, etc.). Thank you very much. The explanation was very easy to understand, and I am grateful.
What about the next video? 🤣
Neat explanation.
Seems like it needed to be 20 minutes. xD
Is this an Unlocked and Unleashed version that you are talking about or is this software just a normal Locked version? Thanks and I look forward to your reply. John E
I have done all the process but in the end stable diffusion don't let me to choose my custom model. One thing I notice is that my model is a ckpt file, and all the other models that I can choose ends with safetensors
Hello. help me with choosing a video card for the difusion stem. I have a choice between rtx 3060 12gb, intel arc 770 16gb and rtx 4060 8gb, what do you advise?
3060 12gb, prioritize what has the most vram
This is amaz8mg!
This was perfect! 👍
what if i wanted to generate pictures of two different real people in the same picture?
where can i get the app?
When I watched this, You said to press Shift + Option. I'm using a Dell computer with no option button and you gave no alternate methods so I'm now just stuck.
thank you, very very much
Thank you.. exactly what I was after..😅
Thanks for the video: after months of trying finally I've managed to install Stable Diffusion on my Mac. Only one problem now, when I try to generate an image this text pops up: AttributeError: 'NoneType' object has no attribute 'lowvram' Time taken: 0.0 sec. Do you know whay is this happening and what can be done?thanks
How do i cast an on-the-ground shadow from the dancer?
lots has changed since then. now there are plenty of ai services that will extract mocap from video and export as fbx
Faboulus video. Will used as ref in my Master thesis.Thanks!!!
"smack lips" Mac is slow. I agree
Congrat ! What a nice job ! I have my own set of test but I think I will integrate your prompts. I noticed that many models change drastically from one version to another (Juggernault XL 9 to 10 for instance, classic or Turbo/Lightning versions)
Has something changed with this process? It seems like the "Safetensors" checkbox in the Model Download section is gone.
It seems great!! i couldn't test because the manager don't give me access to "IPAdapterApplyEncoded" missing for me...
the workflow is a bit old the "IPAdapterApplyEncoded" was replaced by the new "IPAdapter Advanced" just replace the node and reconnect
Thanks for this detailed video. One question: what minimum GPU is needed for the complete workflow, and which of the individual steps need the most GPU RAM?
Which one is better, the Rtx 3060 12GB or 3060Ti 8GB?
12gb. ram matters more for this.
7:57 i been using colab to get to familiarize myself with SD. I feel I am now enough and want to start generating locally. I don't want to invest too much initially, so I am trying to figure out what's the best low with good enough performance and I thought the 3060 would be a good start, nothing less. Your test suggest that, now I trying to decide between RTX 3060 vs RX 6600 vs Arc 580 >. I am next going to see if Arc can generate and if so, at what results. 5700XT should be good too, but for SURE the power consumption will be crazy.
Fantastic job! Well done and thanks. But like you said,, it is getting a liitle bit stale… I would really like an small update on your A and B models and maybe some other up-and-comers. I mostly use Juggernaut, but you have given me nice pointers on other, maybe better models. Thanks again.
I didnt understand about colabs, does it has limit on generated image per month?
Promo-SM
Thank you so much fow this easy and detailed tutorial! im a total noob with computers and very new to stable diffusion and until this video i was just confused and intimidated because i dot know any of these words and couldt follow. but this time i made it and it worked and tbh it was so much easier than expected... thx a lot!!!
Great, thank you, easy to understand visual lecture!