I tested GPU+ CPU rendering with 4 different scenes. and It was way "slower" with both of them turned ON. I asked several people and they all agree they make your renders slower in most cases. So please don't do that. ( I have a similar setup just like yours except I'm rocking an intel CPU)
Ive heard people say the same. MIght be an AMD thing, idk. But its good that you wrote this. At the time when I made this video I didnt know this so its good to hear from other people how their computers behave.
@@VisualPixelsCG hey also, idk if u tested this. Running GTX cards on OPTIX is a joke. I got around 5-18 second differences on test with CUDA. Which is actually alot when u rendering an animation. So definitely use CUDA for non RTX cards.
Ive noticed one thing with the tiling. In the video I said that using a smaller tile will make it easier for the gpu wich isnt fully right. The smaller tiling will make it easier for the Cpu. One big tile will lead to the cpu bottlenecking and the gpu will run on ~10%. and a small tile can bring the usage of the gpu to 100%.
this could be caued by different things, to me it sound like a hardware problem. I think you might render out a animation in mp4 format and then at frame 800 your computer cant handle more data. therefore I suggest changing to rendering in a sequence. So instead of rendering in mp4 format you change to png and select a folder where you want every frame to be saved. Then when you hit animation every frame will be saved in that folder. then in the video editor you can import a sequence and render it again to make it a video. hope that helps
I tested GPU+ CPU rendering with 4 different scenes. and It was way "slower" with both of them turned ON. I asked several people and they all agree they make your renders slower in most cases. So please don't do that. ( I have a similar setup just like yours except I'm rocking an intel CPU)
Ive heard people say the same. MIght be an AMD thing, idk. But its good that you wrote this. At the time when I made this video I didnt know this so its good to hear from other people how their computers behave.
@@VisualPixelsCG I assume you probably done some tests too and got the results in the video? Yeah like you said it might be the AMD thing.
@@VisualPixelsCG hey also, idk if u tested this. Running GTX cards on OPTIX is a joke. I got around 5-18 second differences on test with CUDA. Which is actually alot when u rendering an animation. So definitely use CUDA for non RTX cards.
@@msb8111 Thats intresting. Because before I made the video I tested that and my laptop was way faster with optix
Ive noticed one thing with the tiling. In the video I said that using a smaller tile will make it easier for the gpu wich isnt fully right. The smaller tiling will make it easier for the Cpu. One big tile will lead to the cpu bottlenecking and the gpu will run on ~10%. and a small tile can bring the usage of the gpu to 100%.
Really very thanku bro
Thanks! The only issue is the cycles render stops at frame 800-900, do you know why? I use amd gpu 5500 and Cuda
this could be caued by different things, to me it sound like a hardware problem. I think you might render out a animation in mp4 format and then at frame 800 your computer cant handle more data. therefore I suggest changing to rendering in a sequence. So instead of rendering in mp4 format you change to png and select a folder where you want every frame to be saved. Then when you hit animation every frame will be saved in that folder. then in the video editor you can import a sequence and render it again to make it a video. hope that helps
i have GTX 1650 which render engine i used please help
I used cycles. I dont fully understand what you mean. can you explain more?
@@VisualPixelsCG i mean i have GTX 1650
so what can i use CUDA or OpticX
you should use optix. but since its not a rtx card it wont have a big difference just a little. @@shahmaarbaba
@@VisualPixelsCG thank u
@@shahmaarbaba np