The GPU engine is an alternative to its CPU counterpart. It allows you to use your GPUs (graphics cards) to compute the renders. They are usually much much faster than the CPUs, especially when using more than one graphics card. In this version, many lines of code have changed in order to make the GPU engine even faster and to allow for rendering bigger images.
Not all the graphics cards will work; please check the next section to know about the details.
GPU engine hardware requirements
Graphics cards have to be based onCUDA. The GPU engine is built upon CUDA computing platform, which is developed by Nvidia, so only Nvidiagraphics cards will work. AMD or Intel cards won't work.
Maxwell, Pascal, Volta and Turing micro-architectures are supported (compute capability 5.0 and up). Kepler architecture is not supported but cards based on it could work.
It's very important that graphics card drivers are up to date.
Maxwell now allows you to use all the graphics cards on your computer to calculate the render.
In order to choose which graphics cards you want to use, you will find a gear icon next to Engine selector in Render Options, when the GPU engine is selected. The plugins have a similar button on their equivalent menu.
Customize your GPU preferences here (capture from Maxwell Studio)
This button will open the Maxwell GPU Preferences panel.
It will set Maxwell's behavior on a particular computer for all Maxwell apps and plugins. These settings are not sent within a scene file or distributed across the network as each computer can have very different configurations.
In the case of FIRE (the interactive preview), Maxwell will only use one of the cards (the best one) and will be marked with [FIRE] tag.
Just mark the ones you want to use and leave unmarked the ones you want to spare.
This panel can also be opened from the Maxwell installation folder by running the “mxgpuprefs” application.
GPU Preferences panel
Please, note that the memory is not shared between the different cards, so the graphics card with a smaller amount of memory will be the bottleneck for the render. The whole scene has to fit into each used graphics card's memory.
Will I get the same results using the GPU and the CPU engines?
Our aim is to make them render exactly the same result, however, the nature of the two engines is not the same.
Additionally, not all the features available in the CPU engine are supported yet in the GPU engine. We will keep working to add items to the list of supported features.
Here is a table summarizing the features currently supported or not supported in the GPU engine:
Normal layer mode
Additive layer mode
Layer opacity and masks
Interactive render (FIRE)
Motion Blur *
Hidden to Camera
Displacement (pretess mode)
Hidden to Refl/Refr
Hidden to GI
* Motion blur works for all transformations but not for deformation.
Which card should I buy to use with the GPU engine?
Only Nvidia, no AMD or Intel, sorry.
As usual, the bigger the numbers, the better, but which numbers do really affect the render?
The memory of the graphics is crucial as the whole scene has to fit in it to be able to render, so the higher the better. The amount of memory is not relevant for the speed of the render though; this will affect the size of the image you can render, the amount of geometry you can load and the number and size of the textures that can be loaded to render.
The higher the number of CUDA cores and their speed, the better. This will determine the rendering speed for that particular graphics card.
The size of the memory interface and its bandwidth will also affect speed as it affects how fast the information travels between the memory and the cores. The higher the better.