Topaz Video AI v5.5.0
System Information
OS: Windows v11.23
CPU: AMD Ryzen 7 7800X3D 8-Core Processor 31.604 GB
GPU: NVIDIA GeForce RTX 3080 9.8174 GB
Processing Settings
device: -2 vram: 1 instances: 1
Input Resolution: 1920x1080
Benchmark Results
Artemis 1X: 19.99 fps 2X: 13.73 fps 4X: 04.72 fps
Iris 1X: 19.05 fps 2X: 12.47 fps 4X: 03.81 fps
Proteus 1X: 19.48 fps 2X: 13.68 fps 4X: 05.08 fps
Gaia 1X: 06.57 fps 2X: 04.48 fps 4X: 03.14 fps
Nyx 1X: 07.86 fps 2X: 06.80 fps
Nyx Fast 1X: 16.53 fps
Rhea 4X: 02.72 fps
4X Slowmo Apollo: 24.26 fps APFast: 62.23 fps Chronos: 14.72 fps CHFast: 24.20 fps
16X Slowmo Aion: 35.76 fps
Quite happy with these results, running an off brand HP RTX 3080. Haven’t had TVAi for more than a week or so and have learned a lot about the why’s and how’s regarding the extremely sensitive voltages and clock speeds compared to synthetic tests(a lot of parts, such as tensor cores and CUDA cores don’t get utilized the same, or barely at all simultaneously, compared to AI related rendering. Before this i ran a fully stable undervolt and overclock at around 938mV@2040MHz with memory at +1200MHz and was barely touching 280W, scoring 10-15% higher than other overclocked 3080s. Stable through several different benchmarks and multiple hours long standard stability tests. But that all changed rendering with TVAi due to the nature of AI rendering. It was far from stable.
Comparing my results with most rtx 4090 in here. For the interpolation models Apollo and Chronos, my 3080 seem to render at same/slightly faster speeds. For the enhancement/upscaling models, 4090’s posts in here seem to average about 33-35fps for art, iris and gaia, which is about 40-45% faster than my 3080. Looking at raw specifications regarding the part of the GPU that is optimized for AI rendering, the 4090 stock should render about 80-88% faster than a stock 3080. I’d most likely fall behind running huge upscaling such as Rhea due to less than half the amount of VRAM.
MSI Afterburner is the app I’ve used. Above benchmarks are with a voltage limit set to 875mV, and max boost clock at 1860Mhz, memory clock at +900MHz and a 98% power limit. And of course, both CPU and memory and optimized as well. CPU oc -27 llc mode 4, Ram 2x16Gb@6000MHz 30cl and tightened timings with a latency of 63ns. Haven’t had to change any settings regarding those two from my previous, as loads on cpu and ram rarely surpass 75% power limit for the cpu and memory bandwidth is barely running at 50% of its top speed.
—DISCLAIMER–
My clock speeds might not be optimal for other 3080 cards, but it could be a decent starting point for anyone with a 3080 looking to min-max it’s performance. Do this at your own risk.
With that said, GPU crashes nowadays aren’t as bad as they used to be, as they put a lot of restraints on the cards as is. You’d really have to try hard to cause actual damage to it, eve, from severe crashes causing a full reboot.
After hours of testing different input formats, model combinations and outputformats with crash count surpassing x10 times, it’s now been running stable without any crash for about 3 days with an avg. 15h of rendering per day.
Finding a voltage that works with a corresponding clock is so much more sensitive than a typical benchmark. Slightly too high voltage and clock speeds drop drastically due power limit throttle. Slightly too high clock or slightly too low voltage limit it crashes.
My current voltage limit puts the gpu at an average 94-97% power, or about 300-310W(rtx 3080 power limit is 320W). Total Gpu load hovers around 98-100% with an average at slightly more than 99%. I have however put a limit on power to 98% as some of the models, eg. Apollo massively spikes in power, both by itself and combined with an enhancement model. These spikes were the biggest offender regarding gpu micro crashes(1-2s full freeze, screen flicker, process gets and error, and the gpu reloads - so no full-on crash that causes a reboot, but a crash none the less forcing the process to restart from the beginning). I’ve mostly been running without any power limit but set it to 98% yesterday when i was closely monitoring its power running Apollo. Apollo massively shifts in power, producing spikes between 250W to 330W. Setting a slight power limit should help mitigate that to a degree, since I’ve been rendering with apollo most of the last 24h and had no crashes. It’s the most prominent model to cause issues, due to its power spikes - at least from what I’ve found.
Due to recent thermal pad change as well as adding a few extra thermal pads on a few small ram modules, and repasting the gpu-chip. Core temp doesn’t pass 80C, hotspot 92C, memory conjunction 88C. Everything air cooled, with a conservative fan curve. Fans won’t run at more than 60%(a limit I’ve set, as i render overnight) if reaching >83C core, so they average 50-55%, each fan with slightly different % which is the result of hours upon hours of testing, so it’s a well optimized fan curve(worked as a hvac fitter years ago, now engineering) keeps those temps with a difference of +/- 1C during a +24h constant full load batch render.
If anyone has any good insights or guides regarding gpu accelerated AI rendering, I’d be more than happy to read up and learn more. Everything I’ve done so far has been trial and error testing as well as asking chatgpt a few specific questions as to why loads are different. Nothing about voltage and clock speeds as that can differ from gpu to gpu, even from the same manufacturer, thus asking chatgpt about that would be pointless.
Let me know what you think about my approach, and if you have any tips!