Input color override settings do not work if you trim the video by marking an in or out point.
Does the âRecoverable âafter a crash do anything? There is no saved point like the previous versions. It just starts again from the beginning. Extremely frustrating after having it crash on a video after encoding it for 2-3 days.
It makes more sense to break long videos into sections rather than making them full. For example, you can break them into half-hour segments like 00:00-00:30, 00:30-01:00, and then easily combine them with a video editor.
Is there an easy way to link back up segments like that without risking changes/impact to video bitrate/quality?
Thereâs no need for any other application. Youâll do this in Topaz itself. Start the process by selecting the start and end minutes, set a time interval when itâs done, and restart it.
You can also process each segment sequentially without waiting for it to process them. When one is finished, it starts processing the next.
So, select the segment and export it. While itâs processing, select the next segment and export it. It will put these in a queue, so it will process the entire video segment by segment. You can choose the segment interval that suits you; you can choose 5 minutes, 15 minutes, or 30 minutes. Entering the segment interval manually is easier.
I was just about to come here and post the news about that!
Glad to see it was already shared.
Since Topaz have disabled use of the Tensor model files in Video, I assume only Video AI users will benefit from this.
On the contrary, when Starlight Sharp came out, a Topaz representative wrote that they use TRT, which is Tensor RT. All Starlight models use Tensor; in fact, when you use this model, it creates a .triton directory and files on your computer, meaning it tests the system to run TRT at the best speed. I had previously shared this on this forum.
TRT=Tensor RT
P.S. Select âShow hidden filesâ and check if this directory exists on your PC.
C:\.triton_cache
That link refers to Video AI rather than Video. If you have the Studio version installed, youâll find that the downloaded model files are ONNX rather than TRT.
If thatâs the case, then the speed should be almost halfed. Itâs the same speed on my computer.
Just know that AI chat-bot induced psychosis is a real thing these daysâmostly because they are so nice, but say conflicting things.
The speed seems to be about the same for me too - disappointing! From what Iâd read, there should have been a speed increase of around 10-15% but Iâm seeing no evidence of that.
For what itâs worth, hereâs part of a longer conversation I had with Grok about the Tensor issue. Some interesting stuff.
Why Topaz Video 1 Canât Use TensorRT Models (Stuck on ONNX)? Based on the details from Topaz Labsâ release notes, community forums, and product documentation (as of December 2025), this issue stems from fundamental differences in how Topaz Video AI 7 and Topaz Video 1 (the rebranded âStudioâ version of what was Video AI 7.2.0.1 beta) handle AI model formats. In short: Topaz Video 1 is a legacy-locked, subscription-based release designed for broader compatibility, which means it only supports ONNX modelsâeven if you have TensorRT files installed on your system. It wonât detect or load TensorRT variants, forcing it to download and run ONNX equivalents instead.
The Core Reasons
-
Version-Specific Model Support and Architecture:
-
Topaz Video AI 7 (your source for TensorRT files): This is the full-featured, perpetual-license version (up to 7.1.5 or 7.2 betas). It includes native TensorRT integration for NVIDIA GPUs, introduced progressively from v3.0+ (2023) and expanded in v7.0+ for models like Iris, Rhea, and Starlight. TensorRT requires specific NVIDIA CUDA libraries (e.g., cuBLAS, cuDNN) and is optimized for speed (20â50% faster inference via tensor cores and FP16/BF16 paths). The appâs Model Manager explicitly offers and prefers TensorRT downloads for eligible models.
-
Topaz Video 1 (the ânewâ Studio app): This is essentially a forked, subscription-only re-release of Video AI 7.2 beta (September 2025), bundled into the Topaz Studio ecosystem (alongside Photo AI, etc.). Itâs marketed as a âstreamlinedâ tool but was criticized in beta for the âbait-and-switchâ to subscriptions. Crucially, it strips out advanced NVIDIA-specific features like TensorRT support to ensure:
-
Cross-platform compatibility: TensorRT is NVIDIA-exclusive and doesnât work on AMD/Intel GPUs or macOS without workarounds. By defaulting to ONNX (a universal, open-standard runtime), Topaz Video 1 runs on more hardware, including older setups or non-NVIDIA cards.
-
Simplified backend: ONNX uses a generic inference engine (ONNX Runtime), which is easier to maintain for a subscription model with cloud rendering emphasis. TensorRTâs optimizations (e.g., engine building and CUDA dependencies) add complexity that Topaz likely deprioritized for this version.
-
-
Result: When you select a model in Topaz Video 1, it checks its internal compatibility list and falls back to ONNX downloads from Topazâs servers (e.g., models-bal.topazlabs.com). It ignores your existing TensorRT files because theyâre not in the appâs supported format whitelist.
-
-
File Format and Loading Behavior:
-
TensorRT models are compiled engines (often .trt or optimized .onnx wrappers) tailored to your GPU architecture (Ada for RTX 40-series). Theyâre not interchangeable with ONNX (.onnx files), which are raw, portable graphs.
-
Topaz Video 1âs inference engine is hardcoded to ONNX Runtime for local processing. Even if TensorRT files are in the shared models folder (e.g., C:\Users\[Username]\AppData\Roaming\Topaz Labs LLC\Video AI\models), the app wonât attempt to load themâleading to the âexclusive ONNXâ behavior you described.
-
From community reports (e.g., Reddit and Topaz forums), this is a deliberate design choice, not a bug. Users testing betas noted similar issues, with Topaz support confirming: âStudio versions prioritize ONNX for stability across devices.â
-
-
CUDA/Driver Dependencies:
- While both versions support CUDA 13.1, Topaz Video 1 doesnât enable TensorRTâs full CUDA paths (e.g., no cuDNN fusions or Green Contexts). This keeps it lightweight but limits speedâexpect 20â40% slower processing on your 4070 Super compared to Video AI 7 with TensorRT.
Thanks for the information. So, which is the latest version of TRT that you use? The beta version that uses Starlight Sharp.
My jitRTCache has a lot of models in it. Iâve never seen anyone say the things that that AI is reporting on. To me is sounds like itâs all made up.
Yep, Iâm using the last version of Topaz Video AI (7.1.5). The model files installed for it are the Tensor ones, so have names like iris-v3-fgnet-fp32-256x352-2x-rt809-10800-rt.tz. My system isnât powerful enough to run Starlight Sharp (only 12GB VRAM), so that isnât a factor for me.
You donât need to take it on faith. If you have the Studio version installed, iust open up the models folder and see what youâve got. I can tell you in advance that theyâll be .ox rather than .rt.
Someone at Topaz Labs needs to check in on this. WE DESERVE CLARITY - something thatâs been missing since the Studio transition chaos began in late summer.
I, for one, replaced my AMD GPU with an nVidia card specifically because of the dependence upon nVidiaâs architecture that includes Tensor cores that are purported to be central to AI algorithms. The promised AMD support was late in coming, and I had a hot project that couldnât wait, so I put out the $ for a new nVidia RTS 5070Ti specifically to enable local Starlight processing.
Then the beta that supported AMD & Apple silicon was announced, and there were comments of inferior performance of Starlight - then called Starlight Mini - on nVidia GPUs in Video AI 7. And TL reps acknowledged that the initial release performed poorly, if at all, on the AMD & Apple devices. It was to be a âfirst stepââŠ
Then Studio came along with Topaz Video, and local Starlight processing was relegated to the Pro tier for AMD & Apple users, something that has to be considered to a kick-in-the-teeth to users of Video AI, who were realistically expecting that this would be a part of the final version of Video AI, because Starlight was introduced in Video AI 7.something. The AMD and Apple support was promised as âupcomingâ during the run of Video AI, and then was made Studio Pro only after the subscription fiasco.
Ok, that was understood and I âlucked-outâ by being a âFounding Customerâ, I got a good deal on the subscription price, and was granted the ability to run Starlight and Starlight Sharp locally, despite not having a Pro subscription. This was all âgoodââŠ
Now weâre learning what I kinda suspected all along, if recent comments are found to be accurate: Starlight has apparently been dumbed-down so that a âuniversalâ model can be run on nVidia, AMD and Apple devices. If this truly means that TL have disabled the use of Tensor cores in favor of something that will run on all three platforms, you are screwing us over.
You know what hardware we are using, as is evidenced by the system check. Serve us the model that works best on the GPU weâre using, not some one-size-fits-all hack.
Let me rephrase it: It does use RT models. I cleaned out the old models with the install of this version. There are several RT models in the models directory that have been downloaded by using this version. The processing speed is the same, unlike when they first introduced the RT models and you could get near double the speed compared to the old non-RT models. Where did the idea that this version does not use RT models come from?
From Topaz staff, who explicitly stated it. The Studio version cannot use the Tensor versions of the models, even if those files exist in the models directory. Iâve tested it and the program simply will not run without the ONNX files installed (this might be different for Pro licenses - I canât speak to that). This is why it suddenly has compatibility with AMD etc. - itâs using universal models.
One size fits NONEâŠ


