Yes that’s correct.
I have an Asus Zephyrus G15 gaming laptop.
It’s VERY fast. The market is much bigger for them so the R&D is bigger and the prices are cheaper.
Just be careful and check reviews of screens. And expect you’ll need an i1Display pro type of screen calibrator if you want to do any kind of color accurate work.
I don’t believe GPU memory usage is much of a factor as long as you have 2GB or more. I have 8 and never see substantial memory usage. I said tensor cores but I really can’t say whether it’s tensor cores or CUDA cores.
I’m sure generation by generation there are substantial improvements as well.
Tensor cores are a Nvidia thing , and only since the rtx2000 series.
My gtx1060 without any tensor cores works fine on deep prime :).
Cuda cores are calculations , tensor cores are ai acceleration.
If DxO would ever benefit from ai acceleration it would be a nice thing to implement it. Most Intel CPU cores have hardware for it for the last generations , heir arc gpus have ai acceleration cores (xmx engines ) , apple m1/m2 chips have ai acceleration hardware…
I just don’t know about amd gpus , i just don’t know then (i now realize )
I have been using DeepPRIME XD quite successfully with my NVidia GTX 1050ti on my Windows 10 machine. I’m not at home right now so I can’t tell you the driver version.
The GTX 1060 is a bit outdated. It is (was) one of those graphic adapters, that don’t get too hot. Even the fans seem to shut down when in idle mode.
There Which Video Card? - #30 by Wolfgang I listed some newer models & specs.
A faster one is something like a RTX 3070.
But they are exorbitant expensive, produce more heat, need a strong power supply and in my case also a new motherboard w/ PCIe 4.0 – so not for me.
I have a 35watt rtx 3050ti laptop chip , that in games at least seems to match my older desktop gtx1060. Shows the gains that have been made at the lower end :).
Ting why i ask i PL6 deep prime XD generates artifacts for me on the 3050ti while the normal deep prime seems to work fine like it always has in PL5.4.
Deep prime XD works fine on the same laptop when set to cpu only mode. But… Even the monster laptop cpu i have takes just under 2 minutes for a file
New NVIDIA Studio Driver today, which lists DXO Photo in the release notes as new supported application.
My DeepPrimeXD artefacts in GPU mode seem to be gone now - only quick testing so far. But before I didn’t manage a single export without weird artefacts, and the first image I try now is clean.
Windows 11 laptop with RTX 3050ti 4gb, 35watt model (the lowest wattage model) (beefy Intel 12th gen CPU though).
24mp image exported in 12 seconds with DeepPrimeXD, 4 seconds with DeepPrime.
I just downloaded the latest Nvidia driver for my GTX 1050ti. I was shocked to see that my average processing time for DeepPRIME XD fell from around 50 seconds to around 35 seconds. That around a 30% improvement. The speed of processing with DeepPRIME went from around 21 seconds to 15 seconds, also a 30% improvement. That seems too good to be true, but your results after downloading you latest driver seem to indicate a similar 30% improvement when processing those two files with DPXD. This is very exciting, but how is it possible that we both are seeing such a huge processing improvement? I wonder if the new drivers for other Nvidia cards are also having a similar effect.
Hi Mark,
it was @jorismak who found out about the new driver and details after coming up with questions … Otherwise I wouldn’t have checked the driver then in use and the developing times for DP XD …
Wolfgang
(had always OpenCL enabled, the same like in PL5 – no idea if the former driver didn’t make use of it)
For the past year or so (beginning with driver version 5.11), NVIDIA has been working to implement a vastly-improved OpenCL/CLANG compiler internal to the drivers. This was an end-user opt-in post-install configuration choice for “power users” until the most recent release of the driver packages.
There is some (minor) information available on this in the NVIDIA Driver release notes, section 2.5.3.
The new internal compilers also improve/speeds up CUDA operations.
On other apps (Davinci Resolve Studio) I have seen a 25-40% improvement in GPU-specific rendering and video stream manipulation work.
Beginning this week with the release of driver pack 5.22.30, the use of the new compiler became standard for Windows and Linux users. There is no indication when/if the improved driver sets will be available for Mac.