General GPU/testing discussion from benchmark thread
This discussion has been closed.
Adding to Cart…
Licensing Agreement | Terms of Service | Privacy Policy | EULA
© 2025 Daz Productions Inc. All Rights Reserved.You currently have no notifications.
Licensing Agreement | Terms of Service | Privacy Policy | EULA
© 2025 Daz Productions Inc. All Rights Reserved.
Comments
So it sounds like there's fairly general consensus that when someone posts an Iray log for a render that says "Total Rendering Time: 3 minutes 33.27 seconds" it's "common sense" that the user wasn't running anything that might affect that, and you can fully rely upon and use the 3:33.27 time to compare with another render on another system? And it's actually accurate to 270 milliseconds for comparison purposes?
If that's what you believe, or anything close to that, then I think you are grossly mistaken. And clearly there's nothing I can say to change that belief, so I'll stop trying.
You can never fully rely on any individual result. Look at the big picture and you will surely be able to make conclusions. That's all that counts. Far more valuable than elaborate benchmark tables to me is when I get a few people say "I rendered this on 1080 TI and on 2080 TI and the render time was roughly halfed". Just a few separate claims like this should be enough to reasonably assume that a 2080 TI indeed renders 2x faster in Iray. That is the type of stuff I want to know. Not whether someone benched the same scene at 10 seconds more or less despite having the same hardware. As long as there aren't huge unexplainable differences, nobody cares.
"real" benchmarks are conducted on machines with very carefully audited configurations and procedures. The benchmarks in the thread here are more informal, and so need to be treated with a pinch or two of salt (especially as DS and iray versions change) but that doesn't stop them from being of some use in rough guidance.
Something else to consider I had two identical machines and got different times from each one as much as 1 min when I built them I split a 128 meg matched memory kit between them and all components bought at the same time and from one place
Plus, it's also reasonable to assume that people are not all using GPUs that are reference designs uising the founders edition cooler with stock clock speeds or identical computer towers...
This in itself will lead to variations in GPU temperature causing the clock speed to vary from card to card.
Plus there can be pretty big clock speed and boost clock variation between the same GPU made by the same manufacturer.
example:
GTX 1080 ti with 2 80mm fans in a cramped tower with no extra cooling fans will not be able to maintain anywhere near the full boost clock....
GTX 1080 ti Founders with the blower style cooling in a well designed tower that pulls air in the front and blows it out the back and no card in the slot directly under the GPU will be be able to maintain full boost all day and maybe even handle overclocking past the rated boost speed for extended periods.
In my Dell Workstation tower my Quadro P4000 and GTX 1080 ti both render at full boost speed all day and the computer fans never even get as loud as the air conditioner in my house.
I asked for people to post clock speeds before, because it can make a difference. However, it is not a huge difference in the scope of Iray unless there is a huge difference in clock speed, like more than 100 Mhz, a lot more. This mostly comes into play when looking at laptop models of GPUs. Pascal laptops had the very same chip as the desktop models, however they were vastly downclocked.
Clockspeed is fine because that generally covers the card's cooling as well. Cards that do not clock as well are usually given cheaper coolers and sold cheaper than the higher clocked varients. So the market takes care of itself here. The exception would Founder's Editions which typically price higher, but their clocks still reflect the blower fan they have as Pascal FE cards have lower clocks than others. And another exception would be if users do their own overclocking.
It is very easy to test the impact of clock speed. Simply turn your clock speed down and render. You can tune your cards with software like MSI Afterburner and EVGA PrecisionX. Both of these will work with any brand.
Other components have little impact on render speed. I can point you to benchmarks made with 1070s installed inside a Core 2 QUAD from over a decade ago which ran the SY Bench in practically identical times as a 1070 installed with much newer machines. It just doesn't matter. My own machine has a i5-4690K from 2014 which is only 4 cores and 4 threads. It is showing its age in other software, but for Iray my 1080tis match the speeds of users with brand new CPUs. In fact, I have sightly faster times (by a couple seconds) than some people with brand new machines. So again, it just proves that rest of the machine doesn't matter. Iray is not a video game, and having an "unbalanced" PC is not an issue, as long as that balance favors the GPU. It has been proven time and time again.
The Quadro P4000 is in a very weird space. It seems to be right in between the 1060 and 1070. With only 8GB of VRAM, it offers the same amount that the 1070 does, but the 1070 is much faster AND cheaper. The P4000 only has a couple advantages. You can buy a single slot version, so you can cram a bunch of them into a board, and being a Quadro means they can enable TCC (Tesla Compute Cluster). TCC mode allow you to get around the Windows 10 VRAM issue as all video output is disabled so the card becomes a pure compute device. This allows you to access the full 8GB capacity. It is also possible that TCC mode might enhance the rendering performance of the card. I saw a twitter post that showed a Titan V running Vray with TCC enabled VS disabled, and there was a large difference of speed in favor of TCC enabled. But I have never seen anybody test TCC with Iray. In order to do so, you need another GPU for video output, because again, TCC disables video output.
I would be very curious to know if TCC mode offers any benefit to Iray rendering, or if it can even work with Iray.
However, for the price the P4000 goes for, you could get a 1080ti, which would be much faster than two P4000's put together, and have more VRAM than the P4000 even with TCC enabled. There are certain programs that can benefit greatly by using Quadro, but Daz Studio and Iray are not really amung those, unless TCC mode actually makes a big difference.
It is should be obvious to run benchmarks without running other software. Its a benchmark.
Will be investigating this very subject with the Titan RTX in the next couple weeks. God willing and the creek don't rise, that is (literally - finally got the last bits needed to properly watercool this thing. The stock Titan RTX cooler, while undeniably pretty, I'm sorry to say is largely a joke performance-wise.)
ETA: Also, for what it's worth, TCC compatibility with Iray has been tested here as working (if buggily) at least once already. See this post from JD_Mortal.
Some Infos about the denoiser from Optix docs
So no tool directly avalable or useable to make new training data
Here's the course referenced in that, in case anyone wants to check it out (I'm a bit busy with other things at the moment.)
Just launched a brand new benchmarking thread Daz Studio Iray - Rendering Hardware Benchmarking to coincide with today's fully RTX compatible DS Beta release. I highly recommend checking it out. It's obviously still a work-in-progress. But the sooner I can get data/feedback from people about it the better/more useful I can make it!
Niiiice. There's no RTX on/off feature in the 4.12 beta right? Optix Prime needs to be on to use RT cores?
Don't know but I"ve got that in my logs
Makes sense I guess, they removed Optix Prime and rewrote it in Optix proper. Or... something?
Based on my current testing with the Ttitan RTX, there's no control for it (it "just works" tm) and selecting/deselecting Optix Prime Acceleration under Advanced Render Options makes no difference in render times whatsover. Rendering is also significantly faster in all scenes so far tested.
For me that doesn't make sense at all.
Optix contains Optix Prime for what I've seen
Kicking Optix Prime out means that for non RTX Owner, there is no acceleration ?
Optix runs fine on none RTX too. It's just that Optix Prime was a coding shortcut but couldn't support RT cores because reasons. Huge grain of salt though, I really don't know what the hell I'm talking about. This is what I remember someone a long time ago say about why Iray wouldn't support RT cores for a while.
I can't really test anything until I fix my CMOS battery or get a new board. Almost comical how a simple thing such as system time can screw up stuff.
Optix runs, but as I said, you don't have Prime acceleration for non RTX cards. Now I'm curious and I'd like to see some non RTX bench on this version
A CMOS battery shouldn't be a big problem ?
Boot. Press F1 to accept whatever is asked. Change date and system time and you should be able to use your computer
Or does it generate a bigger problem ?
Yeah the time is lagging behind. Have to set it manually over and over again in Windows
Have you try to hibernate the computer instead of shutting it down ? That may keep the date and time.
And if you have internet, Windows should be able to keep them synchronized then.
I know that by default Windows can't sync if the difference is greater than 2 days (not sure from memory)
Tried lots of things, but yeah let's not do silly bluejaunte tech support stuff here. Get benchin' you fools!
Had a crash at my first test. Just testing stability for the moment
Renders seem quicker
And I've read a post in "Who wants better animation too" thread where DAZ_Steve states that all Nvidia GPU should render quicker
* Edit : Load time seem longer though
On CPU only it now uses Embree instead of Optix Prime:
Which, according to the Iray changelog is supposed to result in significantly faster CPU rendering over Optix Prime. Haven't tested it out to verify for myself yet, though.
ETA: Gonna try it on my GTX 1050 laptop (as soon as I get it updated) and see what that tells me its using (whether OptiX Prime or something else.)
UPDATE: This is what I got:
So OptiX Prime is still being utilized on non-RTX cards.
Yeah... incorrect board time can be a huge PITA - epsecially if you don't realize that's what's at fault. All sorts of weird problems start happening
Some issues to keep in mind while playing around with this beta if you're a Turing owner (copied from the official Iray changelog thread):
FYI, those are notes for the Iray 2019 beta, all of the known issues have been resolved (AFAIK) in the final release.
I let it render while I ate supper and did the dishes. 653 iterations with 2 dressed and posed people as the scene in 50 minutes, that's not all that bad as far as cpu only rendering goes.
One thing I wonder about, with Embree being made by Intel, and here it is inside a Nvidia software...that might mean we have a chance of seeing Intel's future GPU support CUDA. Now that would be cool, you would have a genuine choice in the market, assuming the Intel cards are any good. Intel's GPU is quite a wild card, we really have no idea what they are doing. But Intel has the ability to really shake up things if they want to, given they are so much larger than AMD and Nvidia combined. How will they segment their cards? It may just be possible they attempt to do things outside the box to get attention to their new brand.
That's all for 2020, but I can't help but wonder.
"Oxford Library" from Daz Store shows off RTX-accelaration quite well for me. "Dome and Scene" lighting, "Camera 6", rendered to 95% convergence, 1280x720 output for quick test.
4.11:
CUDA device 0 (GeForce RTX 2080 Ti): 5810 iterations, 2.999s init, 907.709s render
4.12 beta
CUDA device 0 (GeForce RTX 2080 Ti): 5827 iterations, 1.934s init, 574.256s render
So... anybody have recommendations for particular Daz Studio compatible scenes/sets (either from the Daz Store or other sources) that are reeeeeaaally big/difficult to render? I'm looking for ways to study RTX acceleration more thoroughly, and I suspect there is a direct correllation between scene asset size/complexity and how much RTX/RTCores speedup the rendering process (hence why all the freely available benchmarking scenes - including this one just released by me - show such a little difference with it versus without.)