Iray Starter Scene: Post Your Benchmarks!

14344454749

Comments

  • RayDAntRayDAnt Posts: 1,135
    Robinson said:

    Edit: Confirmed, Optix Prime setting is irrelevant, i.e. it was slower because that was the first bench I did (probably cached shader compilation or similar).  Running again I get:

    GPU Only 1 x Geforce RTX 2070  = 1 minutes 27.3 seconds (Optix Prime on)

    Here's what it probably was:

    Iray RTX 2019.1.1, build 317500.2554

    Fixed Bugs

    • Iray Photoreal
      • RTX startup time reduced when using the NVIDIA R430 GA3 (or newer) driver. Then only the very first rendering after an Iray- or NVIDIA-driver update will need some additional startup time.

    Ie. just the very first render after either an Iray update or an nvidia driver update will take slightly longer with a Turing card.

  • RobinsonRobinson Posts: 751
    RayDAnt said:
    Ie. just the very first render after either an Iray update or an nvidia driver update will take slightly longer with a Turing card.

    Sounds about right.  Thanks for that. 

     

  • Robinson said:

    As posted in another thread test scene my previous times were:

    GPU Only 1 x Geforce RTX 2070  = 2 minutes 21.14 seconds (Optix Prime off)
    GPU Only 1 x Geforce RTX 2070  = 1 minutes 49.11 seconds (Optix Prime on)
    GPU Only 1 x Geforce RTX 2070 + 1 x Geforce GTX 970 = 1 minutes 45.78 seconds (Optix Prime off)
    GPU Only 1 x Geforce RTX 2070 + 1 x Geforce GTX 970 = 1 minutes 19.47 seconds (Optix Prime on)

    , with 4.12 (2070 only) they were:

    GPU Only 1 x Geforce RTX 2070  = 1 minutes 29.28 seconds (Optix Prime off)
    GPU Only 1 x Geforce RTX 2070  = 1 minutes 34.58 seconds (Optix Prime on)

    Don't turn on Optix Prime if you've got an RTX GPU (I'm wondering if those 5 seconds were due to some caching or similar, hence Optix Prime setting may be irrelevant).  Also 4.12 beta is around 63% faster for me.

     

    Edit: Confirmed, Optix Prime setting is irrelevant, i.e. it was slower because that was the first bench I did (probably cached shader compilation or similar).  Running again I get:

    GPU Only 1 x Geforce RTX 2070  = 1 minutes 27.3 seconds (Optix Prime on)

    Thank you for these. So:

    Without RT cores before v4.12: GPU Only 1 x Geforce RTX 2070  = 1 minutes 49.11 seconds (Optix Prime on)

    With RT cores in v4.12:  GPU Only 1 x Geforce RTX 2070  = 1 minutes 27.3 seconds (Optix Prime on)

    ~12% improvement with dedicated RT hardware? am I missing something. Some guys were saying 2x-10x improvement.

     

  • timon630timon630 Posts: 37

     

     

    techofied said:

    ~12% improvement with dedicated RT hardware? am I missing something. Some guys were saying 2x-10x improvement.

    Problems with arithmetic? It all depends on the specific scene, but the average gain seems to be 20% (test)

  • RayDAntRayDAnt Posts: 1,135
    techofied said:
    Robinson said:

    As posted in another thread test scene my previous times were:

    GPU Only 1 x Geforce RTX 2070  = 2 minutes 21.14 seconds (Optix Prime off)
    GPU Only 1 x Geforce RTX 2070  = 1 minutes 49.11 seconds (Optix Prime on)
    GPU Only 1 x Geforce RTX 2070 + 1 x Geforce GTX 970 = 1 minutes 45.78 seconds (Optix Prime off)
    GPU Only 1 x Geforce RTX 2070 + 1 x Geforce GTX 970 = 1 minutes 19.47 seconds (Optix Prime on)

    , with 4.12 (2070 only) they were:

    GPU Only 1 x Geforce RTX 2070  = 1 minutes 29.28 seconds (Optix Prime off)
    GPU Only 1 x Geforce RTX 2070  = 1 minutes 34.58 seconds (Optix Prime on)

    Don't turn on Optix Prime if you've got an RTX GPU (I'm wondering if those 5 seconds were due to some caching or similar, hence Optix Prime setting may be irrelevant).  Also 4.12 beta is around 63% faster for me.

     

    Edit: Confirmed, Optix Prime setting is irrelevant, i.e. it was slower because that was the first bench I did (probably cached shader compilation or similar).  Running again I get:

    GPU Only 1 x Geforce RTX 2070  = 1 minutes 27.3 seconds (Optix Prime on)

    Thank you for these. So:

    Without RT cores before v4.12: GPU Only 1 x Geforce RTX 2070  = 1 minutes 49.11 seconds (Optix Prime on)

    With RT cores in v4.12:  GPU Only 1 x Geforce RTX 2070  = 1 minutes 27.3 seconds (Optix Prime on)

    ~12% improvement with dedicated RT hardware? am I missing something. Some guys were saying 2x-10x improvement.

     

    Yeah, to echo others it depends ENTIRELY on the specific composition of each individual scene how much of a boost there is. So far the best rule seems to be that the larger (in terms of VRAM memory footprint/objects complexity) the greater the boost. Unfortunately all the benchmarking scenes currently around are optimized for small VRAM footprints (so that they can be rendered on all levels of hardware.) Making them bad candidates for showing how much of a difference dedicated RTCores can make.

  • RobinsonRobinson Posts: 751
    edited August 2019
    techofied said:
    Without RT cores before v4.12: GPU Only 1 x Geforce RTX 2070  = 1 minutes 49.11 seconds (Optix Prime on)

    With RT cores in v4.12:  GPU Only 1 x Geforce RTX 2070  = 1 minutes 27.3 seconds (Optix Prime on)

    ~12% improvement with dedicated RT hardware? am I missing something. Some guys were saying 2x-10x improvement.

     

    I wasn't comparing the right figures.  I should have chosen the lowest non-RT figure against the 4.12 one.  You're right, it's not much of an improvement.  I think this blog explains it quite well:

     

    In many scenes, we found that ray intersection comprises only 20% of the total work being performed by rendering. This is a very important point. Even if the new RT Cores were to make ray intersection infinitely fast so that it takes no time, 80% of the work still remains in that scene. So a 100 second render would still take 80 seconds with RTX acceleration, giving a speed-up of 1.25x (25%). Of course, ray intersection is not free with RTX, just faster, so the speed-up would be lower than this but this is the hypothetical upper limit.

    If you have a scene where 60% of the work is ray intersection you will naturally see a much more significant speed-up. In that case on a 100 second render, with an infinitely fast ray intersector you still have 40 seconds of rendering, giving a speed-up of 2.5x (250%) at the hypothetical upper limit. In general we have found RTX provides the greatest benefit in very complex scenes with millions of triangles and also scenes that heavily exploit instancing.

    e.g. some of the scenes tested there have a x2 or x3 improvement.  Others only order 10-20%.  The bench scene in this thread is no longer fit for purpose.  Or perhaps it is and Daz users don't generally deal with the kind of scene complexities that would result in a huge speedup.  Anyway it doesn't matter so much to me.  I'm just happy this stuff is getting out there.  Hardware manufacturers will iterate on it and make improvements.  It's been fun playing with RTX in Unreal Engine too.

     

    Post edited by Robinson on
  • Takeo.KenseiTakeo.Kensei Posts: 1,303

    I made a new benchmark scene that will show the power of RTX cards here https://www.daz3d.com/forums/discussion/344451/rtx-benchmark-thread-show-me-the-power#latest

    No need for a big scene

  • Late to the party, but ran the original scene on a seven-year-old Dell T5600 with 2x 1050Ti and dual Xeon 8-core with Daz 4.11. Also did a single run on a two-year-old iMac. Both have 32GB of memory, but the clock speed for the CPU is about twice as fast on the iMac.

    2x 1050Ti with 16-core CPU assist: 3min 6sec, 3min 3sec, 3min 4sec, 3min 6sec. Average is about 3min 5sec.

    Quad-core i7 iMac, CPU only: 45min 5sec. (I have no interest in reruns on the iMac without Nvidia cards.)

    I hate to render on the iMac, but setting up a scene is more comfortable for me on the iMac than the Dell. The Dell was a frustration-purchase because of Apple's Nvidia=Never OSX updates.

  • RayDAntRayDAnt Posts: 1,135
    edited August 2019
    tchas1949 said:

    I hate to render on the iMac, but setting up a scene is more comfortable for me on the iMac than the Dell. The Dell was a frustration-purchase because of Apple's Nvidia=Never OSX updates.

    You should check out Iray Server. It isn't free past the first month, but it would let you both have your cake and eat it too (scene setup on your iMac. Scene renders on any/all Windows or Linux boxes in your house while SIMULTANEOUSLY doing the next scene's setup back on your iMac.)

    Post edited by RayDAnt on
  • Here is a blast from the past.

    I am using a 4 core/8 thread Xeon, E3-1245, 3.4 GHz, 16GB RAM with an Nvidia Quadro 2000. The machine dates to Aug 2013 when it became my work PC, then after it was retired in January the company offered it to me instead of throwing it away.

    Benchmark times: Full scene with graphics card and optimization: 50 mins 56 .82 secs to get to 5000 iterations and 99.99%, so almost there. It was at 90% after 3 mins 46.

    Have not seen times like this since page 4 of this thread. Looks like a secondhand 1060 is under £100, so it may be worth getting for the machine. Dunno, it is a fair bit of money.

    Regards,

    Richard.

     

    Now I have fitted a GTX 1060 6Gb to the same machine, working on GPU only, the benchmark dropped to 4 mins 14 secs. Have to say I'm a little happy.

    Regards,

    Richard

  • brainmuffinbrainmuffin Posts: 1,205

    I have a 2010 i5 iMac. This ought to be interesting. I'll let you know.

  • JD_MortalJD_Mortal Posts: 760
    edited September 2019

    Just a note...

    I was testing the new 4.12-Beta version and the renderings are finishing before the 5000 iteration mark. Seems the new convergance formula is saying that the renders are complete, a lot faster.

    You will have to manually adjust that value, to put it above the default setting, or your benchmarks will not be accurate. (From card to card, or from 4.11 to 4.12-beta, if you are comparing notes.)

    I will post more benchmarks once I have run one with the correct values for 4.12-beta.

    Post edited by JD_Mortal on
  • outrider42outrider42 Posts: 3,679

    Some people have made new benchmark threads that cap at a set iteration count.

  • RayDAntRayDAnt Posts: 1,135
    edited September 2019
    JD_Mortal said:

    Just a note...

    I was testing the new 4.12-Beta version and the renderings are finishing before the 5000 iteration mark. Seems the new convergance formula is saying that the renders are complete, a lot faster.

    You will have to manually adjust that value, to put it above the default setting, or your benchmarks will not be accurate. (From card to card, or from 4.11 to 4.12-beta, if you are comparing notes.)

    I will post more benchmarks once I have run one with the correct values for 4.12-beta.

    Fwiw 4.10's convergence algorithm routinely did this as well. It's one of the main reasons I ended up developing an alternative scene to use for benchmarking going forward. Because of the inherent lack of consistency.

    ETA:Having said that, I'd actually advise AGAINST attempting to tweak Sickleyield's original scene toward more consistency this late in the game. Doing so will essentially mean having two versions of a virtually visually identical benchmarking scene floating around. Which is a guaranteed recipe for confusion down the line (hence why I didn't just tweak an existing benchmarking scene for the newer benchmarking thread I started.)

    SY's scene has already been consistently inconsistent about this for years now.

    Post edited by RayDAnt on
  • JD_MortalJD_Mortal Posts: 760
    edited September 2019

    Yes, I agree with the issues that the scene has, but it is what I keep using as my baseline. I'll try a newer one next...

    Finally, after a long wait, Titan-V cards now work on BOTH 4.11 and 4.12-Beta (Public). Honestly though, they may have worked months ago, but I have not had time to test them out.

    For those looking for Titan-V benchmarks, here are some, in comparison to the Titan-Xp Collectors Edition cards.

    Daz3D V4.11

    Titan-Xp Col Edi, No OptiX (149.342sec), With OptiX (105.732sec)

    Titan-V, No OptiX (78.362sec), With OptiX (63.192sec)

    Daz3D V4.12-Beta (Public) * Seems that OptiX is the only choice... Unselecting it, does not stop it. Render test adjusted to 99%, so it actually hit 5000 iterations.

    Titan-Xp Col Edi, No OptiX (103.636sec), With OptiX (103.636sec)

    Titan-V, No OptiX (63.132sec), With OptiX (63.132sec)

    I tried TCC mode, for both versions, but the numbers were WORSE than WDDM mode, but only by a few seconds. Unlike the suggestion in the LOGs will indicate. The result is not more efficient, at all, in any variation of rendering, on either version. Though, it should be. However, in all fairness, this was only done with ONE card in TCC mode, at a time. I think it may have gains if there are multiple cards in TCC mode, which I will try after I get both my Titan-V's setup with at-least one Titan-Xp Collectors Edition card. I can't fit two of the Collectors Editions cards in, with the Titan-V's, because the cards have an annoying cosmedic guard around the fan which can't be removed unless I use a hack-saw. (That may be an option.) However, I do have a normal Titan-X (The non-pascal version), which I can throw in there, for more comparison. All cards can be put in TCC mode, but I need one in WDDM mode, or I can't see anything. My CPU does not have an internal video-card, being the older/new i9-7980XE.

    Next is Dual Titan-V's... But I have to flip my water cooler around, if I want to fit all four cards in the tower. :)

    UPDATE: I managed to fit all four cards in the machine! Just took some cutting and bending of parts. (Seriously)

    Daz3D V4.12-Beta (Public) * Seems that OptiX is the only choice... Unselecting it, does not stop it. Render test adjusted to 99%, so it actually hit 5000 iterations.

    Titan-V [x2] + Titan-Xp Col Edi [x2] + CPU, No OptiX (21.558sec), With OptiX (21.541sec), Optomized for speed and OptiX (21.119sec)

    P.S. I shaved a lot more time off the Titan-Xp Collectors edition, if I set the optomised settings for "SPEED" and not "MEMORY". From 149.342sec down to 113.425sec. The Titan-V was faster too, but it only went from 78.362sec down to 62.981sec.

    Final note...

    Both of the above cards, together, best times with OptiX, was 39.970sec, on both versions of Daz3D.

    Post edited by JD_Mortal on
  • nonesuch00nonesuch00 Posts: 18,142

    Yes, many of my renders, especially ones with no water or clouds or such amorphous substances are now mostly finishing before the 2000 max iterations I set for them. I set convergence at 95% / 1.0 and CPU renders on so (relatively) fast on this 7 year old laptop now.

    I think it's because it looks like the convergence criteria in the SDK or something has changed. Eg, the render I'm doing now is at 56% after only 1 hour 36 minutes 16 seconds but has done only 138 of 2000 allowed iterations. It can run as long as it takes to do 2000 iterations.   

  • Hi,

     

    I decided to check 4.12 making a test with a heavy load: image size 4096x2048px, max samples 10000 iteractions, max time 14400 secs, rendered quality 5.0, no denoiser.

    Hardware: RTX 2080ti (dedicated for render), Ryzen 7 1700, 64gb 2400mhz RAM, Daz installed and scene saved on SSD WD Blue 1TB sata, Windows 10 on SSD Adata XPG Gammix S11 Pro 512GB M.2 NVMe. GTX 970 (for Windows).

     

    Running on 4.11:
    2019-09-01 23:11:32.275 Total Rendering Time: 2 hours 37 minutes 3.8 seconds

    Running on 4.12:
    2019-09-02 05:27:05.118 Total Rendering Time: 2 hours 3 minutes 25.85 seconds

    I look for the Iray version on the log file, I found:
    2019-09-02 03:21:15.039 Iray [INFO] - API:MISC ::   0.0   API    misc info : Iray RTX 2019.1.3, build 317500.3714, 19 Jul 2019, nt-x86-64

     

    Somebody mention that the first render is slower and the others are faster, so I rendered again but didn't see any difference:
    2019-09-02 07:47:01.392 Total Rendering Time: 2 hours 3 minutes 22.49 seconds

     

     

    In this case I had little more than 30 percent of improvement on the render time. I didn't notice any different on the bright between the two versions, what happened between 4.10 and 4.11, don't know if there will be any disadvantage to keep using 4.12 Beta.

     

     

     

  • Hurdy3DHurdy3D Posts: 1,047
    edited September 2019

    I got a new PC with an Ryzen 7 3700x and a MSI GeForce RTX 2080 Ti GAMING X TRIO

     

    here are my results. First is Daz 4.11 and second is Daz 4.12.67 Beta with the IrayTestSceneStarterEssentialsOnly scene

    Daz 4.11/4.12 GPU Optix
    1 minutes 8.94 seconds/56.28 seconds

    Daz 4.11/4.12 GPU
    1 minutes 21.65 seconds/53.84 seconds

    Daz 4.11/4.12 GPU CPU Optix
    1 minutes 7.26 seconds/57.5 seconds

    Daz 4.11/4.12 GPU CPU
    1 minutes 21.37 seconds/54.74 seconds

     

    one question, who do I know if RTX is used? Is there something in the log?

     

    Post edited by Hurdy3D on
  • FlortaleFlortale Posts: 611
    edited September 2019

    Which is faster to render with IRAY?

    i7-9700k w/32gb memory

    or

    Nvidia GTX 980 4gb

    Post edited by Flortale on
  • RayDAntRayDAnt Posts: 1,135
    edited September 2019
    Flortale said:

    Which is faster to render with IRAY?

    i7-9700k w/32gb memory

    or

    Nvidia GTX 980 4gb

    From what I've read a i7-9700K is about 10% more powerful than an i7-8700K and a GTX 980 about 87% as powerful as a GTX 980Ti. Based on the following performance numbers (found here):
    i7-8700K: 00.414 ips
    GTX 980Ti: 2.094 ips

    That would put the i7-9700K and GTX 980 at roughly:
    i7-9700K: 00.455 ips
    GTX 980: 01.822 ips

    Or the GTX 980 at slightly more than 3x faster than the i7-9700K.

    As a rule, CPU Iray performance pales in comparison to GPU performance across multiple generations. The only reason why CPU might be superior over GPU is if VRAM size is an issue.

    Post edited by RayDAnt on
  • outrider42outrider42 Posts: 3,679
    I would say that 4GB of VRAM might indeed be an issue. A lot of people here, including me, would feel pretty limited by that. I used to a 970 with 4GB, and I felt that way. If you exceed 4GB, then the 980 will sit idle and useless. And if that happens, well the i7 would win by default.

    I'm guessing Flortale has a chance to score a cheap 980, but unless they are really confident that 4GB is enough for them, I would personally pass unless this deal is ridiculously good. Plus a 980 is from way back in 2014, that's pretty old for computer hardware, especially a GPU. A newer 1070 would be faster and double the VRAM to 8GB, and should not cost much more if you look around. A new 1660ti would be faster still, though it drops to 6GB VRAM, that's still 50% more than a 980.
  • RayDAnt said:
    Flortale said:

    Which is faster to render with IRAY?

    i7-9700k w/32gb memory

    or

    Nvidia GTX 980 4gb

    From what I've read a i7-9700K is about 10% more powerful than an i7-8700K and a GTX 980 about 87% as powerful as a GTX 980Ti. Based on the following performance numbers (found here):
    i7-8700K: 00.414 ips
    GTX 980Ti: 2.094 ips

    That would put the i7-9700K and GTX 980 at roughly:
    i7-9700K: 00.455 ips
    GTX 980: 01.822 ips

    Or the GTX 980 at slightly more than 3x faster than the i7-9700K.

    As a rule, CPU Iray performance pales in comparison to GPU performance across multiple generations. The only reason why CPU might be superior over GPU is if VRAM size is an issue.

    Thank you!  I wish this forum had a thank you button feature.

  • RTX 2070 Super 

    DS4.12 GPU (Optix Prime) = 1 minute 11 seconds

    DS4.12 GPU (NO Optix Prime) = 1 minute 07 seconds

     

  • DigiDotz said:

    Amazon Web Services G2.2xlarge instance - Grid K520 8gb(4GB used as dual GPU), 8 vcpu's (from Xeon E5-2670 cpu) with 15GB ram

    Render time disappointing - 21 minutes

    This was the lowest powered GPU instance on AWS  -Still I hoped for better, maybe it wasnt using all cuda cores? will re-check

    EDIT Ok -so you do only get 1 GPU from the K520 = a GK104 GPU with 1536 cuda cores

    @DigitDotz how much did that cost? 

  • inacentaurinacentaur Posts: 109
    edited September 2019

    Intel Core i7-6700K @ 4 GHz
    GeForce GTX980 x2 (Kingpin)
    Total Rendering Time: 3 minutes 1.44 seconds (Two Daz instances open) 

    2 minutes 54.24 seconds (closed the other Daz instance) 

    Daz 4.11 Pro

    Post edited by inacentaur on
  • inacentaurinacentaur Posts: 109
    edited September 2019

    Intel Core i7-6700k @ 4 GHz

    Single GeForce 1080 TI Aorus

    Total Rendering Time: 2 minutes 58.13 seconds

     

    Aorus OC mode 

    2 minutes 54.33 seconds

     

    2x GeForce 1080Ti Aorus

    2019-09-20 23:41:40.798 Total Rendering Time: 1 minutes 39.60 seconds

    Post edited by inacentaur on
  • Intel Core i7-6700k @ 4 GHz

    Single GeForce 1080 TI Aorus

    Total Rendering Time: 2 minutes 58.13 seconds

     

    Daz 4.11 Pro

  • inacentaurinacentaur Posts: 109
    edited September 2019

    Intel Core i7-6700k @ 4 GHz

    Single GeForce 1080 TI Aorus

    Total Rendering Time: 2 minutes 58.13 seconds

     

    2x GeForce 1080 Ti (no SLI) 

    2019-09-20 23:41:40.798 Total Rendering Time: 1 minutes 39.60 seconds

    with 1 SLI bridge: 2019-09-21 00:01:03.278 Total Rendering Time: 1 minutes 36.59 seconds

    Daz 4.11 Pro

    Post edited by inacentaur on
  • TheKDTheKD Posts: 2,691
    edited September 2019

    msi gaming trio x 2080 super with optix on in the beta daz studio

    2019-09-21 09:27:57.616 Total Rendering Time: 52.7 seconds

    with optix off

    2019-09-21 09:30:57.447 Total Rendering Time: 50.97 seconds

    kinda weird optix off rendered faster for this one lol

    Post edited by TheKD on
  • RayDAntRayDAnt Posts: 1,135
    TheKD said:

    msi gaming trio x 2080 super with optix on in the beta daz studio

    2019-09-21 09:27:57.616 Total Rendering Time: 52.7 seconds

    with optix off

    2019-09-21 09:30:57.447 Total Rendering Time: 50.97 seconds

    kinda weird optix off rendered faster for this one lol

    If you dissect the log file you will discover that the OptiX Prime acceleration toggle in 4.12 Beta is a placebo if you are rendering with a Turing GPU. OptiX (non-Prime) and consequently RTCores ALWAYS get used for raytracing rather than OptiX Prime. The only difference the toggle makes is that it generates non-critical error messages in the log file about it no longer meaning anything.

Sign In or Register to comment.