Jump to content

Is Nvidia's Tensor and Ray Tracing cores gimping conventional performance?

atavax

Has been any talk from knowledgeable people about like, If Nvidia is losing any significant performance pushing these two features?

Link to comment
Share on other sites

Link to post
Share on other sites

The cores themselves are taking up die space and power budget which technically could be pushed towards more cuda cores, but the binning selection for that many cuda cores would be ludicrous.

 

Not to mention, RTX cores and tensor cores are actually extremely efficient at what they do.

I WILL find your ITX build thread, and I WILL recommend the SIlverstone Sugo SG13B

 

Primary PC:

i7 8086k - EVGA Z370 Classified K - G.Skill Trident Z RGB - WD SN750 - Jedi Order Titan Xp - Hyper 212 Black (with RGB Riing flair) - EVGA G3 650W - dual booting Windows 10 and Linux - Black and green theme, Razer brainwashed me.

Draws 400 watts under max load, for reference.

 

How many watts do I needATX 3.0 & PCIe 5.0 spec, PSU misconceptions, protections explainedgroup reg is bad

Link to comment
Share on other sites

Link to post
Share on other sites

49 minutes ago, Fasauceome said:

The cores themselves are taking up die space and power budget which technically could be pushed towards more cuda cores, but the binning selection for that many cuda cores would be ludicrous.

Without specifics, I'd hope the tensor/RT parts would be idle when not in use, and essentially take no power. Silicon area is a manufacturing cost. For the same core count with and without these extra features, I'd not expect any notable performance difference.

 

There might be small factors that could swing it either way. It will spread out the heat production and may make cooling a bit easier, in a similar way some older Intel CPUs often run much cooler than AMD CPUs on the same cooling, even though AMD CPUs used less power. The smaller process condensed the heat output making it more difficult to cool. However, you could argue a bigger size might increase internal latencies. I don't feel this is a problem in practical implementations, but a more spread out die could have higher latency overall compared to a more compact one, which would be taken into consideration at design step.

 

If we go the other way, and ask what if we put that used die area into more "non-RTX" stuff, what could we gain? Link below claims 22% increase in area requirement for a TPC from adding those features. Note TPCs aren't the only things in a GPU, there's other parts like L2 cache and external IO that will dilute the RTX cost. Also we have to consider if you increase the core compute, you need to scale VRAM bandwidth if you want perf to scale better. Not doing so would reduce the benefit from the cores. Likewise consider power budget. More cores without increasing power budget will result in much less improvement than might be expected. Take all the above into consideration, dropping RTX stuff to allow more basic stuff probably would make a pretty insignificant difference in performance unless you start scaling everything else to go with it. 

 

https://www.techpowerup.com/254452/nvidia-rtx-logic-increases-tpc-area-by-22-compared-to-non-rtx-turing

Main system: i9-7980XE, Asus X299 TUF mark 2, Noctua D15, Corsair Vengeance Pro 3200 3x 16GB 2R, RTX 3070, NZXT E850, GameMax Abyss, Samsung 980 Pro 2TB, Acer Predator XB241YU 24" 1440p 144Hz G-Sync + HP LP2475w 24" 1200p 60Hz wide gamut
Gaming laptop: Lenovo Legion 5, 5800H, RTX 3070, Kingston DDR4 3200C22 2x16GB 2Rx8, Kingston Fury Renegade 1TB + Crucial P1 1TB SSD, 165 Hz IPS 1080p G-Sync Compatible

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×