Jump to content

[Updated] Oxide responds to AotS Conspiracies, Maxwell Has No Native Support For DX12 Asynchronous Compute

EDIT: Just found out that it was already in OP. Silly me :P

'Fanboyism is stupid' - someone on this forum.

Be nice to each other boys and girls. And don't cheap out on a power supply.

Spoiler

CPU: Intel Core i7 4790K - 4.5 GHz | Motherboard: ASUS MAXIMUS VII HERO | RAM: 32GB Corsair Vengeance Pro DDR3 | SSD: Samsung 850 EVO - 500GB | GPU: MSI GTX 980 Ti Gaming 6GB | PSU: EVGA SuperNOVA 650 G2 | Case: NZXT Phantom 530 | Cooling: CRYORIG R1 Ultimate | Monitor: ASUS ROG Swift PG279Q | Peripherals: Corsair Vengeance K70 and Razer DeathAdder

 

Link to comment
Share on other sites

Link to post
Share on other sites

Again, these are all assumptions. You don't know if Pascal will beat out Arctic Islands.

So far it looks like AMD is winning DX12.

 

Considering Maxwell and Pascal were announced in at least 2010 and 2013, Its a safe bet that Nvidia will not Support Native DX12 Async Compute (out-of-sync to cpu draw calls) for at least 2-3 years unless Nvidia pump billions of reserve cash into a complete redesign that implements the proper hardware. They can do it, but will they? 

R9 3900XT | Tomahawk B550 | Ventus OC RTX 3090 | Photon 1050W | 32GB DDR4 | TUF GT501 Case | Vizio 4K 50'' HDR

 

Link to comment
Share on other sites

Link to post
Share on other sites

snip

designing Pascal without full DX12 hardware support can be quite catastrophic for nVidia; but .. I don't believe they'll do such thing

it was shown time and time again that the vast majority of the people buying computer hardware, that they don't have a fucking clue what they're buying - so, for them to do that, would be suicidal; that market share will turn on it's head so fast that it wouldn't know what hit them

so no, I don't believe nVidia to be stupid

Link to comment
Share on other sites

Link to post
Share on other sites

They want people to buy their cards and really just force them to - AMD put everything into GCN 1.2 - Tessellation, A-sync compute, and more whereas Nvidia likely anticipated the outcome and planned so that their cards get a shorter lifespam and the enthusiasts are forced to upgrade.

Yup. Nvidia gets you the best for the here and now and charges a premium for it. AMD goes cheap and has its tech age well without taking the performance crown which makes them look both cheap and uncompetitive. AMD needs to act and think like a business, not a charity.

Software Engineer for Suncorp (Australia), Computer Tech Enthusiast, Miami University Graduate, Nerd

Link to comment
Share on other sites

Link to post
Share on other sites

I generally agree with your post, but these two points stick out:

If people get the notion that AMD can handle DX12 better, AMD should gain market share both in the near future and long term, as people are looking forward to upcoming titles, utilizing DX12. So I very much doubt AMD will suffer from this. On the contrary.

NVidia hardware is not faster. Not in flops and not in DX12. NVidia's performance gains are due to multithreaded low overhead DX11 drivers. Those are rendered irrelevant in DX12, which is one of the biggest reasons we are seeing this shift.

Nvidia wins in flops, just not DP in consumer cards. Until the Fury X, Nvidia has had the flops lead since Kepler if not Fermi. And games don't use DP.

Software Engineer for Suncorp (Australia), Computer Tech Enthusiast, Miami University Graduate, Nerd

Link to comment
Share on other sites

Link to post
Share on other sites

Nvidia wins in flops, just not DP in consumer cards. Until the Fury X, Nvidia has had the flops lead since Kepler if not Fermi. And games don't use DP.

 

Nvidia hasnt bean in lead in FLOPS since pre fermi. Ever since GCN came out, Nvidia has LOST in flops, in every fucking card to card comparison. Stop spreading bullshit.

 

R9 380 

Pix rate: 31 GPixels/s

Tex Rate: 109 GTexel/s

FLOPS: 3,476 GFLOPS

 

GTX 960

Pix rate: 36.1 GPixels/s

Tex Rate: 72.1 GTexel/s

FLOPS: 2,308 GFLOPS

-------------------------------------------------

 

R9 390

Pix rate: 64 GPixels/s

Tex Rate: 160 GTexel/s

FLOPS: 5,120 GFLOPS

 

 

970

Pix rate: 58.8 GPixels/s

Tex Rate: 109 GTexel/s

FLOPS: 3,494 GFLOPS

-------------------------------------------------

 

R9 390X

Pix rate: 67.2 GPixels/s

Tex Rate: 185 GTexel/s

FLOPS: 5,914 GFLOPS

 

 

GTX 980

Pix rate: 71.2 GPixels/s

Tex Rate: 144 GTexel/s

FLOPS: 4,616 GFLOPS

-------------------------------------------

 

R9 Fury

Pix rate: 64 GPixels/s

Tex Rate: 224 GTexel/s

FLOPS: 7,168 GFLOPS

-------------------------------------------

 

R9 Fury X

Pix rate: 67.2 GPixels/s

Tex Rate: 269 GTexel/s

FLOPS: 8,602 GFLOPS

 

 

980Ti

Pix rate: 96 GPixels/s

Tex Rate: 176 GTexel/s

FLOPS: 5,632 GFLOPS

 

Titan X

Pix rate: 96 GPixels/s

Tex Rate: 192 GTexel/s

FLOPS: 6,144 GFLOPS

------------------------------------------

 

 

Now for the past

GTX 680

Pix rate: 32.2 GPixels/s

Tex Rate: 129 GTexel/s

FLOPS: 3,090 GFLOPS

 

HD 7970GHz Edt.

Pix rate: 32 GPixels/s

Tex Rate: 128 GTexel/s

FLOPS: 4,096 GFLOPS

 

 

Ever since AMD dropped VLIW in favor of GCN, NIVIDIA HAS NEVER. ON ANY ACCOUNT. EVER. CAUGHT UP TO THEM IN FLOPS

Link to comment
Share on other sites

Link to post
Share on other sites

So NVIDIA is working on a driver update for async compute. That explains why the Ark Survival DX12 patch was delayed at the last moment? They said it was a driver thing.

well drivers are only going to do so much.... if the hardware scheduler is not up to the task. No matter how great of a driver, you will lose to the hardware solution with the better scheduler...

 

Unfortunatly, some things, you just cannot get over by simple driver magic.

Link to comment
Share on other sites

Link to post
Share on other sites

So NVIDIA is working on a driver update for async compute. That explains why the Ark Survival DX12 patch was delayed at the last moment? They said it was a driver thing.

Not really - drivers cannot make A-sync compute magically appear. Think of it as hyperthreading.

"It allows jobs to be cycled into the GPU during dormant phases. In can vaguely be thought of as the GPU equivalent of hyper threading. Like hyper threading, it really depends on the workload and GPU architecture for as to how important this is. In this case, it is used for performance. I can’t divulge too many details, but GCN can cycle in work from an ACE incredibly efficiently. Maxwell’s schedular has no analog just as a non hyper-threaded CPU has no analog feature to a hyper threaded one."

If a driver could fix it - then I could simply flash my CPU's software to that of a 4790 and have an i7 at half the price

Archangel (Desktop) CPU: i5 4590 GPU:Asus R9 280  3GB RAM:HyperX Beast 2x4GBPSU:SeaSonic S12G 750W Mobo:GA-H97m-HD3 Case:CM Silencio 650 Storage:1 TB WD Red
Celestial (Laptop 1) CPU:i7 4720HQ GPU:GTX 860M 4GB RAM:2x4GB SK Hynix DDR3Storage: 250GB 850 EVO Model:Lenovo Y50-70
Seraph (Laptop 2) CPU:i7 6700HQ GPU:GTX 970M 3GB RAM:2x8GB DDR4Storage: 256GB Samsung 951 + 1TB Toshiba HDD Model:Asus GL502VT

Windows 10 is now MSX! - http://linustechtips.com/main/topic/440190-can-we-start-calling-windows-10/page-6

Link to comment
Share on other sites

Link to post
Share on other sites

So NVIDIA is working on a driver update for async compute. That explains why the Ark Survival DX12 patch was delayed at the last moment? They said it was a driver thing.

https://docs.unrealengine.com/latest/INT/Programming/Rendering/ShaderDevelopment/AsyncCompute/index.html .... UE4 will use assync compute... so looks like .... they wont be able to fix... cuz its a architecture issue... AoTS dev that works with DX 12 said that and i believe him...he even tried to run on a 980ti... and the 980ti got rekt pretty much...

AMD Rig - (Upgraded): FX 8320 @ 4.8 Ghz, Corsair H100i GTX, ROG Crosshair V Formula, Ghz, 16 GB 1866 Mhz Ram, Msi R9 280x Gaming 3G @ 1150 Mhz, Samsung 850 Evo 250 GB, Win 10 Home

(My first Intel + Nvidia experience  - recently bought ) : MSI GT72S Dominator Pro G ( i7 6820HK, 16 GB RAM, 980M SLI, GSync, 1080p , 2x128 GB SSD + 1TB HDD... FeelsGoodMan

Link to comment
Share on other sites

Link to post
Share on other sites

Not really - drivers cannot make A-sync compute magically appear. Think of it as hyperthreading.

"It allows jobs to be cycled into the GPU during dormant phases. In can vaguely be thought of as the GPU equivalent of hyper threading. Like hyper threading, it really depends on the workload and GPU architecture for as to how important this is. In this case, it is used for performance. I can’t divulge too many details, but GCN can cycle in work from an ACE incredibly efficiently. Maxwell’s schedular has no analog just as a non hyper-threaded CPU has no analog feature to a hyper threaded one."

If a driver could fix it - then I could simply flash my CPU's software to that of a 4790 and have an i7 at half the price

nVIDIA do support Asynchronous Compute on a hardware level but a large part of their scheduler is implemented in software. Therefore what feeds their Asynchronous Warp Schedulers is a series of Software solutions.
 
nVIDIAs solution is also limited to 32 Compute Queues (or 1 Graphic + 31 Compute). Therefore under any sort of high Asynchronous Compute workloads, they run the risk of hitting a bottleneck. This is further compounded by the Latency introduced under Preemption due to their slow context switching.
 
Face it... nVIDIA was not as prepared as AMD for DX12. That's to be expected given that AMD has had a lot of experience with Mantle as well as the console API's.
 
We may need to wait until Pascal in order to see the level of engineering talent over at nVIDIA as it pertains to DX12.

"Once you eliminate the impossible, whatever remains, no matter how improbable, must be the truth." - Arthur Conan Doyle (Sherlock Holmes)

Link to comment
Share on other sites

Link to post
Share on other sites

https://docs.unrealengine.com/latest/INT/Programming/Rendering/ShaderDevelopment/AsyncCompute/index.html .... UE4 will use assync compute... so looks like .... they wont be able to fix... cuz its a architecture issue... AoTS dev that works with DX 12 said that and i believe him...he even tried to run on a 980ti... and the 980ti got rekt pretty much...

 

They will definitely have to use the ACEs in the Xbone to maximize its performance for ARK.

R9 3900XT | Tomahawk B550 | Ventus OC RTX 3090 | Photon 1050W | 32GB DDR4 | TUF GT501 Case | Vizio 4K 50'' HDR

 

Link to comment
Share on other sites

Link to post
Share on other sites

Nvidia hasnt bean in lead in FLOPS since pre fermi. Ever since GCN came out, Nvidia has LOST in flops, in every fucking card to card comparison. Stop spreading bullshit.

R9 380

Pix rate: 31 GPixels/s

Tex Rate: 109 GTexel/s

FLOPS: 3,476 GFLOPS

GTX 960

Pix rate: 36.1 GPixels/s

Tex Rate: 72.1 GTexel/s

FLOPS: 2,308 GFLOPS

-------------------------------------------------

R9 390

Pix rate: 64 GPixels/s

Tex Rate: 160 GTexel/s

FLOPS: 5,120 GFLOPS

970

Pix rate: 58.8 GPixels/s

Tex Rate: 109 GTexel/s

FLOPS: 3,494 GFLOPS

-------------------------------------------------

R9 390X

Pix rate: 67.2 GPixels/s

Tex Rate: 185 GTexel/s

FLOPS: 5,914 GFLOPS

GTX 980

Pix rate: 71.2 GPixels/s

Tex Rate: 144 GTexel/s

FLOPS: 4,616 GFLOPS

-------------------------------------------

R9 Fury

Pix rate: 64 GPixels/s

Tex Rate: 224 GTexel/s

FLOPS: 7,168 GFLOPS

-------------------------------------------

R9 Fury X

Pix rate: 67.2 GPixels/s

Tex Rate: 269 GTexel/s

FLOPS: 8,602 GFLOPS

980Ti

Pix rate: 96 GPixels/s

Tex Rate: 176 GTexel/s

FLOPS: 5,632 GFLOPS

Titan X

Pix rate: 96 GPixels/s

Tex Rate: 192 GTexel/s

FLOPS: 6,144 GFLOPS

------------------------------------------

Now for the past

GTX 680

Pix rate: 32.2 GPixels/s

Tex Rate: 129 GTexel/s

FLOPS: 3,090 GFLOPS

HD 7970GHz Edt.

Pix rate: 32 GPixels/s

Tex Rate: 128 GTexel/s

FLOPS: 4,096 GFLOPS

Ever since AMD dropped VLIW in favor of GCN, NIVIDIA HAS NEVER. ON ANY ACCOUNT. EVER. CAUGHT UP TO THEM IN FLOPS

780TI says hello. Just not in DP.

Software Engineer for Suncorp (Australia), Computer Tech Enthusiast, Miami University Graduate, Nerd

Link to comment
Share on other sites

Link to post
Share on other sites

780TI says hello. Just not in DP.

780Ti was matched against the 290X (at the time). So was the Titan Black...

 

So lets see how these add up, shall we?

 

R9 290X

Pix rate: 64 GPixels/s

Tex Rate: 176 GTexel/s

FLOPS: 5,632 GFLOPS

 

GTX 780Ti

Pix rate: 52.5 GPixels/s

Tex Rate: 210 GTexel/s

FLOPS: 5,040 GFLOPS

 

GTX Titan Black

Pix rate: 53.3 GPixels/s

Tex Rate: 213 GTexel/s

FLOPS: 5,125 GFLOPS

 

 

You were saying?

Time for an apology to all the people you tried to bullshit i guess....

Link to comment
Share on other sites

Link to post
Share on other sites

780Ti was matched against the 290X (at the time). So was the Titan Black...

 

So lets see how these add up, shall we?

 

R9 290X

Pix rate: 64 GPixels/s

Tex Rate: 176 GTexel/s

FLOPS: 5,632 GFLOPS

 

GTX 780Ti

Pix rate: 37.1 GPixels/s

Tex Rate: 210 GTexel/s

FLOPS: 5,345 GFLOPS

 

GTX Titan Black

Pix rate: 39.2 GPixels/s

Tex Rate: 213 GTexel/s

FLOPS: 5,644 GFLOPS

 

 

You were saying?

Time for an apology to all the people you tried to bullshit i guess....

 

Fixed!

Link to comment
Share on other sites

Link to post
Share on other sites

Fixed!

 

Source?

 

Because Wiki says otherwise:

https://en.wikipedia.org/wiki/GeForce_700_series

Watching Intel have competition is like watching a headless chicken trying to get out of a mine field

CPU: Intel I7 4790K@4.6 with NZXT X31 AIO; MOTHERBOARD: ASUS Z97 Maximus VII Ranger; RAM: 8 GB Kingston HyperX 1600 DDR3; GFX: ASUS R9 290 4GB; CASE: Lian Li v700wx; STORAGE: Corsair Force 3 120GB SSD; Samsung 850 500GB SSD; Various old Seagates; PSU: Corsair RM650; MONITOR: 2x 20" Dell IPS; KEYBOARD/MOUSE: Logitech K810/ MX Master; OS: Windows 10 Pro

Link to comment
Share on other sites

Link to post
Share on other sites

When you see a 290x almost reaching a 980 ti of course peeps will be butthurt ;)

 

The benchmark I saw was 290X and 980, not Ti.

Link to comment
Share on other sites

Link to post
Share on other sites

Source?

 

Because Wiki says otherwise:

https://en.wikipedia.org/wiki/GeForce_700_series

yeah, i cannot find those numbers anywere. May be when OCd to kingdom come. but hey, 290X can OC too... so that would be a fun comparison, wouldnt it....

Link to comment
Share on other sites

Link to post
Share on other sites

The benchmark I saw was 290X and 980, not Ti.

there was a comparison between a 980Ti and FuryX....

 

you know, when you note down the FPS numbers for both the 980Ti and 290X in the same game with the same settings, its not hard to compare them really......

 

Also, not hard to imagine 290X being close, it has exactly the same amount of FLOPs as the 980Ti... So computational power is equal. 980Ti has higher pixel rate by far (nearly double)

Link to comment
Share on other sites

Link to post
Share on other sites

To confirm what @Prysin said:

 

https://www.techpowerup.com/gpudb/2512/geforce-gtx-780-ti.html

https://www.techpowerup.com/gpudb/2549/geforce-gtx-titan-black.html

 

5040 GFLOPS (780TI) and

5121 GFLOPS (Titan Black)

 

for reference models at least.

 

https://www.techpowerup.com/gpudb/2460/radeon-r9-290x.html

 

5632 GFLOPS (290x)

R9 3900XT | Tomahawk B550 | Ventus OC RTX 3090 | Photon 1050W | 32GB DDR4 | TUF GT501 Case | Vizio 4K 50'' HDR

 

Link to comment
Share on other sites

Link to post
Share on other sites

https://www.techpowerup.com/gpudb/2512/geforce-gtx-780-ti.html

https://www.techpowerup.com/gpudb/2549/geforce-gtx-titan-black.html

 

5040 GFLOPS (780TI) and

5121 GFLOPS (Titan Black)

 

for reference models at least.

i already said that. We are just calling bullshit on @DuckDodgers more then suspicious numbers.

 

But hey, if he is talking custom models, why not just look at 290X MSI Lighting w/OC.... just to keep things fair i mean....

Link to comment
Share on other sites

Link to post
Share on other sites

The benchmark I saw was 290X and 980, not Ti.

 

Oh? I thought many have seen this? http://arstechnica.com/gaming/2015/08/directx-12-tested-an-early-win-for-amd-and-disappointment-for-nvidia/

 

Review-chart-template-final-full-width-3

 

Review-chart-template-final-full-width-3

i5 2400 | ASUS RTX 4090 TUF OC | Seasonic 1200W Prime Gold | WD Green 120gb | WD Blue 1tb | some ram | a random case

 

Link to comment
Share on other sites

Link to post
Share on other sites

i already said that. We are just calling bullshit on @DuckDodgers more then suspicious numbers.

 

But hey, if he is talking custom models, why not just look at 290X MSI Lighting w/OC.... just to keep things fair i mean....

 

I know, I was just linking to techpowerup. definitely more reliable source than wikipedia for sure. fixed my post to reflect that.

R9 3900XT | Tomahawk B550 | Ventus OC RTX 3090 | Photon 1050W | 32GB DDR4 | TUF GT501 Case | Vizio 4K 50'' HDR

 

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now


×