Jump to content

[UPDATE] AMD Announces Polaris Architecture - 4th Generation GCN Arriving in Mid 2016

HKZeroFive

Apparently the first GPU is launching in two months.

http://wccftech.com/amd-polaris/

 

That's the card being demoed after just 2 months from initial production.

Watching Intel have competition is like watching a headless chicken trying to get out of a mine field

CPU: Intel I7 4790K@4.6 with NZXT X31 AIO; MOTHERBOARD: ASUS Z97 Maximus VII Ranger; RAM: 8 GB Kingston HyperX 1600 DDR3; GFX: ASUS R9 290 4GB; CASE: Lian Li v700wx; STORAGE: Corsair Force 3 120GB SSD; Samsung 850 500GB SSD; Various old Seagates; PSU: Corsair RM650; MONITOR: 2x 20" Dell IPS; KEYBOARD/MOUSE: Logitech K810/ MX Master; OS: Windows 10 Pro

Link to comment
Share on other sites

Link to post
Share on other sites

I like the fact that they didn't say if it was 16nm or 14nm

Ryzen 7 5800X     Corsair H115i Platinum     ASUS ROG Crosshair VIII Hero (Wi-Fi)     G.Skill Trident Z 3600CL16 (@3800MHzCL16 and other tweaked timings)     

MSI RTX 3080 Gaming X Trio    Corsair HX850     WD Black SN850 1TB     Samsung 970 EVO Plus 1TB     Samsung 840 EVO 500GB     Acer XB271HU 27" 1440p 165hz G-Sync     ASUS ProArt PA278QV     LG C8 55"     Phanteks Enthoo Evolv X Glass     Logitech G915      Logitech MX Vertical      Steelseries Arctis 7 Wireless 2019      Windows 10 Pro x64

Link to comment
Share on other sites

Link to post
Share on other sites

I like the fact that they didn't say if it was 16nm or 14nm

It's both IIRC

Archangel (Desktop) CPU: i5 4590 GPU:Asus R9 280  3GB RAM:HyperX Beast 2x4GBPSU:SeaSonic S12G 750W Mobo:GA-H97m-HD3 Case:CM Silencio 650 Storage:1 TB WD Red
Celestial (Laptop 1) CPU:i7 4720HQ GPU:GTX 860M 4GB RAM:2x4GB SK Hynix DDR3Storage: 250GB 850 EVO Model:Lenovo Y50-70
Seraph (Laptop 2) CPU:i7 6700HQ GPU:GTX 970M 3GB RAM:2x8GB DDR4Storage: 256GB Samsung 951 + 1TB Toshiba HDD Model:Asus GL502VT

Windows 10 is now MSX! - http://linustechtips.com/main/topic/440190-can-we-start-calling-windows-10/page-6

Link to comment
Share on other sites

Link to post
Share on other sites

that's quite incorrect

WCCFTech making wild assumption .. again - this site should just get banned from being used as a source ever again

rumor-has-it-that-its-your-birthday-happ

Archangel (Desktop) CPU: i5 4590 GPU:Asus R9 280  3GB RAM:HyperX Beast 2x4GBPSU:SeaSonic S12G 750W Mobo:GA-H97m-HD3 Case:CM Silencio 650 Storage:1 TB WD Red
Celestial (Laptop 1) CPU:i7 4720HQ GPU:GTX 860M 4GB RAM:2x4GB SK Hynix DDR3Storage: 250GB 850 EVO Model:Lenovo Y50-70
Seraph (Laptop 2) CPU:i7 6700HQ GPU:GTX 970M 3GB RAM:2x8GB DDR4Storage: 256GB Samsung 951 + 1TB Toshiba HDD Model:Asus GL502VT

Windows 10 is now MSX! - http://linustechtips.com/main/topic/440190-can-we-start-calling-windows-10/page-6

Link to comment
Share on other sites

Link to post
Share on other sites

I like the fact that they didn't say if it was 16nm or 14nm

It's both IIRC

You're right, they will use both.

 

From the Anandtech article:

As for RTG’s FinFET manufacturing plans, the fact that RTG only mentions “FinFET” and not a specific FinFET process (e.g. TSMC 16nm) is intentional. The group has confirmed that they will be utilizing both traditional partner TSMC’s 16nm process and AMD fab spin-off (and Samsung licensee) GlobalFoundries’ 14nm process, making this the first time that AMD’s graphics group has used more than a single fab. To be clear here there’s no expectation that RTG will be dual-sourcing – having both fabs produce the same GPU – but rather the implication is that designs will be split between the two fabs. To that end we know that the small Polaris GPU that RTG previewed will be produced by GlobalFoundries on their 14nm process, meanwhile it remains to be seen how the rest of RTG’s Polaris GPUs will be split between the fabs.

 

So it seems like they might use GlobalFoundries for the lower end cards, and TSMC for the high end cards.

Link to comment
Share on other sites

Link to post
Share on other sites

Apparently the first GPU is launching in two months.

http://wccftech.com/amd-polaris/

 

81e0af215ea87afb53613a295d79329f.png

i7 7700k @ 4.9ghz | Asus Maximus IX Hero | G.skill 32gb @ 3200 | Gtx 1080 classified | In win 909 | Samsung 960 pro 1tb | WD caviar blue 1tb x3 | Dell u3417w | Corsair H115i | Ducky premier dk9008p (mx reds) | Logitech g900 | Sennheiser hd 800s w/ hdvd 800 | Audioengine a5+ w/ s8

 

Link to comment
Share on other sites

Link to post
Share on other sites

Their assumptions have been accurate lately.

They are going for the shotgun method. They report every single thing and once in a while it turns out to be true.

Even a broken watch is right twice a day...

Link to comment
Share on other sites

Link to post
Share on other sites

Every time I see Polaris I think of Super Motherload

Gamma v2.2 | i7 6700k @ 4.6ghz| Dark Rock TF | ASRock Z170 OC Formula | G-SKILL TridentZ Royal 2x16Gb 3200mhz | MSI GTX 1070 Ti Titanium | Sandisk 120Gb SSD | WD Black 1Tb HDD | Corsair RMx 850w | Corsair Spec Alpha | MSI Optix G27C2/2x19" monitors/34" Insignia tv

Spoiler

Secondary rig status: Blendin Blandin | Xeon E5 2670 E3 ES | Noctua L12s | ASRock X99 OC Formula | 48Gb Ram Smoothie | EVGA 980ti Superclocked+ | ADATA SU800 | SFFTime P-Atx | 

 

 

Link to comment
Share on other sites

Link to post
Share on other sites

No, Fermi was just bad design. GCN isn't bad design. It's just dense (HDL-High Density Libraries and all that). While FinFET voltage control is very different, did we all forget what happened to Intel upon moving to FinFET? Heat went up, and overclocking headroom came down. That was true for both the mainstream and HEDT platforms, where HEDT chips were still soldered to their heat spreaders.

Nvidia didn't get it from nowhere. They underwent a paradigm shift in design philosophy. While Koduri hasn't been with AMD for a while, GCN is still based on his previous work. GCN 2.0/4.0 won't be a drastic architectural change other than allowing more TMUs and ROPs as Fiji should have.

No.

Fermi included hardware scheduling components. This made sense, at the time, because Fermi was geared for GPGPU performance.

Kepler ditched most of the hardware based scheduling components as the architecture was geared towards gamers.

Maxwell expanded upon Kepler.

GCN contains hardware based scheduling components as well as ACEs. These all draw power, hence the power usage. This is why GCN's architecture tends to excel at compute workloads.

Polaris, though I can't speak of details, is like GCN but with some shader optimizations. In other words, Polaris can do more with less in terms of compute performance. This opened the door for caching improvements, tessellation improvements as well as a more robust front end. Polaris is different enough from GCN that it can be considered a new architecture. The ALUs are not the same, the Command Processor is not the same, the caching is not the same, the CUs aren't the same, the RBEs aren't the same, the memory controller is brand new etc etc etc.

This test was conducted with a frame rate cap of 60FPS. The figures, AMD obtained, are entirely within the realm of possibility. We're looking at up to 2.5 times the performance per watt of Fiji with Polaris.

Pascal, given that it is Maxwell version 2, should retain its excellent performance per watt figures but will likely be beaten in terms of pure compute performance by Greenland.

My take? Greenland overtakes Pascal.

"Once you eliminate the impossible, whatever remains, no matter how improbable, must be the truth." - Arthur Conan Doyle (Sherlock Holmes)

Link to comment
Share on other sites

Link to post
Share on other sites

No.

Fermi included hardware scheduling components. This made sense, at the time, because Fermi was geared for GPGPU performance.

Kepler ditched most of the hardware based scheduling components as the architecture was geared towards gamers.

Maxwell expanded upon Kepler.

GCN contains hardware based scheduling components as well as ACEs. These all draw power, hence the power usage. This is why GCN's architecture tends to excel at compute workloads.

Polaris, though I can't speak of details, is like GCN but with some shader optimizations. In other words, Polaris can do more with less in terms of compute performance. This opened the door for caching improvements, tessellation improvements as well as a more robust front end. Polaris is different enough from GCN that it can be considered a new architecture. The ALUs are not the same, the Command Processor is not the same, the caching is not the same, the CUs aren't the same, the RBEs aren't the same, the memory controller is brand new etc etc etc.

This test was conducted with a frame rate cap of 60FPS. The figures, AMD obtained, are entirely within the realm of possibility. We're looking at up to 2.5 times the performance per watt of Fiji with Polaris.

Pascal, given that it is Maxwell version 2, should retain its excellent performance per watt figures but will likely be beaten in terms of pure compute performance by Greenland.

My take? Greenland overtakes Pascal.

Considering Maxwell already loses out in compute performance to GCN, I agree.

"We also blind small animals with cosmetics.
We do not sell cosmetics. We just blind animals."

 

"Please don't mistake us for Equifax. Those fuckers are evil"

 

This PSA brought to you by Equifacks.
PMSL

Link to comment
Share on other sites

Link to post
Share on other sites

So I can assume this new architecture won't be good because they've been alternating between good and bad every generation-

tahiti- good

hawaii- bad

fiji- good

polaris- bad?

Hawaii was/is great without the stock cooler :D

USEFUL LINKS:

PSU Tier List F@H stats

Link to comment
Share on other sites

Link to post
Share on other sites

No.

Fermi included hardware scheduling components. This made sense, at the time, because Fermi was geared for GPGPU performance.

Kepler ditched most of the hardware based scheduling components as the architecture was geared towards gamers.

Maxwell expanded upon Kepler.

GCN contains hardware based scheduling components as well as ACEs. These all draw power, hence the power usage. This is why GCN's architecture tends to excel at compute workloads.

Polaris, though I can't speak of details, is like GCN but with some shader optimizations. In other words, Polaris can do more with less in terms of compute performance. This opened the door for caching improvements, tessellation improvements as well as a more robust front end. Polaris is different enough from GCN that it can be considered a new architecture. The ALUs are not the same, the Command Processor is not the same, the caching is not the same, the CUs aren't the same, the RBEs aren't the same, the memory controller is brand new etc etc etc.

This test was conducted with a frame rate cap of 60FPS. The figures, AMD obtained, are entirely within the realm of possibility. We're looking at up to 2.5 times the performance per watt of Fiji with Polaris.

Pascal, given that it is Maxwell version 2, should retain its excellent performance per watt figures but will likely be beaten in terms of pure compute performance by Greenland.

My take? Greenland overtakes Pascal.

 

You don't need hardware scheduling for good GPGPU performance. The Kepler Teslas prove as much being the top used accelerator in the world for compute tasks. The only thing hardware scheduling is particularly useful for is Asynchronous Compute/Shading pipelines.

 

GCN doesn't excel at compute workloads. They're outclassed by a Knight's Corner Xeon Phi in Linpack and HPC applications despite the Xeon Phi having 1/3 the theoretical performance cap. Either bad drivers or bad design, but they most certainly are not compute monsters as everyone seems to think.

 

Only theoretically. AMD has a lot of catching up to do to make their compute competitive.

Software Engineer for Suncorp (Australia), Computer Tech Enthusiast, Miami University Graduate, Nerd

Link to comment
Share on other sites

Link to post
Share on other sites

You don't need hardware scheduling for good GPGPU performance. The Kepler Teslas prove as much being the top used accelerator in the world for compute tasks. The only thing hardware scheduling is particularly useful for is Asynchronous Compute/Shading pipelines.

GCN doesn't excel at compute workloads. They're outclassed by a Knight's Corner Xeon Phi in Linpack and HPC applications despite the Xeon Phi having 1/3 the theoretical performance cap. Either bad drivers or bad design, but they most certainly are not compute monsters as everyone seems to think.

Only theoretically. AMD has a lot of catching up to do to make their compute competitive.

Where do i buy a xeon phi graphics card to run my games?

Link to comment
Share on other sites

Link to post
Share on other sites

Where do i buy a xeon phi graphics card to run my games?

 

You can't, they are co-processors for certain types of server motherboard.

Intel i7 5820K (4.5 GHz) | MSI X99A MPower | 32 GB Kingston HyperX Fury 2666MHz | Asus RoG STRIX GTX 1080ti OC | Samsung 951 m.2 nVME 512GB | Crucial MX200 1000GB | Western Digital Caviar Black 2000GB | Noctua NH-D15 | Fractal Define R5 | Seasonic 860 Platinum | Logitech G910 | Sennheiser 599 | Blue Yeti | Logitech G502

 

Nikon D500 | Nikon 300mm f/4 PF  | Nikon 200-500 f/5.6 | Nikon 50mm f/1.8 | Tamron 70-210 f/4 VCII | Sigma 10-20 f/3.5 | Nikon 17-55 f/2.8 | Tamron 90mm F2.8 SP Di VC USD Macro | Neewer 750II

Link to comment
Share on other sites

Link to post
Share on other sites

LOL I found these post hilarious... I love how it got so off topic from the video, and also I would guess that if anything the videos are true but the amd system is getting measured at a lower power usage time, and the nvidea system is at a higher load time, but also the cpu is capped at 80%.

another thing @patrickjp93 saying that nvidea crushes or AMD crushes(not you but others) over the past few years they all performed relatively good for their prices, but amd had a lot more price to performance cards in the mid range, and the R9 290 was basically the 980ti in 980ti vs Titain compared to R9 290 vs R9 290x. because they releases a full core, but at a lower clock. also a r9 fury x may not compare to a 980ti very well but a non x fury is $100+ less. 

 

Also @patrickjp93 you are a very opinionated person... you made me laugh the most... its like you have a tin hat for amd, and then had nvidea and intels mistakes surgically removed from your mind... but I guess to each their own, and I'm prepared for your flaming...

Link to comment
Share on other sites

Link to post
Share on other sites

Where do i buy a xeon phi graphics card to run my games?

 

They aren't GPUs, not unless you use a very efficient software-based rendering engine, and even then you need to send the image data through the motherboard graphics output.

 

Xeon Phi are built purely as compute accelerators.

Software Engineer for Suncorp (Australia), Computer Tech Enthusiast, Miami University Graduate, Nerd

Link to comment
Share on other sites

Link to post
Share on other sites

LOL I found these post hilarious... I love how it got so off topic from the video, and also I would guess that if anything the videos are true but the amd system is getting measured at a lower power usage time, and the nvidea system is at a higher load time, but also the cpu is capped at 80%.

another thing @patrickjp93 saying that nvidea crushes or AMD crushes(not you but others) over the past few years they all performed relatively good for their prices, but amd had a lot more price to performance cards in the mid range, and the R9 290 was basically the 980ti in 980ti vs Titain compared to R9 290 vs R9 290x. because they releases a full core, but at a lower clock. also a r9 fury x may not compare to a 980ti very well but a non x fury is $100+ less. 

 

Also @patrickjp93 you are a very opinionated person... you made me laugh the most... its like you have a tin hat for amd, and then had nvidea and intels mistakes surgically removed from your mind... but I guess to each their own, and I'm prepared for your flaming...

I'm not tin foil hat on AMD or forgetful of Nvidia's mistakes. I always base my arguments on what is, not what was.

 

Intel buying off OEMs? Otellini and Huddy stepped down just shy of a decade ago. Same issue with the Pentium 4 (iirc) floating point problem: the perpetrators are long gone.

 

Nvidia melting GPUs? 7 years old now? I don't consider the 970 issues to be either problematic nor maliciously intended. Clerical errors will happen to any company EVENTUALLY. I'd say the same if it happened to Intel or AMD. Gameworks clearly isn't made to cripple AMD since there are a couple GW titles AMD is better at.

 

AMD's complete hype/lies about Barcelona are also that old. Bulldozer is much more recent in history, and the engineers responsible for it still work at AMD. 290/X reference heat problems are also still current. AMD's driver problems are more current, with Crimson having fried a few GPUs as well. AMD didn't deliver on open-source Mantle as promised. AMD has yet to deliver on GPUOpen or fully O-S graphics drivers for Windows or Linux. We'll have to wait and see on that end.

 

No flaming required, just facts.

Software Engineer for Suncorp (Australia), Computer Tech Enthusiast, Miami University Graduate, Nerd

Link to comment
Share on other sites

Link to post
Share on other sites

They aren't GPUs, not unless you use a very efficient software-based rendering engine, and even then you need to send the image data through the motherboard graphics output.

Xeon Phi are built purely as compute accelerators.

Forgot to add /s

Link to comment
Share on other sites

Link to post
Share on other sites

mate, you're saying that both testing are capped - that means the GTX950 would draw a significant amount less power than it's actually rated for

keeping in mind that both systems minus videocards would draw equal amounts of power, that would make system 1 draw power from where!? a parallel universe!?

let's say the GTX950 would draw on it's own, capped, ~50W - that would mean the rest of the system would draw 100W

and this is where all bells, whistles and horns would ring because system1, with video card, draws only 85W - HELLO!

as I said, AMD's testing is bullcrap!

system1 is running with all power caps in place, disabled core, disabled lanes

system2 is running with no caps

How do you know AMD disabled things for one system and not for the other? You'd believe it if it was Nvidia you dumbass fanboy.
Link to comment
Share on other sites

Link to post
Share on other sites

You don't need hardware scheduling for good GPGPU performance. The Kepler Teslas prove as much being the top used accelerator in the world for compute tasks. The only thing hardware scheduling is particularly useful for is Asynchronous Compute/Shading pipelines.

GCN doesn't excel at compute workloads. They're outclassed by a Knight's Corner Xeon Phi in Linpack and HPC applications despite the Xeon Phi having 1/3 the theoretical performance cap. Either bad drivers or bad design, but they most certainly are not compute monsters as everyone seems to think.

Only theoretically. AMD has a lot of catching up to do to make their compute competitive.

I believe I've mentioned this to you on several instances. Software.

AMD lacks proper software optimization tools. AMD rely on OpenCL.

Knights Landing/Xeon Phi can also be powered by OpenCL but its main software stack are in

OpenMP, MPI, C, C++ and Fortran which are all more robust programming languages.

CUDA is still more robust than OpenCL as well.

In terms of the hardware, though, GCN is a monster compared to Kepler and Maxwell in terms of GPGPU. Heck Hawaii, R9 290x, outperforms the Kepler based Titan Black under OpenCL and DirectCompute.

GCN is a compute oriented architecture but which made a few compromises in order to handle Gaming Tasks as well. One such compromise, compared to Tahiti, is Hawaii and Fiji's FP64 performance.

Polaris builds upon GCN. When Greenland arrives, you will see what I mean. Larger registers and cache should mean better FP64 performance.

What AMD needs, in the realm of HTPCs, is more robust OpenCL support. This will take time and will arrive as OpenCL matures.

http://us.hardware.info/reviews/5231/8/nvidia-geforce-gtx-titan-black-4-way-sli-review-gpgpu-benchmarks

"Once you eliminate the impossible, whatever remains, no matter how improbable, must be the truth." - Arthur Conan Doyle (Sherlock Holmes)

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now


×