Jump to content

A REAL 64 Core CPU - For SCIENCE!

Forget Threadripper and Core i9 - Xeon Phi is the go-to for x86 core count. But how does it actually perform?

 

 

More info on the Supermicro SuperServer 5038K-I: http://geni.us/g82JV
Buy a Xeon Phi processor on Amazon(?): http://geni.us/z2IB4Og

Emily @ LINUS MEDIA GROUP                                  

congratulations on breaking absolutely zero stereotypes - @cs_deathmatch

Link to comment
Share on other sites

Link to post
Share on other sites

I bet there is one guy in the world who actually has bought one of these with his own money.

زندگی از چراغ

Intel Core i7 7800X 6C/12T (4.5GHz), Corsair H150i Pro RGB (360mm), Asus Prime X299-A, Corsair Vengeance LPX 32GB (4X4GB & 2X8GB 3000MHz DDR4), MSI GeForce GTX 1070 Gaming X 8G (2.113GHz core & 9.104GHz memory), 1 Samsung 970 Evo Plus 1TB NVMe M.2, 1 Samsung 850 Pro 256GB SSD, 1 Samsung 850 Evo 500GB SSD, 1 WD Red 1TB mechanical drive, Corsair RM750X 80+ Gold fully modular PSU, Corsair Obsidian 750D full tower case, Corsair Glaive RGB mouse, Corsair K70 RGB MK.2 (Cherry MX Red) keyboard, Asus VN247HA (1920x1080 60Hz 16:9), Audio Technica ATH-M20x headphones & Windows 10 Home 64 bit. 

 

 

The time Linus replied to me on one of my threads: 

 

Link to comment
Share on other sites

Link to post
Share on other sites

1 minute ago, Name Taken said:

Wonder how fast it would mine Monero.

probably slower then a GPU at the same power draw and a FRACTION the price, CPU mining is stupid slow.

Link to comment
Share on other sites

Link to post
Share on other sites

Hey Linus, can I borrow that to do my MATLAB assignments? Solving 5000x5000 matrices took a good 1.5 hours to run on the remote Xeon workstation I had available. ;)

Intel® Core™ i7-12700 | GIGABYTE B660 AORUS MASTER DDR4 | Gigabyte Radeon™ RX 6650 XT Gaming OC | 32GB Corsair Vengeance® RGB Pro SL DDR4 | Samsung 990 Pro 1TB | WD Green 1.5TB | Windows 11 Pro | NZXT H510 Flow White
Sony MDR-V250 | GNT-500 | Logitech G610 Orion Brown | Logitech G402 | Samsung C27JG5 | ASUS ProArt PA238QR
iPhone 12 Mini (iOS 17.2.1) | iPhone XR (iOS 17.2.1) | iPad Mini (iOS 9.3.5) | KZ AZ09 Pro x KZ ZSN Pro X | Sennheiser HD450bt
Intel® Core™ i7-1265U | Kioxia KBG50ZNV512G | 16GB DDR4 | Windows 11 Enterprise | HP EliteBook 650 G9
Intel® Core™ i5-8520U | WD Blue M.2 250GB | 1TB Seagate FireCuda | 16GB DDR4 | Windows 11 Home | ASUS Vivobook 15 
Intel® Core™ i7-3520M | GT 630M | 16 GB Corsair Vengeance® DDR3 |
Samsung 850 EVO 250GB | macOS Catalina | Lenovo IdeaPad P580

Link to comment
Share on other sites

Link to post
Share on other sites

26 minutes ago, GabenJr said:

Forget Threadripper and Core i9 - Xeon Phi is the go-to for x86 core count. But how does it actually perform?

 

 

More info on the Supermicro SuperServer 5038K-I: http://geni.us/g82JV
Buy a Xeon Phi processor on Amazon(?): http://geni.us/z2IB4Og

are you guys going to try get ahold of EPYC?

Link to comment
Share on other sites

Link to post
Share on other sites

16 minutes ago, LinusTechTipsFanFromDarlo said:

I bet there is one guy in the world who actually has bought one of these with his own money.

Well, one guy sure. Bet large datacenters or whatever have bought quite a few of these guys.

Main rig on profile

VAULT - File Server

Spoiler

Intel Core i5 11400 w/ Shadow Rock LP, 2x16GB SP GAMING 3200MHz CL16, ASUS PRIME Z590-A, 2x LSI 9211-8i, Fractal Define 7, 256GB Team MP33, 3x 6TB WD Red Pro (general storage), 3x 1TB Seagate Barracuda (dumping ground), 3x 8TB WD White-Label (Plex) (all 3 arrays in their respective Windows Parity storage spaces), Corsair RM750x, Windows 11 Education

Sleeper HP Pavilion A6137C

Spoiler

Intel Core i7 6700K @ 4.4GHz, 4x8GB G.SKILL Ares 1800MHz CL10, ASUS Z170M-E D3, 128GB Team MP33, 1TB Seagate Barracuda, 320GB Samsung Spinpoint (for video capture), MSI GTX 970 100ME, EVGA 650G1, Windows 10 Pro

Mac Mini (Late 2020)

Spoiler

Apple M1, 8GB RAM, 256GB, macOS Sonoma

Consoles: Softmodded 1.4 Xbox w/ 500GB HDD, Xbox 360 Elite 120GB Falcon, XB1X w/2TB MX500, Xbox Series X, PS1 1001, PS2 Slim 70000 w/ FreeMcBoot, PS4 Pro 7015B 1TB (retired), PS5 Digital, Nintendo Switch OLED, Nintendo Wii RVL-001 (black)

Link to comment
Share on other sites

Link to post
Share on other sites

I don't get it. This thing is hella expensive and can't do anything too well. And the stuff it was designed for runs much much better on a 1080ti...?

Folding stats

Vigilo Confido

 

Link to comment
Share on other sites

Link to post
Share on other sites

32 minutes ago, LinusTechTipsFanFromDarlo said:

I bet there is one guy in the world who actually has bought one of these with his own money.

 Some were actually on sale a while ago for 200$.

Link to comment
Share on other sites

Link to post
Share on other sites

So fun fact, I am actually the customer for products like these. And I straight up don't care about Xeon Phis. They were interesting a while ago but from my perspective there is no reason to use one of those chips over your standard GPU. In fact my lab uses off the shelf Titan XPs for most of our simulations. For most of our purposes these cards are the best options with only the new Tesla V100s being more powerful. Standard disclaimer our simulations only utilize single floating point precision so pure FLOPS are fine. That being said I (being the odd man in our lab) want to get a V100 for a personal project.

TL;DR even for HPC applications Xeon Phis are mainly irrelevant.

The award fits...

Link to comment
Share on other sites

Link to post
Share on other sites

So I noticed it has 4 threads per core. I've never seen that technology before. Are there any downsides. Also is there any chance that this will come to consumer or enthusest level CPUs in the near future? 

Link to comment
Share on other sites

Link to post
Share on other sites

Thank Dog for this video!  I've been waiting forever to see reviews of science-driven computers!

 

"unless you like to science for funsies."-Linus (6:45) ....   

 

C'mon Linus, some of us do science "for realsies". I couldn't give to sh--ts about video gaming, which seems to be the focus of this channel (sad poop emoji).  Please do more reviews on computers that are simulation focused like this!

 

Link to comment
Share on other sites

Link to post
Share on other sites

1 hour ago, Nicnac said:

I don't get it. This thing is hella expensive and can't do anything too well. And the stuff it was designed for runs much much better on a 1080ti...?

 

It isn't an apples to apples comparison. The Phi is a CISC x64 processor, whereas the Pascal GPU is considered a RISC processor (though it doesn't follow any declared RISC standard). Yes there are major differences, and they are quite relevant depending on the task at hand.

 

And the Phi being a manycore x64 processor was the main selling point: existing software doesn't have to be significantly modified to run on it. And Intel's tools could help determine performance bottlenecks in multi-threaded applications to ensure performance scales with the core count, which is extremely important when talking about having these in a cluster.

 

One of the major reasons to have that many cores and threads is to spread out all the tasks a node will do. The Titan supercomputer uses 16-core Opteron processors with Tesla GPUs. The CPUs are not used for processing, but for driving the individual nodes, including getting data to and from the GPUs and communicating across the networking backbone, such as with the high speed storage or sending and receiving data from siblings and masters.

 

People who talk computers often forget the very diverse roles CPUs take on. One of which is handling hardware interrupts.

 

Such as those coming from networking equipment. Such as the Infiniband and 100GbE cards used in supercomputers. One of the reasons Threadripper's PCI-E lane count is eye-catching is the ability to support up to four (4) PCI-E 3.0x16 cards at full x16 speed across its 64 lanes. The Phi supports 36 lanes by comparison, still enough to handle a GPU compute card and an Infiniband card. The Phi does support hexa-channel memory configurations, compared to Threadripper's quad-channel.

 

Though it is interesting they made the Phi a primary processor instead of a coprocessor on a daughter card.

Wife's build: Amethyst - Ryzen 9 3900X, 32GB G.Skill Ripjaws V DDR4-3200, ASUS Prime X570-P, EVGA RTX 3080 FTW3 12GB, Corsair Obsidian 750D, Corsair RM1000 (yellow label)

My build: Mira - Ryzen 7 3700X, 32GB EVGA DDR4-3200, ASUS Prime X470-PRO, EVGA RTX 3070 XC3, beQuiet Dark Base 900, EVGA 1000 G6

Link to comment
Share on other sites

Link to post
Share on other sites

1 hour ago, borisdarker said:

So I noticed it has 4 threads per core. I've never seen that technology before. Are there any downsides. Also is there any chance that this will come to consumer or enthusest level CPUs in the near future? 

To give the standard answer, depends on what your work load is. If we are talking about gaming or something that needs high performance of a single core then yes. There are huge drawbacks, because you basically have four threads competing for one core. However for certain applications if you are intelligent on how you do your threads you could get some respectable speed ups.

 

Edit: Forgot to answer the last part of the question, no I don't see this coming to consumer any time soon. It'd be unnecessary for most applications I can think of.

The award fits...

Link to comment
Share on other sites

Link to post
Share on other sites

20 minutes ago, brandishwar said:

 

It isn't an apples to apples comparison. The Phi is a CISC x64 processor, whereas the Pascal GPU is considered a RISC processor (though it doesn't follow any declared RISC standard). Yes there are major differences, and they are quite relevant depending on the task at hand.

 

And the Phi being a manycore x64 processor was the main selling point: existing software doesn't have to be significantly modified to run on it. And Intel's tools could help determine performance bottlenecks in multi-threaded applications to ensure performance scales with the core count, which is extremely important when talking about having these in a cluster.

 

One of the major reasons to have that many cores and threads is to spread out all the tasks a node will do. The Titan supercomputer uses 16-core Opteron processors with Tesla GPUs. The CPUs are not used for processing, but for driving the individual nodes, including getting data to and from the GPUs and communicating across the networking backbone, such as with the high speed storage or sending and receiving data from siblings and masters.

 

People who talk computers often forget the very diverse roles CPUs take on. One of which is handling hardware interrupts.

 

Such as those coming from networking equipment. Such as the Infiniband and 100GbE cards used in supercomputers. One of the reasons Threadripper's PCI-E lane count is eye-catching is the ability to support up to four (4) PCI-E 3.0x16 cards at full x16 speed across its 64 lanes. The Phi supports 36 lanes by comparison, still enough to handle a GPU compute card and an Infiniband card. The Phi does support hexa-channel memory configurations, compared to Threadripper's quad-channel.

 

Though it is interesting they made the Phi a primary processor instead of a coprocessor on a daughter card.

I apologize if this sounds condescending as I am unaware of your knowledge base. However while the statement that the code doesn't have to be significantly modified to run on it is true. If you wanted to get the most of the performance out of it you'd definitely have to rewrite it in order for it to be more vectorizable. Which has its own challenges. Of course if you've already written your code for that purpose then sure slapping one of these in versus a GPU is a huge difference. 

 

That being said I agree with your comment about it is interesting they made it a primary processor instead of a coprocessor. Honestly, that is most of the reason I've lost interest. Also the comment from Linus about it being good for NN is a bit strange considering Intel announced the Nervana Chip to handle that end of the market.

The award fits...

Link to comment
Share on other sites

Link to post
Share on other sites

56 minutes ago, SiverJohn17 said:

I apologize if this sounds condescending as I am unaware of your knowledge base. However while the statement that the code doesn't have to be significantly modified to run on it is true. If you wanted to get the most of the performance out of it you'd definitely have to rewrite it in order for it to be more vectorizable. Which has its own challenges. Of course if you've already written your code for that purpose then sure slapping one of these in versus a GPU is a huge difference. 

 

Not condescending at all. You merely clarified what I said, given I mentioned the Intel analysis tools but didn't mention really why you'd want to use them. Intel advertised that toolset for better multithreading performance regardless of whether you use the Phi coprocessor or not.

 

My knowledge base, to fill in that detail, comes from about 20 years experience with programming, including writing multi-threaded applications, though I have never written anything intended to target the Phi. I've been considering playing around with the various parallelization options available -- just a matter of finding the time and motivation.

 

Plus slapping in a Phi versus using a GPU means not having to rewrite code that's already targeting the x64 instruction set. If you're building something new, however, you're likely much better off targeting a GPU platform -- whether using NVIDIA's library directly or going through OpenCL -- than targeting any CPU platform, though there may be use cases wherein the CPU or a hybrid approach is better than pure GPU computing.

 

1 hour ago, SiverJohn17 said:

That being said I agree with your comment about it is interesting they made it a primary processor instead of a coprocessor. Honestly, that is most of the reason I've lost interest.

 

Not to mention the setup that is needed to power it. I don't think you could've just plugged the daughter card into any mainboard. It had to meet certain requirements with regard to its BIOS, making it inaccessible as an option to most people. Which is half the genius behind CUDA and OpenCL: you can use consumer-grade hardware out of the gate.

Wife's build: Amethyst - Ryzen 9 3900X, 32GB G.Skill Ripjaws V DDR4-3200, ASUS Prime X570-P, EVGA RTX 3080 FTW3 12GB, Corsair Obsidian 750D, Corsair RM1000 (yellow label)

My build: Mira - Ryzen 7 3700X, 32GB EVGA DDR4-3200, ASUS Prime X470-PRO, EVGA RTX 3070 XC3, beQuiet Dark Base 900, EVGA 1000 G6

Link to comment
Share on other sites

Link to post
Share on other sites

4 hours ago, Nicnac said:

 This thing is hella expensive and can't do anything too well

Wrong, it makes a cool looking paperweight. But in all seriousness, I think your right, It gets outperformed by other xeons in single core and by gpus in multicore, I can't think of why someone wouldn't just buy a gpu instead.

phi.JPG

Link to comment
Share on other sites

Link to post
Share on other sites

8 minutes ago, antoipod said:

Wrong, it makes a cool looking paperweight. But in all seriousness, I think your right, It gets outperformed by other xeons in single core and by gpus in multicore, I can't think of why someone wouldn't just buy a gpu instead.

 

GPUs and CPUs fit different use cases, and not every use case can be accelerated with a GPU.

Wife's build: Amethyst - Ryzen 9 3900X, 32GB G.Skill Ripjaws V DDR4-3200, ASUS Prime X570-P, EVGA RTX 3080 FTW3 12GB, Corsair Obsidian 750D, Corsair RM1000 (yellow label)

My build: Mira - Ryzen 7 3700X, 32GB EVGA DDR4-3200, ASUS Prime X470-PRO, EVGA RTX 3070 XC3, beQuiet Dark Base 900, EVGA 1000 G6

Link to comment
Share on other sites

Link to post
Share on other sites

I wonder how this would fare with Folding@Home? 

There is more that meets the eye
I see the soul that is inside

 

 

Link to comment
Share on other sites

Link to post
Share on other sites

Just now, hey_yo_ said:

I wonder how this would fare with Folding@Home? 

It would definitely be beneficial, how beneficial depends on exactly how that code is implemented, something I am ashamed to say I am ignorant of (considering it is close to my field of interest).

The award fits...

Link to comment
Share on other sites

Link to post
Share on other sites

5 minutes ago, hey_yo_ said:

I wonder how this would fare with Folding@Home? 

Probably very poorly, I doubt folding at home can utilize 256 threads. 

Link to comment
Share on other sites

Link to post
Share on other sites

1 minute ago, SiverJohn17 said:

It would definitely be beneficial, how beneficial depends on exactly how that code is implemented, something I am ashamed to say I am ignorant of (considering it is close to my field of interest).

 

1 minute ago, antoipod said:

Probably very poorly, I doubt folding at home can ustilize 256 threads. 

Linus could’ve tested it with F@H. 

There is more that meets the eye
I see the soul that is inside

 

 

Link to comment
Share on other sites

Link to post
Share on other sites

Love these more out there and non traditional stuff. Definitely my favorite kinds of videos! 

Link to comment
Share on other sites

Link to post
Share on other sites

4 minutes ago, antoipod said:

Probably very poorly, I doubt folding at home can utilize 256 threads. I'd like to try it out but don't have a powerful enough power supply or windows server 2016.

Actually, and I didn't realize when I made my first comment, it most certainly can. According to Wikipedia at least (I had trouble finding this on the F@H website) Folding@Home utilizes GROMACS which is known for its strong scalability. As well as its optimization so much so that it gets relatively minor improvements from going from CPU to GPU, making it a perfect application if someone has a couple grand they want to just spend for the hell of it.

The award fits...

Link to comment
Share on other sites

Link to post
Share on other sites

2 hours ago, antoipod said:

Wrong, it makes a cool looking paperweight. But in all seriousness, I think your right, It gets outperformed by other xeons in single core and by gpus in multicore, I can't think of why someone wouldn't just buy a gpu instead.

phi.JPG

 

7 hours ago, Nicnac said:

I don't get it. This thing is hella expensive and can't do anything too well. And the stuff it was designed for runs much much better on a 1080ti...?

PHI brings something(s) to the table that GPUs can't. More instruction sets.

Come Bloody Angel

Break off your chains

And look what I've found in the dirt.

 

Pale battered body

Seems she was struggling

Something is wrong with this world.

 

Fierce Bloody Angel

The blood is on your hands

Why did you come to this world?

 

Everybody turns to dust.

 

Everybody turns to dust.

 

The blood is on your hands.

 

The blood is on your hands!

 

Pyo.

Link to comment
Share on other sites

Link to post
Share on other sites

Note the thumbnail with Linus having the Xeon Phi at the bottom of the cpu cooler. No, the thermal paste isn't super glued to the cpu. That's how it actually is, cpu is first attached to the heatsink with that black bracket piece. Then the entire thing goes onto the socket, with metal rods as guides.

Intel Xeon E5 1650 v3 @ 3.5GHz 6C:12T / CM212 Evo / Asus X99 Deluxe / 16GB (4x4GB) DDR4 3000 Trident-Z / Samsung 850 Pro 256GB / Intel 335 240GB / WD Red 2 & 3TB / Antec 850w / RTX 2070 / Win10 Pro x64

HP Envy X360 15: Intel Core i5 8250U @ 1.6GHz 4C:8T / 8GB DDR4 / Intel UHD620 + Nvidia GeForce MX150 4GB / Intel 120GB SSD / Win10 Pro x64

 

HP Envy x360 BP series Intel 8th gen

AMD ThreadRipper 2!

5820K & 6800K 3-way SLI mobo support list

 

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×