Jump to content

Where Intel is in REAL Trouble...

nicklmg

Becasue server/enterprise grade reliability specs.

F@H
Desktop: i9-13900K, ASUS Z790-E, 64GB DDR5-6000 CL36, RTX3080, 2TB MP600 Pro XT, 2TB SX8200Pro, 2x16TB Ironwolf RAID0, Corsair HX1200, Antec Vortex 360 AIO, Thermaltake Versa H25 TG, Samsung 4K curved 49" TV, 23" secondary, Mountain Everest Max

Mobile SFF rig: i9-9900K, Noctua NH-L9i, Asrock Z390 Phantom ITX-AC, 32GB, GTX1070, 2x1TB SX8200Pro RAID0, 2x5TB 2.5" HDD RAID0, Athena 500W Flex (Noctua fan), Custom 4.7l 3D printed case

 

Asus Zenbook UM325UA, Ryzen 7 5700u, 16GB, 1TB, OLED

 

GPD Win 2

Link to comment
Share on other sites

Link to post
Share on other sites

Why not try SnapRAID with MergerFS?

Link to comment
Share on other sites

Link to post
Share on other sites

  • 1 month later...

Especially as of PCIe Gen 3/4, there are extensive per-lane stats available within the endpoints. It's possible that due to extended cabling and cross-talk within the cable bundles that the bit error-rate (BER) is exceeding a threshold that triggers the bus reset.

 

Also are the Gigabyte CNV3024 riser cards PCIe retimers or only simple redrivers? It would appear that four of the drives are potentially directly connected to CPU PCIe lanes and those are long paths to the front of the chassis. Signal integrity becomes more challenging with PCIe Gen 3 and 4 rates.

Link to comment
Share on other sites

Link to post
Share on other sites

  • 3 years later...

Rewatching these vids.. I seem to remember that the actual issue was found later to be a bad raid card, or backplane.  Does anyone remember this?  A link to the vid would be appreciated

 

Link to comment
Share on other sites

Link to post
Share on other sites

  • 9 months later...
On 3/18/2023 at 8:41 PM, Fr8Train said:

Rewatching these vids.. I seem to remember that the actual issue was found later to be a bad raid card, or backplane.  Does anyone remember this?  A link to the vid would be appreciated

 

Use polling mode, somehow
4:58 in; 

Good engineering. In my experience with our AMD-based NVMe storage server, I learned that AMD has a whole server/data center

platform out there that, as far as I can tell, no one ever thought to go, hey, what happens if I light up all these PCIe lanes with data,

and then actually load up the CPU at the same time? Turns out that the data gets retrieved so fast from that big flash array that,

by the time the CPU comes back around to get it, it's gone.

Ruh-roh.

Wendell made an offhand comment to me that really stood out while we were troubleshooting the issue. I said to him, man, I never saw any of these kinds of stupid issues when we rolled our Intel all in DME storage server, and he goes, "Yeah, Intel's got fewer PCIe lanes, "and they're a generation behind, "but they put some really good engineering into them."

See also


 

 

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×