Testing PCIe 4.0

It's been over a year since the first consumer CPUs and SSDs supporting PCIe 4.0 hit the market, so we're a bit overdue for a testbed upgrade. Our Skylake system was adequate for even the fastest PCIe gen3 drives, but is finally a serious bottleneck.

We have years of archived results from the old testbed, which are still relevant to the vast majority of SSDs and computers out there that do not yet support PCIe gen4. We're not ready to throw out all that work quite yet; we will still be adding new test results measured on the old system until PCIe gen4 support is more widespread, or my office gets too crowded with computers—whichever happens first. (Side note: some rackmount cases for all these test systems would be greatly appreciated.)

AnandTech 2017-2020 Skylake Consumer SSD Testbed
CPU Intel Xeon E3 1240 v5
Motherboard ASRock Fatal1ty E3V5 Performance Gaming/OC
Chipset Intel C232
Memory 4x 8GB G.SKILL Ripjaws DDR4-2400 CL15
Software Windows 10 x64, version 1709
Linux kernel version 4.14, fio version 3.6
Spectre/Meltdown microcode and OS patches current as of May 2018

Since introducing the Skylake SSD testbed in 2017, we have made few changes to our testing configurations and procedures. In December 2017, we started using a Quarch XLC programmable power module (PPM), providing far more detailed and accurate power measurements than our old multimeter setup. In May 2019, we upgraded to a Quarch HD PPM, which can automatically compensate for voltage drop along the power cable to the drive. This allowed us to more directly measure M.2 PCIe SSD power: these drives can pull well over 2A from the 3.3V supply which can easily lead to more than the 5% supply voltage drop that drives are supposed to tolerate. At the same time, we introduced a new set of idle power measurements conducted on a newer Coffee Lake system. This is our first (and for the moment, only) SSD testbed that is capable of using the full range of PCIe power management features without crashing or other bugs. This allowed us to start reporting idle power levels for typical desktop and best-case laptop configurations.

Coffee Lake SSD Testbed for Idle Power
CPU Intel Core i7-8700K
Motherboard Gigabyte Aorus H370 Gaming 3 WiFi
Memory 2x 8GB Kingston DDR4-2666

On the software side, the disclosure of the Meltdown and Spectre CPU vulnerabilities at the beginning of 2018 led to numerous mitigations that affected overall system performance. The most severe effects were to system call overhead, which has a measurable impact on high-IOPS synthetic benchmarks. In May 2018, after the dust started to settle from the first round of vulnerability disclosures, we updated the firmware, microcode and operating systems on our testbed and took the opportunity to slightly tweak some of our synthetic benchmarks. Our pre-Spectre results are archived in the SSD 2017 section of our Bench database while the current post-Spectre results are in the SSD 2018 section. Of course, since May 2018 there have been many further related CPU security vulnerabilities found, and many changes to the mitigation techniques. Our SSD testing has not been tracking those software and microcode updates to avoid again invalidating previous scores. However, our new gen4-capable Ryzen test system is fully up to date with the latest firmware, microcode and OS versions.

AnandTech Ryzen PCIe 4.0 Consumer SSD Testbed
CPU AMD Ryzen 5 3600X
Motherboard ASRock B550 Pro
Memory 2x 16GB Mushkin DDR4-3600
Software Linux kernel version 5.8, fio version 3.23

Our new PCIe 4 test system uses an AMD Ryzen 5 3600X processor and an ASRock B550 motherboard. This provides PCIe 4 lanes from the CPU but not from the chipset. Whenever possible, we test NVMe SSDs with CPU-provided PCIe lanes rather than going through the chipset, so the lack of PCIe gen4 from the chipset isn't an issue. (We had a similar situation back when we were using a Haswell system that supported gen3 on the CPU lanes but only gen2 on the chipset.) Going with B550 instead of X570 also avoids the potential noise of a chipset fan. The DDR4-3600 is a big jump compared to our previous testbed, but is a fairly typical speed for current desktop builds and is a reasonable overclock. We're using the stock Wraith Spire 2 cooler; our current SSD tests are mostly single-threaded, so there's no need for a bigger heatsink.

For now, we are still using the same test scripts to generate the same workloads as on our older Skylake testbed. We haven't tried to control for all possible factors that could lead to different scores between the two testbeds. For this review, we have re-tested several drives on the new testbed to illustrate the scale of these effects. In future reviews, we will be rolling out new synthetic benchmarks that will not be directly comparable to the tests in this review and past reviews. Several of our older benchmarks do a poor job of capturing the behavior of the increasingly common QLC SSDs, but that's not important for today's review. The performance differences between new and old testbeds should be minor, except where the CPU speed is a bottleneck. This mostly happens when testing random IO at high queue depths.

More important for today is the fact that our old benchmarks only test queue depths up to 32 (the limit for SATA drives), and that's not always enough to use the full theoretical performance of a high-end NVMe drive—especially since our old tests only use one CPU core to stress the SSD. We'll be introducing a few new tests to better show these theoretical limits, but unfortunately the changes required to measure those advertised speeds also make the tests much less realistic for the context of desktop workloads, so we'll continue to emphasize the more relevant low queue depth performance.

Samsung 980 Pro Cache Size Effects
POST A COMMENT

137 Comments

View All Comments

  • romrunning - Tuesday, September 22, 2020 - link

    Agreed! Reply
  • Tomatotech - Wednesday, September 23, 2020 - link

    See the StoragePro review for some eyebrow-raising numbers that may change your view.

    https://www.storagereview.com/review/samsung-980-p...
    Reply
  • Tams80 - Tuesday, September 22, 2020 - link

    What's the point of these if they are going to use TLC? They already make TLC SSDs and the market is pretty full.

    So what if using MLC is more expensive? Isn't the whole point of these to be their best of the best? And if an SLC cache benefits TLC SSDs, then why not have an MLC SSD with an SLC cache?
    Reply
  • Kaziglu Bey - Tuesday, September 22, 2020 - link

    Not the hope we're looking for.
    Time for me to buy another 2TB SX8200 Pro.
    Reply
  • vladx - Thursday, September 24, 2020 - link

    Yep, beyond enterprise workloads there's no point buying anything other than the SSD with best price/GB. Reply
  • Makaveli - Tuesday, September 22, 2020 - link

    What were they thinking here.

    600 TBW on this "Pro" drive then a small 114GB SLC cache?

    The E16 offers 1,800 TBW at 1TB and 333GB SLC cache.

    Not impressed with this drive and looking forward to the E18 drives that are slowly coming out.
    Reply
  • Luckz - Thursday, September 24, 2020 - link

    You mean it will just use the entire empty SSD, because leaving the SSD empty is both a 'feature' and a smart use case? I barely even have 10% free on actively used SSDs. Reply
  • Golgatha777 - Tuesday, September 22, 2020 - link

    For $20 more, I'll take a slight performance hit, but gain 4X the endurance by buying a 512GB 970 Pro. There's been nothing worth upgrading to since the 950 Pro when it comes to real world usage anyway though. Reply
  • MDD1963 - Tuesday, September 22, 2020 - link

    I've written 30 TB in 3 years to my 960 EVO...; folks worried about 'only' 600 TBW (60 years at my rate usage?) specs are amusing. Reply
  • Makaveli - Tuesday, September 22, 2020 - link

    And its assuming to me that you think your anecdotal point speaks for the whole market. Reply

Log in

Don't have an account? Sign up now