The Intel Core i9-7980XE and Core i9-7960X CPU Review Part 1: Workstation
by Ian Cutress on September 25, 2017 3:01 AM ESTNew Features in Skylake-X: Cache, Mesh, and AVX-512
We have covered this in previous Skylake-X reviews, but it is worth a refresher about what is new in these Skylake-X processors over previous generations. The interesting thing to note is that the core design in these processors is different to the consumer-grade Skylake (known as Skylake-S) processors. This is a distinct change in policy from previous generations, where we saw parity on the base microarchitecture design. The changes for Skylake-X over Skylake-S are three fold: a change in the L2/L3 cache arrangement, a new routing mechanism for cores to send data, and the inclusion of AVX-512 units.
In our initial Skylake-X review, we tested how the new changes directly effect IPC / performance against Skylake-S in our benchmark suite. Overall they have a positive effect, with one major exception as noted below.
Cache Me If You Can: More Private L2 Cache
Modern Intel x86 processors use a three level cache design, known as L1, L2, and L3. The L1 cache is the one nearest to where the action happens in the processor core, so it is the fastest but also the smallest. When data is not in the L1 cache, a request is made to L2 cache, which is bigger but slightly slower. L3 follows the same route: bigger but slower. An ideal processor would have all of the data it needs always in the L1 cache in order to perform the best, but software is not always that nice: you cannot load that 4MB cat gif into a 1MB L2 cache.
In all of Intel’s Core microarchitecture designs, both the L1 and L2 caches are private to each core, with the L3 shared among all cores. This is achieved by giving each core a ‘slice’ of L3 cache and having appropriate tags, but allowing all the other cores to pull data from it when needed. The L3 level is typically where cores can ‘share’ data without going out to main memory (discussing simply here, rather than going into complex ideas such as snooping). In all previous Core designs, including Skylake-S, these caches were called inclusive caches: in each core, the L2 would contain a copy of L1, and the overriding L3 would contain a copy of every L2. This means that if a cache line in the L2 was invalidated, it would still be present in the L3 if needed later. This improves the ‘hit rate’ (getting data from a lower level cache), but puts demands on cache size: if a 10-core CPU has 10 L2 caches (one per core) at 256 KB each, the L3 cache must at least be 256 KB per core. Cache is not cheap and takes up a lot of die area, so it becomes a balance of performance and cost.
Inclusive Caches in Skylake-S: Green for L1 data, Yellow for L2 data, Red for L3 data
For Skylake-X, Intel does three things: it increases the size of the private L2 cache, which increases the hit rate for loading data, from 256 KB per core to 1 MB per core. To compensate for the increase in die area, Intel reduced the size of the size of the L3 from 2.5 MB per core to 1.375 MB per core, keeping the overall L2+L3 constant. The third element is inclusivity of the L3 cache: with it only being 1.375 MB per core, it no longer made sense for it to also hold 1 MB of the L2 data, and so with Skylake-X the L3 cache is now a non-inclusive cache (basically an exclusive cache) with the inability to obtain pre-fetch data from DRAM.
Moving the L3 to a non-inclusive cache with the inability to obtain data direct from DRAM turns it into what is known as a victim cache. When data comes in from DRAM into the L2, is used and then retired/overwritten, a copy is moved into the L3 (the L3 obtains the ‘victim’ cache line). If the data is then needed again in quick succession, there will be a copy in L3 rather than moving all the way out into DRAM. This situation depends highly on the software being used: software that has a large L2 cache requirement (e.g. 2MB) but frequency reuses data can take advantage of this scenario. Unfortunately not a lot of software is like this, and the L3 cache can become almost a dead-weight. Historically we see victim caches perform best with software that has a memory-speed bottleneck, or with integrated graphics. Skylake-X has no integrated graphics.
The upside is the size of the L2 private cache. Having spoken with engineers from different companies, a doubling of the size of the L2 cache often leads to a 1.414x decrease in cache misses (when data is not present and has to move out to the next level of cache). So by increasing the cache from 256 KB to 1 MB, thereby doubling the cache and doubling it again, the cache miss rate should be halved in total, leading to what most engineers would consider a 4-10% general IPC increase. The downside of a larger cache is that it takes longer to process data from it, adding latency. Keeping the latency low is difficult, but Intel claims that it has spent a lot of R&D effort here, with the L2 latency only moving from 11 cycles in Skylake-S to 13 cycles in Skylake-X. This will have a slight knock on that 4-10% IPC increase, but it does provide an overall benefit in almost all scenarios.
Making a Mesh: Replacing Rings with Routers
When reading about high performance processors, one phrase will always crop up: ‘feeding the beast’. In order for a CPU core to run at peak throughput, it has to continually have data to work on – if it does not have data to work on, it will sit idle and the throughput will drop. A lot of the core design is devoted to this problem: multi-level caches can store megabytes of data ready to go, and each generation brings a smarter pre-fetch engine to pull the right data from memory into cache at the right time. We’ve spoken about caches, but when a core needs data from the cache of another core, or needs to send/receive data from main memory, this also becomes a very important element to the design.
Most almost all of Intel’s processors, except for Xeon Phi, the solution to this problem has been bidirectional rings connecting the cores together (known as ring buses). The ring would allow data to travel in both directions, and it would have stations at each core (technically it grouped cores into pairs) as well as the main memory and system agent (IO). For the largest enterprise cores, in order to keep latency low, Intel separated the cores into two sets with some overlap, and a ring in each that had stations which could communicate with each other. As long as the cores and software kept the data it needed close by, latency was fairly low, and successive generations had led to higher frequencies and optimized design. The downside of a ring is scalability: the more cores you put in, the more variable the latency and more inconsistent performance can occur.
A dual ring design with Intel’s Xeon E5 v4 24-core processors
With Skylake-X, Intel felt that the ring strategy was not the right solution, and implemented its Xeon Phi strategy instead.
Rather than each core being a station on a ring, each core becomes a node in a two-dimensional grid or ‘mesh’. The node acts like a router: it is connected to the core above, the core below, and the core either side, and can direct data in each direction or drop it off at the core the node is on, similar to how a router works. In the scientific literature, this is seen as a decoupled crossbar (a central routing block), with the ‘decoupled’ bit being that every core acts like a localized crossbar.
We believe that Intel is using a design known as MoDe-X, which simplifies the router at each stage to a very small routing mechanism, optimized through wiring lengths as these designs can burn power proportional to wire length. By optimizing in this way, Intel wants to reserve the power more for cores and IO. We go into greater depth on how MoDe-X works in our initial Skylake-X review.
So far the reaction to the mesh method has been mixed. While it offers scalability over using rings, it has not had over a decade of optimization, and some users have pointed to the frequency (usually 2.4 GHz) as being a bottleneck in their software over the faster ring design. Intel is likely to continue with the mesh design for the next few generations, so it will be interesting to see what upgrades are made (if disclosed).
I’ve Got 512 Problems but AVX Ain’t One: Accelerating Compute
Both the cache adjustment and the mesh networking are redesigns of features already implemented in the core. AVX-512 by contrast is a distinct addition to the design, requiring a fairly substantial amount of die area (approx. 20% of a Skylake-S core). Much like AVX and AVX2 before it, the point of AVX-512 is bundling sets of data together and performing the same operation on all the data with one instruction, rather than eight or sixteen instructions.
Bundling data together to unify a single instruction is at the heart of many dense compute tasks, and most prosumer and enterprise software with high compute requirements is designed with using instructions like this to accelerate performance. Most professional software today will take advantage of AVX or AVX2 instructions, with Intel expecting the same companies to update their software with AVX512 support soon.
Due to the density of the computation, the localized energy consumption is often very high (higher than in the standard core logic), and draws a lot of power – to compensate the core will often run at a lower frequency when these instructions are used, but there is an overall net gain. The support of AVX-512 is meant to be a highlight of Intel’s enterprise processors, but due to the unified design between enterprise and HEDT, consumers are able to take advantage as well, as long as their software does. For anyone using these new processors for light tasks, or gaming, are unlikely to see any advantage through AVX-512, announcing for a lot of unused silicon on the shiny new processor.
For the new processors launched today, all of them will support dual 512-bit FMA execution through a single 512-bit FMA on port 5, and a two joined 256-bit FMAs on ports 0 and 1. Intel had noted that the six and eight core parts on Skylake-X were manufactured with the port 5 FMA disabled, however several outlets have reported that both FMAs are detected for these parts.
There is a lot to dissect with AVX-512, such as the different instruction support as well as a few new (to Intel) implementations such as mask registers and small-datatype support. We covered this in detail in our initial Skylake-X review.
152 Comments
View All Comments
Spunjji - Monday, September 25, 2017 - link
Super relevant, because they indicate how badly thermally limited the CPU is - which is hella good info to have if you're, say, considering delidding a $1999 processor because the manufacturer used toothpaste under the IHS.tricomp - Monday, September 25, 2017 - link
Poor AMD...No chance they are going to supply (even more) cpu's demand after posting this article..
I am trying to purchase at list 7 systems for my customers in my country but there's nowhere I can find them beasts here..
iwod - Monday, September 25, 2017 - link
I wish someone could do an article on that too. GF doesn't seems to be the limitation here. GF, should in theory more then enough capacity in their Fab 8 for AMD. Unless GF have some other big customers, otherwise AMD should really be bumping out as much unit as possible.Atom11 - Monday, September 25, 2017 - link
Can we please see one test (!), if you could possible manage, that shows the advantage of AVX-512 in compare to AVX2 when doing:1.) matrix multiply
2.) FFT
3.) convolution
ZeDestructor - Monday, September 25, 2017 - link
Give us a comparison to AVX1 and SSE4 too!Gothmoth - Monday, September 25, 2017 - link
threadripper delivers 80+% of the perfromance for less than 50% of the price.... you don´t have to be a genius to see what the better deal is (price germany: TR 1950x = 950 euro, 7890xe =2300 euro)Spunjji - Monday, September 25, 2017 - link
Don't let that stop them equivocating about how companies who need that power yet somehow have no need for ECC don't care about cost because something something software TCO blah blah.spdragoo - Monday, September 25, 2017 - link
I'm trying really, really hard to think of a company that, at some point or another, doesn't say, "Equipment X may outperform Equipment Y, but the extra cost to buy Equipment X is too much, we'll just make-do with Y instead." Especially since 100% of companies have a limit on their budgets.What's that, you say? Multi-billion dollar corporations don't have to worry about the money they spend? Someone apparently didn't pay attention in their Econ 200 class, or their Introduction to Accounting coursework.
By definition, every business has a *finite* amount of money they can spend, based on a) how much money they collect from their customers, b) how much they can recoup on the sale of assets (tangible or intangible), & c) how much they can get from "other sources" (mostly bank loans or by selling stock shares, or sometimes government grants, but you might find the occasional situation where a generous benefactor just bequeaths money to a company...but I doubt you'll even see that happen to 1% of the companies out there -- & no, venture capitalists pouring money into a company is *not* a situation where they "give the money away", they're getting something for their money, usually stock shares or guarantees of repayment of the loans). Of that money, some of it is earmarked for employee compensation (not just the executives, but the office drones & lower-level employees that do 99% of the actual work), some of it goes towards taxes, some of it pays for rental payments, some for loan payments, some for utilities (telephone, Internet, electricity, gas, water, etc.), some of it may get set aside for "emergencies", some gets earmarked for dividends to the shareholders, etc. That means that a (relatively) small portion is set aside for "equipment replacement". Now, if the company is lucky, the lion's share of that budget is for IT-related equipment...but that covers more than just the office drones' machines, that covers everything: server racks, storage services, cloud vendor payments, etc.
And that is where the price comes into play. For probably 90% of office users out there, not only is Threadripper an overpowered product, so are these products. Heck, we're in the middle of an upgrade from Windows 7 to Windows 10, & they're taking the opportunity to replace our old Sandy Bridge i5 machines with Skylake i7 machines. Sure, they're running faster now...but the main reason they're running faster is because we went from 32-bit Windows to 64-bit Windows, so our PCs now have 8GB of RAM instead of 4GB. That helps with our workload...which primarily revolves around MS Office & using browsers to access & modify a number of massive databases. Having an 8C/16T CPU, let alone a 16C/32T CPU, wouldn't provide any boost for us, since the primary slowdown is on the server side.
These are going to be expensive systems for specialized purposes...& those individual companies are going to look at their budgets very closely, as well as the performance benchmarks, before deciding to purchase these systems. Sure, they may hold the performance crown...but not by that big of a margin, & especially when compared to the margin that gives them the "most expensive price" crown.
BrokenCrayons - Monday, September 25, 2017 - link
Human labor is more expensive than hardware. The 20% additional performance for $1000 more can be earned back quickly by the increased productivity of your workforce (assuming your management staff is effective enough to keep the employees gainfully employed of course and that's certainly not always the case).vladx - Tuesday, September 26, 2017 - link
Indeed the difference in price is pretty much negligible in a professional setting.