Cache Hierarchy Changes: Double L3, Faster Memory

Among the biggest changes of the Ryzen 3000, alongside the improved core microarchitecture, is the chip’s overall cache hierarchy. The new chiplet houses CCXes with double the amount of L3, now 16MB instead of 8MB.

Furthermore the chiplet design with the introduction of the cIO die which houses the new memory controllers are undoubtedly going to have an impact on the memory latency and performance of the overall chip.

On the memory controller side particularly, AMD promises a wholly revamped design that brings new support for a whole lot faster DDR4 modules, with the chip coming by default categorized as supporting DDR4-3200, which is a bump over the DDR-2933 support of the Ryzen 2000 series.

AMD had published an interesting slide in regards to the new faster DDR support that went well above the officially supported 3200 speeds, with AMD claiming that the new controllers are able to support up to DDR4-4200 with ease and overclocking being possible to achieve ever higher speeds. However there’s a catch: in order to support DDR4 above 3600, the chip will automatically change the memory controller to infinity fabric clock ratio from being 1:1 to 2:1.

Whilst this doesn’t bottleneck the bandwidth of the memory to the cores as the new microarchitecture has now doubled the bus width of the Infinity Fabric to 512 bits, it does add a notable amount of cycles to the overall memory latency, meaning for the very vast majority of workloads, you’re better off staying at or under DDR4-3600 with a 1:1 MC:IF ratio. It’s to be noted that it’s still possible to maintain this 1:1 ratio by manually adjusting it at higher MC speeds, however stability of the system is no longer guaranteed as you’re effectively overclocking the Infinity Fabric as well in such a scenario.

For this article we didn’t have enough time to dive into the scaling behaviour of the different DRAM speeds, what we did investigate is a more architectural question of how exactly the new chiplet and cIO die architecture has impacted Zen2’s memory latency and memory performance.

To give better insights, we’re using my custom memory latency test that I use for mobile SoC testing and first covered in our review of the Galaxy S10+ and its two SoCs. Memory latency testing nowadays is a complicated topic as microarchitectures advance at a rapid rate, and in particular prefetchers can cause for sometimes misleading figures. Similarly, more brute-force approaches such as full random tests contain a lot of TLB miss latencies which don’t represent the actual structural latency of the system. Our custom latency suite thus isn’t a single one-number-fits-all test but rather a collection of tests that expose more details of the memory behaviour of the system.

The figures published on this page are run on DDR4-3200CL16 on the Ryzen 3900X and 2700X at timings of 16-16-16-36, and the i9-9900K was run with similar DDR4-3200CL16 at timings of 16-18-18-36.

  

Looking at the memory latency curves in a linear plotted graph, we see that there’s some larger obvious differences between the new Ryzen 3900X and the Ryzen 2700X. What immediately catches the eye when switching between the two results is the new 16MB L3 cache capacity which doubles upon the 8MB of Zen1. We have to remind ourselves that even though the whole chip contains 64MB of L3 cache, this is not a unified cache and a single CPU core will only see its own CCX’s L3 cache before going into main memory, which is in contrast to Intel’s L3 cache where all the cores have access to the full amount.

Before going into more details in the next graph, another thing that is obvious is that seemingly the 3900X’s DRAM latency is a tad worse than the 2700X’s. Among the many test patterns here the one to note is the “Structural Estimate” curve. This curve is actually a simple subtraction of the TLB+CLR Thrash tests minus the TLB Penalty figure. In the former, we’re causing as much cache-line replacement pressure as possible by repeatedly hitting the same cacheline within each memory page, also repeatedly trying to miss the TLB. In the latter, we’re still hitting the TLB heavily, but always using a different cache-line and thus having a minimum of cache-line pressure, resulting in an estimate of the TLB penalty. Subtracting the latter from the former gives us a quite good estimate of the actual structural latency of the chip and memory.

Now the big question is, why do it this way? I’ve found that with increasingly better prefetchers, it’s getting difficult in getting good memory latency numbers. Whilst it’s possible to just outright disable prefetchers on some platforms, that avenue isn’t always available.

Precisely when looking at the other various patterns in the graph, we’re seeing quite a large difference between the 3900X and the 2700X, with the 3900X showcasing notably lower latencies in a few of them. These figures are now a result of the new Zen2’s improved prefetchers which are able to better recognize patterns and pull out data from DRAM before the CPU core will handle that memory address.

  

Plotting the same data on a logarithmic graph, we better see some of the details.

In terms of the DRAM latency, it seems that the new Ryzen 3900X has regressed by around 10ns when compared to the 2700X (Note: Just take into the leading edge of the “Structural Estimate” figures as the better estimate) with ~74-75.5ns versus ~65.7ns.

It also looks like Zen2’s L3 cache has also gained a few cycles: A change from ~7.5ns at 4.3GHz to ~8.1ns at 4.6GHz would mean a regression from ~32 cycles to ~37 cycles. Such as change however was to be expected since doubling of the L3 cache structure has to come with some implementation compromises as there’s never just a free lunch. Zen2’s L3 cache latency is thus now about the same as Intel’s – while it was previously faster on Zen+.

Further interesting characteristics we see here is the increase of the capacity of the L2 TLB. This can be seen in the “TLB Penalty” curve, and the depth here corresponds to AMD’s published details of increasing the structure from 1536 pages to 2048 pages. It’s to be noted that the L3 capacity now exceeds the capacity of the TLB, meaning a single CPU core will have only the best access latencies to up to 8MB in the cache before starting to have to page-walk. A similar behaviour we see in the L2 cache where the L1 TLB capacity only covers 256KB of the cache before having to look up entries in the L2 TLB.

Another very interesting characteristic of AMD’s microarchitecture which contrasts Intel’s, is the fact that AMD prefetches all patterns into the L2 cache, while Intel only does so for the nearest cache-line. Such a behaviour is a double-edged sword, on one hand AMD’s cores have can have better latencies to needed data, but on the other hand in the case of a unneeded prefetch, it puts a lot more pressure on the L2 cache capacity, and could in effect counter-act some of the benefits of having double the capacity over Intel’s design.

  

Switching over to the memory bandwidth of the cache hierarchy, there’s one obvious new chance in the 3900X and Zen2: the inclusion of 256-bit wide datapaths. The new AGU and path changes mean that the core is able to now handle 256-bit AVX instruction once per cycle which is a doubling over the 128-bit datapaths of Zen and Zen+.

So while the bandwidth of 256-bit operations on the Ryzen 2700X looked identical to the 128-bit variants, the wider ops now on Zen2 effectively double the bandwidth of the core. This bandwidth doubling is evident in the L1 cache (The flip test is equal to a memory copy test), however the increase is only about 20% for the L2 and L3 caches.

There’s an interesting juxtaposition between AMD’s L3 cache bandwidth and Intel’s: AMD essentially has a 60% advantage in bandwidth, as the CCX’s L3 is much faster than Intel’s L3 when accessed by a single core. Particularly read-write modifications within a single cache-line (CLflip test) are significantly faster in both the L2 and L3 caches when compared to Intel’s core design.

Deeper into the DRAM regions, however we see that AMD is still lagging behind Intel when it comes to memory controller efficiency, so while the 3900X improves copy bandwidth from 19.2GB/s to 21GB/s, it still remains behind the 9900K’s 22.9GB/s. The store bandwidth (write bandwidth) to memory is also a tad lower on the AMD parts as the 3900X reaches 14.5GB/s versus Intel’s 18GB/s.

 

One aspect that AMD excels in is memory level parallelism. MLP is the ability for the CPU core to “park” memory accesses when they are missing the caches, and wait on them to return back later. In the above graph we see increasing number of random memory accesses depicted as the stacked lines, with the vertical axis showcasing the effective access speedup in relation to a single access.

Whilst both AMD and Intel’s MLP ability in the L2 are somewhat the same and reach 12 – this is because we’re saturating the bandwidth of the cache in this region and we just can’t go any faster via more accesses. In the L3 region however we see big differences between the two: While Intel starts off with around 20 accesses at the L3 with a 14-15x speedup, the TLBs and supporting core structures aren’t able to sustain this properly over the whole L3 as it’s having to access other L3 slices on the chip.

AMD’s implementation however seems to be able to handle over 32 accesses with an extremely robust 23x speedup. This advantage actually continues on to the DRAM region where we still see speed-ups up to 32 accesses, while Intel peaks at 16.

MLP ability is extremely important in order to be able to actually hide the various memory hierarchy latencies and to take full advantage of a CPU’s out-of-order execution abilities. AMD’s Zen cores here have seemingly the best microarchitecture in this regard, with only Apple’s mobile CPU cores having comparable characteristics. I think this was very much a conscious design choice of the microarchitecture as AMD knew their overall SoC design and future chiplet architecture would have to deal with higher latencies, and did their best in order to minimise such a disadvantage.

So while the new Zen2 cores do seemingly have worse off latencies, possibly a combined factor of a faster memory controller (faster frequencies could have come at a cost of latency in the implementation), a larger L3 but with additional cycles, it doesn’t mean that memory sensitive workloads will see much of a regression. AMD has been able to improve the core’s prefetchers, and average workload latency will be lower due to the doubled L3, and this is on top the core’s microarchitecture which seems to have outstandingly good MLP ability for whenever there is a cache miss, something to keep in mind as we investigate performance further.

New Microarchitecture, New 7nm Process Node, New Chiplet Design X570 Motherboards: PCIe 4.0 For Everybody
Comments Locked

447 Comments

View All Comments

  • tamalero - Monday, July 8, 2019 - link

    Because you're testing the CPUs not the video cards you clown.
  • 529th - Monday, July 8, 2019 - link

    Anyone know how well one chiplet Overclocks VS a 2 chiplet? I'm thinking one chiplet would not be as limited by temps verse a 2 chiplet.
  • acme64 - Monday, July 8, 2019 - link

    Is there any word on a performance difference while on x470 vs x570?
  • haukionkannel - Tuesday, July 9, 2019 - link

    There is not difference. Even 350 motherboard works with 12 core 3900. There was one test that did that.
  • CaedenV - Monday, July 8, 2019 - link

    *Slow clap*
    Great work AMD!
    I have always been a snobbish Intel user. Back in the late '90s and early '00s it was because the audio software I used simply was not stable on AMD (heck, it was barely stable on Intel lol). Then after the Core2 came out Intel was weirdly the best AND the cheapest for a very long time. But now, AMD really has a lot going for itself, and I am truly impressed.
    Hoping to rebuild my aging (but still pretty great) i7 2600 this fall... mostly because I need it as a home server, not really because I 'need' an upgrade. But I think I am going AMD this time around. I really can't believe how much they have improved in the last 3 years!
  • GreenReaper - Monday, July 8, 2019 - link

    Guys... I get you might not want to adjust your testing base. But MDS/Zombieload makes a significant difference when it comes to system calls, such as just about any file or network access.
    https://www.phoronix.com/scan.php?page=article&...

    The reason for this is that the CPU has to perform a crazy sequence of events when accessing privileged data when two threads on a core are involved, essentially yanking the other thread into kernel mode as well, performing the privileged access that the original thread wanted, then flushing all relevant buffers before returning the two threads, so that the other thread can't run a timing attack using data from them.

    It's a hack, and the impact is potentially worse the more modern the Intel CPU is, because - aside from the Atom - they have had increasingly bigger output buffers, especially Skylake+.

    The OS fixes were out in mid-May when Zombieload was announced, for both Windows and Linux, so I don't know where you're getting "the patches for that have not been released yet".

    Maybe you're thinking firmware for your motherboard to load new microcode? This is typically more of an issue for Windows; on Linux you'd just load the appropriate package. But even here, this doesn't make sense, because (for example) your Z370 i7 Gaming (used for the critical Intel 8/9th Gen CPUs) *does* have a newer BIOS:
    https://www.asrock.com/MB/Intel/Fatal1ty%20Z370%20...

    In fact, much newer. The 4.00 is from May 13, so presumably is relevant to MDS. You seem to be on 1.70, from March 26... 2018. There have been five updates since then. Why was it not updated?
  • zealvix - Tuesday, July 9, 2019 - link

    Yea, saw several articles from other sites that both microcode update from Intel and OS patches from Microsoft have been released.
  • CityBlue - Tuesday, July 9, 2019 - link

    I quickly scanned the comments to see if the benchmarks had been performed with all relevant mitigations installed and was not surprised in the least to discover they hadn't, so frankly this entire article is pointless and I won't waste my time reading it. All there is left to say about this article is that whatever difference Anadtech determined between Intel and AMD, it would have been even more in favor of AMD had all Intel mitigations been enabled.

    Anandtech, Ryan Smith etc., do yourself a favor and re-test your Intel CPUs with *all* mitigations enabled otherwise your Intel benchmarks are just a sham, and you will start to lack credibility. Based on the comments for this article and others your readership are already staying to lose faith in your integrity. Other sites such a phoronix.com are doing a great job detailing the full impact of the mitigations (including Zombieload which you should have tested) so it's hard to take serially your reasons for not testing with a level, real-world playing field (ie. full mitigations). Or maybe you just didn't want to give out a Gold award? :)
  • TEAMSWITCHER - Tuesday, July 9, 2019 - link

    Really getting tired of comments like this. They should just delete them.
  • CityBlue - Tuesday, July 9, 2019 - link

    @TEAMSWITCHER For people that aren't Intel apologists, this stuff matters. Not just because we as consumers want to get an honest review of how the latest AMD hardware stacks up against Intel in a real world situation with all mitigations applied, but also because this elephant in the room is a core credibility issue that Anandtech need to deal with.

Log in

Don't have an account? Sign up now