AMD 3rd Gen EPYC Milan Review: A Peak vs Per Core Performance Balanceby Dr. Ian Cutress & Andrei Frumusanu on March 15, 2021 11:00 AM EST
Section by Ian Cutress
CPU List and SoC Updates
In the past AMD has promoted its EPYC positioning in terms of single socket and dual socket setups, mostly on the basis that one of its larger processors can enable the same compute as two of Intel’s top-tier offerings. This time around, AMD actually does little direct comparison with Intel as to where its processors stand, instead focusing on the market optimizations for different elements of the market.
AMD market optimizations come in three main flavors: Core Performance, Core Density, Balanced and Optimized.
Every processor in AMD’s lineup will come with the following:
- 8 Channels of DDR4-3200
- 4 TB Memory Support (8 channel, 2DPC)
- 128 lanes of PCIe 4.0
- Simultaneous MultiThreading
- Performance Modes (Fixed Power) and Deterministic Modes (Fixed Frequency)
- 18G Infinity Fabric
- Secure Encrypted Virtualization with Secure Nested Pages
- Syncronized 1:1 Fabric and Memory Clock Speeds
Core Performance Optimized: 7xF3 Series
The Core Optimized processors were segmented out on the last page as a highlighted part of the processor lineup, now known as AMD’s F series of processors. These parts have F in the name, offer up to 32 cores, and the main focus here is on individual core performance peaks as well as sustained performance. Compared to other parts of a similar core count, these have higher TDP values, and charge a premium.
|AMD EPYC 7003 Processors
Core Performance Optimized
|EPYC 75F3||32 / 64||2950||4000||256
|EPYC 74F3||24 / 48||3200||4000||240 W||$2900|
|EPYC 73F3||16 / 32||3500||4000||240 W||$3521|
|EPYC 72F3||8 / 16||3700||4100||180 W||$2468|
For this review, one of the processors we have tested is the EPYC 75F3, the 32-core processor offering 4.0 GHz turbo, 2.95 GHz base frequency, and all 256 MB of L3 cache. This processor has four cores per chiplet active, and no doubt these processors use the best voltage/frequency response cores inside each chiplet as manufactured. The $4860 price point is above the $3761 for the next best 32-core processor, showcasing some of that premium.
Users will notice that the 16-core processor is more expensive ($3521) than the 24 core processor ($2900) here. This was the same in the previous generation, however in that case the 16-core had the higher TDP. For this launch, both the 16-core F and 24-core F have the same TDP, so the only reason I can think of for AMD to have a higher price on the 16-core processor is that it only has 2 cores per chiplet active, rather than three? Perhaps it is easier to bin a processor with an even number of cores active.
At the bottom is AMD’s sole 8-core offering, meaning only 1 core per chiplet, and zero contention for L3 cache. It also has a small range in frequency, which should enable deterministic workloads – despite the 180 W listed TDP, a 4.1 GHz Zen 3 core should not need more than 10 W per core, which leaves a lot of power for any configuration that wants to push the IO a little faster (more on that below).
All of these processors are aimed for systems that run software that is limited by single thread workloads, such as EDA tools, or for software that needs a lot of IO but can be limited by per-core licensing restrictions. All of these processors can be use in dual socket configurations.
Core Density Optimized: 48 Cores and Up
As part of the processor lineup, AMD is highlighting all of its high-core count products as being core density optimized, and systems built to take advantage of lower memory channel interleaving configurations (see below) could be built to the scale of 384 cores or 768 threads per 1U.
The new halo top-of-stack processor, the 64-core EPYC 7763, naturally fits into this segment. At 280 W, it adds another 40 W to the top processor publicly available from AMD, along with increasing both base frequency and turbo frequency as well as offering additional IPC gains. The price has increased by just under $1000 to represent where AMD believes it fits into this market.
|AMD EPYC 7003 Processors
Core Density Optimized
|EPYC 7763||64 / 128||2450||3400||256
|EPYC 7713||64 / 128||2000||3675||225 W||$7060|
|EPYC 7663||56 / 112||2000||3500||240 W||$6366|
|EPYC 7643||48 / 96||2300||3600||225 W||$4995|
|P-Series (Single Socket Only)|
|EPYC 7713P||64 / 128||2000||3675||256||225 W||$5010|
AMD considers the EPYC 7763 as ‘a step up’ from the previous top-of-stack processor, the 7742. The direct replacement to the 64-core 7742 in this case is the 7713. We’ve tested both of the Milan and the 7742 Rome for this review, so we get to see how much of an uplift the new processors are.
Also part of the Core Density family is the first single socket processor, the 7713P. This is identical to the 7713, however does not support 2P configurations. As a result it comes in at $2000 cheaper.
The other interesting member of the family is the 7663, a 56 core processor. This configuration means that AMD is using 7 cores per chiplet, rather than the full 8, which is a configuration that AMD did not have with the previous generation. It would appear that AMD’s customers have requested a model like this, optimized for their workloads where less cache contention might be needed, or for workloads that won’t scale all the way to 64 cores.
Balanced and Optimized Portfolio
The rest of AMD’s portfolio of Milan processors comes under the ‘Balanced and Optimized’ banner, and this is where AMD will offer its 16-32 core processors as well as the rest of the 1P single socket parts.
|AMD EPYC 7003 Processors|
|EPYC 7543||32 / 64||2800||3700||256 MB||225 W||$3761|
|EPYC 7513||32 / 64||2600||3650||128 MB||200 W||$2840|
|EPYC 7453||28 / 56||2750||3450||64 MB||225 W||$1570|
|EPYC 7443||24 / 48||2850||4000||128
|EPYC 7413||24 / 48||2650||3600||180 W||$1825|
|EPYC 7343||16 / 32||3200||3900||190 W||$1565|
|EPYC 7313||16 / 32||3000||3700||155 W||$1083|
|P-Series (Single Socket Only)|
|EPYC 7543P||32 / 64||2800||3700||256 MB||225 W||$2730|
|EPYC 7443P||24 / 48||2850||4000||128 MB||200 W||$1337|
|EPYC 7313P||16 / 32||3000||3700||155 W||$913|
Almost all of these processors half the L3 cache to 128 MB, suggesting that these processors only have four active chiplets inside – this allows AMD to optimize its silicon product rather than providing lots of only-half enabled chiplets. This is seen perhaps in the price, as the 8-chiplet 32-core EPYC 7543 is almost $1000 more expensive than the EPYC 7513.
AMD only goes down to 16 cores here – the only 8 core processor is the 72F3 mentioned earlier. Also perhaps striking is that AMD does not go below 155 W TDP, which as we will see later in the review, might be down to some of the IO.
Bonus points for anyone spending $1337 on a processor like the EPYC 7443P.
For those wanting to see all the processors in one table with all the configurable TDP options, here we are. Processors we are testing in this review are highlighted in bold.
|AMD EPYC 7003 Processors|
min / max
|EPYC 7763||64 / 128||2.45||3.40||256
( 8 x 32 )
|280 / 225 / 280||$7890|
|EPYC 7713||64 / 128||2.00||3.675||225 / 225 / 240||$7060|
|EPYC 7663||56 / 112||2.00||3.50||240 / 225 / 240||$6366|
|EPYC 7643||48 / 96||2.30||3.60||225 / 225 / 240||$4995|
|EPYC 7543||32 / 64||2.80||3.70||225 / 225 / 240||$3761|
|EPYC 7513||32 / 64||2.60||3.65||128
( 4 x 32 )
|200 / 165 / 200||$2840|
|EPYC 7453||28 / 56||2.75||3.45||64
( 4 x 16 )
|225 / 225 / 240||$1570|
|EPYC 7443||24 / 48||2.85||4.00||128
( 4 x 32 )
|200 / 165 / 200||$2010|
|EPYC 7413||24 / 48||2.65||3.60||180 / 165 / 200||$1825|
|EPYC 7343||16 / 32||3.20||3.90||190 / 165 / 200||$1565|
|EPYC 7313||16 / 32||3.00||3.70||155 / 155 / 180||$1083|
|EPYC 75F3||32 / 64||2.95||4.00||256
( 8 x 32 )
|280 / 225 / 280||$4860|
|EPYC 74F3||24 / 48||3.20||4.00||240 / 225 / 240||$2900|
|EPYC 73F3||16 / 32||3.50||4.00||240 / 225 / 240||$3521|
|EPYC 72F3||8 / 16||3.70||4.10||180 / 165 / 200||$2468|
|P-Series (Single Socket Only)|
|EPYC 7713P||64 / 128||2.00||3.675||256
( 8 x 32 )
|225 / 225 / 240||$5010|
|EPYC 7543P||32 / 64||2.80||3.70||225 / 225 / 240||$2730|
|EPYC 7443P||24 / 48||2.85||4.00||128
( 4 x 32 )
|200 / 165 / 200||$1337|
|EPYC 7313P||16 / 32||3.00||3.70||155 / 155 / 180||$913|
To fill in some of the demands in costs and offering, AMD is also planning to supply 2nd Gen EPYC into the market for a good time, enabling both platforms for customers. These include the 32-core 7532, 24-core 7352, 16-core 7282, and a couple of 8-core processors.
Comparing Gen-on-Gen Pricing
Direct comparisons are always made against the previous generation, and these are the suggested comparisons:
|AMD Third Gen EPYC
|AMD Second Gen EPYC
Overall there is an uptick from Rome to Milan on pricing.
New Features and SoC Design
In our interview with AMD’s Forrest Norrod, he explained that the original goal of Milan was to update Rome’s 8-core chiplets from Zen 2 to Zen 3, but a number of factors meant that the central IO die also received updates, both on the power side and for Infinity Fabric performance.
Beyond the core updates from Zen 2 to Zen 3, which we’ve covered extensively in our consumer coverage, AMD has enabled several key features with Milan.
Memory Interleaving for 4/6/8 Channel Configurations
The enterprise and datacenter markets involve a vast array of potential workloads, some are compute-bound, some are memory-bound, but the goal has always been to get the work done as quickly as possible, as efficiently as possible, and as cheaply as possible. The larger companies will often optimize their deployments in every way imaginable, and that includes memory.
AMD admits that there is a portion of the market that is very much memory-bound, either in capacity, latency, or bandwidth. These customers will keep demanding higher capacity support, faster memory, or just more memory channels for both capacity or bandwidth. We’ve been told that there is a holistic split at some level between these sorts of customers, and those that are simply compute-bound, which may come to a head with differentiated products in the future. But for now, we have eight-channel DDR4 processors on the market, which is not-enough for some customers, but plenty for others who don’t need all that capacity or bandwidth.
It’s the latter customers where optimizations may occur. For example, if memory bandwidth or capacity is not a limiting factor, then those customers can design systems with fewer memory channels being used. This helps by increasing density, making physical system design simpler, perhaps cooling, but also by disabling those memory channels, more power is available for their compute-bound systems. We came to a point in the previous generation lifecycle where we were starting to see motherboards enter the market with fewer than eight memory slots for these reasons.
In order to get the best performance, or reliable performance, from the memory (and not experience any NUMA or silo effects), the memory installed needs to have effective memory interleaving enabled on what is installed. AMD EPYC in the past has supported 8-channel interleaving and 4-channel interleaving, but for Milan, AMD is also supporting 6-channel interleaving for customers that build in those configurations. Four and Eight are obvious powers of two for AMD to have included by default, but enough demand for Six was there to enable an update to the IO die to support this mode.
AMD has confirmed that 6-channel memory interleaving, when using one module per channel, will be supported on all Milan processors.
Security for Return Programming and Secure Nested Pages
The effects of Spectre, Meltdown, and its variants have repurposed processor security to be more proactive at all the major microprocessor designers and manufacturers. AMD has implemented fixes in hardware for the Spectre variants to which it was susceptible, and hasn’t needed to add additional protection from Meltdown and its variants as AMD CPUs are not naturally vulnerable.
The new segment of attacks beyond these are based on return address programming, and effectively adjusting the control flow of code on the processors by manipulating stack return addresses or jump/control commands. AMD implemented a fix for these sorts of attacks in its Zen 3 core, and these fixes get passed into EPYC Milan, known as ‘Shadow Stacks’. These are hardware based solutions, enabling a minimal effect to performance, but absolutely required for security.
The other segment to AMD’s security updates is Secure Nested Paging, part of AMD’s Secure Encrypted Virtualization technology. These technologies all deal with hyperscalers enabling multiple host instances on a single system, and being able to protect the instances from each other. In previous generation Rome, Secure Encrypted Virtualization allowed a secure hypervisor to isolate these instances or virtual machines from each other in a secure and encrypted way (hence the name).
Secure Nested Paging takes this one stage further, and enables additional security of the virtual machines from the hypervisor itself. This allows protection against an untrusted hypervisor, should an attacker get access, as well as support for software running on unmodified guest operating systems for which the software was originally not built for.
Enhanced Memory and IO Performance
One of the big updates from the Zen 2 core complex to the Zen 3 core complex in all of AMD’s product families was the size of that complex, moving from four cores to eight cores. This means that in an eight core chiplet, rather than there being two complexes of four cores and a separated L3 cache, there is now a single eight core complex with a unified cache. As seen in our consumer reviews, this matters a lot for memory-heavy operation as well as branch heavy code. Another benefit is that the control electronics for the complex are now at the edge of the chiplet, providing some small improvement.
We’ll go into how exactly the cache structure changes later in the review.
The other angle to Milan’s performance in IO is on the Infinity Fabric and PCIe. For this generation, AMD has enabled 18 Gbps Infinity Fabric links over its SERDES connections, up from the previous generation 16 Gbps links. This might come across as slightly minor, but it should enable better performance in a competitive core-to-core communications environment.
AMD is also supporting Extended Speed Modes on its PCIe links with Milan. This is part of the PCIe 4.0 standard, and allows customers to enable more bandwidth over the PCIe link to discrete accelerator cards if the motherboard is designed to support those speeds. This requires extra power from the processor, which does take power/performance away from the processor cores, however AMD has stated that because its customers have so many different optimization points, a number of workloads will benefit from this option.
Other Improvements and ISA Enhancements
Also on AMD’s list of updates to Milan are the following:
|INVLPGB||New instruction to use instead of inter-core interrupts to broadcast page invalidates, requires OS/hypervisor support|
|VAES / VPCLMULQDQ||AVX2 Instructions for encryption/decryption acceleration|
|SEV-ES||Limits the interruptions a malicious hypervisor may inject into a VM/instance|
|Memory Protection Keys||Application control for access-disable and write-disable settings without TLB management|
|Process Context ID (PCID)||Process tags in TLB to reduce flush requirements|
|INT8||Dual INT8 pipes, up from 1, doubles INT8 performance|
Post Your CommentPlease log in or sign up to comment.
View All Comments
mode_13h - Monday, March 15, 2021 - linkPlease don't paint Nvidia as a victim. They are not. All of these guys will have to support each other, for the foreseeable future, and for purely pragmatic reasons.
Oxford Guy - Monday, March 15, 2021 - linkThey are not 'guys'. They're corporations. Corporations were invented to, to quote Ambrose Bierce, grant 'individual profit without individual responsibility'.
mode_13h - Wednesday, March 17, 2021 - linkNo disagreement, but I'm slightly disheartened you decided to take issue with my use of the term "guys". I'll try harder, next time--just for you.
Oxford Guy - Tuesday, April 6, 2021 - linkPeople humanize corporations all the time. It doesn't lead to good outcomes for societies.
Of course, it's questionable whether corporations lead to good outcomes, considering that they're founded on scamming people (profit being 'sell less for more', needing tricks to get people to agree to that).
chavv - Monday, March 15, 2021 - linkIs it possible to add another "benchmark" - ESX server workload?
Like, running 8-16-32-64 VMs all with some workload...
Andrei Frumusanu - Monday, March 15, 2021 - linkAs we're rebuilding our server test suite, I'll be looking into more diverse benchmarks to include. It's a long process that needs a lot of thought and possibly resources so it's not always evident to achieve.
eva02langley - Monday, March 15, 2021 - linkJust buy EPYC and start your hybridation and your reliance on a SINGLE supplier...
eva02langley - Monday, March 15, 2021 - linkedit: Just buy EPYC and start your hybridation and STOP your reliance on a SINGLE supplier...
mode_13h - Monday, March 15, 2021 - linkYou guys should really include some workloads involving multiple <= 16-core/32-thread VMs, that could highlight the performance advantages of NPS4 mode. Even if all you did was partition up the system into smaller VMs running multithreaded SPEC 2017 tests, at least that would be *something*.
That said, please don't get rid of all system-wide multithreaded tests, because we definitely still want to see how well these systems scale (both single- and multi- CPU).
ishould - Monday, March 15, 2021 - linkYes this seems more useful for my needs as well. We use a grid system for job submission and not all cores will be hammered at the same time