Hot Chips 33 (2021) Schedule Announced: Alder Lake, IBM Z, Sapphire Rapids, Ponte Vecchioby Dr. Ian Cutress on May 18, 2021 8:00 AM EST
Once a year the promise of super hot potatoes graces the semiconductor world. Hot Chips in 2021 is set to be held virtually for the second successive year, and the presentation schedule has just been announced. Coming this August, there will be deeper disclosures on next-generation processor architectures, infrastructure compute platforms, new enabling technologies such as processing-in-memory, a number of upcoming AI solutions, as well as a deeper look into custom accelerators.
If you thought last year's Hot Chips was a good conference, this one is a strong competitor. Hot Chips is an annual two-day semiconductor disclosure event where the latest processor technology from around the industry (except Apple*, see at the bottom) is presented by the engineers behind the projects. A number of key players use Hot Chips as the first opportunity to promote key details of their designs in the market for potential customers, and startups with enough backing also get to talk about what makes their new chips unique in a crowded market.
A highlight of each event is also the keynotes, with previous years involving Dr. Lisa Su, CEO AMD discussing the companies success, Dr. Phillip Wong of TSMC giving the lay of the land at the leading edge of manufacturing, Jon Masters of Red Hat going through a number of the issues stemming from Spectre and Meltdown, and Raja Koduri covering Intel's vision of a full scalar-vector-matrix-spatial XPU strategy.
TSMC Keynote, Hot Chips 31 (2019)
Normally Hot Chips is an on-location event, but similar to last year due to restrictions on travel it will be an all-virtual event again. This means that most presentations are pre-recorded, but there was a lot of interaction at the event last year. Anyone can attend, and the virtual prices are low, at most $160, which provides the attendee with live presentations, a chance to ask Q&A, access to all the slide decks, and continued access to the talks for several months before they are made public. Last year's online solution went really well.
This year's event will be held August 22-24th, and will run to the Pacific Time Zone. All times below are in PT.
|Hot Chips 33 (2021) Schedule|
|AnandTech||Time and Session||Session Title|
|08h30||Tutorial 1||ML Performance|
|14h00||Tutorial 2||Advanced Packaging|
|11h30||Session 2||Academic Spinout Chips|
|14h30||Session 3||Infrastructure and Data Processors|
|17h30||Session 4||Enabling Technologies|
|08h30||Session 1||ML Inference for the Cloud|
|10h00||Keynote||Department of Energy|
|11h30||Session 2||ML and Computation Platforms|
|14h30||Session 3||Graphics and Video|
|17h00||Session 4||New Technologies|
Here's a quick overview of the Hot Chips 2021 schedule.
Day 0: Tutorial Day
Because Hot Chips caters to both professionals and students, the pseudo-first day of the event is typically a chance for attendees to get to grips with new topics in the industry. Of late these sessions have covered topics such as building scaleout systems, quantum computing, new networking paradigms, and security.
|Hot Chips 33 (2021): Tutorial Day
August 22nd, Sunday
|08h30 - 13h00||Machine Learning Performance|
|Hardware and software co-optimization of the industry-standard MLPerf benchmarks, as well as applications, performance characteristics, key challenges, and considerations for those deploying unique workloads|
|14h00 - 17h15||Advanced Packaging|
|How advanced packaging techniques enable performance and density improvements, covering how current technologies in the market work, how they are used, and the cutting edge of packaging and chip design by the industry leaders|
The first tutorial here is an expansion of previous talks by MLCommons, the incorporated industry body behind MLPerf. Over the past year we have seen the benchmark reach a full v1.0 with respect to inference, and much in the same way that the industry-standard SPEC benchmarks are optimized to the n-th degree, this session is here to assist with how companies can optimize their hardware and software stack to get the best MLPerf results.
The second tutorial sounds really interesting. Packaging (and interconnect) are the next frontiers of scaled computational resources, with lots of research from the big players already put to use in modern mobile processors to big AI chips. This session is likely to cover TSMC's 3DFabric family of packaging methods, along with associated roadmaps that were disclosed last year, but also Intel's EMIB, Foveros, and ODI packaging. Other companies with advanced packaging products are also likely to get involved in how they use TSMC's and Intel's designs.
Day One: Morning
Day One is going to be very busy, and is split into six sessions, from 8:45 am to 7pm PT.
As with any conference, the opening minutes are spent detailing the conference, what’s new for the year, and some of the rules (such as no streaming). I suspect there will be a large discussion about how the COVID situation will affect the presentations, what to do if one of the presentations fails, or such. I actually hope that the presentations are pre-recorded so that doesn’t happen.
The first session is on Processors.
|Hot Chips 33 (2021): Day One, Session 1
|09h00||Efraim Rotem||Intel||Intel Alder Lake CPU Architectures|
|09h30||Mark Evers||AMD||AMD Next Generation Zen 3 Core|
|10h00||Christian Jacobi||IBM||The > 5 GHz next-generation IBM Z processor chip|
|10h30||Arijit Biswas||Intel||Next-Generation Intel Xeon CPU
The first official day of presentations always starts with discussing leading-edge processors, and this year looks to be a stellar set of talks.
First up is Intel discussing Alder Lake, its second-generation heterogeneous processor architecture (after Lakefield) that is set to be the next-generation processor for both desktops and high-end laptops. We already know that Alder Lake will use both Golden Cove and Gracemont microarchitectures, to the hope here is that Intel will spend time going deep into both. Normally this is the sort of thing they would have disclosed over several hours at an Intel-specific event, and given 30 minutes for the talk I wonder how much will actually be disclosed - it might instead be a talk solely about the SoC and we won't get microarchitecture detail at all.
Second is an AMD talk about its latest Zen 3 core microarchitecture. As Zen 3 was launched into the market in Q4 last year, with an updated back-end and unified L3 cache structure, I doubt we will see anything new in this talk. The hardware has been thoroughly tested; AMD typically uses Hot Chips to refresh what's already out in the market, and there's an RDNA2 talk on the second day which is expected to be of a similar nature.
Third is IBM discussing its next-generation mainframe architecture and product line, the Z processor. We've covered IBM discussing z14 and z15 in previous Hot Chips events, and so this is either a deeper dive into z15 (which was presented last year in lots of detail) or a new look at an upcoming z16 design. The Z mainframe solution usually consists of compute processors and control/cache processors across a unified multi-rack approach - because this talk is titled 'processor chip', I suspect it is more about the compute processor than the solution, but hopefully there will be a slide or two on how it all fits together.
The final talk of the session is another Intel talk, this time discussing the upcoming next-generation Sapphire Rapids platform, set to launch either sometime at the end of this year or early next year (Intel has a contract with DoE for the Aurora supercomputer it needs to fill by the end of the year with this part, so general availability might be after). Sapphire Rapids is using Intel's 10nm Enhanced SuperFin process, and the same Golden Cove cores mentioned in Alder Lake, though perhaps cache optimized for server use. I suspect this talk will be heavy on the die configuration and new elements, such as PCIe 5.0 and DDR5.
After a short break, we get to the Academic section.
|Hot Chips 33 (2021): Day One, Session 2
Academic Spinout Chips
|11h30||Karu Sankaralingam||University of Wisconsin-Madison||Mozart, Designing for Software Maturity and the Next Paradigm for Chip Architectures|
|12h00||Todd Austin||University of Michigan||Morpheus II: A RISC-V Security Extension for Protecting Vulnerable Software and Hardware|
It's not often we pay too much attention here given the research nature of the devices, however the second talk on a RISC-V security extension is rather interesting. Coming from the University of Michigan, the Morpheus II core has been reported as being the target for 500+ cybersecurity researchers for 3 months as part of the DARPA red-teaming challenge, and had zero penetrations in that time.
Day One: Keynote One
|12h30||Aart de Geus||Synopsys||Synopsys Keynote|
This year is a little different to most, with the first day having two separate keynotes. First up is an untitled presentation from the CEO of Synopsys, a company well known in the industry for its EDA (electronic design automation) tools. This includes logic synthesis, place and route, static timing analysis, hardware language simulators, and transistor-level circuit simulation. This talk should give the company a chance to discuss its next-generation technologies, especially as we move into an era of 3D design and packaging.
When we get the full title of the talk, this segment will be updated.
Day One: Afternoon
After lunch the next series of talks are on non-standard processor designs, typically optimized for infrastructure or data processing.
|Hot Chips 33 (2021): Day One, Session 3
Infrastructure and Data Processors
|14h30||Andrea Pellegrini||Arm||Arm Neoverse N2: Arm's second-generation high-performance infrastructure CPUs and system products|
|15h00||Idan Burstein||NVIDIA||NVIDIA Data Center Processing Unit (DPU) Architecture|
|15h30||Bradley Burres||Intel||Intel's Hyperscale-Ready Smart-NIC for Infrastructure Processing|
The first talk is regarding Arm's Neoverse N2 core, the upgraded model of the N1, which was announced in late April. N2 products aren't due out until next year, however companies working with N2 are likely already designing their SoCs with the core. Arm didn't disclose many of the pipeline details in the April announcement, and so we might see more information along these lines, however there is the potential for it to just be the same announcement as April. We shall wait and see.
The second talk from NVIDIA is about its Data Processor Unit architecture, also known as Bluefield. Coming from its acquisition of Mellanox, the Bluefield line of DPUs enables Smart-NIC like network acceleration by integrating a network controller, general-purpose compute cores, and PCIe connectivity into the same device. The latest product is Bluefield-2, however murmurings of Bluefield-3 have been made as to the future roadmap replacement. The title of the talk does not specifically state NVIDIA will talk about current generation architectures or next generation.
Third up is the Intel SmartNIC solution which has been announced previously as the C5000X platform. A number of Intel's customers and OEM partners, are already shipping the hardware, based on Intel Altera FPGAs, to key customers such as Baidu. Partners such as Silicom are selling the parts under their own brand. There hasn't been much presentational material about the architecture of the SmartNIC, so this might be an interesting insight into one of Intel's new market pushes.
Day One: Keynote Two
In an interesting turn of events, the first day of Hot Chips has two keynotes.
|16h00||Abraham Bachrach||Skydio||Skydio Autonomy Engine: Enabling the Next Generation of Autonomous Flight|
The second keynote of the day comes from Skydio, a company I had not heard of before the announcement, but appears to be on the leading edge of AI-based pilot technology. The Skydio 2 platform for example seems to be powered by an NVIDIA Tegra TX2, and the concept of autonomous flight / drone technology has always been on the cusp in the evolution on AI. Skydio looks set to talk about its next-generation platform, given that its Skydio 2 was released in 2019 and since then it has had another round of VC funding.
Day One: Even More Talks
To end the first day, Hot Chips will discuss future technologies. After a long day, I do wonder why they don't extend Hot Chips out into a third day of talks - normally at this point I am truly on information overload. Out of the three talks in this session, the one I'm most familiar with is the final one, by Samsung, about its compute-in-memory solution.
|Hot Chips 33 (2021): Day One, Session 4
|17h30||Ramanujan Venkatadri||Infineon||Heterogeneous computing to enable the highest level of safety in automotive systems|
|18h00||Sriram Rajagopal||EdgeQ||Architecting an Open RISC-V 5G and AI SoC for Next Generation 5G Open Radio Acess Network|
|18h30||Jin Hyun Kim||Samsung||Aquabolt-XL: Samsung HBM2-PIM with in-memory processing for machine learning accelerators|
The Aquabolt-XL was unveiled earlier this year, with Samsung able to add compute cores per memory bank into its HBM2 memory without any hardware modifications on the host side. Aquabolt-XL works by sending commands to specific memory addresses and can perform simple compute tasks on its in-order cores on the data within that memory bank. The idea is that energy will be saved by not having to move data from memory to the core for the simplest operations. At the time, I asked Samsung if this requires extra power, as other compute-in-memory solutions, and Samsung said no - this is very much a simple drop-in replacement for any HBM2 solution today, and requires software modification for use.
Day Two: Morning
The second day starts early, at 8:30am, and runs until 7pm. The first session starts with discussing Machine Learning inference processors from some of the big companies looking for cloud deployment.
|Hot Chips 33 (2021): Day Two, Session 1
ML Inference for the Cloud
|08h30||David Ditzel||Esperanto Technologies||Accelerating ML Recommendation with over a Thousand RISC-V Tensor Processors on Esperanto's ET-SoC-1 Chip|
|09h00||Ryan Liu||Enflame Technology||AI Compute Chip from Enflame|
|09h30||Karam Catha||Qualcomm||Qualcomm Cloud AI 100: 12 TOPs/W Scalable, High Performance and Low Latency Deep Learning Accelerator|
The first talk in this session is from Esperanto, a startup with $58m+ in funding to create recommendation engine processors for the could. All the major hyperscalers and retailers use recommendation engines - showing users what product would most interest them at any given time. Esperanto's solution seems to be a 1000-core RISC-V solution, with combined tensor cores for inference acceleration. This talk should be the first disclosure of the chip and the architecture underneath.
Second is another AI semiconductor startup, this time from China with a large backing from Tencent. The last funding round in January this year was for some $279m+ in Series C, although exactly what Enflame is producing hasn't been announced.
Qualcomm's AI 100 solution is third, a product that was announced just after Hot Chips last year. This AI inference processor is the top version of a range of AI inference hardware built on the same architecture underneath. The Qualcomm Cloud AI 100 has a reported speed of 400 TOPs for 75 W, although the most efficient version is aiming for 12 TOPs per Watt. I expect this talk to go into the architecture details of the processor family.
Day Two: Keynote
The second day only gets a single keynote earlier than usual, but this one should be interesting as it is from the US Department of Energy.
|10h00||Dimitri Kusnezov||DoE||DoE AI and Technology|
In this instance, the Deputy Under Secretary for AI and Technology will present an unnamed talk, which could cover a number of talks from the use of AI in current DoE deployed systems, or the DoE approach to adopting AI at scale, in both training and inference. A number of companies at this event have contracts with the DoE in some form, making it an interesting talk from our point of view.
Day Two: Home Stretch
Continuing the machine learning theme, the second session of the day before lunch is about bigger AI chips built more for training, as well as a special surprise.
|Hot Chips 33 (2021): Day Two, Session 2
ML and Computation Platforms
|11h30||Simon Knowles||Graphcore||Graphcore Colossus Mk2 IPU|
|12h00||Sean Lie||Cerebras Systems||The Multi-Million Core, Multi-Wafer AI Cluster|
|12h30||Raghu Prabhakar||SambaNova Systems Inc||SambaNova SN10 RDU: Accelerating Software 2.0 with Dataflow|
|13h00||J. Adam Butts||D.E. Shaw Research||The Anton 3 ASIC: a Fire-Breathing Monster for Molecular Dynamics Simulations|
|David E. Shaw|
The first talk is from Graphcore, talking about its Mk2 IPU product family. Graphcore has been talking about the Mk2 IPU deployment for over a year, including its four-IPU single server 1U solution all the way up to an IPU Pod and beyond. This talk seems to be a recap of the underlying architecture just at a different event, although fingers crossed we see something about a roadmap here as well.
The second talk is Cerebras Systems with its second-generation Wafer Scale Engine - a CPU the size of your head with 850,000 cores, and one wafer becomes one chip. Having announced WSE-2 at the beginning of the year, one of the highlights of those presentations was using multiple WSE-2 systems in the same rack to scale out the solution. That seems to be the focus of this talk.
Third is SambaNova, one of the AI companies that has recently announced hundreds of millions in their latest rounds of funding. The Cardinal AI solution from SambaNova is targeting AI training, and scaling across many systems in a reconfigurable gate array architecture, very similar to that of an FPGA but geared towards AI. SambaNova has presented its architecture at a few select conferences, however this will be the first time I've seen a talk on the topic.
Finally we have the surprise talk of the event, at least from my perspective. Here's a question - what do you do if there isn't any hardware on the market to solve your problem? Simple, build your own! The Anton 3 ASIC is next-generation dedicated hardware for molecular dynamics for all those molecular modeling problems that take too long to complete on conventional hardware. The story behind how the Anton 3 came into being seems fascinating from my short research, so I hope it becomes a part of the talk. The Anton 2 chip was presented at Hot Chips 26, in 2014, before I started covering the event.
After another lunch, the next session will excite a number of users, as they breach the topic of graphics disclosures. It looks like we're going to get a couple of good ones this year, despite the current state of the graphics market.
|Hot Chips 33 (2021): Day Two, Session 3
Graphics and Video
|14h30||David Blythe||Intel||Intel's Ponte Vecchio GPU Architecture|
|15h00||Andrew Pomianowski||AMD||AMD RDNA 2 Graphics Architecture|
|15h30||Aki Kuusela||Google's Video Coding Unit (VCU) Accelerator|
|16h00||Juanjo Noguera||Xilinx||Xilinx 7nm Edge Processors|
First up is Intel discussing its Xe-HPC architecture, or rather specifically its Ponte Vecchio HPC chip that uses 47 tiles from multiple process nodes on a single substrate. Ponte Vecchio was initially designed for the Aurora Exascale Supercomputer project, and will be partnered with Sapphire Rapids. As the chip is due to be shipped to production systems later this year, an expose on the architecture is widely welcomed.
The second talk will be AMD recapping its RDNA2 architecture, which is now at the heart of many of its graphics products.
The third talk is one that interests me comes from Google - as with all things at Youtube, it has to be done at scale, and making sure that every video on the platform is available at multiple resolutions requires a lot of computational resources. To help speed that up, Google built its own video co-processor, and through 2021 has been disclosing bits of how it works. Currently, the VCU only caters for a couple of the most regular codecs, but hopefully we will get an insight into what future versions might bring with AV1 and newer codecs.
Finally for this session, Xilinx will talk about its 7nm processor designs for the edge which it announced earlier this year.
The final session of the event is filed under 'New Technologies'. Everything here isn't strictly to do with processors, but certainly involves Hot Chips.
|Hot Chips 33 (2021): Day Two, Session 4
|17h00||Michael Wiemer||Mojo Vision||Mojo Lens - AR Contact Lenses for Real People|
|17h30||Sukki Yoon||Samsung||World's Largest Mobile Image Sensor with All Directional Phase Detection and Auto Focus Function|
|18h00||Hidekuni Takao||Kagawa University||New Value Creation by Nano-Tactile Sensor Chip Exceeding our Fingertip Discrimination Ability|
|18h30||Christopher Monroe||IonQ Inc||The IonQ Trapped Ion Quantum Computer Architecture|
*A good number of engineers from Apple attend the event every year, and ask a lot of questions, however they have never presented a talk. These events are often a collaborative industry disclosure mechanism, and gives a chance for companies to present their best technology as well as quiz their competitors. Apple's engineers are happy to ask lots of questions, but so far they have never had a talk good enough to be presented (I'm pretty sure they've never submitted a talk) or sponsored the event in any capacity.
Post Your CommentPlease log in or sign up to comment.
View All Comments
LordSojar - Tuesday, May 18, 2021 - linkAn exciting conference this year, indeed. So many great speakers and topics; I hope coverage is as great as this synopsis, Ian! Are you planning to attend the entirety of the conference?
Andrei Frumusanu - Tuesday, May 18, 2021 - linkWe'll be covering the whole show, yes.
LordSojar - Tuesday, May 18, 2021 - linkExcellent! Thanks Andrei, Ian, the rest of and team Anandtech for doing so. Looking forward to seeing more about rDNA2 and Bluefield-2, especially!
SarahKerrigan - Tuesday, May 18, 2021 - linkHaving been there for the totally content-free SKL presentation in 2016, I don't really have much confidence that Intel will say anything of substance about Alder Lake.
AdrianBc - Tuesday, May 18, 2021 - linkLast year's presentations about Tiger Lake and Ice Lake Server were informative enough.
Obviously they were excessively optimistic, but if you looked at them critically it was possible to assess how the real products will be.
AdrianBc - Tuesday, May 18, 2021 - linkThe only important information that was missing from the last year's Intel presentations and which became visible only after launch, was that even the SuperFin 10-nm Intel process cannot reach the efficiency of the TSMC 7-nm process, in the sense that whenever the CPUs reach their power limits and are constrained by them, at equal power consumption and equal number of active cores, an Intel CPU will have a lower clock frequency than an AMD CPU.
Except for this important detail, the future performance of Tiger Lake and Ice Lake Server could be evaluated from the Hot Chips presentations.
Targon - Tuesday, May 18, 2021 - linkIntel: We have the fastest processors, but they need 350W for the CPU alone to even have a chance to compete with AMD.
JayNor - Thursday, May 20, 2021 - linkWe had a glimpse of the Xe-HP last year on 10esf ... 42 TFlops FP32 ... so looks like Intel 10esf is doing fine. Perhaps this will become clear this year with Sapphire Rapids, which has been sampling broadly since Nov 2020, according to Intel's SC20 slides.
Otritus - Saturday, May 22, 2021 - linkIt's true that Zen 3 is more efficient than Willow Cove, but it should be noted that different CPU architectures can have differences in efficiency not related to node. Otherwise, Ryzen 5000 would be 0% faster than Ryzen 3000, since they are built on the same node and have the same power limits. By evaluating Cypress Cove, Comet Lake, Willow Cove, and Zen 3, we can assume that 10SF is not as performant as 14+++, 10SF is less efficient than N7, and Willow Cove is less efficient than Zen 3, meaning that Intel is at a node and architecture deficit. In a way this actually makes sense because Zen was more efficient than Kaby Lake and Coffee Lake (in spite of an inferior node), N7 is roughly 3x as efficient as 14LPP, and Zen 3 is ~1.2x as efficient as Zen 2.
eastcoast_pete - Tuesday, May 18, 2021 - linkIf they don't say anything substantial this year, they're in big trouble. Intel needs to build confidence that they got their version of bigLITTLE ready for prime time, or they will lose a lot of customers and also share value. The latter would bode really poorly for Intel's Chief Revenue Officer, which I was astonished to learn exists just recently.