IBM Introduces its First Power10-based Server, the Power E1080; Targets Hybrid Cloud

By John Russell

September 8, 2021

IBM today introduced the Power E1080 server, its first system powered by a Power10 IBM microprocessor. The new system reinforces IBM’s emphasis on hybrid cloud markets and the new chip beefs up its inference capabilities. IBM – like other CPU makers – is hoping to make inferencing a core capability of host CPUs and diminish the need for separate AI accelerators. IBM’s Power8 and Power9 were usually paired with Nvidia GPUs to deliver AI (and HPC) capabilities.

“When we were designing the E1080, we had to be cognizant of how the pandemic was changing not only consumer behavior, but also our customers’ behavior and needs from their IT infrastructure,” said Dylan Boday, vice president of product management for AI and hybrid cloud, in the official announcement. “The E1080 is IBM’s first system designed from the silicon up for hybrid cloud environments, a system tailor-built to serve as the foundation for our vision of a dynamic and secure, frictionless hybrid cloud experience.”

Few details about the Power10 chip were discussed, nor was a more detailed spec sheet for the Power E1080 presented at an analyst/press pre-briefing last week. IBM instead chose to cite new key functional capabilities that blended the boundary between system and chip and to highlight favorable benchmarks. General availability for E1080 is scheduled for later this month. No timetable was given for direct sales (if any) of the Power10 chips.

Here are the highlights as reported by IBM:

  • Enhancements for hybrid cloud such as by the minute metering of Red Hat software including Red Hat OpenShift and Red Hat Enterprise Linux, 4.1x greater OpenShift containerized throughput per core vs x86-based servers, and “architectural consistency and cloud-like flexibility across the entire hybrid cloud environment to drive agility and improve costs without application refactoring.”
  • New hardware-driven performance improvements that deliver up to 50 percent more performance and scalability than its predecessor the [Power9-based] IBM Power E980, while also reducing energy use and carbon footprint of the E980. The E1080 also features four matrix math accelerators per core, enabling 5x faster inference performance as compared to the E980.
  • New security tools designed for hybrid cloud environments including transparent memory encryption “so there is no additional management setup,” 4x the encryption engines per core, allowing for 2.5x faster AES encryption as compared to the IBM Power E980, and “security software for every level of the system stack.”
  • Robust ecosystem of ISVs, Business Partners, and support to broaden the capabilities of the IBM Power E1080 and how customers can build their hybrid cloud environment, including record-setting performance for SAP applications in an 8-socket system. IBM is also launching a new tiered Power Expert Care service to help clients as they protect their systems against the latest cybersecurity threats while also providing hardware and software coherence and higher systems availability.”

In recent years IBM’s positioning of its Power platforms and Power CPU line has significantly transitioned from HPC-centricity to enterprise-centricity with a distinct hybrid cloud focus. Introduction of the E1080 server seems to complete the journey. Standing the Summit supercomputer (at ORNL) in 2018 based on IBM’s AC922 nodes with Power9 CPUs was probably the high-water mark for IBM HPC. Summit was the fastest supercomputer in the world for a couple cycles of the Top500.

However, Power9-based IBM systems achieved lackluster traction in the broader HPC market. IBM shifted gears gears a few times trying to find the right fit. The IBM purchase of Red Hat for $34 billion in 2019 marked a massive shift for IBM strategy to the cloud and stirred uncertainty about IBM’s plans for Power-based platforms and its role in the OpenPOWER Foundation. The integration of E1080 server line into IBM’s hybrid-cloud strategy now seems to remove the ambiguity surrounding IBM plans for the IBM Power product line.

Patrick Moorhead, founder and president of Moor Insights & Strategy, noted, “IBM has changed its focus on Power over the past few years. Power10 is focused on enterprise big data and AI inference workloads delivered in a secure, hybrid cloud model. It looks to really scream on SAP, Oracle, and OpenShift environments when compared to Cascade Lake. The performance numbers IBM touted make sense given the chip’s architecture.” 

“On-chip ML inference makes lots of sense when latency is of the upmost importance and being off-chip versus going through PCIe delivers just that in an open (ONNX supported) way. Some enterprises will even train models on these systems if they’re underutilized,” added Moore who said he thought IBM could gain traction, “if it aggressively markets and sells these systems against x86-systems. … I’d say the past few generations were marketed and sold to current clients as replacements for older IBM systems versus ‘going after’ Intel.”

Analyst Peter Rutten of IDC also thinks IBM’s E1080 is a good move. “Keep in mind that this is the 8- or 16-socket enterprise class system that runs AIX first and foremost, as well as OS i and Linux. This is IBM’s transactional/analytics processing system that offers 99.999% availability, high security, and a lot of performance for such traditional workloads as database. The new chip offers several benefits for this system – higher performance (versus Power9), less energy, more bandwidth, lower latency, greater security with baked-in encryption, and the MMA for AI inferencing on the chip, which is something enterprises increasingly want to be able to do on their traditional workloads. The way I see it, IBM hit sort of a sweet spot with this system.”

Rutten also doesn’t think IBM is easing out of HPC and AI. “I don’t see this as IBM meandering, but as parallel tracks. There is the scale-out portfolio that is all Linux and that’s focused on AI training, HPC, big data analytics. These are the one- and two-socket systems that include the AC922 which was used for Summit. IBM didn’t win the latest supercomputer RFPs but they revealed some very interesting features with Power10 for those workloads. The E1080 is based on a single chip module. But forthcoming is a Dual-Chip Module (DCM), which takes two Power10 chips and puts them (1200 mm2 combined) into the same form factor where there used to be just one Power9 processor. This DCM is targeting compute-dense, energy-dense, volumetric space-dense cloud-type configurations with systems ranging from 1 to 4 sockets. I think we’re going to see some screaming performance from these systems when they arrive.”

Top down view of IBM E1080

Not a lot was said about the specific Power10 chip inside the E1080 at IBM’s pre-launch briefing.

Responding to a question about the physical components of the system and new chip during Q&A at the briefing, Boday said, “The E1080 will scale to 240 cores in the entire system itself. The Power10 processor will have 15 cores, the prior generation (Power9-E980) was maxed out at 12. That allows us to scale up to the 240 cores (16 Power10s). We’re also improving the overall number of DIMM slots to where we can actually do 256 DDR4 DIMMs in the system. The overall memory bandwidth [is increased] to over 400 gigs per second, per socket. We’ve introduced a Gen5 PCIe slots and we’re allowing you to connect all of this together, [the] individual drawers and nodes of the system through a faster fabric that we call our SMP fabric.”

Ken King, general manager of IBM Power, “You will see more announcements coming later this year for more of our Power10 family coming to the market, and we’ll be rolling additional ones into early 2022 as well.” IBM disclosed last year that Power10 would be its first 7nm process part and was being fabbed by Samsung.

A fuller picture of the Power10 chip lineup and associated systems will emerge over the next few months. One interesting point is the inclusion on-chip inference capability. At the briefing, Satya Sharma, IBM fellow and CTO of IBM Power, emphasized the practice of “not requiring exotic accelerators” is a growing trend in the market. Indeed, IBM showcased such capabilities in its new Z series chip (Telum) at the recent Hot Chips conference. Intel has also announced plans to incorporate similar capabilities in its Sapphire Rapids CPU (Intel’s next-generation “Intel 7” processor).

Given IBM’s newer focus on adding inferencing capabilities to Power10, it would be interesting to see how the E1080 fares on the MLPerf inferencing competition. Boday was non-committal and said, “We’re excited about the number of MM (matrix multiply) engines per core that Power10 delivers and how those are going to be very advantageous. As we continue to build out those benchmarks, such as MLPerf, those are things that will be on the radar for us to deliver.” (See HPCwire article on IBM Power10 presentation last year at Hot Chips 2020.)

Mostly IBM stuck tightly to a script touting the new system’s functionality and favorable benchmarks versus physical specs. IBM is strongly promoting E1080’s security features and tools. The entire memory is encrypted, with no performance penalty or management set-up, said Satya Sharma, IBM fellow and CTO of IBM Power.

As an example, he said, “We are providing Forex crypto engines in every core. As a result, customers can get 2.5x more crypto performance. [Using this engine], you can do either end-to-end decryption or encryption. Or you can do [this] file systems or databases or applications. You can go from server all the way to network to the storage. With this crypto engine capability, you can implement full stack and end to end encryption.”

There is also a centralized dashboard for managing security on the E1080. “Customers can implement a number of different compliance automation tools [including] PCI, HIPAA-readiness, GDPR, and we would ensure that all of the servers in the server farm all comply with these security compliance profiles. At the same time, we are monitoring the entire server farm, if any of these servers go out of compliance,” added Sharma. He ticked through several security elements such as libraries of algorithms for so-called “post-quantum security” as well and isolation measure taken at the CPU and system level.

Certainly, SAP is a big factor in the enterprise market and IBM’s reported results again its own prior generation Power9 as well as x86 rivals will draw attention. It will be interesting to keep watching the Power10 family’s development and how many Power10 skus IBM ends up offering.

Analyst Shahin Kahn, of OrionX noted, “AI inference will be the tail that is wagging the Deep Learning dog. It is about infusing apps with AI models and feeding new data back to AI learning. So AI inference is a very large market attracting many new chip and system players. While increased focus on AI is to be expected, IBM’s innovations with memory really also stand out: Open Memory Interface, shared memory, large address space, memory bandwidth, memory clustering, and memory encryption are all very cool and very useful. In an interesting twist, Arm’s success helps expand the market for Power10 since developers who have already re-targeted their app once will find it a lot easier to do so a second or third time.”

Addison Snell, CEO of Intersect360 Research, thought IBM’s latest system and chip fit well into IBM’s expanding enterprise AI focus. “The Power E1080 is an interesting step in IBM’s continued focus on enterprise services and hybrid cloud. Power10 has features that would be useful in HPC, such as its Matrix Math Accelerator (MMA) engines, but IBM is focusing these exclusively on AI inference now—a whiplash-inducing abandonment of HPC since the installations of Summit and Sierra, which are still among the most powerful supercomputers in the world. For enterprise AI, it makes sense to move inferencing capabilities onto the CPU, and this will be part of a general trend among CPU providers,” said Snell.

Stay tuned.

Link to IBM announcement: https://www.hpcwire.com/off-the-wire/ibm-unveils-new-generation-of-ibm-power-servers-for-frictionless-scalable-hybrid-cloud/

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Natcast/NSTC Issues Roadmap to Implement CHIPS and Science Act

May 29, 2024

Yesterday, CHIPS for America and Natcast, the operator of the National Semiconductor Technology Center (NSTC), released a roadmap of early steps for implementing portions of the ambitious $5 billion program. Natcast is t Read more…

Scientists Use GenAI to Uncover New Insights in Materials Science

May 29, 2024

With the help of generative AI, researchers from MIT and the University of Basel in Switzerland have developed a new machine-learning framework that can help uncover new insights about materials science. The findings of Read more…

Microsoft’s ARM-based CPU Cobalt will Support Windows 11 in the Cloud

May 29, 2024

Microsoft's ARM-based CPU, called Cobalt, is now available in the cloud for public consumption. Cobalt is Microsoft's first homegrown CPU, which was first announced six months ago. The cloud-based Cobalt VMs will support Read more…

2024 Winter Classic Finale! Gala Awards Ceremony

May 28, 2024

We wrapped up the competition with our traditional Gala Awards Ceremony. This was an exciting show, given that only 40 points or so separated first place from fifth place after the Google GROMACS Challenge and heading in Read more…

IBM Makes a Push Towards Open-Source Services, Announces New watsonx Updates

May 28, 2024

Today, IBM declared that it is releasing a number of noteworthy changes to its watsonx platform, with the goal of increasing the openness, affordability, and flexibility of the platform’s AI capabilities. Announced Read more…

ISC 2024 Takeaways: Love for Top500, Extending HPC Systems, and Media Bashing

May 23, 2024

The ISC High Performance show is typically about time-to-science, but breakout sessions also focused on Europe's tech sovereignty, server infrastructure, storage, throughput, and new computing technologies. This round Read more…

Scientists Use GenAI to Uncover New Insights in Materials Science

May 29, 2024

With the help of generative AI, researchers from MIT and the University of Basel in Switzerland have developed a new machine-learning framework that can help un Read more…

watsonx

IBM Makes a Push Towards Open-Source Services, Announces New watsonx Updates

May 28, 2024

Today, IBM declared that it is releasing a number of noteworthy changes to its watsonx platform, with the goal of increasing the openness, affordability, and fl Read more…

ISC 2024 Takeaways: Love for Top500, Extending HPC Systems, and Media Bashing

May 23, 2024

The ISC High Performance show is typically about time-to-science, but breakout sessions also focused on Europe's tech sovereignty, server infrastructure, storag Read more…

ISC 2024 — A Few Quantum Gems and Slides from a Packed QC Agenda

May 22, 2024

If you were looking for quantum computing content, ISC 2024 was a good place to be last week — there were around 20 quantum computing related sessions. QC eve Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's l Read more…

Europe’s Race towards Quantum-HPC Integration and Quantum Advantage

May 16, 2024

What an interesting panel, Quantum Advantage — Where are We and What is Needed? While the panelists looked slightly weary — their’s was, after all, one of Read more…

The Future of AI in Science

May 15, 2024

AI is one of the most transformative and valuable scientific tools ever developed. By harnessing vast amounts of data and computational power, AI systems can un Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Leading Solution Providers

Contributors

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

The NASA Black Hole Plunge

May 7, 2024

We have all thought about it. No one has done it, but now, thanks to HPC, we see what it looks like. Hold on to your feet because NASA has released videos of wh Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Intel Plans Falcon Shores 2 GPU Supercomputing Chip for 2026  

August 8, 2023

Intel is planning to onboard a new version of the Falcon Shores chip in 2026, which is code-named Falcon Shores 2. The new product was announced by CEO Pat Gel Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

How the Chip Industry is Helping a Battery Company

May 8, 2024

Chip companies, once seen as engineering pure plays, are now at the center of geopolitical intrigue. Chip manufacturing firms, especially TSMC and Intel, have b Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire