Nvidia emerging as toughest Arm wrestle for Intel in 5GNvidia emerging as toughest Arm wrestle for Intel in 5G

The Grace part of Nvidia's superchip has generated less buzz than the AI-training Hopper, but it could be the strongest competitor to Intel in virtual RAN.

Iain Morris, International Editor

June 5, 2024

7 Min Read
Nvidia CEO Jensen Huang
Jensen Huang's Nvidia is attracting the interest of companies building virtual RAN products.(Source: Nvidia)

Leather-jacketed, bespectacled and grey-haired, Jensen Huang has the appearance of an intellectual biker. "Accelerate everything," the Nvidia boss told attendees at this week's Computex show in Taipei. He was, of course, not talking about two-wheeled modes of transportation but urging companies to continue their splurge on Nvidia's graphics processing units (GPUs), the chips that have serendipitously turned out to be good for artificial intelligence (AI). Customers seem to be fighting over constrained supplies, like bog-roll shoppers in a pandemic. Gaining 14 percentage points in the last year, Nvidia's gross margin is an obese 78%. Systemic risk is made of such things.

But it's another, more apparently mundane product that is the cause of some excitement in telecom. As well as building GPUs, Nvidia is also a customer of Arm, the UK-headquartered chips designer whose blueprints feature in most of the world's phones. Various chipmakers are now using these Arm blueprints to produce central processing units (CPUs), general-purpose silicon that forms the brain of a laptop or server. The Grace part of Nvidia's Grace Hopper superchip is an Arm-based CPU. In the future, it could power 5G networks.

This overarching CPU market today is dominated by Intel. Estimates of its market share vary, but in the subsector for server CPUs it claimed 71% of all sales in 2022, according to Counterpoint Research. Like AMD, the second-biggest vendor by far, Intel bases its chips on a so-called complex instruction set computing (CISC) architecture dubbed x86. Reduced instruction set computing (RISC), the main alternative, is the province of Arm (whose name literally stands for Advanced RISC Machines).

For years, CISC and RISC have been like the armies of WWI, barely advancing beyond their established lines. Intel's x86 muscle is a valuable powerlifter in the data center but judged too resource hungry for a mobile device. Arm has long seemed the opposite – supremely energy efficient but too puny to cope with any workload needs outside a smartphone chassis. Yet this perception is changing in Arm's favor.

Exploding silos

This matters in the world of telecom because operators are increasingly interested in using general-purpose processors and IT platforms to host network software. With such virtualization, or even cloudification, they could theoretically explode the silos that come with customized products and have commonality across all workloads. But Intel's CPU dominance is an obvious barrier. Numerous telco sources still call it a monopoly.

In the virtual radio access network (RAN), the label seems apt. While virtual RAN still accounts for only a sliver of the total RAN market, just about every rollout so far is based on Intel's Xeon-branded family of CPU products. Judging by telco feedback, the lack of options is partly to blame for the slow progress. Upset by government curbs on the use of Huawei, a Chinese vendor, operators are nervier than most about concentrations of market power.

Despite this, competitors to Intel have focused mainly on Layer 1, a slice of the technology that includes the most resource-hungry network software. In a virtual RAN, Layer 1 processing happens chiefly in a server box called the distributed unit (DU). Intel has stumped up its own accelerators – hardware dedicated to Layer 1's specific needs. Yet so have chipmakers including AMD, Marvell Technology, Qualcomm and Nvidia. In most cases, these accelerators live on smart network interface cards that can be plugged into a server. But the Intel CPU still does everything else.

Changing that means cultivating CPU alternatives to Intel. AMD, given its established presence in server CPUs, is perhaps the most obvious candidate. Because it also hails from the x86 world, it could necessitate less redesign of the network software originally intended to run on Intel.

Yet several companies have latched onto Arm's designs for virtual RAN CPUs. Amazon, most prominently, has built a Graviton-branded chip for its public cloud data centers. In principle, this could also host RAN workloads, although it does not yet appear to have been employed this way. Ampere Computing, an Arm licensee backed by Oracle, is developing servers that could be deployed in a more conventional RAN environment.

For Nvidia, another member of this group, the big pitch to operators is about the Hopper GPU as an accelerator for Layer 1. In typical style, Nvidia even has a set of Layer 1 software, branded Aerial. So far, however, relatively few telcos have looked eager to bite. Specs for Nvidia's Hopper-based accelerator cards last year put maximum power at a range of between 230 and 350 watts. That compared unfavorably with a 40-watt Layer 1 card then being marketed by Qualcomm. Unfortunately, the economics of GPUs look questionable unless they support both AI and Layer 1 workloads. And while that is what Nvidia intends, the opportunity for AI outside public cloud facilities is unclear.

Grace in favor

By contrast, experimentation with the Grace CPU has been far more positive. Take customized accelerators out of the equation and a previous drawback of Arm noted by Ericsson was its lack of support for vector processing. Critical to Layer 1, this can effectively tackle a whole array of data in one go rather than working sequentially through individual data points. Intel caters to it with an instruction set known as AVX512. Thanks to a recent v9 architectural update, including a library called SVE2, Arm has been adapting to the same needs. And Nvidia's implementation of vector processing in Grace CPUs appears to have brought success in recent third-party trials, the results of which were obtained by Light Reading.

Those trials used a Grace CPU with 72 Arm "cores" (the building blocks of a chip) and ranked Nvidia ahead of other Arm licensees on several important metrics. The points of comparison were Intel CPUs mixed with separate accelerators for the forward error correction (FEC), an especially gluttonous bit of Layer 1 software.

With vector processing, Grace was able to handle the FEC and still have cores available. When Grace was combined with an accelerator card from Qualcomm, rather than Nvidia's Hopper GPU, this "free" core count rocketed, leaving Intel far behind. While results also put Intel slightly ahead on CPU performance, they scored Grace – even with its higher overall number of cores – better on power efficiency.

There are total cost of ownership trade-offs to consider in all this, but the main black mark for Intel is still likely to be its dominant position in virtual RAN, small as that market may be. In newer products, Intel has stopped putting Layer 1 accelerators on separate cards and started to host them on the same die as the CPU. It has been scathing about cards as an extra cost and complication. But the move could backfire. Some operators are still using its cards, for one thing. And an "integrated" product, as the chipmaker calls it, makes replacing Intel in any area even harder. Given that open and virtual RAN is supposed to be all about "disaggregation," it is an odd marketing tactic.

None of this means Nvidia or Arm generally is an ideal choice. The Arm ecosystem for CPUs is far less developed, just as the AI one is for anyone except Nvidia. This makes it hard for some telcos to regard any Arm-based product as a common, off-the-shelf server, housed in a data center and good for any workload. The cloud universe remains an x86 constellation.

But at the same time, the servers built around Intel's latest "integrated" products arguably feature so much customization that they are unsuitable for anything except the RAN. Given a mix of regulatory, technical and cost concerns, most RAN compute equipment is still found at the base of antennas. Where centralization has occurred, it has not usually resulted in the sharing of equipment between RAN and other workloads. And the public cloud has had a negligible influence on the RAN. In this world, finding alternatives to Intel is perhaps even more important.

Read more about:

AIAsiaEurope

About the Author

Iain Morris

International Editor, Light Reading

Iain Morris joined Light Reading as News Editor at the start of 2015 -- and we mean, right at the start. His friends and family were still singing Auld Lang Syne as Iain started sourcing New Year's Eve UK mobile network congestion statistics. Prior to boosting Light Reading's UK-based editorial team numbers (he is based in London, south of the river), Iain was a successful freelance writer and editor who had been covering the telecoms sector for the past 15 years. His work has appeared in publications including The Economist (classy!) and The Observer, besides a variety of trade and business journals. He was previously the lead telecoms analyst for the Economist Intelligence Unit, and before that worked as a features editor at Telecommunications magazine. Iain started out in telecoms as an editor at consulting and market-research company Analysys (now Analysys Mason).

Subscribe and receive the latest news from the industry.
Join 62,000+ members. Yes it's completely free.

You May Also Like