New Heavy Reading report highlights the evolving role of hyperscalers in edge computing deployments and other telco workloads

Heavy Reading's Telcos and Hyperscalers report evaluates the willingness of CSPs to partner with the hyperscalers.

Jennifer Clark, Principal Analyst – Cloud Infrastructure & Edge Computing

January 18, 2024

4 Min Read
Edge computing photo illustration
(Source: Kirill Ivanov/Alamy Stock Photo)

Heavy Reading collaborated with Ericsson, Google Cloud, Intel and Viavi to conduct a survey of 101 global CSPs that have launched edge computing solutions or are planning to do so within 24 months. The aim was to understand how communications service providers (CSPs) and hyperscale cloud providers (HCPs) work together to enable edge computing. One of the drill-downs of the survey was to evaluate the willingness of CSPs to partner with the hyperscalers, specifically at the edge but also for other telco workloads.

It is clear from the overall survey results that carriers' attitudes toward partnering with the hyperscalers have become more optimistic. However, many perceived challenges remain (see the figure below). Unsurprisingly, "security" and "cost" are the most prevalent concerns, garnering by far the greatest percentages of "significant concern" votes. However, it is the third line item, "operational control," that has the greatest percentage of combined "significant concern" and "somewhat of a concern" votes from a whopping 98% of respondents. Carriers are more comfortable with the resiliency, recoverability and overall performance available with the HCPs, but continue to be unwilling to relinquish control of the network. HCP solutions that enable CSPs to monitor, maintain and scale their own environment and workloads are essential to the success of these joint ventures. As CSPs become more comfortable with working with HCPs, this may change. However, CSPs will always need to retain control of the workload, service or application to ensure that the HCP solution continues to meet both their needs and those of their customers and to give them the ability to change HCPs if need be.

"HCP lock-in" has the smallest percentage of "significant" concern responses, but conversely, it has the greatest percentage of "somewhat of a concern" responses. Heavy Reading explores this issue of HCP lock-in further in the report; it is clear that lock-in is a source of persistent angst among service providers. For now, it is considered a necessary evil to reap the benefits of agility, scalability, lower capex and AI capabilities that the hyperscalers bring to these CSP partnerships. The CSPs would like to add "lower opex "to this list of benefits, paired with more latitude in shifting workloads from one HCP to another ― or even, dare we say it, using multiple HCPs for individual workloads. For now, however, vendor lock-in is on a slow simmer as an issue, and CSPs are resigned to relating specific workloads with specific HCPs.

Loss of control is the most pervasive concern in moving network functions to the HCP cloud

Chart showing deployment concern levels in HCP clouds.

(Source: Heavy Reading)

What level of concern do you have about considerations and obstacles that are limiting your deployments of network functions in HCP clouds? (Source: Heavy Reading, 2023)

Where on the network continuum are edge deployments taking place?

These concerns reflect the changing demographics of edge computing. Edge computing means different things to different people. When Heavy Reading asked CSPs where they felt edge platforms should be deployed, a majority (56%) pointed to the enterprise edge as the most appropriate location for deployment. This suggests that those surveyed are largely focused on reducing latency, improving overall application performance and/or supporting massive Internet of Things (IoT) implementation. Meanwhile, 30% pointed to the network far edge as their preferred location, and the remaining 14% suggested the network near edge would meet their needs for edge deployments.

Where is the edge?

Chart showing where a managed edge platform makes the most sense.

(Source: Heavy Reading)

Q: Where does a managed edge platform make the most sense? (Source: Heavy Reading, 2023)

The percentages above show a shift in how service providers think about their edge deployments. When edge computing implementations first began to emerge roughly four years ago, the bulk of CSP implementations focused on operational efficiencies. The latency targets were about 20ms round-trip delay, which was satisfactory for most use cases at the time but insufficient for AR/VR-enabled applications for either the consumer or enterprise. Instead, the interrelated use cases were intended to provide a relief valve for network pressure during times of traffic stress, keep more traffic local to reduce traffic hair-pinning and reduce traffic volume at the network core. These use cases can be justified with a straightforward ROI analysis. On the other hand, as deployments reach further toward the edge of the network, implementation costs increase along with the number of edge nodes, and the spread creates end-to-end service management challenges. As Heavy Reading's survey results show, CSPs are increasingly likely to turn to the HCPs to assist in scaling their edge services due to the speed and agility with which they can add or remove edge nodes, compute and storage capacity and use cases.

Read the full report for more edge computing insights

We asked respondents about the timeframe of deploying on HCP cloud infrastructure at scale for a variety of telco workloads beyond edge computing, including packet core, cloud RAN, BSS and OSS. It is interesting to note that edge computing did not have the most immediate timeframe. For a description of those results and to gain more in-depth details of service providers' perspectives on edge computing deployment and their hyperscaler partners, download and read the full report now.

This blog is sponsored by Ericsson.

About the Author(s)

Jennifer Clark

Principal Analyst – Cloud Infrastructure & Edge Computing

Jennifer Pigg Clark is Principal Analyst with Heavy Reading covering Cloud Infrastructure and Edge Computing. Clark provides actionable insight into service provider evolution, examining the challenges and opportunities facing network operators as they move towards 5G and IoT with an increasingly virtualized and cloud native infrastructure. Clark examines the solutions and technology reshaping the telco data center, technologies such as Edge Computing, Open Source, OpenStack, container networking, Network Orchestration, Software Defined Networks (SDN), Network Functions Virtualization (NFV), and SD-WAN. Clark started her industry research career with the Yankee Group, which was acquired by 451 Research in 2013. She held the role of Sr. Vice President at Yankee Group. Prior to joining Yankee Group, Clark was Manager of Network Planning and Strategy for Wang Laboratories'corporate data network. She began her career at Wang with responsibility for the domestic and international roll-out of Wang's packet network, connecting more than 250 locations in 14 countries. Before joining Wang, she was a member of the IT research and development division of Commercial Union Insurance Companies. Clark is a highly regarded speaker at industry seminars and conferences and is frequently cited by the commercial and trade press. She has been a guest lecturer at the MIT Sloan School of Management and is a member of the IEEE. She holds a B.A. degree from Mount Holyoke College.

Subscribe and receive the latest news from the industry.
Join 62,000+ members. Yes it's completely free.

You May Also Like