Hardware acceleration 5G vRAN -a case for GPU

RAN Virtualization and 5G use-cases

Ericsson has been finding out and prototyping designs for a virtualized NR baseband distributed unit (vDU or DU performance applied as software program on a generic {hardware} platform). We have lately taken further steps in direction of such a product in collaboration with modern suppliers of complementary applied sciences, together with NVIDIA for GPUs and Intel for CPUs. This will allow options to rising challenges, e.g. deployment in knowledge center-like environments and industrial and enterprise personal networks. Benefits embody roll-out and operational synergies resulting from the usage of open-source cloud platforms and business off-the-shelf (COTS) {hardware}. 

In common, rising use-cases for 5G will be classed into two teams: these requiring a low latency round-trip time between system and software service (e.g. cloud gaming, server-side video rendering and deep studying) and people characterised by localized/site-bound software deployment (e.g. Industry 4.0, controller software program for machines, sensors and actuators in factories). The former class includes software providers transferring nearer to the entry community (RAN) from hyperscale clouds (AWS, Azure, GCP). The latter includes RAN and Core being deployed in stand-alone personal networks. Both circumstances suggest a convergence of connectivity (5G) and IT providers.

But there’s a catch: the community could not evolve to help such functions (low-latency or localized/personal) earlier than the application-side demand emerges. At the identical time, lack of help within the community means software builders don’t come to the desk in huge numbers. To circumvent this, we – as business leaders – have to make a transfer. To paraphrase Kevin Costner from the movie Field of Dreams, “Build it and they’ll come.” The logic is evident: an preliminary “base” workload is required for what’s loosely known as “edge.” NR baseband processing can fill that position—compelling a shift in structure and expertise stack—serving as a basis for new providers over an extended timeframe.

Harmonizing Virtualization with Acceleration

vRAN is about 5G changing into software-defined and programmable. By definition, vRAN includes disaggregation of software program and {hardware}. There is an expectation that an operator will be capable of run the identical 5G software program stack on a wide range of servers and evolve capability by swapping out the {hardware}, as we do with our PCs. There’s a value to this: a decrease diploma of system integration implying decrease efficiency per watt of energy. But there’s additionally an upside the place Ericsson is taking the lead: vastly bettering the tempo of latest community options and the adaptability of 5G to rising use-cases.

Many 5G use-cases run nicely on pure CPU platforms, however as bandwidths improve and superior antenna methods are deployed, x86 cores wrestle to maintain up and begin driving impractical ranges of energy consumption. Hardware acceleration will likely be wanted for the compute-heavy bodily layer and scheduling workloads in 5G NR.

From desktop gaming to supercomputer benchmarking, {hardware} acceleration is attained utilizing GPUs. While there are different paths to {hardware} acceleration, they often contain customization utilizing FPGAs or ASICs, subsequently ruling out COTS {hardware}. Concerning compute density, GPUs dedicate a bigger fraction of their chip space to arithmetic than CPUs do. As applied sciences evolve, GPUs are free to maintain optimizing for a slender workload of high-performance computing, whereas CPUs goal a extra numerous workload, similar to databases and workplace functions.

To get excessive efficiency in a CPU implementation of NR one has to massively optimize single-core, single-thread efficiency with SIMD (single instruction a number of knowledge). This is mostly enjoyable and thrilling work however leads down a dangerous path. Once the final optimization is finished with no cache misses – what do you do? Wait for quicker CPU clock charges? Or try to go multi-core within the CPU? The former different is a useless finish. The latter case seems to be difficult within the NR bodily layer exactly as a result of the single-threaded code will be so extremely optimized. Going to a number of cores requires sacrificing the very knowledge locality you labored so onerous to engineer, that means the primary few cores past the preliminary one could value extra in misplaced per-core efficiency than is gained in parallelization. Bad information.

On the opposite hand, with NVIDIA GPUs you may take an excessive multi-threaded method from the beginning. Since efficient core counts per chip space have a tendency to extend quicker than clock charges, this units you up for an ongoing impact over time. And GPUs are in all the main cloud platforms and run numerous AI workloads throughout all industries (making them appropriate for COTS {hardware} platform design).

Requirements

NVIDIA GPUs provide an acceleration platform that delivers efficiency and programmability with out sacrificing the “COTS” of x86 structure. Here we define our key necessities which the platform fulfills:

  • Cloud-native and well-integrated with x86 virtualization
  • Programmable in C and C++ with a versatile programming mannequin
  • An ecosystem of instruments and libraries
  • Availability of design professionals throughout many industries and educational analysis
  • Mass-market and mass-production volumes
  • The potential to scale from very small to very massive deployments
  • A roadmap of accelerating compute functionality

The CUDA ecosystem encompasses a big and vibrant developer neighborhood with entry to excellent libraries of widespread computational constructing blocks. In this respect, CUDA and NVIDIA GPUs haven’t any rivals in high-performance computing. And NVIDIA is scaled for massive quantity manufacturing and logistics since GPUs additionally goal PC gaming.

Our method to RAN virtualization as a complement to our present portfolio will present other ways of deploying NR baseband. Basing it on COTS {hardware} permits a variety of rising edge functions together with augmented actuality, AI and deep studying, and server-side video rendering to be co-located. In addition, it permits the answer to scale down to satisfy the wants of personal networks. We see clear benefits in implementing the baseband processing and new rising 5G use-cases in shut proximity, with each benefiting from NVIDIA GPU-based {hardware} acceleration.

[1] – Ericsson Mobility Report, November 2019 version and 2019 This autumn replace.

This is the second of two articles on 5G RAN virtualization. The first, entitled “Virtualized 5G RAN: why, when and how?” 

Related Posts