High-Performance Server System with CXL based on the COM-HPC standard and Switchtec™ Gen 5 PCIe switches by Microchip, GPUs, and Mellanox NIC PCIe cards by Nvidia
High-Performance Server System with CXL based on the COM-HPC standard and Switchtec™ Gen 5 PCIe switches by Microchip, GPUs, and Mellanox NIC PCIe cards by Nvidia

Development of Scalable High-Performance Server System Based on CXL Technology

Project in Nutshell:

We developed a scalable server system concept with the CXL technology at the request of a leading provider of high-performance computing and cloud infrastructure solutions. In such a solution, data transfer rates can reach 800 Gbps and a storage capacity of 8192 TB. If implemented, this system will allow our client to establish itself as an industry leader.

Client & Challenge 

Our client is a leading provider of high-performance computing and cloud infrastructure solutions, serving companies across a variety of industries, including big data analytics, scientific research, and AI solutions. Its solutions help systems process massive amounts of data and tolerate high computational loads. 

The client faced a challenge: the growing demand for high-speed data processing and the inability to meet it with the existing infrastructure based on traditional PCIe connections. Here's what the challenge was related to: 

  • bottlenecks in PCIe connections slowed down data processing; 
  • high costs to maintain and scale existing infrastructure; 
  • limited flexibility and adaptability to support different computing architectures based on CPU, GPU, and FPGA; 
  • data growth and the need for scalability began to lead to downtime and increased costs. 


As a result, the client approached Promwad with a request to develop a CXL-based server system that would solve these problems. 


Compute Express Link (CXL) is an advanced technology that improves communication between computer components and enables high-speed and coherent memory access between processors, accelerators, memory buffers, and non-volatile memory.

CXL offers high bandwidth and low latency, and CXL 2.0 switches improve system efficiency by allowing multiple devices to be connected to a single host.

The technology reduces power consumption by optimising memory access and resource sharing. In the context of working with different types of memory, CXL provides several benefits:

  • CXL DRAM: flexible and scalable DRAM integration with high-speed access and reduced latency compared to traditional memory architectures.
  • CXL persistent memory: non-volatile memory that retains data during system power outages, which is critical for applications requiring high availability and fast recovery.

Companies can use CXL technology to create new solutions and services:

  • CXL memory extenders extend system memory beyond traditional DIMM modules for greater flexibility and scalability.
  • CXL Mem, a CXL protocol that provides consistent memory access between processors and devices, simplifying data sharing and memory management.
  • Entrepreneurial CXL devices: memory modules, accelerators, and interconnects.
  • CXL services: service providers use CXL to enhance cloud services, AI processing and enterprise solutions that require high-speed, low-latency data access.

 

Solution

1. Concept Development

Promwad scrutinised the customer's request and proposed a solution based on COM-HPC, an open standard for embedded computing modules, and Switchtec™ Gen 5 switches with firmware-level control from Microchip. 

The PFX and PAX Switchtec™ Gen 5 switches are designed to be versatile, allowing for the construction of various systems with the ability to adapt the usage vector to any possible fractionalisation.
 

 Scalable multi-host+PFX/PAX+PFX CXL switching developed at Promwad

2. Hardware Design

The Promwad engineers select the following components for designing a multi-host system concept and fulfilling customer requirements: 

  • A COM-HPC carrier board manages communication and data flow between different components of the system.  
  • Switchtec™-based high-speed, scalable backplane to provide high-speed connectivity between multiple devices.  
  • The Nvidia Mellanox NIC PCIe card connects the system to high-speed networks and enables fast data transfer.  
  • Two Nvidia GPU cards (full-size PCIe format) are used to perform compute-intensive tasks such as modelling, artificial intelligence, and big data analytics in HPC systems.
     

3D model of a fully functional system cluster for a multi-host system 

3. Software Development

 Our engineers also considered and incorporated the following software components into the concept: 

  • a file manager based on DPDK/SPDK and low-level firmware control for the Switchtec™ components; 
  • low-level firmware for power systems and system health control; 
  • web UI/UX;  
  • system control middleware and software;  
  • middleware layer for the file manager (capturing/recording); 
  • and other software components encapsulated in Docker and Kubernetes: server system, server user control interface, user data source subsystem control, user service (capturing/recording/transcoding/streaming), and backend. 

4. System Specifications

The newly designed system has the following characteristics:

  • Networking bandwidth:
    • 200Gb/s for single host architecture; 
    • 400-800Gbs for multi-host architecture.
       
  • Storage recording bandwidth:
    • 16GB/s for single host architecture;
    • 32GB/s for multi-host architecture.
       
  • Storage recording capacity:
    • from 16 hot-pluggable NVMe to 32, up to 1024TB for single host architecture;
    • from 32 hot-pluggable NVMe to 256, up to 8192TB for multi-host architecture.
       

There is a comparison of the main key parameters for various system configuration in the table below:

 

Half System

Advanced Half System

Full System

Networking bandwidth, Gb/s

200 

400 

800 

Storage recording bandwidth, GB/s

16 

24 

32 

Storage recording capacity, TB

1024 

4096 

8192 

CPU threads amount, virtual thread quantity estimated

32 

64 

128 

GPU/CUDA threads amount, virtual thread quantity estimated

2048 

3072 

4096 

Business Value 

As a result, our client now has a robust and scalable server system concept that can deliver up to 800 Gbps transfer speeds and 8192 TB of storage capacity. 

Modernisation of the system will allow the company to reach a new market position, retain current clients, and attract new ones due to its reliability and high speed of data transmission. 

More of What We Do for High-Performance Data Processing

  • Enterprise NAS: a case study of developing a NAS storage system with DPDK / SPDK, improving data transfer speed and network efficiency. 
  • SPDK & DPDK: our experience in applying SPDK and DPDK technology to create complete solutions for storage and other networking applications. 

Other Case Studies

Tell us about your project!

All submitted information will be kept confidential.