l Introduction
Goal: An innovative architecture for small-size edge cloud using data processor
To meet the demanding increase of 5G data and the utilize the advantage of low latency and high bandwidth of 5G technologies, the number of small-size datacenter is increasing dramatically. Research shows there shall be more than 2.8 million cloudlet datacenters with less than 200 servers connected near the 5G tower. In IEC Type 5 project, we will build an innovative architecture for small-size edge cloud computing using latest and greatest data processor.
Fig. IEC and Cloudlet
l Release 6
1. Brief Introduction of R6
In R6, we introduce an innovative networking architecture based on PCIe data fabric to lower both the cost (CAPEX) and power consumption (OPEX) in small clusters for edge cloud computing. Based-on innovative data processor (DPU and XPU), the next-generation networking features with:
- New networking architecture to lower the TCO of edge infrastructure
- TCP/IP compatible and cloud native for develops and developers
- Green to protect the environment for lasting development
- Scalable and composable to meet the dynamical workload
— IEC Type 5 | Latest update |
— System Architecture | R6 |
— Software & Platform API | - |
— Hardware Design | R6 |
— CD log | R6 |
2. System Architecture
2.1 Extending PCIe Transport
To achieve low cost and low power, in R6, we introduce the latest and greatest technology to connect two servers directly with PCIe links via innovative data processors. In this scenario, we eliminate the legacy network adapters (such as NIC, optics and legacy network switches). In typical application, with PCIe networking, we can reduce the number of connection components by 75%.
2.2 PCIe Networking and Cloud-on-Board
By the advantage of PCIe networking, we can unified the system-on-board (SoB) connection and the cloud cluster topologies into one single and simple architecture, which we named as Cloud-on-Board (CoB) Architecture.
In the CoB architecture, we can connect CPU directly without additional adapters.
2.3 PCIe Extending DPU Cluster
Since DPU is PCIe-compatible device, we can further combine DPU and PCIe Networking together. In R6, we introduce a hardware layer or physical link/fabric layer between the DPU and the CPUs as below. With this layer, we extend the DPU cluster size and also use the DPU management features as well.
IEC Type 5 System Architecture
2.4 Roadmap
In R6, we introduce PCIe based data fabric. In the future, we will include CXL or UCIe based as well.
3. Software & Platform API
No changes from previous Release 5.
4. Hardware Design
4.1 Cloud-on-Board (CoB) Architecture
For general computing, we introduce the CoB with three major components:
1) Edge Computing Module (ECM): CPU+RAM+OS SSD combo,
2) Edge Base Board (EB2): for PCIe Data Fabric,
3) Edge Adapter Module (EAM): PCIe-compatible Device, such as GPU, NIC, SSD etc.
All in PCIe Fabric
4.2 Networking Topology
In CoB design, we have multiple networks. At least one PCIe networking for multiple CPUs. Also we can introduce more connections as well as traditional RJ45 as the management ports as well. Hence all CoB Hardware are cloud native compatible at the beginning.
Networking in CoB system for Cloud Native Applications
4.3 Cloud Native Server Reference Design
At least, we pack all components together, and give a reference design as below.
4.4 Test Environment for CoB Prototype
Please refer to https://dev.socnoc.ai. Send email to demo@socnoc.ai for test account.
5. CD log
- System bringup log
Link:
- Docker installation log
Link:
- Docker cluster status log
Link:
- Web portal screenshot
Link:
- Demo video
Link: