X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=docs%2Foverview.rst;h=56de5a6acea3967280efe6321498e7edd00d609f;hb=1d6d734a03cc2bf3ddbf401e0cb15a50e825c3be;hp=a00f1be1f273bd999871969c2a89fac0ef9b006b;hpb=15cbb2128368816a939b0bb815716a42e64b4cbb;p=pti%2Frtp.git diff --git a/docs/overview.rst b/docs/overview.rst index a00f1be..56de5a6 100644 --- a/docs/overview.rst +++ b/docs/overview.rst @@ -2,38 +2,173 @@ .. SPDX-License-Identifier: CC-BY-4.0 .. Copyright (C) 2019 Wind River Systems, Inc. -RTP Overview -============ +Infrastructure Overview (INF) +============================= -This project implements a real time platform to deploy the O-CU and O-DU. +This project is a reference implementation of O-Cloud infrastructure and it implements a real time platform (rtp) to deploy the O-CU and O-DU. -In O-RAN architecture, the O-DU and O-CU could have different deployed scenarios. -In general the performance sensitive parts of the 5G stack require real time platform, +In O-RAN architecture, the O-DU and O-CU could have different deployed scenarios. +The could be container based or VM based, which will be both supported in the release. +In general the performance sensitive parts of the 5G stack require real time platform, +especially for O-DU, the L1 and L2 are requiring the real time feature, the platform should support the Preemptive Scheduling feature. - -For example, from implementation perspective, the non-virtualized DU, -VM based DU and container based DU are requiring a real time host system. + Following requirements are going to address the container based solution: -1.Support Node Feature Discovery +1. Support the real time kernel +2. Support Node Feature Discovery +3. Support CPU Affinity and Isolation +4. Support Dynamic HugePages Allocation + +And for the network requirements, the following should be supported: +1. Multiple Networking Interface +2. High performance data plane including the DPDK based vswitch and PCI pass-through/SR-IOV. -2.Support CPU Affinity and Isolation +O-Cloud Components +------------------ +In this project, the following O-Cloud components and services are enabled: -3.Support Dynamic HugePages Allocation +1. Fault Management + - Framework for infrastructure services to raise and persist alarm and event data. + + - Set, clear and query customer alarms + - Generate customer logs for significant events -And for the network requirements, the following should be supported: -1.Multiple Networking Interface + - Maintains an Active Alarm List + - Provides REST API to query alarms and events, also available through SNMP traps + - Support for alarm suppression + - Operator alarms + + - On platform nodes and resources + - On hosted virtual resources + + - Operator logs - Event List + + - Logging of sets/clears of alarms + - Related to platform nodes and resources + - Related to hosted virtual resources + +2. Configuration Management + + - Manages Installation and Commissioning + + - Auto-discover of new nodes + - Full Infrastructure management + - Manage installation parameters (i.e. console, root disks) + + - Nodal Configuration + + - Node role, role profiles + - Core, memory (including huge page) assignments + - Network Interfaces and storage assignments + + - Hardware Discovery + + - CPU/cores, SMT, processors, memory, huge pages + - Storage, ports + - GPUs, storage, Crypto/compression H/W + +3. Software Management + + - Manages Installation and Commissioning + + - Auto-discover of new nodes + - Full Infrastructure management + - Manage installation parameters (i.e. console, root disks) + + - Nodal Configuration + + - Node role, role profiles + - Core, memory (including huge page) assignments + - Network Interfaces and storage assignments + + - Hardware Discovery + + - CPU/cores, SMT, processors, memory, huge pages + - Storage, ports + - GPUs, storage, Crypto/compression H/W + +4. Host Management + + - Full life-cycle and availability management of the physical hosts + - Detects and automatically handles host failures and initiates recovery + - Monitoring and fault reporting for: + + - Cluster connectivity + - Critical process failures + - Resource utilization thresholds, interface states + - H/W fault / sensors, host watchdog + - Activity progress reporting + + - Interfaces with board management (BMC) + + - For out of band reset + - Power-on/off + - H/W sensor monitoring +5. Service Management -2.High performance data plane including the DPDK based vswitch and PCI pass-through/SR-IOV. + - Manages high availability of critical infrastructure and cluster services + - Supports many redundancy models: N, or N+M + - Active or passive monitoring of services + - Allows for specifying the impact of a service failure and escalation policy + - Automatically recovers failed services -This is based on Yocto/OpenEmbedded, so it includes a Yocto/OpenEmbedded compatible -layers meta-oran and wrapper scripts to pull all required Yocto/OE layers to build -out the reference platform. + - Uses multiple messaging paths to avoid split-brain communication failures + + - Up to 3 independent communication paths + - LAG can also be configured for multi-link protection of each path + - Messages are authenticated using HMAC + - SHA-512 if configured / enabled on an interface by-interface basis + +6. Support the ansible bootstrap to implement the zero touch provisioning + +Enable the ansible configuration functions for infrastructure itself including the image installation and service configuration. + +NOTE: These features leverage the StarlingX (www.starlingx.io). And in current release, these features are only avalaible for IA platform. + +Multi OS and Deployment Configurations +-------------------------------------- + +The INF project supports Multi OS and currently the following OS are supported: + + - CentOS 7 + - Yocto 2.6 + +A variety of deployment configuration options are supported: + +1. **All-in-one Simplex** + + A single physical server providing all three cloud functions (controller, worker and storage). + +2. **All-in-one Duplex** + + Two HA-protected physical servers, both running all three cloud functions (controller, worker and storage), optionally with up to 50 worker nodes added to the cluster. + +3. **All-in-one Duplex + up to 50 worker nodes** + + Two HA-protected physical servers, both running all three cloud functions (controller, worker and storage), plus with up to 50 worker nodes added to the cluster. + +4. **Standard with Storage Cluster on Controller Nodes** + + A two node HA controller + storage node cluster, managing up to 200 worker nodes. + +5. **Standard with Storage Cluster on dedicated Storage Nodes** + + A two node HA controller node cluster with a 2-9 node Ceph storage cluster, managing up to 200 worker nodes. + +6. **Distributed Cloud** + + Distributed Cloud configuration supports an edge computing solution by providing central management and orchestration for a geographically distributed network of StarlingX systems. + +**NOTE:** + + - For CentOS based image, all the above deployment configuration are supported + - For Yocto Based image, only deployment 1 - 3 are supported, and only container based solution is supported, VM based is not supprted yet. About Yocto and OpenEmbedded ---------------------------- @@ -48,8 +183,13 @@ OpenEmbedded is the recommended build system of the Yocto Project, which is a Li Foundation workgroup that assists commercial companies in the development of Linux-based systems for embedded products. + +About StarlingX +--------------- +StarlingX is a complete cloud infrastructure software stack for the edge used by the most demanding applications in industrial IOT, telecom, video delivery and other ultra-low latency use cases. With deterministic low latency required by edge applications, and tools that make distributed edge manageable, StarlingX provides a container-based infrastructure for edge implementations in scalable solutions that is ready for production now. + Contact info ------------ If you need support or add new features/components, please feel free to contact the following: -- Jackie Huang -- Xiaohua Zhang + + - Jackie Huang