Introduction
ICN strives to automate the process of installing the local cluster controller to the greatest degree possible–"zero touch installation". Most of the work is done simply by booting up the jump host (Local Controller). Once booted, the controller is fully provisioned and begins to inspect and provision baremetal servers, until the cluster is entirely configured.
This document show step by step to configure the network, and deployment architecture for ICN BP.
License
Apache license v2.0
Deployment Architecture
The local controller is provisioned with the Metal3 Baremetal Operator and Ironic, which enable provisioning of Baremetal servers. The controller has three network connections to the baremetal servers: network A connects baremetal servers, network B is a private network used for provisioning the baremetal servers, and network C is the IPMI network, used for control during provisioning. In addition, the baremetal hosts connect to the network D, the SRIOV network.
In some deployment model, you can combine Net C and Net A to be the same networks, but developer should take care of IP Address management between Net A and IPMI address of the server.
Pre-installation Requirements
There are two main components in ICN Infra local controller - Local controller and Compute K8s cluster
Local controller:
Local controller will reside in the jump server to run the Metal3 operator, Binary provisioning agent operator and Binary provisioning agent restapi controller.
Compute k8s cluster:
Compute K8s cluster will actually run the workloads and it installed on Baremetal nodes
Hardware Requirements
Minimum Hardware Requirement:
All-in-one VM based deployment required at least 32 GB RAM and 32 CPU servers
Recommended Hardware Requirements
Recommended Hardware requirements 64GB Memory and 32 CPU servers, QAT card and SRIOV network cards
Software Prerequisites
Jump server required to be pre-installed with Ubuntu 18.04
Database Prerequisites
No Prerequisites for ICN BP
Other Installation Requirements
Jump Host Requirements
Jump server required to be installed with Ubuntu 18.04 server, and have 3 distinguished networks as shown in figure 1
Jump server Hardware Requirements
Local controller: at least three network interfaces.
Baremetal hosts: four network interfaces, including one IPMI interface.
Four or more hubs, with cabling, to connect four networks.
Hostname | CPU Model | Memory | Storage | 1GbE: NIC#, VLAN, (Connected extreme 480 switch) | 10GbE: NIC# VLAN, Network (Connected with IZ1 switch) |
---|---|---|---|---|---|
Jump | Intel 2xE5-2699 | 64GB | 3TB (Sata) | IF0: VLAN 110 (DMZ) | IF2: VLAN 112 (Private) |
Jump server Software Requirements:
ICN R2 release support Ubuntu 18.04 - ICN BP install all required software during "make install"
Network Requirements
Please refer the figure 1, for all the network requirement in ICN BP
Please sure you have 3 distinguished networks net A, Net B and Net C as mentioned in figure 1. Local controller uses the Net B and Net C to provision the Baremetal servers to do the OS provisioning.
Bare Metal Node Requirements
Compute k8s cluster:
Compute server Hardware Requirements:
(Tested as below)
Hostname | CPU Model | Memory | Storage | 1GbE: NIC#, VLAN, (Connected extreme 480 switch) | 10GbE: NIC# VLAN, Network (Connected with IZ1 switch) |
---|---|---|---|---|---|
node1 | Intel 2xE5-2699 | 64GB | 3TB (Sata) | IF0: VLAN 110 (DMZ) | IF2: VLAN 112 (Private) |
node2 | Intel 2xE5-2699 | 64GB | 3TB (Sata) | IF0: VLAN 110 (DMZ) | IF2: VLAN 112 (Private) |
node3 | Intel 2xE5-2699 | 64GB | 3TB (Sata) | IF0: VLAN 110 (DMZ) | IF2: VLAN 112 (Private) |
Compute server Software Requirements:
The local controller will install all the software in compute servers right from OS, the software required to bring up the Kubernetes cluster
Execution Requirements (Bare Metal Only)
ICN BP check all the precondition and execution requirements for both Baremetal and VM deployment
Installation High-Level Overview
Installation is two-step process and everything starts with one command "make install"
- Installation of the local controller in the edge location
- Installation of Compute cluster to run the workload invoked by the local controller in the edge location
Baremetal Deployment Guide
Install Bare Metal Jump Host
Creating a Node Inventory File
Preconfiguration for the local controller.
User required to provide the IPMI information of the edge server they required to connect to the local controller by editing node JSON sample file in the directory icn/deploy/metal3/scripts/nodes.json.sample as below. If you want to increase nodes, just add another array
{ "nodes": [ { "name": "edge01-node01", "ipmi_driver_info": { "username": "admin", "password": "admin", "address": "10.10.10.11" }, "os": { "image_name": "bionic-server-cloudimg-amd64.img", "username": "ubuntu", "password": "mypasswd" } }, { "name": "edge01-node02", "ipmi_driver_info": { "username": "admin", "password": "admin", "address": "10.10.10.12" }, "os": { "image_name": "bionic-server-cloudimg-amd64.img", "username": "ubuntu", "password": "mypasswd" } } ] }
Local controller Metal3 configuration Reference:
- node: The array of nodes required to add to local controller
- name: Name of the Baremetal to be provisioned by Metal3, and this name will be the hostname for the machine, once it is provisioned
- ipmi_driver_info: IPMI driver info is a json field, currently holds the IPMI information required for Ironic to send the IPMI tool command
- username: BMC username required to be provided for Ironic
- password: BMC password required to be provided for Ironic
- address: BMC server IPMI LAN IP address
- os: Baremetal machine OS information is a json field, currently holds the image name to be provisioned, username name and password for the login.
- image_name: images name should be in qcow2 format
- username: login username for the OS provisioned
- password: login password for the OS provisioned
Creating the Settings Files
Local controller network configuration Reference:
User will find the network configuration file named as "user_config.sh" in the icn parent folder
#!/bin/bash #Local controller - Bootstrap cluster DHCP connection #BS_DHCP_INTERFACE defines the interfaces, to which ICN DHCP deployment will bind #e.g. export BS_DHCP_INTERFACE="ens513f0" export BS_DHCP_INTERFACE= #BS_DHCP_INTERFACE_IP defines the IPAM for the ICN DHCP to be managed. #e.g. export BS_DHCP_INTERFACE_IP="172.31.1.1/24" export BS_DHCP_INTERFACE_IP= #Edge Location Provider Network configuration #Net A - Provider Network #If provider having specific Gateway and DNS server details in the edge location #export PROVIDER_NETWORK_GATEWAY="10.10.110.1" export PROVIDER_NETWORK_GATEWAY= #export PROVIDER_NETWORK_DNS="8.8.8.8" export PROVIDER_NETWORK_DNS= #Ironic Metal3 settings for provisioning network #Interface to which Ironic provision network to be connected #Net B - Provisioning Network #e.g. export IRONIC_INTERFACE="enp4s0f1" export IRONIC_INTERFACE= #Ironic Metal3 setting for IPMI LAN Network #Interface to which Ironic IPMI LAN should bind #Net C - IPMI LAN Network #e.g. export IRONIC_IPMI_INTERFACE="enp4s0f0" export IRONIC_IPMI_INTERFACE= #Interface IP for the IPMI LAN, ICN verfiy the LAN Connection is active or not #e.g. export IRONIC_IPMI_INTERFACE_IP="10.10.110.20" #Net C - IPMI LAN Network export IRONIC_IPMI_INTERFACE_IP=
Running
After configuring, Node inventory file and setting files. Please run "make install" from the ICN parent directory as shown below:
root@pod11-jump:# git clone "https://gerrit.akraino.org/r/icn" Cloning into 'icn'... remote: Counting objects: 69, done remote: Finding sources: 100% (69/69) remote: Total 4248 (delta 13), reused 4221 (delta 13) Receiving objects: 100% (4248/4248), 7.74 MiB | 21.84 MiB/s, done. Resolving deltas: 100% (1078/1078), done. root@pod11-jump:# cd icn/ root@pod11-jump:# vim Makefile root@pod11-jump:# make install
Following steps occurs once the "make install" command is given.
- All the software required to run the bootstrap cluster is being downloaded and installed
- Kubernetes cluster to maintain the Bootstrap cluster and all the servers in the edge location is installed
- Metal3 specific network configuration such as local DHCP server networking for each edge location, Ironic networking for both provisioning network and IPMI LAN network are identified and created
- Metal3 is launched with IPMI configuration as configured in "user_config.sh" and provision the Baremetal servers using IPMI LAN network. For more information refer the Debugging Failure section
- Metal3 launch verification run without a timeout of 60 mins, by checking the status of all the servers being provisioned or not,
- All servers are provisioned parallelly. For example, if your deployment is having 10 servers in the edge location. All the 10 servers are provisioned at the same time
- Metal3 launch verification take care of checking all the servers are provisioned, the network interfaces are up and provisioned with a provider network gateway and DNS server
- Metal3 launch verification checks the status of all servers given in user_config.sh to make sure all the servers are provisioned. For example, if 8 servers are provisioned and 2 servers are not provisioned, Launch verification make sure all servers are provisioned before launch Kubernetes clusters on those servers
- BPA Baremetal components are invoked with the mac address of the servers provisioned by metal3, BPA Baremetal components decide the cluster size and also the number of clusters required in the edge location
- BPA Baremetal runs the containerized KUD as a job for each cluster. KUD install the kubernetes cluster on the slice of servers and install ONAP4k8s and all other default plugins such as Multus, OVN, OVN4NFV, NFD, Virtlet, SRIOV, QAT
- BPA rest agent installed in the bootstrap cluster or jump server, and this install rest-api, rook/ceph, Mimio as the cloud storage. This provides a way for user to upload their own software, container images or os image to jump server
Virtual Deployment Guide
Standard Deployment Overview
Virtual deployment is used for the dev env using metal3 virtual deployment to create VM with PXE boot. VM Ansible scripts the node inventory file in the /opt/ironic. No setting is required from the user to deploy the virtual deployment. Virtual deployment is used for dev works.
Verifying the Setup - VMs
"make verify_all" install two VMs with name master-0 and worker-0 with 8GB RAM and 8vCPUs, And install k8s cluster on the VMs using the ICN - BPAoperator and install the ICN - BPA rest API verifier. BPA operator installs the Multi cluster KUD to bring up the kubernetes with all addons and plugins.
Verifying the Setup
ICN blueprint checks all the setup in both baremetal and VM deployment. Verify script will check the metal3 provision the OS in each baremetal nodes by checking with a timeout period of 60 sec and interval of 30. BPA operator verifies will check, whether the KUD is installation is complete by doing plain curl command to the Kubernetes cluster installed in baremetal and VM setup.
Baremetal Verifier: Run the "make bm_verifer", it will verify the bare-metal deployment
VM Verifier: Run the "make vm_verifier", it will verify the Virtual deployment