This page captures the information and steps
...
to deploy ONAP using OOM on VIO 4.0 with Kubernetes.
This pages assumes that
...
VIO 4.0 is deployed successfully with Kubernetes
...
and 'OpenStack' was selected as the Cloud Provider during configuration.
VMware VIO 4.0 Kubernetes Architecture
a. Prerequisites -
...
Prerequisites
Mgmt IP Address, username and password
...
to create Kubernetes cluster .
...
K8S Master and K8S Node instances should have a OpenStack Flavor attached as per the requirement below
...
vCPU | 48 |
RAM | 96GB |
Storage | 256GB |
Steps to create a Kubernetes cluster
Follow the below steps 1
...
- 4 to create a Kubernetes cluster
...
...
Step
...
5 onwards are needed for getting the Kubernetes Host IP so that user can login to Kubernetes Host and ONAP
...
using OOM can be deployed.
...
TODO 20171205 - steps that needs to be done prior to git Clone of OOM .
...
Step-1 - https://
...
MGMT_IP_ADDRESS/LOGIN
Step2
Step-
...
2 Create the Cloud Provider before creating a kubernetes cluster
Cloud Provider creation is prerequistie a prerequisite to Kubernetes cluster creation. VMware Integrated OpenStack VIO with Kubernetes uses the cloud provider to create the infrastructure required to deploy all your Kubernetes clusters. VMware currently supports 2 options for infrastructure provider. VMware SDDC (vSphere + NSX + VSAN) or VIO OpenStack (i.e. VMware Integrated OpenStack). When choosing the type of provider to create, consider the following:
- With an existing VMware Integrated OpenStack VIO deployment, you can create an OpenStack provider.
- Without an existing VMware Integrated OpenStack VIO deployment, you can create an SDDC provider Im.
Step3
Step-
...
3 Create the
...
Step4 - Creating the Kubernetes cluster
...
Kubernetes cluster
3.1 - Click '+NEW' to create a kubernetes cluster
3.2 - Click NEXT
43.2 3 - Select an Infrastructure Provider for creating the kubernetes
Before you deploy a Kubernetes cluster, you must have created create the cloud provider. VMware Integrated OpenStack with Kubernetes uses the cloud provider to create the infrastructure required to deploy all your Kubernetes clusters. VMware currently supports 2 options for infrastructure provider. VMware SDDC (vSphere + NSX + VSAN) or VIO (VMware Integrated OpenStack).
- Provider name can be custom name given by user where as provider Type has to be openstack or sddc .
4.3 Cloud provides can be SDDC or OpenStack. Select the option as appropriate.
Here, with VIO with Kubernetes, we select 'OpenStack' as cloud provider
3.4 - Select a Node Profile. If you have more than one node profiles, uncheck the box "Use default node profile" to see the list.
43.4 5 - Provide the Input for the Cluster as indicated in Example Data below
...
Cluster type: Exclusive Cluster
43.5 - Fill the details in User & Group page6 - Add Users and Groups for this cluster
Once a Kubernetes cluster is created, you can authorize users or groups for the cluster. The users and groups belong to the SDDC or OpenStack provider where the cluster was created..
In the Configure user and group for cluster dialogue box, check the boxes for users or groups that you want to authorize for access to the cluster.
Or check off the boxes for users or groups that you no longer want to authorize for access to the cluster.
43.6 7 - Click on FINISH and wait for few minutes for the kubernetes cluster to get created .
Step5. Verifying
Step-4 Verifying the VIO Kubernetes Cluster -
If everything in step 3.1 to step 3.6 has been done successfully .The Summary information for the cluster will be filled as given in the example below .
Step6. Step-5 How to get Kubernetes Host host IP Address and login to Kubernetes Host host .
65.1 - Login via console window to BLUESHIFT_MGMT_IP_ADDRESS . user name and password is same as used in step1 .
65.2 - Once logged into BLUESHIFT_MGMT_IP_ADDRESS session use the command "vkube login --insecure" . use the user name and password same as in step1 .
65.3 - Get the list of clusters using command "vkube cluster list --insecure" . make a note of cluster Id in the output of the command
65.4 - Get the cluster node details using command "vkube cluster show <cluster Id > --insecure " .make note of worker ip address .Worker IP Address KUBERNETES_HOST_IP.
65.5 - Login to app-api docker using docker exec it app-api bash
65.6 - Once inside the app-api docker go to /var/lib/vrc/terraform/<cluserId>
65.7 - ssh to KUBERNETES_HOST_IP_ADDRESS using below command
ssh -i private.key -F ssh.bastion.conf ubuntu@KUBERNETES_HOST_IP_ADDRESS
65.8 - Once inside KUBERNETES_HOST_IP_ADDRESS (to be reviewed)
Step7. Step-6 Installing kubectl to manage Kubernetes cluster
TBC20171207
7cluster
6.1 - Download the kubectl using below command
curl -LO https://storage.googleapis.com/kubernetes-release/release/v1.7.0/bin/linux/amd64/kubectl
76.2 - Make the kubectl binary executable.
chmod +x ./kubectkubectl
76.3 - Move the kubectl to PATH
sudo mv ./kubectl /usr/local/bin/kubectl
Step8. Step-7 Verifying that kubernetes kubectl config is good good
7.1 - On Kubernetes cluster
root@localhost:
...
~# kubectl cluster-info
Kubernetes master is running
...
at ....
Heapster is running at....
KubeDNS is running at ....
kubernetes-dashboard is running at ...
monitoring-grafana is running at ....
monitoring-influxdb is running at ...
tiller-deploy is running at....
7.2 - On client from Where Kubernetes cluster can be managed Remotely
root@localhost:~# kubectl version
Client Version: version.Info{Major:"1", Minor:"7", GitVersion:"v1.7.0", GitCommit:"d3ada0119e776222f11ec7945e6d860061339aad", GitTreeState:"clean", BuildDate:"2017-06-29T23:15:59Z", GoVersion:"go1.8.3",
Compiler:"gc", Platform:"linux/amd64"}
Server Version: version.Info{Major:"1", Minor:"7+", GitVersion:"v1.7.7-rancher1", GitCommit:"a1ea37c6f6d21f315a07631b17b9537881e1986a", GitTreeState:"clean", BuildDate:"2017-10-02T21:33:08Z",GoVersion:"go1.8.3" Compiler:"gc", Platform:"linux/amd64"}
Step-8 Verifying that kube config is good
8.1 - On Kubernetes cluster
root@localhost:~# cat ~/.kube/config
apiVersion: v1
kind: Config
clusters:
- cluster:
api-version: v1
insecure-skip-tls-verify: true
server: "<SERVER_IP_ADDRESS:8080/r/projects/CLUSTER_NAME/kubernetes:SERVER_PORT_NUMBER"
name: "(CLUSTER_NAME)"
contexts:
- context:
cluster: "(CLUSTER_NAME)"
user: "(CLUSTER_NAME)"
name: "(CLUSTER_NAME)"
current-context: "(CLUSTER_NAME)"
users:
- name: "(CLUSTER_NAME/USER_NAME)"
user:
token: "<SECURITY_TOKEN>"
8.2 - On client from where Kubernetes cluster can be managed Remotely
root@localhost:~# cat ~/.kube/config
current-context: default-context
apiVersion: v1
clusters:
- cluster:
api-version: v1
server: https://SERVER_IP_ADDRESS:SERVER_PORT_NUMBER/
insecure-skip-tls-verify: true
name: CLUSTER_NAME
contexts:
- context:
cluster: CLUSTER_NAME
namespace: default
user: user1
name: default-context
users:
- name: user1
user:
username: "<USERNAME>"
password: "<PASSWORD"
kind: Config
preferences:
colors: true
Step-9 Installing Helm
9.1 - Download the helm using below command
wget http://storage.googleapis.com/kubernetes-helm/helm-v2.3.0-linux-amd64.tar.gz
9.2 - Untar the fileDownload the helm using below command
tar -zxvf helm-v2.3.0-linux-amd64.tar.gz
...
sudo mv linux-amd64/helm /usr/local/bin/helm
Step-10 Verifying Helm
10.1 - Type the below command
helm help
10.2 - Helm version .
Client: &version.Version{SemVer:"v2.3.0", GitCommit:"d83c245fc324117885ed83afc90ac74afed271b4", GitTreeState:"clean"}
Server: &version.Version{SemVer:"v2.3.0", GitCommit:"d83c245fc324117885ed83afc90ac74afed271b4", GitTreeState:"clean"}