Overview
In its early releases, ONAP DCAE deployed all data collection and analysis components into a single Kubernetes cluster in a single central site. Since some operators will use ONAP to deploy network functions at multiple geographically distinct sites, and since it will often make sense to deploy DCAE data collection and analysis components close to the network functions the tools are monitoring, it will necessary for DCAE to be able to deploy components to multiple Kubernetes clusters at multiple sites.
Several things are needed to support deployment of DCAE components to remote sites.
- DCAE deploys components using blueprints that are processed by the Cloudify Manager orchestration tools. The blueprints must provide a way to specify the target site for a deployment. (Typically the site will be set at the time of deployment using an input to the blueprint, rather than being hardcoded into the blueprint.)
- The DCAE Kubernetes plugin that Cloudify Manager invokes must be able to extract the target site from a blueprint and obtain information about the Kubernetes cluster at the target site (such as the address of the Kubernetes API server and the credentials needed to access the API server).
- Because we are not deploying separate, independent instances of the DCAE platform into each site, DCAE components running in remote locations will need a way to access services running at the central site. These include the config binding service, which provides components with their configurations, and, in some scenarios, the Consul server and the DMaaP DR and MR servers.
Dublin Scope
In the Dublin release, we are providing an interim solution for remote deployment that requires no development outside of the DCAE project. Ideally, there would be an ONAP-wide approach to support multiple sites, including a mechanism for bringing up new clusters and adding information about them to an ONAP-wide data store. We do not have this in Dublin–instead, DCAE keeps its own store of data about remote Kubernetes clusters.
In Dublin, we delivered the following changes into DCAE:
- We modified the Cloudify node type definitions for Kubernetes components to include a new node property, called
location_id
, that specifies the name of the site where the component should be deployed. - We modified the DCAE Kubernetes plugin for Cloudify to read the
location_id
for a component from the blueprint and to use thelocation_id
to find the target Kubernetes cluster and to deploy the component into the target cluster. - We adopted the Kubernetes "kubeconfig" file format for storing information about the Kubernetes clusters available as deployment targets. During the initial deployment of DCAE using Helm, we create a Kubernetes ConfigMap to hold the cluster information and automatically populate it with the data for the central site. In the Dublin release, the ConfigMap must be edited manually to add clusters. (As noted above, we believe there should be an ONAP-wide store for this data, and we hope that when we have such a store, the process of adding data for a cluster can be automated.)
- We allow components deployed into remote sites to access central site services through proxies (using nginx as the server). We created a Helm chart to deploy and configure the proxy.
There is more information about these changes in this presentation.
The remaining sections of this document describe how to add information to the cluster ConfigMap and how to use the Helm chart to deploy the proxy into remote sites.