...
- DCAE deploys components using blueprints that are processed by the Cloudify Manager orchestration tools. The blueprints must provide a way to specify the target site for a deployment. (Typically the site will be set at the time of deployment using an input to the blueprint, rather than being hardcoded into the blueprint.)
- The DCAE Kubernetes plugin that Cloudify Manager invokes must be able to extract the target site from a blueprint and obtain information about the Kubernetes cluster at the target site (such as the address of the Kubernetes API server and the credentials needed to access the API server).
- Because we are not deploying separate, independent instances of the DCAE platform into each site, DCAE components running in remote locations will need a way to access services running at the central site. These include the config binding service, which provides components with their configurations, and, in some scenarios, the Consul server and the DMaaP DR and MR servers.
Dublin Scope
In the Dublin release, we are providing an interim solution for remote deployment that requires no development outside of the DCAE project. Ideally, there would be an ONAP-wide approach to support multiple sites, including a mechanism for bringing up new clusters and adding information about them to an ONAP-wide data store. We do not have this in Dublin–instead, DCAE keeps its own store of data about remote Kubernetes clusters.
In Dublin, we delivered the following changes into DCAE:
- We modified the Cloudify node type definitions for Kubernetes components to include a new node property, called
location_id
, that specifies the name of the site where the component should be deployed. - We modified the DCAE Kubernetes plugin for Cloudify to read the
location_id
for a component from the blueprint and to use thelocation_id
to find the target Kubernetes cluster and to deploy the component into the target cluster. - We adopted the Kubernetes "kubeconfig" file format for storing information about the Kubernetes clusters available as deployment targets. During the initial deployment of DCAE using Helm, we create a Kubernetes ConfigMap to hold the cluster information and automatically populate it with the data for the central site. In the Dublin release, the ConfigMap must be edited manually to add clusters. (As noted above, we believe there should be an ONAP-wide store for this data, and we hope that when we have such a store, the process of adding data for a cluster can be automated.)
- We allow components deployed into remote sites to access central site services through proxies (using nginx as the server). We created a Helm chart to deploy and configure the proxy.
The remaining sections of this document describe how to add information to the cluster ConfigMap and how to use the Helm chart to deploy the proxy into remote sites.
Assumptions
Pre-requisite
Artifacts
...