Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Table of Contents

Draft

This is the initial draft of the design page.

Much of the information here has been pulled from this source document: Promethus-aggregation.pdf

Requirements

  • Collect hardware, infrastructure (Linux OS), Hypervisor and Kubelet statistics
  • Aggregate (&Silencing) them for consumption
    • Via rules
  • Provide visualization of statistics
    • At site granularity level
    • At compute node granularity level.
    • Historical level
    • By Applying user defined filters
  • Counter collection & aggregation service – Offload collection and aggregation at various levels.
    • At ONAP level
    • At Regional Site level
    • At Edge Site level
  • Placement decisions based on HPA resource health/availability.

...

For HPA telemetry capture use case, the idea is to create a client to receive alerts from the Alert Manager and update HPA information in A&AI.  See HPA - Telemetry OOF and A&AI

Grafana

Configured to recieve Grafana is configured to query data from Prometheus and display information on HPA telemetry.

Compute Host Components

There are many projects which can be used to export data to the Prometheus server.  The initial components listed here are expected to support an initial list of HPA attributes.

Node Exporter

The node exporter runs on the compute hosts and sends a variety of statistics about the node.  Some of these will be used to identify the HPA attributes and capacity of the compute host.

collectd

The collectd component is another exporter that can be executed on the compute host to collect information.  collectd has support for SR-IOV information that will be used to support the HPA telemetry.

HPA Additions

It's possible that some HPA attributes may not be supported by any existing exporter.  Some additions to existing or new components may be developed to provide support.

Sequence Diagram

Image Added

Development Phases

Phase 1

The initial phase of HPA Platform Telemetry will involve setting up and operating the basic components.

  1. Running Node Exporter and collectd on a set of compute hosts
  2. Setting up Prometheus and Grafana to run in ONAP (e.g. Helm charts, etc.)
  3. Configuring a set of Prometheus query rules that will expose interesting HPA information
  4. Configure Grafana to display interesting HPA information pulled from Prometheus

Design topics to resolve:

  1. Identify methods for doing target discovery - that is, adding compute hosts to the list that Prometheus monitors.
  2. Identify source, configurations and queries required to capture interesting HPA capabilities and capacities.  The initial target list of HPA information is:
    1. CPUs
    2. Memory (DDR)
    3. Disk
    4. Huge page
    5. PCIe VFs

Phase 2

With phase 2, the goal is to export HPA information into ONAP A&AI so that OOF can use the data during placement decisions.

Additional components needed:

  1. Set up Alert Manager
  2. Develop an HPA Info Export component that can interface with A&AI

Phase 3 and beyond

Some items that could be included here:

  1. Extend beyond the initial short list of HPA attributes to additional attributes that are relevant to ONAP operations (e.g. Closed Loop)
  2. A VES Proxy or Exporter capability
  3. ...