Contents
Table of Contents |
---|
Summary
The OOF plans to provide optimization capability as a service for ONAP R2 and beyond. OOF uses a typical optimization construct:
- Objective: Maximize/minimize a metric, measured by appropriate key performance indicators (KPIs)
- Technology and operating constraints, such as:
- Parameter change limits (such as power)
- Frequency of changes permitted
- Number of parameters that can be changed simultaneously
- Data latencies (typically in percentile)
- DC compute, network, storage, energy capacity
- Location based and time based energy cost
The objective metrics could be throughput (maximize), interference levels (minimize), accessibility/retainability (maximize), cost (minimize) etc. KPIs could be infrastructure utilization statistics provided by ONAP-MC.
The OOF is developed based on the following core ideas:
- Most optimization problems can be solved in a declarative manner using a high-level modeling language.
- Recent advances in open source optimization platforms allow the solution process to be mostly solver-independent.
- By leveraging the library of standard/global constraints, optimization models can be rapidly developed.
- By developing a focused set of platform components, we can realize a policy-driven, declarative system that allows ONAP optimization applications be composed rapidly and managed easily
- Policy and data adapters
- Execution and management environment
- Curated "knowledge base" and recipes to provide information on typical optimization examples and how to use the OOF
- More importantly, by providing a way to support both "traditional" optimization applications and model-driven applications, we can provide a choice for users to adapt the platform based on their business needs and skills/expertise.
The OOF aims to realize these via a set of initial applications for ONAP use cases that are being developed collaboratively across a broad team.
Rationale and Motivation
Traditional Optimization
...
Traditionally, optimization applications are tailor-made for specific requirements, and the process for developing an optimization application often involves substantial application-specific "custom code". Any changes in the problem (e.g. new optimization constraints, objectives, or data sources) requires development effort involving code changes in various components of the application. These changes can span aspects such as (a) Optimization Model Specification, (b) Request Handler, (c) Adapters for data and parameters, (d) Application Configuration, (e) Code in custom solver, etc., and involve long development cycles even for simple changes in requirements.
Declarative, Policy- and Model-Driven Architecture for Optimization Applications
The goal of the Optimization Framework is to drastically reduce the amount of such code changes by providing platform-level functionality.
...
Features and Utility of OOF
The main features of the ONAP Optimization Framework (OOF) are:
- It provides a robust, scalable optimization framework for rapidly developing new optimization applications independent of how the underlying optimization modules are implemented.
- It enables reusability of optimization engines, addresses problems arising due to different applications using custom optimization codes, adapter libraries, and configuration logic. This is achieved via a policy-driven configuration system, and a library of generalized optimizers that can be configured via policies.
- OOF-based solutions can be quickly on-boarded onto ONAP, linked to various data collectors, databases, and microservices in ONAP/DCAE, and can dynamically be scaled at run-time.
- Overall, the OOF eliminates software redundancy and inconsistencies arising from variations in quality and configurability of different optimizers. The unified approach of OOF reduces the overhead associated with managing different optimization applications.
Advantages of the Unified Approach of OOF
- OOF is policy driven.
- OOF provides mechanisms to specify optimization constraints as policies that are configurable by service designers or operators. In contrast, legacy optimization applications include such information inside configuration files and sometimes in the code.
- Constraints and other policies are available for multiple uses, which encourages reusability. This reduces inconsistencies in constraints or policies across services, and helps reduce duplication of effort for common tasks.
- OOF provides reusable, model-driven adapters for data sources and external systems
- Data formats and API calls are model-driven in OOF, so likely errors are identified very early in the request-response sequence. OOF provides adapters to different data sources (DCAE as well as external systems), which can be directly reused.
- OOF is agnostic towards service, application, and optimization engine technology/language
- Optimizers that can be chained together in a technology and programming-language agnostic manner (e.g. general purpose mathematical solvers such as GLPK and CPLEX can co-exist with custom-implementations of algorithms). New optimization solutions can be composed by chaining existing optimizers (e.g. by linking placement, networking, and licensing optimizations).
- OOF provides a dynamically scalable, fault-tolerant environment with resource pooling
- Runtime environment is based on ONAP/DCAE's Hadoop/Yarn technologies (and can be easily adapted to other cluster technologies such as kubernetes).
- OOF uses a queue-based system with independent ``workers'' processing optimization tasks. These workers can be dynamically scaled and jobs are picked up by the next available worker if a worker container fails.
Contributors
Overall Architecture/Design: AT&T, VMWare, Intel
Core OOF Components (Adapters, architecture, seed code): AT&T
Underlying optimization platform: Code developed by University of Melbourne and Monash University
Packaging and verification of OOF System: VMWare, NetCracker, AT&T
Policies and interpretation: Intel, AT&T
Homing and Allocation Service: AT&T, Intel
CI/CD and Test Coverage: NetCracker, AT&T
Modification of Adapters: Huawei, AT&T, VMWare
Functional Architecture
Technology Choices
Minizinc provides an open source constraint modeling language/platform for specifying optimization applications. It contains direct interfaces to COIN-OR CBC, Gurobi and IBM ILOG CPLEX. Additionally, many optimization projects support minizinc via FlatZinc interfaces. The Minizinc standard library provides a subset of constraints form the global constraint catalogue as a high-level abstraction that have efficient algorithms implemented by several solvers.
...
Contents
Table of Contents |
---|
Summary
The OOF plans to provide optimization capability as a service for ONAP R2 and beyond. OOF uses a typical optimization construct:
- Objective: Maximize/minimize a metric, measured by appropriate key performance indicators (KPIs)
- Technology and operating constraints, such as:
- Parameter change limits (such as power)
- Frequency of changes permitted
- Number of parameters that can be changed simultaneously
- Data latencies (typically in percentile)
- DC compute, network, storage, energy capacity
- Location based and time based energy cost
The objective metrics could be throughput (maximize), interference levels (minimize), accessibility/retainability (maximize), cost (minimize) etc. KPIs could be infrastructure utilization statistics provided by ONAP-MC.
The OOF is developed based on the following core ideas:
- Most optimization problems can be solved in a declarative manner using a high-level modeling language.
- Recent advances in open source optimization platforms allow the solution process to be mostly solver-independent.
- By leveraging the library of standard/global constraints, optimization models can be rapidly developed.
- By developing a focused set of platform components, we can realize a policy-driven, declarative system that allows ONAP optimization applications be composed rapidly and managed easily
- Policy and data adapters
- Execution and management environment
- Curated "knowledge base" and recipes to provide information on typical optimization examples and how to use the OOF
- More importantly, by providing a way to support both "traditional" optimization applications and model-driven applications, we can provide a choice for users to adapt the platform based on their business needs and skills/expertise.
The OOF aims to realize these via a set of initial applications for ONAP use cases that are being developed collaboratively across a broad team.
Rationale and Motivation
Traditional Optimization
Gliffy | ||||||||
---|---|---|---|---|---|---|---|---|
|
Traditionally, optimization applications are tailor-made for specific requirements, and the process for developing an optimization application often involves substantial application-specific "custom code". Any changes in the problem (e.g. new optimization constraints, objectives, or data sources) requires development effort involving code changes in various components of the application. These changes can span aspects such as (a) Optimization Model Specification, (b) Request Handler, (c) Adapters for data and parameters, (d) Application Configuration, (e) Code in custom solver, etc., and involve long development cycles even for simple changes in requirements.
Declarative, Policy- and Model-Driven Architecture for Optimization Applications
The goal of the Optimization Framework is to drastically reduce the amount of such code changes by providing platform-level functionality.
Gliffy | ||||||||
---|---|---|---|---|---|---|---|---|
|
Features and Utility of OOF
The main features of the ONAP Optimization Framework (OOF) are:
- It provides a robust, scalable optimization framework for rapidly developing new optimization applications independent of how the underlying optimization modules are implemented.
- It enables reusability of optimization engines, addresses problems arising due to different applications using custom optimization codes, adapter libraries, and configuration logic. This is achieved via a policy-driven configuration system, and a library of generalized optimizers that can be configured via policies.
- OOF-based solutions can be quickly on-boarded onto ONAP, linked to various data collectors, databases, and microservices in ONAP/DCAE, and can dynamically be scaled at run-time.
- Overall, the OOF eliminates software redundancy and inconsistencies arising from variations in quality and configurability of different optimizers. The unified approach of OOF reduces the overhead associated with managing different optimization applications.
Advantages of the Unified Approach of OOF
- OOF is policy driven.
- OOF provides mechanisms to specify optimization constraints as policies that are configurable by service designers or operators. In contrast, legacy optimization applications include such information inside configuration files and sometimes in the code.
- Constraints and other policies are available for multiple uses, which encourages reusability. This reduces inconsistencies in constraints or policies across services, and helps reduce duplication of effort for common tasks.
- OOF provides reusable, model-driven adapters for data sources and external systems
- Data formats and API calls are model-driven in OOF, so likely errors are identified very early in the request-response sequence. OOF provides adapters to different data sources (DCAE as well as external systems), which can be directly reused.
- OOF is agnostic towards service, application, and optimization engine technology/language
- Optimizers that can be chained together in a technology and programming-language agnostic manner (e.g. general purpose mathematical solvers such as GLPK and CPLEX can co-exist with custom-implementations of algorithms). New optimization solutions can be composed by chaining existing optimizers (e.g. by linking placement, networking, and licensing optimizations).
- OOF provides a dynamically scalable, fault-tolerant environment with resource pooling
- Runtime environment is based on ONAP/DCAE's Hadoop/Yarn technologies (and can be easily adapted to other cluster technologies such as kubernetes).
- OOF uses a queue-based system with independent ``workers'' processing optimization tasks. These workers can be dynamically scaled and jobs are picked up by the next available worker if a worker container fails.
Contributors
Overall Architecture/Design: AT&T, VMWare, Intel
Core OOF Components (Adapters, architecture, seed code): AT&T
Underlying optimization platform: Code developed by University of Melbourne and Monash University
Packaging and verification of OOF System: VMWare, NetCracker, AT&T
Policies and interpretation: Intel, AT&T
Homing and Allocation Service: AT&T, Intel
CI/CD and Test Coverage: NetCracker, AT&T
Modification of Adapters: Huawei, AT&T, VMWare
Functional Architecture
Technology Choices
Minizinc provides an open source constraint modeling language/platform for specifying optimization applications. It contains direct interfaces to COIN-OR CBC, Gurobi and IBM ILOG CPLEX. Additionally, many optimization projects support minizinc via FlatZinc interfaces. The Minizinc standard library provides a subset of constraints form the global constraint catalogue as a high-level abstraction that have efficient algorithms implemented by several solvers.
Components of the Core Framework
Gliffy | ||||||||
---|---|---|---|---|---|---|---|---|
|
An overview of the components of the core optimization framework. The OOF utilizes the open source project Minizinc, which has a solver-independent modeling language and has interfaces to various open source and commercial solvers. One of the additional benefits of this approach is that by developing a focused set of ONAP-related components, we can utilize ongoing advances in optimization technologies, as well as adapt other currently available extensions to Minizinc and related projects. The OOF project aims to build these components with a focus on minimal viable product for Beijing Release in order to support initial applications and use cases, with subsequent focus on expanding the platform.
Data Adapter Library
The OOF will provide a library of adapters for common ONAP systems. These can be directly used in data specification templates of the applications. In the initial release, these will include adapters to Policy, A&AI, SDN-C, Multi-Cloud, and SDC. As new use cases are implemented, this library will be augmented by new adapters to other services.
Translation Modules
The OOF will provide modules for translating policies into constraints for the optimization environment. When an underlying minizinc model is used for optimization, it is translated into a minizinc constraint (either via a data specification template or directly from policy). For custom optimizers, these constraints will be translated to the input format expected by the optimizer via the data specification template. The Homing and Allocation Service (HAS; described in the next section) uses a custom optimization module and hence uses the data specification template approach.
Modeling Support
The OOF provides simple templating system through which users can specify links to different ONAP components, including policy, A&AI, SDC, etc.
Execution Environment
The OOF execution environment contains the minizinc system, along with data/template rendering system that leverages the adapters to various systems. The flow of execution can be configured via the configuration file for the application and supports a choice of specific solver or invocations to custom/external solvers.
Simple Example: Budget Constrained Maximum Network Flow Problem
Problem Description
Gliffy
Gliffy | ||||||||
---|---|---|---|---|---|---|---|---|
|
An overview of the components of the core optimization framework. The OOF utilizes the open source project Minizinc, which has a solver-independent modeling language and has interfaces to various open source and commercial solvers. One of the additional benefits of this approach is that by developing a focused set of ONAP-related components, we can utilize ongoing advances in optimization technologies, as well as adapt other currently available extensions to Minizinc and related projects. The OOF project aims to build these components with a focus on minimal viable product for Beijing Release in order to support initial applications and use cases, with subsequent focus on expanding the platform.
Data Adapter Library
The OOF will provide a library of adapters for common ONAP systems. These can be directly used in data specification templates of the applications. In the initial release, these will include adapters to Policy, A&AI, SDN-C, Multi-Cloud, and SDC. As new use cases are implemented, this library will be augmented by new adapters to other services.
Translation Modules
The OOF will provide modules for translating policies into constraints for the optimization environment. When an underlying minizinc model is used for optimization, it is translated into a minizinc constraint (either via a data specification template or directly from policy). For custom optimizers, these constraints will be translated to the input format expected by the optimizer via the data specification template. The Homing and Allocation Service (HAS; described in the next section) uses a custom optimization module and hence uses the data specification template approach.
Modeling Support
The OOF provides simple templating system through which users can specify links to different ONAP components, including policy, A&AI, SDC, etc.
Execution Environment
The OOF execution environment contains the minizinc system, along with data/template rendering system that leverages the adapters to various systems. The flow of execution can be configured via the configuration file for the application and supports a choice of specific solver or invocations to custom/external solvers.
Simple Example: Budget Constrained Maximum Network Flow Problem
Problem Description
|
Illustration of the budget-constrained max-flow problem. Given a set of starting nodes and destination nodes, the objective is to calculate the maximum amount of "flow" across the nodes, subject to capacity constraints of nodes, constraints on the capacity of the connections (edges), as well as costs associated with utilization of each connection (edge). The objective is to maximize the flow subject to the budget constraint.
This model can be composed from different components, each developed by contrbutors with different expertise and roles. The development can happen at different times (initial problem definition, service design time, run-time, and continuous improvement stages). Contributors can have different roles and expertise. Examples include:
- An optimization modeler (with expertise on modeling and mathematical aspects of optimization)
- A software developer with knowledge of data modeling and model/data templates
- A developer with expertise in creating policy models in the context of specific business domains
- An Ops team member with knowledge of applicable policies for a set of services and applications and knowledge on run-time configuration
An optimization model can therefore evolve from a mathematical concept to a computer program to a policy-driven, dynamically re-configurable application through the following stages:
An optimization modeler identifies the key concepts, comes up with a mathematical model, creates an optimization model, and tests with a small example dataset (a prototype corresponding to the model shown in the problem description above.
Then, a developer with an understanding of the API requirements, basic understanding of the model and data sources/adapters links the model to a data template (template file shown below). The developer needs a basic understanding of key variables of a model such as nodes, bandwidths, capacities in this example. While most adapters will be available from OSDF library, additional adapters/libraries can be developed and can be contributed back to the ODSF library.
A service designer, with basic understanding of key model concepts (key variables of a model such as nodes, bandwidths, capacities in this example) can create policy models (or use/extend existing constraint policy models from the OSDF library). In this example, this introduces a new constraint reflecting the maximum amount of flow that can go through a single network link, in order to reduce service disruption risks.
On top of this, an Ops person can enable or configure an applicable run-time operational constraint policy. In this example, this introduces a more stringent budget constraint (from all allowed budget to only 80% of the budget).
Thus, a new service can be created by extending existing optimization models, policies, and adapters, and using them as building blocks/ingredients.
Section | |||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| |||||||||||||||||||||
Minizinc Model
Minizinc Data Template
Minizinc Data File
|
...