Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 162 Next »

vCPE Use Case - Customer Service Instantiation - 171103.pdf


11/11/2017

  1. Regression test in SB01
    1. SO API handler dropped source from VID request, causing error in BPMN. Fixed. See  SO-340 - Getting issue details... STATUS
    2. SO has cannot pass AAI authentication. The problem is that AAI replaced the old cert (exp 11/30/2017) with a new one (exp dec. 2018). The solution is to use the CA in SO. 
      1. In SO docker, find aai.crt, replace the content with this ca_bundle_for_openecomp.txt (which includes both root and intermediate CA).
      2. run "update-ca-certificates -f"
      3. restart SO docker.
    3. SO failed to query AAI cloud region, the problem is the default config in /etc/mso/config.d/mso.bpmn.urn.properties. The correct lines are below:

      mso.workflow.default.aai.v11.tenant.uri=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/CloudOwner/RegionOne/tenants/tenant
      mso.workflow.default.aai.v11.cloud-region.uri=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/CloudOwner
      mso.workflow.default.aai.v11.cloud-region.uri=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/CloudOwner
    4. Preloading works!
    5. Successfully created infrastructure!

11/10/2017

  1. Regression test in SB01
    1. An ONAP based on the latest build was installed on 11/12. Need to go to appc_vm and change /opt/config/docker_version.txt: 1.1-STAGING-latest to 1.2-STAGING-latest.

    2. Onboarding, service design, service distribution completed in SB01.
    3. SO does not add recipe automatically. Manual insertion was done.
    4. SDNC does not insert correct data to its DB. Tracked by  SDNC-194 - Getting issue details... STATUS
      1. For ALLOTTED_RESOURCE_MODEL, BRG and TunnelXConn do not have 'Y' for its ecomp_generated_naming. Their types are "VF", and their 'allotted resource type' fields are null.
      2. VF_MODEL is good.
      3. SERVICE_MODEL is good.
      4. VF_MODULE_MODEL is good.
    5. VID has a problem to create services. Tracked by  VID-91 - Getting issue details... STATUS

  2. vCpeResCust custom workflow:
    1. Jim fixed xml parsing problem for service delete and now the flow can pick up vgw, tunnelxconn AR, and vbrg AR from the service info returned from AAI. An error was captured when SO requests SDNC for brg deactivate. The request to SDNC is below


      <sdncadapterworkflow:SDNCAdapterWorkflowRequest xmlns:ns5="http://org.openecomp/mso/request/types/v1"
                                                                                                              xmlns:sdncadapterworkflow="http://org.openecomp/mso/workflow/schema/v1"
                                                                                                              xmlns:sdncadapter="http://org.openecomp/workflow/sdnc/adapter/schema/v1">
                                         <sdncadapter:RequestHeader>
                                                              <sdncadapter:RequestId>b6106b00-b9f1-418c-9dc3-aca97664cd05</sdncadapter:RequestId>
                                                              <sdncadapter:SvcInstanceId>41235b16-5c38-4c29-82e9-6784d9decf46</sdncadapter:SvcInstanceId>
                                                              <sdncadapter:SvcAction>deactivate</sdncadapter:SvcAction>
                                                              <sdncadapter:SvcOperation>brg-topology-operation</sdncadapter:SvcOperation>
                                                              <sdncadapter:CallbackUrl>http://mso:8080/mso/SDNCAdapterCallbackService</sdncadapter:CallbackUrl>
                                              </sdncadapter:RequestHeader>
                                      <sdncadapterworkflow:SDNCRequestData>
                                              <request-information>
                                                      <request-id>2ba02fea-7a07-4641-87fb-abb10a305b44</request-id>
                                                      <request-action>DeleteBRGInstance</request-action>
                                                      <source>MSO</source>
                                                      <notification-url/>
                                                      <order-number/>
                                                      <order-version/>
                                              </request-information>
                                              <service-information>
                                                      <service-id></service-id>
                                                      <subscription-service-type>vCPE</subscription-service-type>
                                                      <onap-model-information></onap-model-information>
                                                      <service-instance-id>41235b16-5c38-4c29-82e9-6784d9decf46</service-instance-id>
                                                      <subscriber-name/>
                                                      <global-customer-id>SDN-ETHERNET-INTERNET</global-customer-id>
                                              </service-information>
                                              <allotted-resource-information>
                                                      <allotted-resource-id>99dc7978-3efe-4074-805c-2ae5dc785c88</allotted-resource-id>
                                                      <allotted-resource-type>brg</allotted-resource-type>
                                                      <parent-service-instance-id>e565bb6b-de14-4a5c-a992-65a681771a7a</parent-service-instance-id>
                                                      <onap-model-information>
                                                              <model-invariant-uuid></model-invariant-uuid>
                                                              <model-uuid></model-uuid>
                                                              <model-customization-uuid></model-customization-uuid>
                                                              <model-version></model-version>
                                                              <model-name></model-name>
                                                      </onap-model-information>
                                              </allotted-resource-information>
                                              <brg-request-input>
                                              </brg-request-input>
                                      </sdncadapterworkflow:SDNCRequestData>
                                      </sdncadapterworkflow:SDNCAdapterWorkflowRequest>
      
      
      

      The AAI request from SDNC is below. Note that a few values are 'null'

      2017-11-13 17:07:01,916 | INFO  | SvcLogicGraph [module=GENERIC-RESOURCE-API, rpc=brg-topology-operation-deactivate, mode=sync, version=1.2.0-SNAPSHOT] | Request URL : https://aai.api.simpledemo.openecomp.org:8443/aai/v11/business/customers/customer/null/service-subscriptions/service-subscription/null/service-instances/service-instance/null/allotted-resources/allotted-resource/99dc7978-3efe-4074-805c-2ae5dc785c88

11/10/2017

  1. vCpeResCust custom workflow:
    1. Brian made changes on the SDNC side. Now SDNC can pass a list of parameters to SO for vG assign call. Then SO passes those parameters to HEAT to instantiate vG.
    2. Jim fixed the DoDeleteVfModule flow to use generic-resource-api and construct the corresponding request body when performing SDNC deactivate call. Note that the flow checks the configuration variable sdncversion to determine what request body to construct. This is something not fully understood by the team.
    3. Delete of vG succeeded.
    4. Jim continues to work on  SO-325 - Getting issue details... STATUS .
    5. Eric has modified vGMUX to add a workaround, see below:

      vGMUX set sourceName
      There is a vgmux image in the ONAP-vCPE project space called:  vgmux2-base-ubuntu-16-04
      
      This one has the ability to configure the sourceName in the VES event to something different than the default value (which is the vnf-id present in the vm’s openstack metadata).
      
      Some documentation:
      
      Configuring the VES mode - via REST
       
                     This will set the ‘demo’ mode and packet loss to 40%, but does ‘not’ change the sourceName:
      curl -i -H "Content-Type:application/json" --data '{"mode":{"working-mode":"demo","base-packet-loss":40,"source-name":""}}' -X POST -u admin:admin http://127.0.0.1:8183/restconf/config/vesagent:vesagent
       
      Delete the config in order to change it via REST:
      curl -i -H "Content-Type:application/json"  -X DELETE -u admin:admin http://127.0.0.1:8183/restconf/config/vesagent:vesagent/mode
       
      curl -i -H "Content-Type:application/json" --data '{"mode":{"working-mode":"demo","base-packet-loss":88,"source-name":"testing-123-ABC"}}' -X POST -u admin:admin http://127.0.0.1:8183/restconf/config/vesagent:vesagent
       
       
      Configuring the VES mode - via CLI
                     QUERY:  
      # vppctl show ves mode
        Mode   Base Packet Loss Rate   Source Name
        Demo                   88.0%   testing-123-ABC
       
      SET:
      vppctl set ves mode demo base 77 source hello-there
       
      This sets the sourceName to “hello-there”
      Leave off the 'source <name>' arguments to set back to default (i.e. vnf-id from openstack metadata)
       
       
      Sample event with an overwritten sourceName:
      {
          "event": {
              "commonEventHeader": {
                  "domain": "measurementsForVfScaling",
                  "eventId": "Generic_traffic",
                  "eventName": "Measurement_vGMUX",
                  "eventType": "HTTP request rate",
                  "lastEpochMicrosec": 1510347222243201,
                  "priority": "Normal",
                  "reportingEntityId": "No UUID available",
                  "reportingEntityName": "zdcpe1cpe01mux01",
                  "sequence": 23,
                  "sourceId": "vCPE_Infrastructure_vGMUX_demo_app",
                  "sourceName": "testing-123-ABC",
                  "startEpochMicrosec": 1510347212243201,
                  "version": 1.2
              },
              "measurementsForVfScalingFields": {
                  "additionalMeasurements": [
                      {
                          "arrayOfFields": [
                              {
                                  "name": "Packet-Loss-Rate",
                                  "value": "88.0"
                              }
                          ],
                          "name": "ONAP-DCAE"
                      }
                  ],
                  "cpuUsageArray": [
                      {
                          "cpuIdentifier": "cpu1",
                          "cpuIdle": 100.0,
                          "cpuUsageSystem": 0.0,
                          "cpuUsageUser": 0.0,
                          "percentUsage": 0.0
                      }
                  ],
                  "measurementInterval": 10,
                  "measurementsForVfScalingVersion": 2.1,
                  "requestRate": 2567,
                  "vNicUsageArray": [
                      {
                          "receivedOctetsDelta": 0.0,
                          "receivedTotalPacketsDelta": 0.0,
                          "transmittedOctetsDelta": 0.0,
                          "transmittedTotalPacketsDelta": 0.0,
                          "vNicIdentifier": "eth0",
                          "valuesAreSuspect": "true"
                      }
                  ]
              }
          }
      }
      
      
      
      
      
      
       




11/9/2017

  1. vCpeResCust custom workflow:
    1. Closed loop: manually send VES event to DCAE, APPC successfully restarted VNF via two options: directly talking to Openstack, calling MultiCloud API.
      1. To restart through Openstack, we need the following vserver info in AAI. This is added by heat bridge. Note that if identity-url is missing then APPC will use the default one from its properties. 


        "vserver-selflink": "http://10.12.25.2:8774/v2.1/466979b815b5415ba14ada713e6e1846/servers/9e17027d-c09d-45c4-9a7a-84f50a2246fd",
      2. In the appc container, /opt/openecomp/appc/data/properties/appc.properties should have either of the following lines.

        provider1.identity=http://10.12.25.2:5000/v3 #this is for openstack
        
        provider1.identity=http://10.0.14.1:9005/api/multicloud-titanium_cloud/v0/pod25_RegionOne/identity/v2.0 #this is for multicloud



      3. To restart through MultiCloud, the following ESR should be added, note that the identity-url and  the vserver self link is pointing to MultiCloud. Note that the vserver selflink has a different IP and port. APPC picks up the user name and password from ESR and use them to access MultiCloud API.

        ESR and vserver data from AAI
        https://aai1:8443/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/CloudOwner/RegionOne?depth=all
        
        
        {
            "cloud-owner": "CloudOwner",
            "cloud-region-id": "RegionOne",
            "cloud-type": "SharedNode",
            "owner-defined-type": "OwnerType",
            "cloud-region-version": "v1",
            "identity-url": "http://10.0.14.1:9005/api/multicloud-titanium_cloud/v0/pod25_RegionOne/identity/v3",
            "cloud-zone": "CloudZone",
            "sriov-automation": false,
            "resource-version": "1510239649440",
            "tenants": {
                "tenant": [
                    {
                        "tenant-id": "466979b815b5415ba14ada713e6e1846",
                        "tenant-name": "Integration",
                        "resource-version": "1509709609665",
                        "vservers": {
                            "vserver": [
                                {
                                    "vserver-id": "9e17027d-c09d-45c4-9a7a-84f50a2246fd",
                                    "vserver-name": "zdcpe1cpe01dns01_110306",
                                    "vserver-name2": "zdcpe1cpe01dns01_110306",
                                    "prov-status": "ACTIVE",
                                    "vserver-selflink": "http://10.0.14.1:80/v2.1/466979b815b5415ba14ada713e6e1846/servers/9e17027d-c09d-45c4-9a7a-84f50a2246fd",
                                    "in-maint": false,
                                    "is-closed-loop-disabled": false,
                                    "resource-version": "1510183808208",
        ....
        
        
            "esr-system-info-list": {
                "esr-system-info": [
                    {
                        "esr-system-info-id": "432ac032-e996-41f2-84ed-9c7a1766eb29",
                        "system-name": "example-system-name-val-29070",
                        "type": "example-type-val-85254",
                        "vendor": "example-vendor-val-94515",
                        "version": "example-version-val-71880",
                        "service-url": "http://10.12.25.2:5000/v3",
                        "user-name": "demo",
                        "password": "onapdemo",
                        "system-type": "VIM",
                        "ssl-cacert": "example-ssl-cacert-val-75021",
                        "ssl-insecure": true,
                        "ip-address": "example-ip-address-val-44431",
                        "port": "example-port-val-93234",
                        "cloud-domain": "Default",
                        "default-tenant": "Integration",
                        "resource-version": "1510239649503"
                    }
                ]
            }
        }
    2. Tested DeleteVcpeResCust flow.
      1. After a few bug fixes, the flow was able to delete the vG stack. But the subsequent SDNC request returned an error:  SDNC-184 - Getting issue details... STATUS .
      2. The flow failed to pick up allotted resources:  SO-325 - Getting issue details... STATUS

11/8/2017

  1. vCpeResCust custom workflow:
    1. Successfully passed custom workflow test: created tunnelxconn-allotted-resource, vG, brg-allotted-resourece, configured vGMUX and vBRG. Note that currently VxLAN on vG is manually configured.
    2. Task: DG change to configure vG VxLAN:  SDNC-182 - Getting issue details... STATUS
    3. Task: DG change to return the following vG heat parameters to SO. This is in vfmodule assign.

      vG heat parameters to be returned by SDNC
        mux_gw_private_net_id: zdfw1muxgw01_private
        mux_gw_private_subnet_id: zdfw1muxgw01_sub_private
        mux_gw_private_net_cidr: 10.5.0.0/24
        cpe_public_net_id: vCPEInfraCPEPUBLIC110306
        cpe_public_subnet_id: vCPEInfraCPEPUBLICSUB110306
        cpe_public_net_cidr: 10.2.0.0/24
        onap_private_net_id: oam_onap_hUnI
        onap_private_subnet_id: oam_onap_hUnI
        onap_private_net_cidr: 10.0.0.0/16
        vgw_private_ip_0: 10.5.0.22
        vgw_private_ip_1: 10.0.101.30
        vgw_private_ip_2: 10.2.0.2
        vgw_name_0: zdcpe1cpe01gw01
        vnf_id: vCPE_Infrastructure_GW_demo_app
        vf_module_id: vCPE_Customer_GW
        mux_ip_addr: 10.5.0.21
        vg_vgmux_tunnel_vni: 100
    4. Brian will fix the HEAT bridge in robot, which will be used to add vserver info to AAI.
    5. Brian will add a DHCP server in the Webserver.

11/7/2017

  1. vCpeResCust custom workflow:
    1. SO configuration update needed:  SO-315 - Getting issue details... STATUS SO-278 - Getting issue details... STATUS .
    2. Distribution of updated service from SDC to SO problem identified:  SO-316 - Getting issue details... STATUS .
    3. vG heat template updated with two more parameters added in env and yml. Image also updated.
    4. vGMux heat updated.
    5. Error in SDNC on brg-allocated-resource activate. Will fix by Wed.
  2. Closed loop:
    1. vServer data is missing from AAI. May need to add manually.
    2. VES data reporting from vGMUX had incorrect sourceName, need to be updated.

11/6/2017

  1. vCpeResCust custom workflow:
    1. Test almost finished. Successfully created tunnel-x-connect, vG, and brg-allotted-resource. When SO finally queries SDC with the parent service instance id of brg-allotted-resource (which is the service instance id of vbrg service), SDNC failed to find it in its catalog. This will be fixed. Once this query passes, SO service flow will finish successfully.
    2. SO calling SDNC to create a vG instance failed. SDNC queried AAI to find the availability zone but such a zone did not exist. Brian fixed this by manually adding such a zone and a complex in AAI. Jerry will add this in demo.sh init. 
      1. Add an availability zone

        Add availability zone
        https://aai1:8443/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/CloudOwner/RegionOne/availability-zones/availability-zone/nova
        
        
        {
            "availability-zone-name": "nova",
            "hypervisor-type": "KVM",
            "operational-status": "Active"
        }
      2. Add an complex

        Add a complex
        https://aai1:8443//aai/v11/cloud-infrastructure/complexes/complex/clli1
        
        
        {
            "physical-location-id": "clli1",
            "data-center-code": "example-data-center-code-val-5556",
            "complex-name": "clli1",
            "identity-url": "example-identity-url-val-56898",
            "resource-version": "1509928831089",
            "physical-location-type": "example-physical-location-type-val-7608",
            "street1": "example-street1-val-34205",
            "street2": "example-street2-val-99210",
            "city": "example-city-val-27150",
            "state": "example-state-val-59487",
            "postal-code": "example-postal-code-val-68871",
            "country": "example-country-val-94173",
            "region": "example-region-val-13893",
            "latitude": "example-latitude-val-89101",
            "longitude": "example-longitude-val-66229",
            "elevation": "example-elevation-val-30253",
            "lata": "example-lata-val-46073"
        }
      3. Add relationship in CloudOwner to the above complex

        Add relationship to the complex
        https://aai1:8443/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/CloudOwner/RegionOne/
        
        Do a GET of CloudOwner/RegionOne and add this relationship then do a PUT back to CloudOwner/RegionOne
        
        
                "relationship": [
                    {
                        "related-to": "complex",
                        "related-link": "/aai/v11/cloud-infrastructure/complexes/complex/clli1"
        	}
        
        Example: 
        {     
        "cloud-owner": "CloudOwner",     
        "cloud-region-id": "RegionOne",     
        "cloud-type": "SharedNode",     
        "owner-defined-type": "OwnerType",     
        "cloud-region-version": "v1",     
        "cloud-zone": "CloudZone",     
        "sriov-automation": false,     
        "resource-version": "1509931173816",     
        "relationship-list": {         
        	"relationship": [             
                         {
                         "related-to": "complex",
                         "related-link": "/aai/v11/cloud-infrastructure/complexes/complex/clli1"
                         },
                         {
                         "related-to": "l3-network",
        		.... more data ...
        
    3. SO calling SDNC during vG instantiation used incorrect format for vG. What defined in Yang uses dash, e.g., vg-ip. SO used vg_ip. SO fixed this onsite.
    4. vG instantiation completed successfully. Currently the IP address facing vGMUX is using the one in HEAT. SO workflow should take the one allocated by SDNC. 
    5. SO: preProcessSDNCAssign workflow didn't include modelCustomizationUuid. This was fixed onsite.
    6. A routing entry needs to be added to the SDNC VM (not docker) to allow reaching BRG through BNG:

      ip route add 10.3.0.0/24 via 10.0.101.10 dev eth0
    7. SDNC assign of brg-allotted-resource failed. Note that unassign, activate, deactive all may have similar problems.
      1. The original DG had an error and was not loaded to SDNC. Fixed.
      2. The assign DG didn't set global-customer-id. Fixed.
      3. The assign DG didn't set service instance uuid. Fixed.
    8. The latest config for SO are: mso.sdnc.properties and mso.bpmn.urn.properties. Note that these config files are dynamically generated when SO docker container is started/restarted. After the latest fixes, there should be no need to use these files anymore.

11/4/2017

  1. vCpeResCust custom workflow: 

    1. Postman body to invoke SO updated

      to invoke SO
      http://so:8080/ecomp/mso/infra/serviceInstances/v5
      user/pass: InfraPortalClient:password1$
      Content-Type and Accept are both: application/json
      {
         "requestDetails" : {
            "requestInfo" : {
               "productFamilyId" : "a9a77d5a-123e-4ca2-9eb9-0b015d2ee0fb",
               "suppressRollback" : "false",
               "instanceName" : "vcperescust-1104-19",
               "requestorId": "Kang",
               "source" : "Postman-Kang"
            },
            "requestParameters" : {
               "subscriptionServiceType" : "vCPE",
               "userParams" : [{
                  "name":"BRG_WAN_MAC_Address",
                  "value":"fa:16:3e:8f:ea:68"
               }],
               "aLaCarte" : "false"
            },
            "subscriberInfo" : {
               "subscriberName" : "Kaneohe",
               "globalSubscriberId" : "SDN-ETHERNET-INTERNET"
            },
            "cloudConfiguration" : {
               "lcpCloudRegionId" : "RegionOne",
               "tenantId" : "466979b815b5415ba14ada713e6e1846"
            },
            "modelInfo" : {
               "modelType" : "service",
               "modelVersionId" : "2da136af-510d-457e-b3dc-f1c3e6dab0c3",
               "modelName" : "vCpeResCust110403",
               "modelVersion" : "1.0",
               "modelInvariantId" : "35f93cd5-17a8-4735-9531-a63df867d776"
            }
         }
      }
      
      
    2. Made much progress by modifying multiple DGs with temp fixes. Marcus is working to fix them the right way and commit to the repo. Passed TunnelXConn steps: assign, create. Wait to be tested: activate. Some of the problems are:
      1. Incorrect username/password for authentication;
      2. Correct username/password in configuration file, but they are referenced incorrectly in DG.
      3. DG does not obtain globalcustomerID and gmux service ID properly, and thus failed to query AAI.
      4. Use incorrect IP to config vGMUX.
      5. Restconf request to vGMUX has mal-formatted body.
      6. DG has error (use incorrect resource) when allocating IP addresses for vG.
    3. Created a new vCpeResCust110403 service. The difference is that now the vG model is based on the latest heat and env: vgw.zip. The heat and env have been validated using openstack stack create. Note that in order to correctly distribute to SO, all the components were created from scratch, including vG (from onboarding), tunnelXconn VF, vBRG Allotted Resource VF. After distribution, added service recipe in SO DB and allotted resources to SDNC DB. Verified that this service model can be executed through SO API.

11/3/2017

  1. vCpeResCust custom workflow: 

    1. Modified the vcperescust service and then redistribution to AAI failed. It looks like SDC has a problem handling update but it is not easy to pinpoint the problem. The workaround is to recreate the service from scratch. Distribution passed.

    2. Brian created the infrastructure including BNG BRG vGMUX, which are to be used for vcperescust flow.

    3. Found a bug in TunnelXConn flow where the content send to SDNC assign is incorrect. Fixed with code change onsite. Tracked by  SO-302 - Getting issue details... STATUS

    4. Found a few missing configurations and mistakes in mso.sdnc.properties, fixed onsite. The file being used now is mso.sdnc.properties.

    5. Found a problem in SDNC handling TunnelXConn create operation. Dan fixed it onsite. 

    6. TunnelXConn flow has a bug: the SDNC assign request misses some info:  SO-304 - Getting issue details... STATUS

    7. SDNC ueb cannot parse service template when distributed. Reopen this ticket:  SDC-564 - Getting issue details... STATUS .

    8. Manually inserted the following into SDNC DB tables:

      Insert into sdnc db ALLOTTED_RESOURCE_MODEL
      INSERT INTO SERVICE_MODEL (`service_uuid`, `model_yaml`,`invariant_uuid`,`version`,`name`,`description`,`type`,`category`,`ecomp_naming`,`service_instance_name_prefix`,`filename`,`naming_policy`) values ('467676e2-bcd9-4bf5-b9fd-03afea76be17', null, '4f90a895-d9f5-472c-8f65-afc5ca28628d',null,'vCPEService', 'vCPEService', 'Service','Network L1-3', 'N', 'vCPEService', 'vCpeResCust110602/service-Vcperescust110602-template.yml',null);
      
      INSERT INTO ALLOTTED_RESOURCE_MODEL (`customization_uuid`,`model_yaml`,`invariant_uuid`,`uuid`,`version`,`naming_policy`,`ecomp_generated_naming`,`depending_service`,`role`,`type`,`service_dependency`,`allotted_resource_type`)  VALUES ( '48e00a09-4184-4689-b690-3baabaf24838', NULL,  'a99d5de7-0707-4f4c-b7b5-f92101ce9ddc', '4fb476c3-3516-4433-8300-47690e48ea5c', '1.0', NULL,'Y',NULL,NULL,'TunnelXConnect',NULL, 'TunnelXConnect');
      
      INSERT INTO ALLOTTED_RESOURCE_MODEL (`customization_uuid`,`model_yaml`,`invariant_uuid`,`uuid`,`version`,`naming_policy`,`ecomp_generated_naming`,`depending_service`,`role`,`type`,`service_dependency`,`allotted_resource_type`)  VALUES ( '0f2d8c72-6470-42cf-9b2f-f30041ffc78d', NULL,  '8415033a-3c4c-4de5-b38f-3e31be32ddf2', '33094f6f-d19f-451f-b307-201b18400a45', '1.0', NULL,'Y',NULL,NULL,'BRG',NULL, 'TunnelXConnect');
      
      INSERT INTO VF_MODEL (`customization_uuid`,`model_yaml`,`invariant_uuid`,`uuid`,`version`,`name`,`naming_policy`,`ecomp_generated_naming`,`avail_zone_max_count`,`nf_function`,`nf_code`,`nf_type`,`nf_role`,`vendor`,`vendor_version`)  VALUES ( '02f251be-3941-4c4d-9510-71f6cc620e41', NULL,  'b7773de9-b4af-4593-9802-1d7512f8fa38', '7e53fec2-8032-43c0-b41c-ad3f564f42d4', '1.0', 'vgw-kang-110602',NULL,'Y',1,NULL,NULL,NULL,NULL,'vCPE','1.0');
      
      INSERT INTO VF_MODULE_MODEL (`customization_uuid`,`model_yaml`,`invariant_uuid`,`uuid`,`version`,`vf_module_type`,`availability_zone_count`,`ecomp_generated_vm_assignments`)  VALUES ( 'feac8fbd-7de0-4b84-a194-98b1ab6e4750', NULL,  '54d3bef4-48be-47e1-8529-da2a160f40a9', 'eaf47c0a-287f-4d5f-a8f7-f608ac4d4aae', '1.0', 'Base',NULL,NULL);
      
    9. A few other tickets: 

      key summary status
      Loading...
      Refresh

11/2/2017

  1. vCpeResCust custom workflow: 
    1. vcperescust service failed to distribute from SDC to SO for several reasons:
      1. For the vBRG EMU allotted resource, we first designed it using "Allotted Resource" subcategory, SO did not accept it. We then re-designed it using "Tunnel XConn" subcategory, SO accepted it. But it is not clear if this will cause problem later on.
      2. Each VF in the service is assigned a customizationUUID. That ID is not changed unless the VF is removed and recreated in SDC composition. SO throws an error if a service being redistributed includes a component with the same customizationUUID. Basically it checks the DB to see if the ID already exist, if yes then reports a duplicate customizationUUID error in /var/log/ecomp/MSO/ASDC.../msodebug.log. If we need to change and redistribute a service, the workaround is to delete each component and add back in.
      3. Solutions:
        1. We recreated the mso container and db container to start with empty DB tables. Note that only recreate the DB does not work as it will lack the camunda records, which are created when the mso container is initialized.
        2. To recreate, delete the mso container, change /opt/test_lab/deploy.sh to enable "--force-recreate mariadb". Remember to turn it back off afterwards.
        3. After new mso container is created, do the following.
          1. Run cpjars to copy the updated jar and war files into the docker. This is not needed if docker is updated.
          2. Restart the container to load the above files.
          3. Run cpproperties to copy the properties in mso.bpmn.urn.properties, mso.sdnc.properties. Note that this properties got reset each time the container is restarted. This is not needed once docker is updated. But to enable debug log for specific items, some of the flags in mso.bpmn.urn.properties are still needed.
    2. SNIRO is updated to meet the needs of the Homing query. Now the data to be loaded to SNIRO is this (updated 11/7): sniro.json. Note that some of the contents needs to be updated, such as gmux uuid.
    3. To clear the sniro content: GET http://robot:8080/__admin/mappings/reset
    4. To check sniro logs: docker logs -f sniroemulator
    5. TunnelXConn flow queried AAI using the vCpeResCust service instance uuid but got nothing. This was caused by racing condition: it was too soon to query after the service flow put the same uuid in AAI. The temp solution is to let the flow sleep for 30 seconds before performing query. It worked. Note that sleeping for 5 seconds was proven to be insufficient.

11/1/2017

  1. vCpeResCust custom workflow: 
    1. Michael Lando found a solution for  SDC-564 - Getting issue details... STATUS . SDNC should be able to parse service template properly after the fix is committed.

    2. Brian manually configured SDNC to bypass service template parsing and checking as a temp hacking so that SDNC could pass "service assign".
    3. SO flow continued to create TunnelXConn. An exception occurred. It is determined that the cause is homing. The problem is tracked by INT-317 - Getting issue details... STATUS .
    4. The team determined that vCpeResCust needs to include a BRG allotted resource. We were able to modify the service model in SDC but distribution to SO ended in error. Waiting to be solved.

10/31/2017

  1. vCpeResCust custom workflow: 
    1. Service assign request from SO to SDNC has a format error. Fixed it by modifying the following line in mso docker containter: /etc/mso/config.d/mso.sdnc.properties. Note that the above is lost once the container is restarted. A JIRA ticket is opened to request a permanent fix: SO-295 - Getting issue details... STATUS .

      org.openecomp.mso.adapters.sdnc..service-topology-operation.assign=POST|270000|sdncurl8|sdnc-request-header|org:onap:sdnc:northbound:generic-resource
    2. In SDNC ueb listener docker container, /opt/onap/sdnc/data/properties/ueb-listener.properties does not have the correct sdc address, fixed. But the user name and password to authenticate with SDC are incorrect. So it's not able to get service model distributed from SDC. This is now fixed, see the file content below. Dan has committed the new config file to the repo.

      ueb-listener.properties
      org.onap.ccsdk.sli.northbound.uebclient.asdc-address=c2.vm1.sdc.simpledemo.openecomp.org:8443
      org.onap.ccsdk.sli.northbound.uebclient.consumer-group=sdc-OpenSource-Env1-sdnc-dockero
      org.onap.ccsdk.sli.northbound.uebclient.consumer-id=sdc-COpenSource-Env11-sdnc-dockero
      org.onap.ccsdk.sli.northbound.uebclient.environment-name=AUTO
      org.onap.ccsdk.sli.northbound.uebclient.password=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U
      org.onap.ccsdk.sli.northbound.uebclient.user=sdnc
      org.onap.ccsdk.sli.northbound.uebclient.sdnc-user=admin
      org.onap.ccsdk.sli.northbound.uebclient.sdnc-passwd=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U
      org.onap.ccsdk.sli.northbound.uebclient.asdc-api-base-url=http://sdnchost:8282/restconf/operations/
      org.onap.ccsdk.sli.northbound.uebclient.asdc-api-namespace=org:onap:ccsdk
      org.onap.ccsdk.sli.northbound.uebclient.spool.incoming=/opt/onap/sdnc/ueb-listener/spool/incoming
      org.onap.ccsdk.sli.northbound.uebclient.spool.archive=/opt/onap/sdnc/ueb-listener/spool/archive
      org.onap.ccsdk.sli.northbound.uebclient.polling-interval=30
      org.onap.ccsdk.sli.northbound.uebclient.polling-timeout=15
      org.onap.ccsdk.sli.northbound.uebclient.client-startup-timeout=900
      org.onap.ccsdk.sli.northbound.uebclient.relevant-artifact-types=YANG_XML,VF_LICENSE,TOSCA_TEMPLATE,TOSCA_CSAR,UCPE_LAYER_2_CONFIGURATION
      org.onap.ccsdk.sli.northbound.uebclient.activate-server-tls-auth=false
      org.onap.ccsdk.sli.northbound.uebclient.keystore-path=
      org.onap.ccsdk.sli.northbound.uebclient.keystore-password=
      org.onap.ccsdk.sli.northbound.uebclient.xslt-path-list=
      org.onap.ccsdk.sli.northbound.uebclient.artifact-map=/opt/onap/sdnc/data/properties/artifact.map
    3. Now SDNC is able to pick up service template distributed from SDC. But reports an error when Cambria parses yml template. Waiting to be fixed:  SDC-564 - Getting issue details... STATUS
    4. service-Vcperescust-template.yml
    5. service-Vcperescust-template.yml

10/30/2017

  1. vCpeResCust custom workflow: 
    1. Identified configuration mistake in mso.bpmn.urn.properties: both hostname and url path needs to be corrected to provide correct callback url to SNIRO. The updated file is here: mso.bpmn.urn.properties
    2. Identified a bug. When SO service level flow requests SDNC for service instance assignment, the request does not have correct format and got rejected. Dan provides a sample and Yang model in  SDNC-153 - Getting issue details... STATUS . A new ticket is created.  SO-289 - Getting issue details... STATUS .
  2. Eric and team have the following update on VNFs:
    1. There should be vnf snapshot images in the Integration environment now.  Note, that the vbng image Brian made today needs to be replaced with one Matt is in process of building. 
    2. Per Kang’s question below, the last line of each vnf yaml file runs an install script – this is what sets up the interface addresses, per the env file.  The snapshot images skip over need for time consuming compilation of the vpp/honeycomb code.
    3. We have tested manually configuring vxlan tunnels, etc. in the VCPE project and verified that the data path is working.
    4. we would like to remove and rebuild the vnfs in the vcpe project tomorrow morning.
      1. one item we’re in process of testing is control plane connectivity from sdnc to the vbrg
    5. A couple areas where there may be some unkowns that could come up during integration:
      1. interaction with the onap vdhcp vnf (different than what was tested in our local lab)
      2. interaction with the aaa vnf  (did not have this in our local lab setup)

10/27/2017

  1. vCpeResCust custom workflow: 

    1. Passed: Get and decompose service template from catalog.
    2. Passed: Query SNIRO emulator to get homing information. The current config files for SO after manual changes are:  mso.bpmn.urn.propertiesmso-docker.json. They are supposed to be updated by SO so that no more manual changes are needed in the future. The callback URL provided by SO is incorrect, needs to be fixed ( SO-278 - Getting issue details... STATUS ). SNIRO emulator needs to be modified to use the callback URL from SO request ( INT-311 - Getting issue details... STATUS ). Currently we use the following hacking to send the required info to SO.

      Manually send SNIRO results to SO
      curl -i -X POST -d @sniro.json -H Content-Type=application/json --user 'MSOClient:password1$' http://so:8080/workflows/messages/message/SNIROResponse/c39fe43a-8f31-4d70-957d-b1c11e161636
      
      
      sniro.json is preloaded into SNIRO emulator and will be used to feed SO request.
      {
         "solutionInfo" : {
            "placement" : [
               {
                  "serviceResourceId" : "61d563e8-e714-4393-8f99-cc480144a05e",
                  "resourceModuleName" : "TunnelXConn",
                  "serviceInstanceId" : "GMuxInfra-UUID",
                  "cloudRegionId" : "RegionOne",
                  "inventoryType" : "service"
               },
               {
                  "resourceModuleName" : "vG",
                  "serviceResourceId" : "91d563e8-e714-4393-8f99-cc480144a05e",
                  "cloudRegionId" : "RegionOne",
                  "serviceInstanceId" : "vG-service-instance-id",
                  "inventoryType" : "cloud"
               },
               {
                  "inventoryType" : "service",
                  "serviceInstanceId" : "BRG_EMU_UUID",
                  "cloudRegionId" : "RegionOne",
                  "resourceModuleName" : "BRG",
                  "serviceResourceId" : "71d563e8-e714-4393-8f99-cc480144a05e"
               }
            ]
         },
         "requestId" : "111-111-1111",
         "statusMessage" : "",
         "transactionId" : "111-111-1111",
         "requestState" : "complete"
      }
      
      
      
    3. Passed: SO queries AAI to get service and other info include globalCustomerID. We preload AAI with the following info. Note that the ASDC_TOSCA_UUID part is questionable. It seems not necessary. It is tracked by SO-279 - Getting issue details... STATUS .

      Preload AAI for vCpeResCust flow
      PUT /aai/v11 /business/customers/customer/SDN-ETHERNET-INTERNET
      
      {
          "global-customer-id": "SDN-ETHERNET-INTERNET",
          "subscriber-name": "SDN-ETHERNET-INTERNET",
          "subscriber-type": "INFRA",
          "service-subscriptions": {
              "service-subscription": [
                  {
                      "service-type": "123456789",
                      "service-instances": {
                          "service-instance": [
                              {
                                  "service-instance-id": "fbe9ad27-7ddd-49a6-ab2f-8c08e31fe12c",
                                  "service-instance-name": "fbe9ad27-7ddd-49a6-ab2f-8c08e31fe12c",
                                  "service-type": "vcpe"
                              }
                          ]
                      }
                  }
              ]
          }
      }
      
      
      After the sdc models were loaded, I added this model into AAI:
      
      PUT /aai/v11 /service-design-and-creation/models/model/1963dd8b-9375-4cab-aa59-0ee06e8333fa/model-vers/model-ver/ASDC_TOSCA_UUID
      {
          "model-version-id": "ASDC_TOSCA_UUID",
          "model-name": "vCpeResCust",
          "model-version": "1.0",
          "model-description": "Some ASDC Tosca Model"
      }
    4. Passed: SO creates a service instance UUID and  put it in AAI. 
    5. Blocked: SO calls SDNC assign service (type=vCpeResCust, UUID), see  SDNC-153 - Getting issue details... STATUS .
  2. General Infrastructure
    1. Brian has successfully instantiated general infrastructure, hahahaha~~~~~~~~~~
  3. Notes for upcoming test
    1. SO allows only one workflow to executive at a time. To clear the current one: 

      delete from mso_requests.infra_active_requests;
    2. To manually send event to DMaaP to invoke SDNC to create BRG record in AAI (this emulates the event from DHCP), do the following

      http://{{mr}}:3904/events/VCPE-DHCP-EVENT/group1/C1?timeout=5000
      [
          "{\"msg_name\":\"DHCPACK\",\"macaddr\":\"e2:91:8c:7a:1e:9d\",\"yiaddr\":\"10.3.0.2\"}"
      ]


10/26/2017

  1. vCpeResCust custom workflow: 
    1. The config changes for sniro end point should be visible at : /shared/mso-docker.json. We should look at "sniroEndpoint". Also James Hahn found out that the following line needs to be added to the same file. A ticket is opened:  SO-274 - Getting issue details... STATUS . Important: It seems that the mso container needs to be restarted to pick up the change in this file.

      "adaptersWorkflowMessageEndpoint": "http://mso:8080/workflows/messages/Resteasy",
    2. SNIRO emulator config tested. The related wiki page is here: SNIRO Emulator
    3. Add the following to /etc/mso/config.d/mso.bpmn.urn.properties to enable debug. Important: The changes are lost each time the container is restarted.

      Debug mode
      log.debug.CreateVcpeResCustService=true
      log.debug.DeleteVcpeResCustService=true
      log.debug.DoCreateAllottedResourceBRG=true
      log.debug.DoCreateAllottedResourceBRGRollback=true
      log.debug.DoCreateAllottedResourceTXC=true
      log.debug.DoCreateAllottedResourceTXCRollback=true
      log.debug.DoDeleteAllottedResourceBRG=true
      log.debug.DoDeleteAllottedResourceTXC=true
      log.debug.Homing=true
      
    4. A bug was reported when SO tried to create a SNIRO request. It is tracked by  SO-273 - Getting issue details... STATUS .
  2. General Infrastructure:
    1. Brian created preload as following

      VNF-API Preload
      POST http://10.12.25.64:8282/restconf/operations/VNF-API:preload-vnf-topology-operation
      Content-Type: application/json
      
      {
         "VNF-API:input": {
           "VNF-API:request-information": {
             "VNF-API:request-id": "robot12",
             "VNF-API:notification-url": "https://so.onap.org",
             "VNF-API:order-number": "robot12",
             "VNF-API:request-sub-action": "SUPP",
             "VNF-API:request-action": "PreloadVNFRequest",
             "VNF-API:source": "VID",
             "VNF-API:order-version": "1.0"
           },
           "VNF-API:sdnc-request-header": {
             "VNF-API:svc-action": "reserve",
             "VNF-API:svc-notification-url": "https://son.onap.org",
             "VNF-API:svc-request-id": "robot12"
           },
           "VNF-API:vnf-topology-information": {
                   "vnf-topology-identifier": {
                     "service-type": "3c8da74-9225-4b9d-8915-214f8ab273a3",
                     "vnf-type": "VcpeGenInfra..base_vcpe_infra..module-0",
                     "generic-vnf-name": "vcpe-gen-infra 0",
                     "generic-vnf-type": "VcpeGenInfra..base_vcpe_infra..module-0",
                     "vnf-name": "vCPEInfraVF102604"
             },
             "VNF-API:vnf-parameters": [
                    {
                    "vnf-parameter-name": "vcpe_image_name",
                    "vnf-parameter-value": "ubuntu-14-04-cloud-amd64"
                    },
                    {
                    "vnf-parameter-name": "vcpe_flavor_name",
                    "vnf-parameter-value": "m1.medium"
                    },
                    {
                    "vnf-parameter-name": "public_net_id",
                    "vnf-parameter-value": "external"
                    },
                    {
                    "vnf-parameter-name": "cpe_signal_net_id",
                    "vnf-parameter-value": "vCPEInfraCPESIGNAL102604"
                    },
                    {
                    "vnf-parameter-name": "cpe_signal_subnet_id",
                    "vnf-parameter-value": "vCPEInfraCPESIGNALSUB102604"
                    },
                    {
                    "vnf-parameter-name": "cpe_public_net_id",
                    "vnf-parameter-value": "vCPEInfraCPEPUBLIC102604"
                    },
                    {
                    "vnf-parameter-name": "cpe_public_subnet_id",
                    "vnf-parameter-value": "vCPEInfraCPEPUBLICSUB102604"
                    },
                    {
                    "vnf-parameter-name": "onap_private_net_id",
                    "vnf-parameter-value": "oam_onap_hUnI"
                    },
                    {
                    "vnf-parameter-name": "onap_private_subnet_id",
                    "vnf-parameter-value": "oam_onap_hUnI"
                    },
                    {
                    "vnf-parameter-name": "onap_private_net_cidr",
                    "vnf-parameter-value": "10.0.0.0/16"
                    },
                    {
                    "vnf-parameter-name": "cpe_signal_net_cidr",
                    "vnf-parameter-value": "10.4.0.0/24"
                    },
                    {
                    "vnf-parameter-name": "cpe_public_net_cidr",
                    "vnf-parameter-value": "10.2.0.0/24"
                    },
                    {
                    "vnf-parameter-name": "vdhcp_private_ip_0",
                    "vnf-parameter-value": "10.4.0.1"
                    },
                    {
                    "vnf-parameter-name": "vdhcp_private_ip_1",
                    "vnf-parameter-value": "10.0.101.1"
                    },
                    {
                    "vnf-parameter-name": "vaaa_private_ip_0",
                    "vnf-parameter-value": "10.4.0.4"
                    },
                    {
                    "vnf-parameter-name": "vaaa_private_ip_1",
                    "vnf-parameter-value": "10.0.101.2"
                    },
                    {
                    "vnf-parameter-name": "vdns_private_ip_0",
                    "vnf-parameter-value": "10.2.0.1"
                    },
                    {
                    "vnf-parameter-name": "vdns_private_ip_1",
                    "vnf-parameter-value": "10.0.101.3"
                    },
                    {
                    "vnf-parameter-name": "vweb_private_ip_0",
                    "vnf-parameter-value": "10.2.0.10"
                    },
                    {
                    "vnf-parameter-name": "vweb_private_ip_1",
                    "vnf-parameter-value": "10.0.101.40"
                    },
                    {
                    "vnf-parameter-name": "mr_ip_addr",
                    "vnf-parameter-value": "10.0.11.1"
                    },
                    {
                    "vnf-parameter-name": "vaaa_name_0",
                    "vnf-parameter-value": "zdcpe1cpe01aaa01_102604"
                    },
                    {
                    "vnf-parameter-name": "vdns_name_0",
                    "vnf-parameter-value": "zdcpe1cpe01dns01_102604"
                    },
                    {
                    "vnf-parameter-name": "vdhcp_name_0",
                    "vnf-parameter-value": "zdcpe1cpe01dhcp01_102604"
                    },
                    {
                    "vnf-parameter-name": "vweb_name_0",
                    "vnf-parameter-value": "zdcpe1cpe01web01_102604"
                    },
                    {
                    "vnf-parameter-name": "vnf_id",
                    "vnf-parameter-value": "vCPE_Infrastructure_demo_app_102604"
                    },
                    {
                    "vnf-parameter-name": "vf_module_id",
                    "vnf-parameter-value": "vCPE_Intrastructure_102604"
                    },
                    {
                    "vnf-parameter-name": "dcae_collector_ip",
                    "vnf-parameter-value": "10.0.4.102"
                    },
                    {
                    "vnf-parameter-name": "dcae_collector_port",
                    "vnf-parameter-value": "8080"
                    },
                    {
                    "vnf-parameter-name": "repo_url_blob",
                    "vnf-parameter-value": "https://nexus.onap.org/content/sites/raw"
                    },
                    {
                    "vnf-parameter-name": "repo_url_artifacts",
                    "vnf-parameter-value": "https://nexus.onap.org/content/groups/staging"
                    },
                    {
                    "vnf-parameter-name": "demo_artifacts_version",
                    "vnf-parameter-value": "1.1.0"
                    },
                    {
                    "vnf-parameter-name": "install_script_version",
                    "vnf-parameter-value": "1.1.0-SNAPSHOT"
                    },
                    {
                    "vnf-parameter-name": "key_name",
                    "vnf-parameter-value": "vaaa_key"
                    },
                    {
                    "vnf-parameter-name": "pub_key",
                    "vnf-parameter-value": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDQXYJYYi3/OUZXUiCYWdtc7K0m5C0dJKVxPG0eI8EWZrEHYdfYe6WoTSDJCww+1qlBSpA5ac/Ba4Wn9vh+lR1vtUKkyIC/nrYb90ReUd385Glkgzrfh5HdR5y5S2cL/Frh86lAn9r6b3iWTJD8wBwXFyoe1S2nMTOIuG4RPNvfmyCTYVh8XTCCE8HPvh3xv2r4egawG1P4Q4UDwk+hDBXThY2KS8M5/8EMyxHV0ImpLbpYCTBA6KYDIRtqmgS6iKyy8v2D1aSY5mc9J0T5t9S2Gv+VZQNWQDDKNFnxqYaAo1uEoq/i1q63XC5AD3ckXb2VT6dp23BQMdDfbHyUWfJN"
                    },
                    {
                    "vnf-parameter-name": "cloud_env",
                    "vnf-parameter-value": "Openstack"
                    }
             ],
             "VNF-API:vnf-assignments": {
               }
             }
           }
      }
      
      
      
      

10/25/2017

  1. vCpeResCust custom workflow: 
    1. Re-created SO maria DB to reflect the latest change to the DB tables. This done by adding a line to /opt/test_lab/deploy.sh: "$DOCKER_COMPOSE_CMD up -d --force-recreate mariadb".

    2. The workflow successfully obtained service template from the catalog and decomposed it.

    3. A problem was observed when the workflow tried to call the homing service. It is tracked by SO-268 - Getting issue details... STATUS .
  2. General Infrastructure:
    1. Brian modified the neutron heat template and parameters in SO DB, pre-loaded SDNC, and successfully created virtual networks. (The scripts and template on 10/24 notes have been updated accordingly.) SDNC preload data for CPE_SIGNAL is below.

      SDNC Preload for CPE_SIGNAL
      {
        "VNF-API:input": {
          "VNF-API:request-information": {
            "VNF-API:request-id": "robot0012",
            "VNF-API:notification-url": "http://so.onap.org",
            "VNF-API:order-number": "robot0012",
            "VNF-API:request-sub-action": "SUPP",
            "VNF-API:request-action": "PreloadNetworkRequest",
            "VNF-API:source": "robot",
            "VNF-API:order-version": "1.0"
          },
          "VNF-API:network-topology-information": {
            "VNF-API:network-topology-identifier": {
              "VNF-API:network-role": "cpe-public",
              "VNF-API:network-technology": "neutron",
              "VNF-API:service-type": "vCPE",
              "VNF-API:network-name": "vCPEInfraCPEPUBLICL1025BDF01",
              "VNF-API:network-type": "Generic NeutronNet"
            },
            "VNF-API:provider-network-information": {
              "VNF-API:is-external-network": "true",
              "VNF-API:physical-network-name": "vCPEInfraCPEPUBLIC1025BDF01",
              "VNF-API:is-provider-network": "true",
              "VNF-API:is-shared-network": "true"
            },
            "VNF-API:subnets": [
              {
                "VNF-API:start-address": "10.2.0.2",
                "VNF-API:cidr-mask": "24",
                "VNF-API:subnet-name": "vCPEInfraCPEPUBLIC1025BDF01",
                "VNF-API:ip-version": "4",
                "VNF-API:dhcp-enabled": "N",
                "VNF-API:gateway-address": "10.2.0.1"
              }
                    ]
          },
          "VNF-API:sdnc-request-header": {
            "VNF-API:svc-action": "reserve",
            "VNF-API:svc-notification-url": "http://so.onap.org",
            "VNF-API:svc-request-id": "reobot0012"
          }
        }
      }
    2. During VNF instantiation, a bug was discovered. The content of DB table vnf_component_recipe is incorrect. It is tracked by  SO-267 - Getting issue details... STATUS .
    3. Brian found a few bugs related to SDNC: 

      key summary status
      Loading...
      Refresh

10/24/2017

  1. vCpeResCust custom workflow: 
    1. Use the following to add the custom workflow into the recipe table:

      insert into recipe
      INSERT INTO `service_recipe` (`ACTION`, `VERSION_STR`, `DESCRIPTION`, `ORCHESTRATION_URI`, `SERVICE_PARAM_XSD`, `RECIPE_TIMEOUT`, `SERVICE_TIMEOUT_INTERIM`, `CREATION_TIMESTAMP`, `SERVICE_MODEL_UUID`) VALUES ('createInstance','1','vCpeRestCust110301','/mso/async/services/CreateVcpeResCustService',NULL,181,NULL,'2017-11-03 13:48:00','b12d36da-459f-41be-aadc-7034568690eb');
      
      
    2. Successfully invoked the custom workflow from SO NBI using curl (note that VID currently only support a la carte so cannot invoke this flow). 

      invoke vCpeResCust workflow
      curl -X POST \
        http://so:8080/ecomp/mso/infra/serviceInstances/v5 \
        -H 'accept: application/json' \
        -H 'authorization: Basic SW5mcmFQb3J0YWxDbGllbnQ6cGFzc3dvcmQxJA==' \
        -H 'cache-control: no-cache' \
        -H 'content-type: application/json' \
        -H 'postman-token: 0c4f0ea7-736f-4999-3399-982de75ceecf' \
        -d '{
         "requestDetails" : {
            "requestInfo" : {
               "productFamilyId" : "a9a77d5a-123e-4ca2-9eb9-0b015d2ee0fb",
               "suppressRollback" : "false",
               "instanceName" : "vcperescust-102404",
               "requestorId": "demo",
               "source" : "VID"
            },
            "requestParameters" : {
               "subscriptionServiceType" : "123456789",
               "userParams" : [{
                  "name":"BRG_WAN_MAC_Address",
                  "value":"brgmac"
               }],
               "aLaCarte" : "false"
            },
            "subscriberInfo" : {
               "subscriberName" : "Kaneohe",
               "globalSubscriberId" : "SDN-ETHERNET-INTERNET"
            },
            "cloudConfiguration" : {
               "lcpCloudRegionId" : "RegionOne",
               "tenantId" : "466979b815b5415ba14ada713e6e1846"
            },
            "modelInfo" : {
               "modelType" : "service",
               "modelVersionId" : "ASDC_TOSCA_UUID",
               "modelName" : "vCpeResCust",
               "modelVersion" : "1.0",
               "modelInvariantId" : "1963dd8b-9375-4cab-aa59-0ee06e8333fa"
            }
         }
      }
      '
    3. A bug is discovered for the workflow and is being worked on (tracked by  SO-262 - Getting issue details... STATUS .
  2. Genera Infrastructure: With the following manual fix we are able to distribute the service to SO.
    1. The generic neutron network HEAT template is missing from the SO DB. Brian found a way to manually fix it. SO will include this in the repo. It is tracked by  SO-265 - Getting issue details... STATUS . The scipt and heat template are given below (updated on 10/25 based on Brian Freeman's comments to  SO-265 - Getting issue details... STATUS ).

      insert HEAT into SO DB
      INSERT INTO `heat_template_params` (`HEAT_TEMPLATE_ARTIFACT_UUID`,`PARAM_NAME`,`IS_REQUIRED`,`PARAM_TYPE`,`PARAM_ALIAS` ) VALUES ('efee1d84-b8ec-11e7-abc4-cec278b6b50a','network_name',1 ,'string', NULL);
      INSERT INTO `heat_template_params` (`HEAT_TEMPLATE_ARTIFACT_UUID`,`PARAM_NAME`,`IS_REQUIRED`,`PARAM_TYPE`,`PARAM_ALIAS` ) VALUES ('efee1d84-b8ec-11e7-abc4-cec278b6b50a','shared',0 ,'string', NULL);
      INSERT INTO `heat_template` (`ARTIFACT_UUID`, `NAME`,`VERSION` , `BODY`, `TIMEOUT_MINUTES`,`DESCRIPTION`, `CREATION_TIMESTAMP`, `ARTIFACT_CHECKSUM` ) VALUES ('efee1d84-b8ec-11e7-abc4-cec278b6b50a', 'Generic NeutronNet', '1', LOAD_FILE('/tmp/generic_neutron.yaml') , 10 ,'Generic Neutron Template',NOW(), 'MANUAL RECORD');
       
      heat_template_version: 2013-05-23
      description: A simple Neutron network
      parameters:
        network_name:
          type: string
          description: Name of the Neutron Network
          default: ONAP-NW1
        shared:
           type: boolean
           description: Shared amongst tenants
           default: True
      outputs:
        network_id:
          description: Openstack network identifier
          value: { get_resource: network }
      resources:
        network:
          type: OS::Neutron::Net
          properties:
            name: { get_param: network_name }
            shared: { get_param: shared }
      
      
      
      
    2. network_resource table model_invariant_uuid was too short (20 instead of at least 36). Manually increased to 120. It is tracked by  SO-266 - Getting issue details... STATUS
  3. Genera Infrastructure: Instantiation
    1. Need to run "/opt/demo.sh init" in the robot VM first.
    2. A service was created using VID.
    3. Tried to add a neutron network to it. SO received an error from SDNC. This is tracked by  SDNC-143 - Getting issue details... STATUS
    4. The 404 error in VID on deploy is due to missing zone data in AAI. Preload AAI with the following:

      Preload AAI with zone data
      PUT https://{{aai}}:8443/aai/v11/network/zones/zone/nova1
      {
          "zone-id": "nova1",
          "zone-name": "nova",
          "design-type": "integration",
          "zone-context": "labs",
          "status": "Active"
      }
  4. VNFs:
    1. Updated doc is available: ONAP vCPE VPP-based VNF Installation and Usage Information
  5. Closed loop: APPC-MultiCloud:
    1. The team is debugging the API request from APPC to MultiCloud. It is tracked by  MULTICLOUD-119 - Getting issue details... STATUS

10/23/2017

  1. MultiCloud found a bug and fixed it, tracked by  MULTICLOUD-118 - Getting issue details... STATUS .
  2. vGMUX: Eric has created a snapshot image for vGMUX and tested instantiation without ONAP. He and his team is working on the other VNFs.
  3. SDC: David Shadmi and Kang finished TunnelXConn VF design and vcperescust service design and distribution.
    1. To successfully distributed vcperescust, we need to first distribute vgmux service due to dependency.

10/19/2017

  1. APPC was able to reboot VM using legacy provider API. The team is currently debugging using the LCM framework where there is an issue with CDP PAL (Cloud Delivery Platform-Provider Abstraction Layer) interaction with MultiCloud.
  2. vGMUX: Eric and the team choose to build a image for each working VNF to avoid building and installation during instantiation. The team will also start to test the other VNFs including vBRG, vBNG, and vG.

10/18/2017

  1. APPC and MultiCloud located the problem to be API call request prefix. Bin Yang provided the following sample code to call MultiCloud API to do VM restart. APPC will fix the problem accordingly.

    export MULTICLOUD_PLUGIN_ENDPOINT=http://10.0.14.1:9005/api/multicloud-titanium_cloud/v0/pod25_RegionOne
    export TOKEN=$(curl -v -s -H "Content-Type: application/json" -X POST -d '{ }' $MULTICLOUD_PLUGIN_ENDPOINT/identity/v3/auth/tokens 2>&1 | grep X-Subject-Token | sed "s/^.*: //")
    export PROJECT_ID=466979b815b5415ba14ada713e6e1846
    curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X GET $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers
    curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X POST -d '{"os-stop":null}' $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers/0a06842a-4ec4-4918-b046-399f6b38f5f9/action
    curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X POST -d '{"os-start":null}' $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers/0a06842a-4ec4-4918-b046-399f6b38f5f9/action

  2. vGMUX can be instantiated using the HEAT but VPP code build failed. Eric was able to manual build using the same script. He is looking into the problem.

10/17/2017

  1. APPC has made good progress to fix a few bugs. The execution has reached MultiCloud API. A call is set up on 10/18 to look into this problem with MultiCloud people.

10/16/2017

  1. DCAE to Policy: Vijay Kumar has fixed TCA policy to put in the correct closed loop name. However, multiple problems exist in the TCA output, tracked by  DCAEGEN2-164 - Getting issue details... STATUS .
    1. Fixed vCPE policies on the wiki to use VNF as the controlLoopSchemaType. 
    2. Fixed TCA jar to include "version" as is needed by Policy.
  2. Passed test: → DCAE → Policy → . Manually fed a VES data to DCAE, observed an event from Policy to APPC.
    1. Note that currently Policy does not check the time stamp in the event so for testing purpose there is no need to increase the time stamp in the VES data.
    2. Once Policy sends out a restart event to APPC, it will not send another restart event until the current operation is completed or time expires even if it receives subsequent ONSET events. To retest in case operation fails in APPC, Policy needs to be restarted.
  3. Eric Multanen has fixed the even data content problem to include the correct vnf_id passed from the HEAT. See details appended by Eric to to 10/13/2017 notes.

10/13/2017

  1. DCAE to Policy: Vijay Kumar updated TCA and VES data is now processed by collector and TCA and output from TCA is observed on unauthenticated.DCAE_CL_OUTPUT. Policy does observe that event from TCA. But the closed-loop name is incorrect. Vijay Kumar will load the correct TCA Policy listed on Policy R1 Amsterdam Functional Test Cases and the test will continue next week.
  2. APPC to MultiCloud: Problems from the previous days have been fixed. Scott Seabolt tested through Swagger UI and identified a new problem caused by prefix spaces in DG. He will fix it soon.
  3. vGMUX: Eric Multanen has identified how to configure VES collector: one has to delete the current VES collector setting before setting a new one. This applies to both command line and API configuration. A question raised is that the current VES event uses VM name and VM ID. What we use in AAI is actually the VNF ID. Need to figure out how to set that VNF ID in vGMUX.
    1. From Eric: The "sourceName" and "sourceId" in the event data are generated by 'libevel.so' which comes from source code in demo/vnfs/VES5.0/evel/evel-library/code/evel_library.  These elements are currently populated by openstack metadata corresponding to the vm name and vm id.  The vG-MUX server has a property vnf_id='vCPE_Infrastructure_vGMUX_demo_app' - so it is presumably feasible to modify the libevel.so code to use this value (the same one for both 'sourceName' and 'sourceId' ?).  The question is: should this be a quick and dirty hack for the vG-MUX VNF specifically?  or, a more general fix in the library code?  I'd be worried about changing the library code in the 'demo' repository and breaking some other use case or demo.
    2. From Kang
      1. A quick hack should be good enough. In this use case, we only need vGMUX to report events so the other VNFs do not need a similar hack.
      2. Let's give vnf_id to both sourceName and sourceId. In principle only sourceId should be used by DCAE. But my test shows that only sourceName is used instead. Since vnf_id is specified in the HEAT and there is no sourceName at all in the HEAT, the above approach should work well to avoid potential inconsistency.
      3. Update based on communication with John Choma: VNF UUID is the same as VNF ID used in the HEAT.
    3. From Eric:

      Below is output of the updated event data - check the bold items to verify this is what is desired (note:  "vCPE_Infrastructure_vGMUX_demo_app" is the value of the "vnf_id" property from the OpenStack metadata of the vG-MUX VM):

      {"event": {"commonEventHeader": {"domain": "measurementsForVfScaling", "eventId": "Generic_traffic", "eventName": "Measurement_vGMUX", "lastEpochMicrosec": 1508188493486856, "priority": "Normal", "reportingEntityName": "zdcpe1cpe01mux01", "sequence": 55, "sourceName": "vCPE_Infrastructure_vGMUX_demo_app", "startEpochMicrosec": 1508188483486856, "version": 1.2, "eventType": "HTTP request rate", "reportingEntityId": "No UUID available", "sourceId": "vCPE_Infrastructure_vGMUX_demo_app"}, "measurementsForVfScalingFields": {"measurementInterval": 10, "cpuUsageArray": [{"cpuIdentifier": "cpu1", "cpuIdle": 100.000000, "cpuUsageSystem": 0.000000, "cpuUsageUser": 0.000000, "percentUsage": 0.000000}], "requestRate": 9956, "vNicUsageArray": [{"receivedOctetsDelta": 0.000000, "receivedTotalPacketsDelta": 0.000000, "transmittedOctetsDelta": 0.000000, "transmittedTotalPacketsDelta": 0.000000, "valuesAreSuspect": "true", "vNicIdentifier": "eth0"}], "additionalMeasurements": [{"name": "ONAP-DCAE", "arrayOfFields": [{"name": "Packet-Loss-Rate", "value": "0.0"}]}], "measurementsForVfScalingVersion": 2.1}}}

  4. Meeting recordings: 
    1. GMT20171013-135015_Kang-Xi-s-_1920x1200.mp4
    2. GMT20171013-191357_Kang-Xi-s-_1920x1200.mp4

10/12/2017

  1. AAI has a problem to pre-load data using this script closed-loop-biny993.zip, tracked by  AAI-433 - Getting issue details... STATUS .
  2. ethanlynnl has created a doc to show to register VMWare Openstack Fake Cloud into AAI and how to use it: https://gerrit.onap.org/r/#/c/18493/1/docs/Multicloud-Fake_Cloud-Guide.rst
  3. ESR installation guide: http://onap.readthedocs.io/en/latest/submodules/aai/esr-server.git/docs/platform/installation.html
  4. DCAE TCA is tracked by  DCAEGEN2-158 - Getting issue details... STATUS .
  5. Eric Multanen continues to fix vGMUX and made good progress. vGMUX in the ONAP-vCPE work space is now reporting VES data. Some fixes are needed. This is tracked by  INT-275 - Getting issue details... STATUS .

10/11/2017

  1. Test plan for CLAMP (Ron Shacham, Gervais-Martial Ngueko, Xue Gao, Lusheng Ji, Vijay Kumar):
    1. Integration team will install ONAP in a separate work space and let CLAMP do the testing. This is tracked in  INT-271 - Getting issue details... STATUS .
    2. CLAMP will start by test with Policy and SDC on the design part. The goal is to configure Policy and create Blueprint template for DCAE.
    3. Currently DCAEGEN2 cannot be installed in the open lab due to the lack of Openstack Designate support. This is expected to be fixed in a few days. In the meanwhile, CLAMP will do pairwise test with DCAE in the AT&T internal environment in parallel.
  2. MultiCloud (Bin Yangethanlynnl)
    1. MultiCloud sending query to AAI to obtain cloud info passed.
    2. MultiCloud performing VM restart (two commands: stop followed by start) passed. This was done by manually posting to the MultiCloud broker API.
    3. A VMWare VIM is simulated and passed VM restart test.
    4. Normally VIM is registered to ESR, and ESR will populate AAI with the cloud info. In this test, Bin Yang used a script to preload AAI with such info. MultiCloud queries AAI using VIM-ID to get the cloud info to do actual cloud operations.
  3. APPC→MultiCloud to restart VNF (Scott Seabolt, Ryan Young, Randa Maher)
    1. A bug is discovered inside APPC, tracked by APPC-268 - Getting issue details... STATUS .
  4. DCAE (Vijay Kumar)
    1. With the sampe VES provided by Eric Multanen on 10/10, the team has updated the TCA policy accordingly and will load it to TCA later today. It is tracked by  DCAEGEN2-135 - Getting issue details... STATUS .
    2. Note that in the final release, this TCA policy will be designed either using CLAMP or Policy GUI and distributed. No manual loading is needed.
  5. vGMUX (Eric Multanen)
    1. Eric is debugging vGMUX in his environment. Once it is done, he will see if it is feasible to create a VM image and use the image in the open lab instead.
  6. VxLAN config using SDNC
    1. We suggest that SDNC generates the VxLAN config request and send them to Eric Multanen to verify the correctness. It will help the upcoming pairwise test once the VNFs are ready in the open lab. This is tracked by  SDNC-120 - Getting issue details... STATUS .
  7. Customer flow testing in SO
    1. Update from Saryu Shah and BORALE, SHAILENDRA S <sb8915@att.com>: 

      The custom flow testing was performed using Junit testing.

      “mvn install” in so/bpmn/MSOInfrastructureBPMN will runs all the tests.

      Test is self-contained and all the required data files and simulator code was checked in. As of now, Jim is in the process of checking in bug-fixes.

  8. Meeting recording: GMT20171011-140734_Kang-Xi-s-_1920x1200.mp4

10/10/2017

  1. APPC: Scott Seabolt used Swagger API to invoke VNF restart. APPC passed AAI named query. The DG in APPC failed to query AAI because it did not point to the correct AAI. This is tracked by  APPC-267 - Getting issue details... STATUS .
  2. vGMUX: Eric has created a sample VES data as shown below. DCAE needs to be adjusted to process the data. This is tracked by DCAEGEN2-147 - Getting issue details... STATUS . Eric is also working to fix vGMUX.

    vCPE VES data
    {"event": {"commonEventHeader": {"domain": "measurementsForVfScaling", "eventId": "Generic_traffic", "eventName": "Measurement_vGMUX", "lastEpochMicrosec": 1507676920903343, "priority": "Normal", "reportingEntityName": "vg-mux", "sequence": 1, "sourceName": "Dummy VM name - No Metadata available", "startEpochMicrosec": 1507676910903343, "version": 1.2, "eventType": "HTTP request rate", "reportingEntityId": "No UUID available", "sourceId": "Dummy VM UUID - No Metadata available"}, "measurementsForVfScalingFields": {"measurementInterval": 10, "cpuUsageArray": [{"cpuIdentifier": "cpu1", "cpuIdle": 85.700000, "cpuUsageSystem": 14.300000, "cpuUsageUser": 0.000000, "percentUsage": 0.000000}], "requestRate": 9383, "vNicUsageArray": [{"receivedOctetsDelta": 0.000000, "receivedTotalPacketsDelta": 0.000000, "transmittedOctetsDelta": 0.000000, "transmittedTotalPacketsDelta": 0.000000, "valuesAreSuspect": "true", "vNicIdentifier": "eth0"}], "additionalMeasurements": [{"name": "ONAP-DCAE", "arrayOfFields": [{"name": "Packet-Loss-Rate", "value": "49.0"}]}], "measurementsForVfScalingVersion": 2.1}}}
  3. Meeting recording: GMT20171010-181134_Kang-Xi-s-_1920x1200.mp4

10/9/2017

Closed loop control test is decomposed into the following items.

  1. Design and load policy rules to policy engine. Jorge Hernandez created a closed loop for vCPE. Policy is tested by posting a DCAE TCA event on DMaaP. Policy reacted by sending an event to APPC to trigger VNF restart. The message format needs to be fixed, see POLICY-300 - Getting issue details... STATUS . And the topics need to be added to APPC properties, see  APPC-265 - Getting issue details... STATUS .

  2. Scripts are created to preload AAI. Scott Seabolt tested them in the integration environment. Both preloading and named query works.
    closed-loop.zip contains the necessary data (json) as well as shell scripts to load the Integration vGMUX.  Note the attached files load data specific the the "vCPE_Infrastructure_vGMUX_demo_app" VNF.  If you'd like to load a different VNF each of the json files owuld need to be modified to reflect the new VNF.
    1. put_closed_loop.sh

      • usage: put_closed_loop.sh A&AI-IP
    2. verify.sh
      • usage: verify.sh A&AI-IP
  3. VES collector is working. Waiting for sample VES data from Eric Multanen. See  DCAEGEN2-135 - Getting issue details... STATUS . Once the data is available, the TCA policy on Policy R1 Amsterdam Functional Test Cases may need to be adjusted and then TCA testing can be performed.
  4. vGMUX cannot be properly instantiated. Eric MultanenMarco Platania, and Kang Xi are working to solve it.
  5. Docker info shows that the multi-cloud services are running inside the open-o VM as multiple containers
    1. The VM is vm1-openo-server with IP 10.12.25.113
    2. Service ports are listed below. More details of each service are described on Setup MultiCloud Development Env 
      1. broker: 9001
      2. VIO plugin: 9004
      3. Ocata plugin: 9006
      4. WindRiver plugin: 9005


Meeting recording is GMT20171009-170725_Kang-Xi-s-_1920x1200.mp4

10/6/2017

Closed loop control test is decomposed into the following items.

  1. Design and load policy rules to policy engine. Ideally this should be done by CLAMP. Currently CLAMP is working on CLAMP-59 - Getting issue details... STATUS . The workaround is to use Policy GUI. Jorge Hernandez will work on this.

  2. Preload AAI with vCPE data. The purpose is to lookup vServer using vGMUX VNF ID. Scott Seabolt (APPC) will provide a sample data to Venkata Harish Kajur (AAI) to show the named query. Venkata Harish Kajur will then create a script to preload such data to AAI for the testing.

  3. vGMUX sends packet loss VES to DCAE VES collector.

    • Kang Xi created a vGMUX instance in the open lab but the VPP-based VNF inside the VM is not functioning. According to Eric Multanen, most likely it is caused by network issues. Eric is working on this. Kang will also inform Danny Zhou and lee xun about this.

    • The following VES data caused an error. The most possible reason is that the data itself is incorrect. Vijay Kumar is looking into that.

      VES Data
      curl -H "Accept: application/json" -H "Content-Type: application/json" -d '{"event":{"measurementsForVfScalingFields":{"measurementInterval":10,"measurementsForVfScalingVersion":1.1,"vNicUsageArray":[{"multicastPacketsIn":0,"bytesIn":4300,"unicastPacketsIn":0,"multicastPacketsOut":0,"broadcastPacketsOut":0,"packetsOut":0,"bytesOut":0,"broadcastPacketsIn":0,"packetsIn":101,"unicastPacketsOut":0,"vNicIdentifier":"eth1"}]},"commonEventHeader":{"reportingEntityName":"zdfw1fwl01fwl01","startEpochMicrosec":1500379662497999,"lastEpochMicrosec":1500379672497999,"eventId":"1","sourceName":"zdfw1fwl01fwl01","sequence":1,"priority":"Normal","functionalRole":"vFirewall","domain":"measurementsForVfScaling","reportingEntityId":"No UUID available","sourceId":"75ec15e4-1a9a-4ee3-bb3c-31556903558d","version":1.2}}}' 10.12.25.84:8080/eventListener/v5
      
      
  4. DCAE TCA sends an event to Policy through DMaaP. This is tested as described in  DCAEGEN2-124 - Getting issue details... STATUS

  5. Policy process the event from DCAE, and then sends a “restart” event to APPC through DMaaP. Jorge Hernandez will work on this.

  6. APPC captures the event from Policy, performs a named query to AAI to get the vServer ID based on the VNF ID, and then restarts the VM through MultiVIM. 

    • APPC is working on some changes and will be able to test next week. 
    • There should be a MultiVIM VM in the stack but we don't see any. Refer INT-258 - Getting issue details... STATUS  to track this problem.

In addition, the bottom of Policy R1 Amsterdam Functional Test Cases shows important sample data for vCPE test cases.

Update:

  • Scott Seabolt created the data pertaining to the vGMUX in POD25 which needs to be inventoried in POD25 A&AI. Venkata Harish Kajur then created a script to preload AAI with the following instructions: "Unzip this closed-loop.zip and go into closed-loop directory and run the ./put_closed_loop.sh ${AAI_VM1_IP_ADDRESS} And then you can verify that it works by doing a ./verify.sh ${AAI_VM1_IP_ADDRESS} and it should return you that named query response that you sent me."
  • Meeting recording

Related Documents:

  1. VNF Doc: ONAP vCPE VPP-based VNF Installation and Usage Information
  2. DCAE mS Deployment (Standalone instantiation)
  3. SNIRO Emulator


export MULTICLOUD_PLUGIN_ENDPOINT=http://10.0.14.1:9005/api/multicloud-titanium_cloud/v0/pod25_RegionOne

 

export TOKEN=$(curl -v -s -H "Content-Type: application/json" -X POST -d '{ }'  $MULTICLOUD_PLUGIN_ENDPOINT/identity/v3/auth/tokens 2>&1 | grep X-Subject-Token | sed "s/^.*: //")

 

export PROJECT_ID=466979b815b5415ba14ada713e6e1846

 

curl -v -s  -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X GET $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers

 

curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X POST -d '{"os-stop":null}'  $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers/0a06842a-4ec4-4918-b046-399f6b38f5f9/action

 

curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X POST -d '{"os-start":null}'  $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers/0a06842a-4ec4-4918-b046-399f6b38f5f9/action

  • No labels