1. UUI Configuration
Configure CST template UUID and Invariant UUID in slicing.properties file of uui-server microservice
In uui-server microservice, modify the following configuration file, /home/UUI/config/slicing.properties
Add or update the parameters slicing.serviceInvariantUuid and parameter slicing.serviceUuid.
The values of these two parameters come from CST template which can be find on SDC page.
2. MSB Configuration
Register so-orchestrationTasks and so-serviceInstances interface to MSB.
Interface registration can be done through portal.
Steps(Portal):
Link: https://{{master server ip}}:30284/iui/microservices/default.html
1.Select ‘’ in the left pane
2. Click 'Service Register' button.
3. Input the basic info as the picture shows(also refer to the registration info provided above)
4. Click Add Host button.
Input IP Address and Port then click the 'SAVE' button. (Use cmd ’kubectl get svc -n onap so‘ to confirm IP and port.)
6. You should add a aai-business service for MSB.
Steps:
- Got to msb https://{{master server ip}}:30284/iui/microservices/default.html
- Select "Service Discover" from left panel
- Click "Service Register" button
- ADD the following info:
Service Name: aai-business
Url: /aai/v13/business
Protocol: REST
Enable SSL to True
Version: v13
Load balancer: round-robin
Visualranfe: InSystem - Add host:
AAI service ip and port (8443)
- ADD the following info:
- Save all
3. AAI Configuration
Create customer id :
curl --user AAI:AAI -X PUT -H "X-FromAppId:AAI" -H "X-TransactionId:get_aai_subscr" -H "Accept:application/json" -H "Content-Type:application/json" -k -d '{
"global-customer-id":"5GCustomer",
"subscriber-name":"5GCustomer",
"subscriber-type":"INFRA"
}' "https://<worker-vm-ip>:30233/aai/v21/business/customers/customer/5GCustomer"
Create service type:
curl --user AAI:AAI -X PUT -H "X-FromAppId:AAI" -H "X-TransactionId:get_aai_subscr" -H "Accept:application/json" -H "Content-Type:application/json" -k https://<worker-vm-ip>:30233/aai/v21/business/customers/customer/5GCustomer/service-subscriptions/service-subscription/5G
4. SO
Copy subnetCapability.json to SO-API Handler pod to configure subnet capabilities at run time.
{
"AN_NF": {
"latency": 5,
"maxNumberofUEs": 200,
"maxThroughput": 90,
"termDensity": 40
},
"AN": {
"latency": 20,
"maxNumberofUEs": 100,
"maxThroughput": 150,
"termDensity": 50
},
"CN": {
"latency": 10,
"maxThroughput": 50,
"maxNumberofConns": 100
},
"TN_FH": {
"latency": 10,
"maxThroughput": 90
},
"TN_MH": {
"latency": 5,
"maxThroughput": 90
},
"TN_BH": {
"latency": 10,
"maxThroughput": 100
}
}
You can copy the file to the pod using the following command
|
SO Database Update
Insert ORCHESTRATION_URI into service_recipe, SERVICE_MODEL_UUID replaced by CST.ModelId.
|
Insert ORCHESTRATION_URI into service_recipe, SERVICE_MODEL_UUID is ServiceProfile.ModelId
|
5. Policy Creation Steps
Refer Optimization Policy Creation Steps for optimization policy creation and deployment steps
Please find the policies for Option1 below, generate the policies outside the OOF pod and push the policies from inside OOF pod, since it has python3 and necessary libraries already installed
Copy the policy files
unzip policies.zip |
NOTE: For NST Selection based on latency constraint, please make sure you have updated the latency constraint as property in the design time template of NST as below,
Refer Policy Models and Sample policies - NSI selection for sample policies
Updated slice/service profile mapping - https://gerrit.onap.org/r/gitweb?p=optf/osdf.git;a=blob;f=config/slicing_config.yaml;h=179f54a6df150a62afdd72938c2f33d9ae1bd202;hb=HEAD
HAS-API/HAS-DATA - Add data dictionary
Go to (/opt/has/conductor/conductor/data/plugins/inventory_provider/candidates/slice_profiles_candidate.py) in OOF HAS pod update the following : "max_bandwidth": copy_first, "jitter": sum, "sst": copy_first, "latency": sum, "resource_sharing_level": copy_first, "s_nssai": copy_first, "s_nssai_list": copy_first, "plmn_id_list": copy_first, "plmn_id_List": copy_first, "availability": copy_first, "throughput": min, "reliability": copy_first, "max_number_of_ues": copy_first, "exp_data_rate_ul": copy_first, "exp_data_rate_dl": copy_first, "ue_mobility_level": copy_first, "activity_factor": copy_first, "survival_time": copy_first, "max_number_of_conns": copy_first, "coverage_area_ta_list": copy_first, "max_number_of_pdu_session": copy_first, "max_throughput": copy_first, "perf_req": copy_first, "terminal_density": copy_first update those and restart the container
NOTE:
- The service name given for creating the policy must match with the service name in the request
- The scope fields in the policies should match with the value in the resourceSharingLevel(non-shared/shared). Do modify the policy accordingly.
- Check the case of the attributes with the OOF request with the attribute map (camel to snake and snake to camel) in config/slicing_config.yaml, if any mismatch found modify the attribute map accordingly.
You need to restart the OOF docker container once you updated the slicing_config.yaml, you can do it using the following steps,
- Login to the worker VM where the OOF container is running. You can find the worker node by running (kubectl get pods -n onap -o wide | grep dev-oof)
- Find the container using docker ps | grep optf-osdf
- Restart the container using docker restart <container id>
6. Core NF Simulator Setup
- We need a KUD vm with kubernetes installed, please follow below link to setup a KUD VM.
https://wiki.onap.org/display/DW/Kubernetes+Baremetal+deployment+setup+instructions/
Please find below script with required data to be populated in AAI, provide the tenant id of the openstack project
- After KUD setup and AAI data entries, update the connectivity Info: <port: msb-iag port>
curl -k -i -F "metadata=<post.json;type=application/json" -F file=@/root/.kube/config -X POST https://{{k8s}}:32330/api/multicloud-k8s/v1/v1/connectivity-info
Please find post.json below,
{
"cloud-region": "k8sregionfour",
"cloud-owner": "k8scloudowner4"
}
- Load the dictionary data using the below dd.sh script,
Refer Core NF Simulator setup to setup Core NF Simulator.
7. RAN Simulator Setup
7a) RANSIM Controller
Refer RAN-Sim setup to setup RAN-simulator
Refer RAN-Sim for ransim impacts done for Guilin Release
7b) Honeycomb:
Repo URL: https://github.com/onap-oof-pci-poc/ran-sim
Pre-requisite: SDNC & Ransim controller should be up and running.
Build Steps:
If there is a new yang model, build the gnbsim-api with new yang model
Step 1: copy the new yang model to ~/ran-sim/hcsim-content/gnbsim/gnbsim-api/src/main/yang
Step 2: Go to directory ~/ran-sim/hcsim-content/gnbsim/gnbsim-api/
Step 3: Issue the build command "JAVA_HOME=/usr/lib/jvm/java-1.11.0-openjdk-amd64 mvn clean install -Dcheckstyle.skip"
For code changes in gnbsim-impl directory
Step 1: Do necessary code changes in gnbsim-impl
Step 2: Issue the build command "JAVA_HOME=/usr/lib/jvm/java-1.11.0-openjdk-amd64 mvn clean install -Dcheckstyle.skip"
Once the mvn build is done, build the docker image
Step 1: Go to directory ~/ran-sim/hcsim-content/gnbsim
Step 2: JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64/ mvn clean install -Dcheckstyle.skip
Step 3: cp -r gnbsim-distribution/jssecacerts gnbsim-distribution/target/gnbsim-distribution-1.19.08-SNAPSHOT-hc/gnbsim-distribution-1.19.08-SNAPSHOT/
Step 4: cp -r gnbsim-distribution/honeycomb gnbsim-distribution/target/gnbsim-distribution-1.19.08-SNAPSHOT-hc/gnbsim-distribution-1.19.08-SNAPSHOT/
Step 5: cp -r gnbsim-distribution/Dockerfile gnbsim-distribution/target/gnbsim-distribution-1.19.08-SNAPSHOT-hc/gnbsim-distribution-1.19.08-SNAPSHOT/
Step 6: docker build -t gn gnbsim-distribution/target/gnbsim-distribution-1.19.08-SNAPSHOT-hc/gnbsim-distribution-1.19.08-SNAPSHOT/
Cross check the docker image
docker images | grep gnbsim
To run the Netconf Servers:
- Update ransim and honeycom IP address at ~/ran-sim/hcsim-content/gnbsim/hc/config/gnbsim.json
2. Run the below command from ~/ran-sim/hcsim-content/gnbsim/hc directory.
docker-compose up -d
For slicing setup, 12 netconf servers are required. Copy the hc directory 11 more time and rename it.
Change the external port in the docker-compose files for every netconf server.
While running the docker-compose up -d command, these servers will be mounted in SDNC.
Setup SFPT:
- Run the below docker command in the VM where you have setup the netconf servers:
docker run -v /home/$USER/upload:/home/$USER/sftptest/upload -p 2222:22 -d atmoz/sftp $USER:pass:1001
2. Give chmod 777 to upload folder which will be created automatically in the home directory.
8. ConfigDB
Config DB is a spring boot application that works with mariaDB. DB schema details are available at Config DB.
Install config DB application in a separate VM. MariaDB container should be up and running to access the config DB APIs.
Refer https://wiki.onap.org/display/DW/Config+DB+setup for configDB setup. Latest source is available at Image versions, preparation steps and useful info-Config DB Preload Info Section.
Necessary RAN network functions data are preloaded in config DB while booting the maria DB container.
Note: Refer the latest templates from gerrit which are committed in June 2021. https://gerrit.onap.org/r/gitweb?p=ccsdk/distribution.git;a=commit;h=8b86f34f6ea29728e31c4f6799009e8562ef3b6f
9. ACTN Simulator
Refer ACTN Simulator User Guide to setup ACTN-simulator or follow the below steps to launch and initialize domain controllers,
Step 1. Fetch the simulator docker image
First, pull down the simulator image from public docker hub.
|
Step 2. Start the simulator container(s)
After compiling the pnc-simulator image locally or fetching the docker image from remote registry, execute the command below to start the container:
|
|
For demonstration purposes, two domain controllers are required, in our case we use 18181 as SERVER_PORT for domain1 and 18182 for domain2.
After starting the container, you should be able to verify the running container by:
|
Step 3. Initialize the simulator(s)
The simulator offers a set of administrative APIs which are designed for initialization and management purposes.
To initialize the simulator for each domain:
|
For demonstration purposes, two domain controllers are required, both need to be properly launched and initialized. The initialization payloads are attached below.
ESR Registration using AAI ESR Url:
Once a third-party controller is being registered in ONAP, a "Topology Discovery" operation need be triggered and the domain controller will be called in order to retrieve the topology. SDNC then, will store the topology in AAI. Please execute the below curl command for the ESR registration of initialized two domain controllers,
NOTE: Replace <simulator-ip> and <simulator-port> with the IP Address and port of initialized simulators.
E2E network slicing test(using resource occupancy details)
Before creating shared communication service for testing NSI/NSSI selection using resource occupancy details, below manual configurations must be done:
Deployment Prerequisite/dependencies
- VES, PM mapper and DataFile Collector, dcae-mongodb, Datalake Extraction Service(DES) and Datalake-feeder services should be running.
Datalake Extraction Service(DES):
- Build presto image and push the images to a exsting repository
- Download and extract presto package version v0.0.2: presto-v0.0.2.tar.gz
- docker build -t presto:v0.0.2 .
- docker tag presto:v0.0.2 registry.baidubce.com/onap/presto:v0.0.2
- docker push registry.baidubce.com/onap/presto:v0.0.2
- Note: Replace the repository path with your own repository.
- Install presto service
- kubectl -n onap run dl-presto --image=presto:v0.0.2 --env="MongoDB_IP=dcae-mongohost" --env="MongoDB_PORT=27017" --image-pull-policy=IfNotPresent
- kubectl expose pod dl-presto --name=dl-presto --port=9000 --target-port=9000 --type=NodePort -n onap
- Login to the dev-dcae-datalake-postgres-primary container and execute following sql commands:
Load sample data in mongo db(refer DCAE mongodb section)
- Execute below request to verify: curl --header "Content-type: application/json" --request POST --data '{"snssai":"SM.PrbUsedDl.01-910EF6","time":"1649853949111"}' http://dl-des:1681/datalake/v1/exposure/pm_data -i
Datalake Feeder:
- Restart datalake-feeder container after loading above configurations in the DB.
POLICY :
The following policies are to be included for the OOF optimisation based on resource occupancy . Refer Optimization Policy Creation Steps for optimization policy creation and deployment steps.
RANSIM :
- RANSIM and honeycomb containers should be running
- You can update/add/delete data(throughput, PRBs, maxNoConns etc) in mariadb container(credentials in docker-compose file) if required before generation of PM data.
- RANSIM must contain 'plmninfo' of the service which is to be shared with other services. Then generate closed-loop PM data which will be stored in mongo-db by datalake-feeder.
- Sample closed loop data: closed-loop-data.txt
- Run the following command to generate Closed Loop PM data:
- curl -X POST http://localhost:8081/ransim/api/generateClosedLoopPmData -i
- Run the following command to stop Closed Loop PM data(When enough PM data is generated):
- curl -X GET http://localhost:8081/ransim/api/stopClosedLoopPmData -i
- Access the DU honeycomb containers to view the logs:
- docker container logs -f 110
- Check the folder(where SFTP setup is done) to view the generated PM data files.
- Access VES, data file collector, dmaap-dr, PM mapper and slice-analysis logs to verify.
DCAE-MONGO DB:
- Login to dcae-mongodb container and execute following commands to check if PM data was saved in DB:
- mongo
- use datalake
- db.unauthenticatedperformancemeasurements.find({})
- to load sample data:
E2E network slicing issues
Portal UI not working . Alternative for accessing to SDC UI .
If you cannot access to SDC UI via Portal UI, you need to do the following steps in order to be authenticated for accessing SDC UI.
- Go to https://<<sdc_ip>>:<<sdc_port>/ with Google Chrome Dev
- Open Inspect and go to Application section
- On left column, open Cookies
- Add the following cookie key/value
USER_ID cs0008 for designer role
USER_ID demo for admin role
This hack can be possible since SDC UI authentication mode is cookie based. So the attacker can override the cookie to force its request.