/
Holmes Component
Holmes Component
See Guangrong's wiki Installation of Holmes (for Testing Purpose)
Test Resource
alarm_on.json
{
"event": {
"commonEventHeader":{
"sourceId":"example-vserver-name-val-75044",
"startEpochMicrosec":1413378172000000,
"eventId":"ab305d54-85b4-a31b-7db2-fb6b9e546015",
"sequence":0,
"domain":"fault",
"lastEpochMicrosec":1413378172000000,
"eventName":"Fault_MultiCloud_VMFailure",
"sourceName":"example-vserver-name-val-75044",
"priority":"High",
"version":3.0,
"reportingEntityName":"Multi-Cloud"
},
"faultFields": {
"eventSeverity":"CRITICAL",
"alarmCondition":"Guest_Os_Failure",
"faultFieldsVersion":2.0,
"specificProblem":"Fault_MultiCloud_VMFailure",
"alarmInterfaceA":"aaaa",
"eventSourceType":"other",
"vfStatus":"Active"
}
}
}
alarm_off.json
{
"event": {
"commonEventHeader":{
"sourceId":"example-vserver-name-val-75044",
"startEpochMicrosec":1413378172000000,
"eventId":"ab305d54-85b4-a31b-7db2-fb6b9e546015",
"sequence":0,
"domain":"fault",
"lastEpochMicrosec":1413378172000000,
"eventName":"Fault_MultiCloud_VMFailureCleared",
"sourceName":"example-vserver-name-val-75044",
"priority":"High",
"version":3.0,
"reportingEntityName":"Multi-Cloud"
},
"faultFields": {
"eventSeverity":"CRITICAL",
"alarmCondition":"Guest_Os_Failure",
"faultFieldsVersion":2.0,
"specificProblem":"Fault_MultiCloud_VMFailure",
"alarmInterfaceA":"aaaa",
"eventSourceType":"other",
"vfStatus":"Active"
}
}
}
policy_onset_input.json
{
"closedLoopEventClient": null,
"policyVersion": null,
"policyName": null,
"policyScope": null,
"target_type": "VM",
"AAI": {
"vserver.prov-status": "example-prov-status-val-9845",
"vserver.resource-version": "1509546979412",
"vserver.is-closed-loop-disabled": "false",
"vserver.vserver-name2": "example-vserver-name2-val-55230",
"vserver.vserver-id": "example-vserver-id-val-64888",
"vserver.vserver-selflink": "example-vserver-selflink-val-91298",
"vserver.in-maint": "true",
"vserver.vserver-name": "example-vserver-name-val-75044",
"generic-vnf.vnf-id": "example-vnf-id-val-31366",
"generic-vnf.service-instance-id": ""
},
"closedLoopAlarmStart": 1413378172000000,
"closedLoopEventStatus": "ONSET",
"version": "1.0.2",
"closedLoopControlName": "ControlLoop-VOLTE-2179b738-fd36-4843-a71a-a8c24c70c55b",
"target": "vserver.vserver-name",
"closedLoopAlarmEnd": 0,
"requestID": "a0d549da-cd11-4425-af1f-fa40fdfc44ff",
"from": "DCAE"
}
policy_abated_input.json
{
"closedLoopEventClient": null,
"policyVersion": null,
"policyName": null,
"policyScope": null,
"target_type": "VM",
"AAI": {
"vserver.vserver-name2": "example-vserver-name2-val-55230",
"vserver.vserver-id": "example-vserver-id-val-64888",
"vserver.vserver-selflink": "example-vserver-selflink-val-91298",
"vserver.vserver-name": "example-vserver-name-val-75044"
},
"closedLoopAlarmStart": 1413378172000000,
"closedLoopEventStatus": "ABATED",
"version": "1.0.2",
"closedLoopControlName": "ControlLoop-VOLTE-2179b738-fd36-4843-a71a-a8c24c70c55b",
"target": null,
"closedLoopAlarmEnd": 1413378172000000,
"requestID": "a0d549da-cd11-4425-af1f-fa40fdfc44ff",
"from": "DCAE"
}
sampleves
{
"event":{
"commonEventHeader":{
"version": 3.0,
"eventName": "Slave MPU is offline",
"domain": "fault",
"eventId": "1501489595451",
"eventType": "applicationVnf",
"nfcNamingCode": "",
"nfNamingCode": "",
"sourceId": "example-vnf-id-val-31366",
"sourceName": "example-vnf-name-val-51172",
"reportingEntityId": "0000ZTHX1",
"reportingEntityName": "0000ZTHX1",
"priority": "High",
"startEpochMicrosec": 1501518702,
"lastEpochMicrosec": 1501518702,
"sequence": 960
},
"faultFields":{
"faultFieldsVersion": 2.0,
"eventSeverity": "CRITICAL",
"eventSourceType": "PgwFunction",
"eventCategory": "equipmentAlarm",
"alarmCondition": "The slave MPU board is offline or abnormal",
"specificProblem": "Slave MPU is offline",
"vfStatus": "Active",
"alarmInterfaceA": "VNF_194.15.13.138",
"alarmAdditionalInformation": [{
"name":"specificProblemID",
"value": "315"
}, {
"name":"objectUID",
"value":"0000ZTHX1PGWGJI6V1"
}, {
"name":"locationInfo",
"value":"MPU_22_20_0"
}, {
"name":"addInfo",
"value":"Aid:17;AlarmCode:110010;AlarmReasonCode:110010;Remark:\"DeployUnit=22,Node=21,SubNode=0\";"
}]
}
}
}
abatedves
{
"event":{
"commonEventHeader":{
"version": 3.0,
"eventName": "Slave MPU is offlineCleared",
"domain": "fault",
"eventId": "1501489595451",
"eventType": "applicationVnf",
"nfcNamingCode": "",
"nfNamingCode": "",
"sourceId": "example-vnf-id-val-31366",
"sourceName": "example-vnf-name-val-51172",
"reportingEntityId": "0000ZTHX1",
"reportingEntityName": "0000ZTHX1",
"priority": "High",
"startEpochMicrosec": 1501518702,
"lastEpochMicrosec": 1501518702,
"sequence": 960
},
"faultFields":{
"faultFieldsVersion": 2.0,
"eventSeverity": "CRITICAL",
"eventSourceType": "PgwFunction",
"eventCategory": "equipmentAlarm",
"alarmCondition": "The slave MPU board is offline or abnormal",
"specificProblem": "The slave MPU board is offline or abnormal",
"vfStatus": "Active",
"alarmInterfaceA": "VNF_194.15.13.138",
"alarmAdditionalInformation": [{
"name":"specificProblemID",
"value": "315"
}, {
"name":"objectUID",
"value":"0000ZTHX1PGWGJI6V1"
}, {
"name":"locationInfo",
"value":"MPU_22_20_0"
}, {
"name":"addInfo",
"value":"Aid:17;AlarmCode:110010;AlarmReasonCode:110010;Remark:\"DeployUnit=22,Node=21,SubNode=0\";"
}]
}
}
}
holmesrule
{
"ruleName": "VoLTE",
"loopControlName": "ControlLoop-VOLTE-2179b738-fd36-4843-a71a-a8c24c70c55b",
"description": "This rule is designed for the correlation analysis for the VoLTE use case.",
"content": "package org.onap.holmes.droolsRule;\n\nimport org.onap.holmes.common.dmaap.DmaapService;\nimport org.onap.holmes.common.api.stat.VesAlarm;\nimport org.onap.holmes.common.aai.CorrelationUtil;\nimport org.onap.holmes.common.dmaap.entity.PolicyMsg;\nimport org.onap.holmes.common.dropwizard.ioc.utils.ServiceLocatorHolder;\nimport org.onap.holmes.common.utils.DroolsLog;\n \n\nrule \"Relation_analysis_Rule\"\nsalience 200\nno-loop true\n when\n $root : VesAlarm(alarmIsCleared == 0,\n $sourceId: sourceId, sourceId != null && !sourceId.equals(\"\"),\n\t\t\t$sourceName: sourceName, sourceName != null && !sourceName.equals(\"\"),\n\t\t\t$startEpochMicrosec: startEpochMicrosec,\n specificProblem in (\"Fault_MultiCloud_VMFailure\"),\n $eventId: eventId)\n $child : VesAlarm( eventId != $eventId, parentId == null,\n CorrelationUtil.getInstance().isTopologicallyRelated(sourceId, $sourceId, $sourceName),\n specificProblem in (\"Slave MPU is offline\"),\n startEpochMicrosec < $startEpochMicrosec + 60000 && startEpochMicrosec > $startEpochMicrosec - 60000 )\n then\n\t\t$child.setParentId($root.getEventId());\n\t\tupdate($child);\n\t\t\nend\n\nrule \"root_has_child_handle_Rule\"\nsalience 150\nno-loop true\n\twhen\n\t\t$root : VesAlarm(alarmIsCleared == 0, rootFlag == 0, $eventId: eventId)\n\t\t$child : VesAlarm(eventId != $eventId, parentId == $eventId)\n\tthen\n\t\tDmaapService dmaapService = ServiceLocatorHolder.getLocator().getService(DmaapService.class);\n\t\tPolicyMsg policyMsg = dmaapService.getPolicyMsg($root, $child, \"org.onap.holmes.droolsRule\");\n dmaapService.publishPolicyMsg(policyMsg, \"unauthenticated.DCAE_CL_OUTPUT\");\n\t\t$root.setRootFlag(1);\n\t\tupdate($root);\nend\n\nrule \"root_no_child_handle_Rule\"\nsalience 100\nno-loop true\n when\n $root : VesAlarm(alarmIsCleared == 0, rootFlag == 0,\n sourceId != null && !sourceId.equals(\"\"),\n\t\t\tsourceName != null && !sourceName.equals(\"\"),\n specificProblem in (\"Fault_MultiCloud_VMFailure\"))\n then\n\t\tDmaapService dmaapService = ServiceLocatorHolder.getLocator().getService(DmaapService.class);\n\t\tPolicyMsg policyMsg = dmaapService.getPolicyMsg($root, null, \"org.onap.holmes.droolsRule\");\n dmaapService.publishPolicyMsg(policyMsg, \"unauthenticated.DCAE_CL_OUTPUT\");\n\t\t$root.setRootFlag(1);\n\t\tupdate($root);\nend\n\nrule \"root_cleared_handle_Rule\"\nsalience 100\nno-loop true\n when\n $root : VesAlarm(alarmIsCleared == 1, rootFlag == 1)\n then\n\t\tDmaapService dmaapService = ServiceLocatorHolder.getLocator().getService(DmaapService.class);\n\t\tPolicyMsg policyMsg = dmaapService.getPolicyMsg($root, null, \"org.onap.holmes.droolsRule\");\n dmaapService.publishPolicyMsg(policyMsg, \"unauthenticated.DCAE_CL_OUTPUT\");\n\t\tretract($root);\nend\n\nrule \"child_handle_Rule\"\nsalience 100\nno-loop true\n when\n $child : VesAlarm(alarmIsCleared == 1, rootFlag == 0)\n then\n\t\tretract($child);\nend",
"enabled": 1
}
To illustrate, the content field in the above rule is presented as readable format in the following. You need to convert it to valid json string when uploading to Holmes.
Rule content field
package org.onap.holmes.droolsRule;
import org.onap.holmes.common.dmaap.DmaapService;
import org.onap.holmes.common.api.stat.VesAlarm;
import org.onap.holmes.common.aai.CorrelationUtil;
import org.onap.holmes.common.dmaap.entity.PolicyMsg;
import org.onap.holmes.common.dropwizard.ioc.utils.ServiceLocatorHolder;
import org.onap.holmes.common.utils.DroolsLog;
rule "Relation_analysis_Rule"
salience 200
no-loop true
when
$root : VesAlarm(alarmIsCleared == 0,
$sourceId: sourceId, sourceId != null && !sourceId.equals(""),
$sourceName: sourceName, sourceName != null && !sourceName.equals(""),
$startEpochMicrosec: startEpochMicrosec,
specificProblem in ("Fault_MultiCloud_VMFailure"),
$eventId: eventId)
$child : VesAlarm( eventId != $eventId, parentId == null,
CorrelationUtil.getInstance().isTopologicallyRelated(sourceId, $sourceId, $sourceName),
specificProblem in ("Slave MPU is offline"),
startEpochMicrosec < $startEpochMicrosec + 60000 && startEpochMicrosec > $startEpochMicrosec - 60000 )
then
$child.setParentId($root.getEventId());
update($child);
end
rule "root_has_child_handle_Rule"
salience 150
no-loop true
when
$root : VesAlarm(alarmIsCleared == 0, rootFlag == 0, $eventId: eventId)
$child : VesAlarm(eventId != $eventId, parentId == $eventId)
then
DmaapService dmaapService = ServiceLocatorHolder.getLocator().getService(DmaapService.class);
PolicyMsg policyMsg = dmaapService.getPolicyMsg($root, $child, "org.onap.holmes.droolsRule");
dmaapService.publishPolicyMsg(policyMsg, "unauthenticated.DCAE_CL_OUTPUT");
$root.setRootFlag(1);
update($root);
end
rule "root_no_child_handle_Rule"
salience 100
no-loop true
when
$root : VesAlarm(alarmIsCleared == 0, rootFlag == 0,
sourceId != null && !sourceId.equals(""),
sourceName != null && !sourceName.equals(""),
specificProblem in ("Fault_MultiCloud_VMFailure"))
then
DmaapService dmaapService = ServiceLocatorHolder.getLocator().getService(DmaapService.class);
PolicyMsg policyMsg = dmaapService.getPolicyMsg($root, null, "org.onap.holmes.droolsRule");
dmaapService.publishPolicyMsg(policyMsg, "unauthenticated.DCAE_CL_OUTPUT");
$root.setRootFlag(1);
update($root);
nend
rule "root_cleared_handle_Rule"
salience 100
no-loop true
when
$root : VesAlarm(alarmIsCleared == 1, rootFlag == 1)
then
DmaapService dmaapService = ServiceLocatorHolder.getLocator().getService(DmaapService.class);
PolicyMsg policyMsg = dmaapService.getPolicyMsg($root, null, "org.onap.holmes.droolsRule");
dmaapService.publishPolicyMsg(policyMsg, "unauthenticated.DCAE_CL_OUTPUT");
retract($root);
end
rule "child_handle_Rule"
salience 100
no-loop true
when
$child : VesAlarm(alarmIsCleared == 1, rootFlag == 0)
then
retract($child);
end
Before install Holmes dockers, you need to install postgres
Install Postgres
root@vm03-dcae-controller:~# docker login -u docker -p docker nexus3.onap.org:10001
Login Succeeded
root@vm03-dcae-controller:~# docker run --name postgres -p 5432:5432 -e POSTGRES_USER=holmes -e POSTGRES_PASSWORD=holmespwd -d postgres:9.5
Holmes can be deployed as standalone component by using the following steps
Holmes Manual Setup Steps
# Remove the current running instances
docker rm -f holmes-rule-management
docker rm -f holmes-engine-management
# Download latest images
docker login -u docker -p docker nexus3.onap.org:10001
docker pull nexus3.onap.org:10001/onap/holmes/rule-management
docker pull nexus3.onap.org:10001/onap/holmes/engine-management
# Start two Holmes docker containers
sudo docker run --name holmes-rule-management -p 9101:9101 -d -e URL_JDBC=10.0.4.1 -e MSB_ADDR=10.0.14.1 -e TESTING=1 -e HOST_IP=10.0.4.1 nexus3.onap.org:10001/onap/holmes/rule-management
sudo docker run --name holmes-engine-management -p 9102:9102 -d -e URL_JDBC=10.0.4.1 -e MSB_ADDR=10.0.14.1 -e TESTING=1 -e HOST_IP=10.0.4.1 nexus3.onap.org:10001/onap/holmes/engine-management
# Start with an older image
#sudo docker run --name holmes-engine-management -p 9102:9102 -d -e URL_JDBC=10.0.4.1 -e MSB_ADDR=10.0.14.1 -e TESTING=1 -e HOST_IP=10.0.4.1 e7d17bb9a869
# Check health
curl http://10.0.14.1/api/holmes-rule-mgmt/v1/healthcheck
curl http://10.0.14.1/api/holmes-engine-mgmt/v1/healthcheck
# Configure sub and pub topics for Holmes by using registered msb url
curl -X PUT -H 'Content-Type: application/json' -d '{"name":"unauthenticated.SEC_FAULT_OUTPUT", "url":"http://10.0.11.1:3904/events/unauthenticated.SEC_FAULT_OUTPUT"}' http://10.0.14.1/api/holmes-engine-mgmt/v1/dmaap/sub
curl -X PUT -H 'Content-Type: application/json' -d '{"name":"unauthenticated.DCAE_CL_OUTPUT", "url":"http://10.0.11.1:3904/events/unauthenticated.DCAE_CL_OUTPUT"}' http://10.0.14.1/api/holmes-engine-mgmt/v1/dmaap/pub
# Upload Holmes rules defined in file holmesrule
curl -X PUT -H 'Content-Type: application/json' -d @holmesrule http://10.0.14.1/api/holmes-rule-mgmt/v1/rule
# Simulate alarm_on event from MultiVIM to VES
curl -i -X POST -d @alarm_on.json --header "Content-Type: application/json" http://localhost:8081/eventListener/v5 -k
Update Holmes rules by deleting the existing rule with rule_id and upload the new rule
Update Holmes Rules
curl -X GET http://10.0.14.1/api/holmes-rule-mgmt/v1/rule |python -m json.tool
curl -X DELETE http://10.0.14.1/api/holmes-rule-mgmt/v1/rule/rule_1508902811935
curl -X PUT -H 'Content-Type: application/json' -d @holmesrule http://10.0.14.1/api/holmes-rule-mgmt/v1/rule
, multiple selections available,