from osm_mon.core.database import DatabaseManager
from osm_mon.core.message_bus.producer import KafkaProducer
from osm_mon.core.settings import Config
+from osm_mon.plugins.OpenStack.Gnocchi.metrics import METRIC_MAPPINGS
from osm_mon.plugins.OpenStack.common import Common
from osm_mon.plugins.OpenStack.response import OpenStack_Response
log = logging.getLogger(__name__)
-ALARM_NAMES = {
- "average_memory_usage_above_threshold": "average_memory_utilization",
- "disk_read_ops": "disk_read_ops",
- "disk_write_ops": "disk_write_ops",
- "disk_read_bytes": "disk_read_bytes",
- "disk_write_bytes": "disk_write_bytes",
- "net_packets_dropped": "packets_dropped",
- "packets_in_above_threshold": "packets_received",
- "packets_out_above_threshold": "packets_sent",
- "cpu_utilization_above_threshold": "cpu_utilization"}
-
-METRIC_MAPPINGS = {
- "average_memory_utilization": "memory.percent",
- "disk_read_ops": "disk.read.requests",
- "disk_write_ops": "disk.write.requests",
- "disk_read_bytes": "disk.read.bytes",
- "disk_write_bytes": "disk.write.bytes",
- "packets_dropped": "interface.if_dropped",
- "packets_received": "interface.if_packets",
- "packets_sent": "interface.if_packets",
- "cpu_utilization": "cpu_util",
-}
-
SEVERITIES = {
"warning": "low",
"minor": "low",
resource_id = values['resource_uuid']
if metric_name not in METRIC_MAPPINGS.keys():
- log.warn("This metric is not supported.")
+ log.warning("This metric is not supported.")
return None, False
# Check for the required metric
url, auth_token, req_type="post", payload=payload)
return json.loads(new_alarm.text)['alarm_id'], True
else:
- log.warn("The required Gnocchi metric does not exist.")
+ log.warning("The required Gnocchi metric does not exist.")
return None, False
except Exception as exc:
- log.warn("Failed to create the alarm: %s", exc)
+ log.warning("Failed to create the alarm: %s", exc)
return None, False
- def alarming(self, message):
+ def alarming(self, message, vim_uuid):
"""Consume info from the message bus to manage alarms."""
try:
values = json.loads(message.value)
values = yaml.safe_load(message.value)
log.info("OpenStack alarm action required.")
- vim_uuid = values['vim_uuid']
auth_token = Common.get_auth_token(vim_uuid)
alarm_endpoint, metric_endpoint, auth_token, alarm_details, vim_config)
# Generate a valid response message, send via producer
+ if alarm_status is True:
+ log.info("Alarm successfully created")
+ self._database_manager.save_alarm(alarm_id,
+ vim_uuid,
+ alarm_details['threshold_value'],
+ alarm_details['operation'].lower(),
+ alarm_details['metric_name'].lower(),
+ alarm_details['vdu_name'].lower(),
+ alarm_details['vnf_member_index'].lower(),
+ alarm_details['ns_id'].lower()
+ )
try:
- if alarm_status is True:
- log.info("Alarm successfully created")
- self._database_manager.save_alarm(alarm_id, vim_uuid)
-
resp_message = self._response.generate_response(
'create_alarm_response', status=alarm_status,
alarm_id=alarm_id,
cor_id=alarm_details['correlation_id'])
log.info("Response Message: %s", resp_message)
self._producer.create_alarm_response(
- 'create_alarm_response', resp_message,
- 'alarm_response')
+ 'create_alarm_response', resp_message)
except Exception:
log.exception("Response creation failed:")
cor_id=list_details['correlation_id'])
log.info("Response Message: %s", resp_message)
self._producer.list_alarm_response(
- 'list_alarm_response', resp_message,
- 'alarm_response')
+ 'list_alarm_response', resp_message)
except Exception:
log.exception("Failed to send a valid response back.")
cor_id=request_details['correlation_id'])
log.info("Response message: %s", resp_message)
self._producer.delete_alarm_response(
- 'delete_alarm_response', resp_message,
- 'alarm_response')
+ 'delete_alarm_response', resp_message)
except Exception:
log.exception("Failed to create delete response: ")
if response is True:
log.info("Acknowledged the alarm and cleared it.")
else:
- log.warn("Failed to acknowledge/clear the alarm.")
+ log.warning("Failed to acknowledge/clear the alarm.")
elif message.key == "update_alarm_request":
# Update alarm configurations
status=status)
log.info("Response message: %s", resp_message)
self._producer.update_alarm_response(
- 'update_alarm_response', resp_message,
- 'alarm_response')
+ 'update_alarm_response', resp_message)
except Exception:
log.exception("Failed to send an update response: ")
try:
resource = list_details['resource_uuid']
except KeyError as exc:
- log.warn("Resource id not specified for list request: %s", exc)
+ log.warning("Resource id not specified for list request: %s", exc)
return None
# Checking what fields are specified for a list request
try:
name = list_details['alarm_name'].lower()
- if name not in ALARM_NAMES.keys():
- log.warn("This alarm is not supported, won't be used!")
- name = None
except KeyError as exc:
log.info("Alarm name isn't specified.")
name = None
resource_id = rule['resource_id']
metric_name = [key for key, value in six.iteritems(METRIC_MAPPINGS) if value == rule['metric']][0]
except Exception as exc:
- log.warn("Failed to retrieve existing alarm info: %s.\
- Can only update OSM alarms.", exc)
+ log.exception("Failed to retrieve existing alarm info. Can only update OSM alarms.")
return None, False
# Generates and check payload configuration for alarm update
return json.loads(update_alarm.text)['alarm_id'], True
except Exception as exc:
- log.warn("Alarm update could not be performed: %s", exc)
- return None, False
+ log.exception("Alarm update could not be performed: ")
return None, False
def check_payload(self, values, metric_name, resource_id,
'alarm_actions': [cfg.OS_NOTIFIER_URI], })
return payload
except KeyError as exc:
- log.warn("Alarm is not configured correctly: %s", exc)
+ log.warning("Alarm is not configured correctly: %s", exc)
return None
def get_alarm_state(self, endpoint, auth_token, alarm_id):
url, auth_token, req_type="get")
return json.loads(alarm_state.text)
except Exception as exc:
- log.warn("Failed to get the state of the alarm:%s", exc)
+ log.warning("Failed to get the state of the alarm:%s", exc)
return None
def check_for_metric(self, auth_token, metric_endpoint, m_name, r_id):