X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=osm_policy_module%2Fcore%2Fagent.py;h=34b852aed996f4816446c8c8c9da2a0b8f8a4b6e;hb=1a0af1c6928ce7269d25ab10be8aac6c21f69d8b;hp=a194b08151f3f3469deb34d88b22dec94c558f1e;hpb=29d590a038ce5c8d745f97ab935393c0bd6a8228;p=osm%2FPOL.git diff --git a/osm_policy_module/core/agent.py b/osm_policy_module/core/agent.py index a194b08..34b852a 100644 --- a/osm_policy_module/core/agent.py +++ b/osm_policy_module/core/agent.py @@ -21,217 +21,115 @@ # For those usages not covered by the Apache License, Version 2.0 please # contact: bdiaz@whitestack.com or glavado@whitestack.com ## -import json +import asyncio import logging -import threading -from json import JSONDecodeError -import yaml -from kafka import KafkaConsumer +import peewee -from osm_policy_module.common.db_client import DbClient -from osm_policy_module.common.lcm_client import LcmClient -from osm_policy_module.common.mon_client import MonClient -from osm_policy_module.core import database +from osm_policy_module.alarming.service import AlarmingService +from osm_policy_module.autoscaling.service import AutoscalingService +from osm_policy_module.common.common_db_client import CommonDbClient +from osm_policy_module.common.message_bus_client import MessageBusClient from osm_policy_module.core.config import Config -from osm_policy_module.core.database import ScalingGroup, ScalingAlarm, ScalingPolicy, ScalingCriteria log = logging.getLogger(__name__) -ALLOWED_KAFKA_KEYS = ['instantiated', 'scaled', 'notify_alarm'] +ALLOWED_KAFKA_KEYS = ['instantiated', 'scaled', 'terminated', 'notify_alarm'] class PolicyModuleAgent: - def __init__(self): - cfg = Config.instance() - self.db_client = DbClient() - self.mon_client = MonClient() - self.lcm_client = LcmClient() - self.kafka_server = '{}:{}'.format(cfg.OSMPOL_MESSAGE_HOST, - cfg.OSMPOL_MESSAGE_PORT) + def __init__(self, config: Config, loop=None): + self.conf = config + if not loop: + loop = asyncio.get_event_loop() + self.loop = loop + self.msg_bus = MessageBusClient(config) + self.db_client = CommonDbClient(config) + self.autoscaling_service = AutoscalingService(config, loop) + self.alarming_service = AlarmingService(config, loop) def run(self): - consumer = KafkaConsumer(bootstrap_servers=self.kafka_server, - key_deserializer=bytes.decode, - value_deserializer=bytes.decode, - group_id='pol-consumer') - consumer.subscribe(["ns", "alarm_response"]) - - for message in consumer: - t = threading.Thread(target=self._process_msg, args=(message.topic, message.key, message.value,)) - t.start() - - def _process_msg(self, topic, key, msg): + self.loop.run_until_complete(self.start()) + + async def start(self): + topics = [ + "ns", + "alarm_response" + ] + await self.msg_bus.aioread(topics, self._process_msg) + log.critical("Exiting...") + + async def _process_msg(self, topic, key, msg): + log.debug("_process_msg topic=%s key=%s msg=%s", topic, key, msg) try: - log.debug("Message arrived with topic: %s, key: %s, msg: %s", topic, key, msg) if key in ALLOWED_KAFKA_KEYS: - try: - content = json.loads(msg) - except JSONDecodeError: - content = yaml.safe_load(msg) - if key == 'instantiated' or key == 'scaled': - self._handle_instantiated_or_scaled(content) + if key == 'instantiated': + await self._handle_instantiated(msg) + + if key == 'scaled': + await self._handle_scaled(msg) + + if key == 'terminated': + await self._handle_terminated(msg) if key == 'notify_alarm': - self._handle_alarm_notification(content) + await self._handle_alarm_notification(msg) else: log.debug("Key %s is not in ALLOWED_KAFKA_KEYS", key) + except peewee.PeeweeException: + log.exception("Database error consuming message: ") + raise except Exception: log.exception("Error consuming message: ") - def _handle_alarm_notification(self, content): - alarm_id = content['notify_details']['alarm_uuid'] - metric_name = content['notify_details']['metric_name'] - operation = content['notify_details']['operation'] - threshold = content['notify_details']['threshold_value'] - vdu_name = content['notify_details']['vdu_name'] - vnf_member_index = content['notify_details']['vnf_member_index'] - ns_id = content['notify_details']['ns_id'] - log.info( - "Received alarm notification for alarm %s, \ - metric %s, \ - operation %s, \ - threshold %s, \ - vdu_name %s, \ - vnf_member_index %s, \ - ns_id %s ", - alarm_id, metric_name, operation, threshold, vdu_name, vnf_member_index, ns_id) - try: - alarm = ScalingAlarm.select().where(ScalingAlarm.alarm_id == alarm_id).get() - log.info("Sending scaling action message for ns: %s", alarm_id) - self.lcm_client.scale(alarm.scaling_record.nsr_id, alarm.scaling_record.name, alarm.vnf_member_index, - alarm.action) - except ScalingAlarm.DoesNotExist: - log.info("There is no action configured for alarm %s.", alarm_id) - - def _handle_instantiated_or_scaled(self, content): + async def _handle_alarm_notification(self, content): + log.debug("_handle_alarm_notification: %s", content) + alarm_uuid = content['notify_details']['alarm_uuid'] + status = content['notify_details']['status'] + await self.autoscaling_service.handle_alarm(alarm_uuid, status) + await self.alarming_service.handle_alarm(alarm_uuid, status, content) + + async def _handle_instantiated(self, content): + log.debug("_handle_instantiated: %s", content) nslcmop_id = content['nslcmop_id'] nslcmop = self.db_client.get_nslcmop(nslcmop_id) if nslcmop['operationState'] == 'COMPLETED' or nslcmop['operationState'] == 'PARTIALLY_COMPLETED': nsr_id = nslcmop['nsInstanceId'] - log.info("Configuring scaling groups for network service with nsr_id: %s", nsr_id) - self._configure_scaling_groups(nsr_id) + log.info("Configuring nsr_id: %s", nsr_id) + await self.autoscaling_service.configure_scaling_groups(nsr_id) + await self.alarming_service.configure_vnf_alarms(nsr_id) else: log.info( + "Network_service is not in COMPLETED or PARTIALLY_COMPLETED state. " + "Current state is %s. Skipping...", + nslcmop['operationState']) + + async def _handle_scaled(self, content): + log.debug("_handle_scaled: %s", content) + nslcmop_id = content['nslcmop_id'] + nslcmop = self.db_client.get_nslcmop(nslcmop_id) + if nslcmop['operationState'] == 'COMPLETED' or nslcmop['operationState'] == 'PARTIALLY_COMPLETED': + nsr_id = nslcmop['nsInstanceId'] + log.info("Configuring scaled service with nsr_id: %s", nsr_id) + await self.autoscaling_service.configure_scaling_groups(nsr_id) + await self.autoscaling_service.delete_orphaned_alarms(nsr_id) + await self.alarming_service.configure_vnf_alarms(nsr_id) + else: + log.debug( "Network service is not in COMPLETED or PARTIALLY_COMPLETED state. " "Current state is %s. Skipping...", nslcmop['operationState']) - def _configure_scaling_groups(self, nsr_id: str): - # TODO(diazb): Check for alarm creation on exception and clean resources if needed. - with database.db.atomic(): - vnfrs = self.db_client.get_vnfrs(nsr_id) - log.info("Checking %s vnfrs...", len(vnfrs)) - for vnfr in vnfrs: - vnfd = self.db_client.get_vnfd(vnfr['vnfd-id']) - log.info("Looking for vnfd %s", vnfr['vnfd-id']) - scaling_groups = vnfd['scaling-group-descriptor'] - vnf_monitoring_params = vnfd['monitoring-param'] - for scaling_group in scaling_groups: - try: - scaling_group_record = ScalingGroup.select().where( - ScalingGroup.nsr_id == nsr_id, - ScalingGroup.name == scaling_group['name'] - ).get() - except ScalingGroup.DoesNotExist: - log.info("Creating scaling group record in DB...") - scaling_group_record = ScalingGroup.create( - nsr_id=nsr_id, - name=scaling_group['name'], - content=json.dumps(scaling_group) - ) - log.info("Created scaling group record in DB : nsr_id=%s, name=%s, content=%s", - scaling_group_record.nsr_id, - scaling_group_record.name, - scaling_group_record.content) - for scaling_policy in scaling_group['scaling-policy']: - if scaling_policy['scaling-type'] != 'automatic': - continue - try: - scaling_policy_record = ScalingPolicy.select().join(ScalingGroup).where( - ScalingPolicy.name == scaling_policy['name'], - ScalingGroup.id == scaling_group_record.id - ).get() - except ScalingPolicy.DoesNotExist: - log.info("Creating scaling policy record in DB...") - scaling_policy_record = ScalingPolicy.create( - nsr_id=nsr_id, - name=scaling_policy['name'], - scaling_group=scaling_group_record - ) - log.info("Created scaling policy record in DB : name=%s, scaling_group.name=%s", - scaling_policy_record.nsr_id, - scaling_policy_record.scaling_group.name) - for vdu in vnfd['vdu']: - vdu_monitoring_params = vdu['monitoring-param'] - for scaling_criteria in scaling_policy['scaling-criteria']: - try: - scaling_criteria_record = ScalingCriteria.select().join(ScalingPolicy).where( - ScalingPolicy.id == scaling_policy_record.id, - ScalingCriteria.name == scaling_criteria['name'] - ).get() - except ScalingCriteria.DoesNotExist: - log.info("Creating scaling criteria record in DB...") - scaling_criteria_record = ScalingCriteria.create( - nsr_id=nsr_id, - name=scaling_policy['name'], - scaling_policy=scaling_policy_record - ) - log.info( - "Created scaling criteria record in DB : name=%s, scaling_criteria.name=%s", - scaling_criteria_record.name, - scaling_criteria_record.scaling_policy.name) - vnf_monitoring_param = next( - filter(lambda param: param['id'] == scaling_criteria['vnf-monitoring-param-ref'], - vnf_monitoring_params)) - # TODO: Add support for non-nfvi metrics - vdu_monitoring_param = next( - filter( - lambda param: param['id'] == vnf_monitoring_param['vdu-monitoring-param-ref'], - vdu_monitoring_params)) - vdurs = list(filter(lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param['vdu-ref'], - vnfr['vdur'])) - for vdur in vdurs: - try: - ScalingAlarm.select().where( - ScalingAlarm.vdu_name == vdur['name'] - ).where( - ScalingAlarm.scaling_criteria.name == scaling_criteria['name'] - ).get() - log.debug("VDU %s already has an alarm configured") - continue - except ScalingAlarm.DoesNotExist: - pass - alarm_uuid = self.mon_client.create_alarm( - metric_name=vdu_monitoring_param['nfvi-metric'], - ns_id=nsr_id, - vdu_name=vdur['name'], - vnf_member_index=vnfr['member-vnf-index-ref'], - threshold=scaling_criteria['scale-in-threshold'], - operation=scaling_criteria['scale-in-relational-operation'], - statistic=vnf_monitoring_param['aggregation-type'] - ) - ScalingAlarm.create( - alarm_id=alarm_uuid, - action='scale_in', - vnf_member_index=int(vnfr['member-vnf-index-ref']), - vdu_name=vdur['name'], - scaling_criteria=scaling_criteria_record - ) - alarm_uuid = self.mon_client.create_alarm( - metric_name=vdu_monitoring_param['nfvi-metric'], - ns_id=nsr_id, - vdu_name=vdur['name'], - vnf_member_index=vnfr['member-vnf-index-ref'], - threshold=scaling_criteria['scale-out-threshold'], - operation=scaling_criteria['scale-out-relational-operation'], - statistic=vnf_monitoring_param['aggregation-type'] - ) - ScalingAlarm.create( - alarm_id=alarm_uuid, - action='scale_out', - vnf_member_index=int(vnfr['member-vnf-index-ref']), - vdu_name=vdur['name'], - scaling_criteria=scaling_criteria_record - ) + async def _handle_terminated(self, content): + log.debug("_handle_deleted: %s", content) + nsr_id = content['nsr_id'] + if content['operationState'] == 'COMPLETED' or content['operationState'] == 'PARTIALLY_COMPLETED': + log.info("Deleting scaling groups and alarms for network autoscaling_service with nsr_id: %s", nsr_id) + await self.autoscaling_service.delete_scaling_groups(nsr_id) + await self.alarming_service.delete_vnf_alarms(nsr_id) + else: + log.info( + "Network service is not in COMPLETED or PARTIALLY_COMPLETED state. " + "Current state is %s. Skipping...", + content['operationState'])