X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=osm_policy_module%2Fautoscaling%2Fservice.py;h=09b78b288888beac1ef782f9171045f2c3561cbd;hb=f7451f8879d188c9fef9c976d82aa39141601938;hp=e1f0704748cba284a7283dfd5e5951ede98e590b;hpb=fbb5d6d5882461137bf47c313ead6bb913f15e56;p=osm%2FPOL.git diff --git a/osm_policy_module/autoscaling/service.py b/osm_policy_module/autoscaling/service.py index e1f0704..09b78b2 100644 --- a/osm_policy_module/autoscaling/service.py +++ b/osm_policy_module/autoscaling/service.py @@ -31,14 +31,15 @@ from osm_policy_module.common.lcm_client import LcmClient from osm_policy_module.common.mon_client import MonClient from osm_policy_module.core import database from osm_policy_module.core.config import Config -from osm_policy_module.core.database import ScalingGroup, ScalingAlarm, ScalingPolicy, ScalingCriteria +from osm_policy_module.core.database import ScalingGroup, ScalingAlarm, ScalingPolicy, ScalingCriteria, \ + ScalingAlarmRepository, ScalingGroupRepository, ScalingPolicyRepository, ScalingCriteriaRepository from osm_policy_module.core.exceptions import VdurNotFound from osm_policy_module.utils.vnfd import VnfdUtils log = logging.getLogger(__name__) -class Service: +class AutoscalingService: def __init__(self, config: Config, loop=None): self.conf = config @@ -50,223 +51,206 @@ class Service: self.lcm_client = LcmClient(config, loop=self.loop) async def configure_scaling_groups(self, nsr_id: str): - log.debug("_configure_scaling_groups: %s", nsr_id) + log.info("Configuring scaling groups for network service with nsr_id: %s", + nsr_id) alarms_created = [] database.db.connect() - with database.db.atomic() as tx: - try: - vnfrs = self.db_client.get_vnfrs(nsr_id) - for vnfr in vnfrs: - log.info("Processing vnfr: %s", vnfr) - vnfd = self.db_client.get_vnfd(vnfr['vnfd-id']) - log.info("Looking for vnfd %s", vnfr['vnfd-id']) - if 'scaling-group-descriptor' not in vnfd: - continue - scaling_groups = vnfd['scaling-group-descriptor'] - vnf_monitoring_params = vnfd['monitoring-param'] - for scaling_group in scaling_groups: - try: - scaling_group_record = ScalingGroup.select().where( - ScalingGroup.nsr_id == nsr_id, - ScalingGroup.vnf_member_index == int(vnfr['member-vnf-index-ref']), - ScalingGroup.name == scaling_group['name'] - ).get() - log.info("Found existing scaling group record in DB...") - except ScalingGroup.DoesNotExist: - log.info("Creating scaling group record in DB...") - scaling_group_record = ScalingGroup.create( - nsr_id=nsr_id, - vnf_member_index=vnfr['member-vnf-index-ref'], - name=scaling_group['name'], - content=json.dumps(scaling_group) - ) - log.info( - "Created scaling group record in DB : nsr_id=%s, vnf_member_index=%s, name=%s", - scaling_group_record.nsr_id, - scaling_group_record.vnf_member_index, - scaling_group_record.name) - for scaling_policy in scaling_group['scaling-policy']: - if scaling_policy['scaling-type'] != 'automatic': - continue + try: + with database.db.atomic() as tx: + try: + vnfrs = self.db_client.get_vnfrs(nsr_id) + for vnfr in vnfrs: + log.debug("Processing vnfr: %s", vnfr) + vnfd = self.db_client.get_vnfd(vnfr['vnfd-id']) + if 'scaling-group-descriptor' not in vnfd: + log.debug("No scaling group present in vnfd") + continue + scaling_groups = vnfd['scaling-group-descriptor'] + vnf_monitoring_params = vnfd['monitoring-param'] + for scaling_group in scaling_groups: try: - scaling_policy_record = ScalingPolicy.select().join(ScalingGroup).where( - ScalingPolicy.name == scaling_policy['name'], - ScalingGroup.id == scaling_group_record.id - ).get() - log.info("Found existing scaling policy record in DB...") - except ScalingPolicy.DoesNotExist: - log.info("Creating scaling policy record in DB...") - scaling_policy_record = ScalingPolicy.create( + scaling_group_record = ScalingGroupRepository.get( + ScalingGroup.nsr_id == nsr_id, + ScalingGroup.vnf_member_index == int(vnfr['member-vnf-index-ref']), + ScalingGroup.name == scaling_group['name'] + ) + log.debug("Found existing scaling group record in DB...") + except ScalingGroup.DoesNotExist: + log.debug("Creating scaling group record in DB...") + scaling_group_record = ScalingGroupRepository.create( nsr_id=nsr_id, - name=scaling_policy['name'], - cooldown_time=scaling_policy['cooldown-time'], - scaling_group=scaling_group_record + vnf_member_index=vnfr['member-vnf-index-ref'], + name=scaling_group['name'], + content=json.dumps(scaling_group) ) - log.info("Created scaling policy record in DB : name=%s, scaling_group.name=%s", - scaling_policy_record.name, - scaling_policy_record.scaling_group.name) - - for scaling_criteria in scaling_policy['scaling-criteria']: + log.debug( + "Created scaling group record in DB : nsr_id=%s, vnf_member_index=%s, name=%s", + scaling_group_record.nsr_id, + scaling_group_record.vnf_member_index, + scaling_group_record.name) + for scaling_policy in scaling_group['scaling-policy']: + if scaling_policy['scaling-type'] != 'automatic': + continue try: - scaling_criteria_record = ScalingCriteria.select().join(ScalingPolicy).where( - ScalingPolicy.id == scaling_policy_record.id, - ScalingCriteria.name == scaling_criteria['name'] - ).get() - log.info("Found existing scaling criteria record in DB...") - except ScalingCriteria.DoesNotExist: - log.info("Creating scaling criteria record in DB...") - scaling_criteria_record = ScalingCriteria.create( + scaling_policy_record = ScalingPolicyRepository.get( + ScalingPolicy.name == scaling_policy['name'], + ScalingGroup.id == scaling_group_record.id, + join_classes=[ScalingGroup] + ) + log.debug("Found existing scaling policy record in DB...") + except ScalingPolicy.DoesNotExist: + log.debug("Creating scaling policy record in DB...") + scaling_policy_record = ScalingPolicyRepository.create( nsr_id=nsr_id, - name=scaling_criteria['name'], - scaling_policy=scaling_policy_record + name=scaling_policy['name'], + cooldown_time=scaling_policy['cooldown-time'], + scaling_group=scaling_group_record, ) - log.info( - "Created scaling criteria record in DB : name=%s, scaling_policy.name=%s", - scaling_criteria_record.name, - scaling_criteria_record.scaling_policy.name) + if 'scale-in-operation-type' in scaling_policy: + scaling_policy_record.scale_in_operation = scaling_policy[ + 'scale-in-operation-type'] + if 'scale-out-operation-type' in scaling_policy: + scaling_policy_record.scale_out_operation = scaling_policy[ + 'scale-out-operation-type'] + if 'enabled' in scaling_policy: + scaling_policy_record.enabled = scaling_policy['enabled'] + scaling_policy_record.save() + log.debug("Created scaling policy record in DB : name=%s, scaling_group.name=%s", + scaling_policy_record.name, + scaling_policy_record.scaling_group.name) - vnf_monitoring_param = next( - filter( - lambda param: param['id'] == scaling_criteria[ - 'vnf-monitoring-param-ref' - ], - vnf_monitoring_params) - ) - if 'vdu-monitoring-param' in vnf_monitoring_param: - vdurs = list( - filter( - lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param - ['vdu-monitoring-param'] - ['vdu-ref'], - vnfr['vdur'] + for scaling_criteria in scaling_policy['scaling-criteria']: + try: + scaling_criteria_record = ScalingCriteriaRepository.get( + ScalingPolicy.id == scaling_policy_record.id, + ScalingCriteria.name == scaling_criteria['name'], + join_classes=[ScalingPolicy] ) - ) - elif 'vdu-metric' in vnf_monitoring_param: - vdurs = list( - filter( - lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param - ['vdu-metric'] - ['vdu-ref'], - vnfr['vdur'] + log.debug("Found existing scaling criteria record in DB...") + except ScalingCriteria.DoesNotExist: + log.debug("Creating scaling criteria record in DB...") + scaling_criteria_record = ScalingCriteriaRepository.create( + nsr_id=nsr_id, + name=scaling_criteria['name'], + scaling_policy=scaling_policy_record ) - ) - elif 'vnf-metric' in vnf_monitoring_param: - vdu = VnfdUtils.get_mgmt_vdu(vnfd) - vdurs = list( + log.debug( + "Created scaling criteria record in DB : name=%s, scaling_policy.name=%s", + scaling_criteria_record.name, + scaling_criteria_record.scaling_policy.name) + + vnf_monitoring_param = next( filter( - lambda vdur: vdur['vdu-id-ref'] == vdu['id'], - vnfr['vdur'] - ) + lambda param: param['id'] == scaling_criteria[ + 'vnf-monitoring-param-ref' + ], + vnf_monitoring_params) ) - else: - log.warning( - "Scaling criteria is referring to a vnf-monitoring-param that does not " - "contain a reference to a vdu or vnf metric.") - continue - for vdur in vdurs: - log.info("Creating alarm for vdur %s ", vdur) - try: - (ScalingAlarm.select() - .join(ScalingCriteria) - .join(ScalingPolicy) - .join(ScalingGroup) - .where( - ScalingAlarm.vdu_name == vdur['name'], - ScalingCriteria.name == scaling_criteria['name'], - ScalingPolicy.name == scaling_policy['name'], - ScalingGroup.nsr_id == nsr_id - ).get()) - log.debug("vdu %s already has an alarm configured", vdur['name']) + if 'vdu-monitoring-param' in vnf_monitoring_param: + vdurs = list( + filter( + lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param + ['vdu-monitoring-param'] + ['vdu-ref'], + vnfr['vdur'] + ) + ) + elif 'vdu-metric' in vnf_monitoring_param: + vdurs = list( + filter( + lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param + ['vdu-metric'] + ['vdu-ref'], + vnfr['vdur'] + ) + ) + elif 'vnf-metric' in vnf_monitoring_param: + vdu = VnfdUtils.get_mgmt_vdu(vnfd) + vdurs = list( + filter( + lambda vdur: vdur['vdu-id-ref'] == vdu['id'], + vnfr['vdur'] + ) + ) + else: + log.warning( + "Scaling criteria is referring to a vnf-monitoring-param that does not " + "contain a reference to a vdu or vnf metric.") continue - except ScalingAlarm.DoesNotExist: - pass - alarm_uuid = await self.mon_client.create_alarm( - metric_name=vnf_monitoring_param['id'], - ns_id=nsr_id, - vdu_name=vdur['name'], - vnf_member_index=vnfr['member-vnf-index-ref'], - threshold=scaling_criteria['scale-in-threshold'], - operation=scaling_criteria['scale-in-relational-operation'], - statistic=vnf_monitoring_param['aggregation-type'] - ) - alarm = ScalingAlarm.create( - alarm_uuid=alarm_uuid, - action='scale_in', - vnf_member_index=int(vnfr['member-vnf-index-ref']), - vdu_name=vdur['name'], - scaling_criteria=scaling_criteria_record - ) - alarms_created.append(alarm) - alarm_uuid = await self.mon_client.create_alarm( - metric_name=vnf_monitoring_param['id'], - ns_id=nsr_id, - vdu_name=vdur['name'], - vnf_member_index=vnfr['member-vnf-index-ref'], - threshold=scaling_criteria['scale-out-threshold'], - operation=scaling_criteria['scale-out-relational-operation'], - statistic=vnf_monitoring_param['aggregation-type'] - ) - alarm = ScalingAlarm.create( - alarm_uuid=alarm_uuid, - action='scale_out', - vnf_member_index=int(vnfr['member-vnf-index-ref']), - vdu_name=vdur['name'], - scaling_criteria=scaling_criteria_record - ) - alarms_created.append(alarm) + for vdur in vdurs: + log.debug("Creating alarm for vdur %s ", vdur) + try: + ScalingAlarmRepository.get(ScalingAlarm.vdu_name == vdur['name'], + ScalingCriteria.name == scaling_criteria['name'], + ScalingPolicy.name == scaling_policy['name'], + ScalingGroup.nsr_id == nsr_id, + join_classes=[ScalingCriteria, + ScalingPolicy, + ScalingGroup]) + log.debug("vdu %s already has an alarm configured", vdur['name']) + continue + except ScalingAlarm.DoesNotExist: + pass + alarm_uuid = await self.mon_client.create_alarm( + metric_name=vnf_monitoring_param['id'], + ns_id=nsr_id, + vdu_name=vdur['name'], + vnf_member_index=vnfr['member-vnf-index-ref'], + threshold=scaling_criteria['scale-in-threshold'], + operation=scaling_criteria['scale-in-relational-operation'], + statistic=vnf_monitoring_param['aggregation-type'] + ) + alarm = ScalingAlarmRepository.create( + alarm_uuid=alarm_uuid, + action='scale_in', + vnf_member_index=int(vnfr['member-vnf-index-ref']), + vdu_name=vdur['name'], + scaling_criteria=scaling_criteria_record + ) + alarms_created.append(alarm) + alarm_uuid = await self.mon_client.create_alarm( + metric_name=vnf_monitoring_param['id'], + ns_id=nsr_id, + vdu_name=vdur['name'], + vnf_member_index=vnfr['member-vnf-index-ref'], + threshold=scaling_criteria['scale-out-threshold'], + operation=scaling_criteria['scale-out-relational-operation'], + statistic=vnf_monitoring_param['aggregation-type'] + ) + alarm = ScalingAlarmRepository.create( + alarm_uuid=alarm_uuid, + action='scale_out', + vnf_member_index=int(vnfr['member-vnf-index-ref']), + vdu_name=vdur['name'], + scaling_criteria=scaling_criteria_record + ) + alarms_created.append(alarm) - except Exception as e: - log.exception("Error configuring scaling groups:") - tx.rollback() - if len(alarms_created) > 0: - log.info("Cleaning alarm resources in MON") - for alarm in alarms_created: - await self.mon_client.delete_alarm(alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, - alarm.vnf_member_index, - alarm.vdu_name, - alarm.alarm_uuid) - raise e - database.db.close() + except Exception as e: + log.exception("Error configuring scaling groups:") + tx.rollback() + if len(alarms_created) > 0: + log.info("Cleaning alarm resources in MON") + for alarm in alarms_created: + await self.mon_client.delete_alarm( + alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, + alarm.vnf_member_index, + alarm.vdu_name, + alarm.alarm_uuid) + raise e + finally: + database.db.close() async def delete_scaling_groups(self, nsr_id: str): + log.debug("Deleting scaling groups for network service %s", nsr_id) database.db.connect() - with database.db.atomic() as tx: - try: - for scaling_group in ScalingGroup.select().where(ScalingGroup.nsr_id == nsr_id): - for scaling_policy in scaling_group.scaling_policies: - for scaling_criteria in scaling_policy.scaling_criterias: - for alarm in scaling_criteria.scaling_alarms: - try: - await self.mon_client.delete_alarm( - alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, - alarm.vnf_member_index, - alarm.vdu_name, - alarm.alarm_uuid) - except ValueError: - log.exception("Error deleting alarm in MON %s", alarm.alarm_uuid) - alarm.delete_instance() - scaling_criteria.delete_instance() - scaling_policy.delete_instance() - scaling_group.delete_instance() - - except Exception as e: - log.exception("Error deleting scaling groups and alarms:") - tx.rollback() - raise e - database.db.close() - - async def delete_orphaned_alarms(self, nsr_id): - database.db.connect() - with database.db.atomic() as tx: - try: - for scaling_group in ScalingGroup.select().where(ScalingGroup.nsr_id == nsr_id): - for scaling_policy in scaling_group.scaling_policies: - for scaling_criteria in scaling_policy.scaling_criterias: - for alarm in scaling_criteria.scaling_alarms: - try: - self.db_client.get_vdur(nsr_id, alarm.vnf_member_index, alarm.vdu_name) - except VdurNotFound: - log.info("Deleting orphaned alarm %s", alarm.alarm_uuid) + try: + with database.db.atomic() as tx: + try: + for scaling_group in ScalingGroupRepository.list(ScalingGroup.nsr_id == nsr_id): + for scaling_policy in scaling_group.scaling_policies: + for scaling_criteria in scaling_policy.scaling_criterias: + for alarm in scaling_criteria.scaling_alarms: try: await self.mon_client.delete_alarm( alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, @@ -276,36 +260,104 @@ class Service: except ValueError: log.exception("Error deleting alarm in MON %s", alarm.alarm_uuid) alarm.delete_instance() + scaling_criteria.delete_instance() + scaling_policy.delete_instance() + scaling_group.delete_instance() - except Exception as e: - log.exception("Error deleting orphaned alarms:") - tx.rollback() - raise e - database.db.close() + except Exception as e: + log.exception("Error deleting scaling groups and alarms:") + tx.rollback() + raise e + finally: + database.db.close() - async def scale(self, alarm): + async def delete_orphaned_alarms(self, nsr_id): + log.info("Deleting orphaned scaling alarms for network service %s", nsr_id) database.db.connect() - with database.db.atomic(): - delta = datetime.datetime.now() - alarm.scaling_criteria.scaling_policy.last_scale - if delta.total_seconds() > alarm.scaling_criteria.scaling_policy.cooldown_time: - log.info("Sending scaling action message for ns: %s", - alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id) - await self.lcm_client.scale(alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, - alarm.scaling_criteria.scaling_policy.scaling_group.name, - alarm.vnf_member_index, - alarm.action) - alarm.scaling_criteria.scaling_policy.last_scale = datetime.datetime.now() - alarm.scaling_criteria.scaling_policy.save() - else: - log.info("Time between last scale and now is less than cooldown time. Skipping.") - database.db.close() + try: + with database.db.atomic() as tx: + try: + for scaling_group in ScalingGroupRepository.list(ScalingGroup.nsr_id == nsr_id): + for scaling_policy in scaling_group.scaling_policies: + for scaling_criteria in scaling_policy.scaling_criterias: + for alarm in scaling_criteria.scaling_alarms: + try: + self.db_client.get_vdur(nsr_id, alarm.vnf_member_index, alarm.vdu_name) + except VdurNotFound: + log.debug("Deleting orphaned scaling alarm %s", alarm.alarm_uuid) + try: + await self.mon_client.delete_alarm( + alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, + alarm.vnf_member_index, + alarm.vdu_name, + alarm.alarm_uuid) + except ValueError: + log.exception("Error deleting alarm in MON %s", alarm.alarm_uuid) + alarm.delete_instance() - def get_alarm(self, alarm_uuid: str): - database.db.connect() - with database.db.atomic(): - alarm = ScalingAlarm.select().where(ScalingAlarm.alarm_uuid == alarm_uuid).get() - database.db.close() - return alarm + except Exception as e: + log.exception("Error deleting orphaned alarms:") + tx.rollback() + raise e + finally: + database.db.close() def get_nslcmop(self, nslcmop_id): return self.db_client.get_nslcmop(nslcmop_id) + + async def handle_alarm(self, alarm_uuid: str, status: str): + await self.update_alarm_status(alarm_uuid, status) + await self.evaluate_policy(alarm_uuid) + + async def update_alarm_status(self, alarm_uuid: str, status: str): + database.db.connect() + try: + with database.db.atomic(): + alarm = ScalingAlarmRepository.get(ScalingAlarm.alarm_uuid == alarm_uuid) + alarm.last_status = status + alarm.save() + except ScalingAlarm.DoesNotExist: + log.debug("There is no autoscaling action configured for alarm %s.", alarm_uuid) + finally: + database.db.close() + + async def evaluate_policy(self, alarm_uuid): + database.db.connect() + try: + with database.db.atomic(): + alarm = ScalingAlarmRepository.get(ScalingAlarm.alarm_uuid == alarm_uuid) + vnf_member_index = alarm.vnf_member_index + action = alarm.action + scaling_policy = alarm.scaling_criteria.scaling_policy + if not scaling_policy.enabled: + return + if action == 'scale_in': + operation = scaling_policy.scale_in_operation + elif action == 'scale_out': + operation = scaling_policy.scale_out_operation + else: + raise Exception('Unknown alarm action {}'.format(alarm.action)) + alarms = ScalingAlarmRepository.list(ScalingAlarm.scaling_criteria == alarm.scaling_criteria, + ScalingAlarm.action == alarm.action, + ScalingAlarm.vnf_member_index == vnf_member_index, + ScalingAlarm.vdu_name == alarm.vdu_name) + statuses = [] + for alarm in alarms: + statuses.append(alarm.last_status) + if (operation == 'AND' and set(statuses) == {'alarm'}) or (operation == 'OR' and 'alarm' in statuses): + delta = datetime.datetime.now() - scaling_policy.last_scale + if delta.total_seconds() > scaling_policy.cooldown_time: + log.info("Sending %s action message for ns: %s", + alarm.action, + scaling_policy.scaling_group.nsr_id) + await self.lcm_client.scale(scaling_policy.scaling_group.nsr_id, + scaling_policy.scaling_group.name, + vnf_member_index, + action) + scaling_policy.last_scale = datetime.datetime.now() + scaling_policy.save() + + except ScalingAlarm.DoesNotExist: + log.debug("There is no autoscaling action configured for alarm %s.", alarm_uuid) + finally: + database.db.close()