From: Benjamin Diaz Date: Mon, 4 Mar 2019 16:59:53 +0000 (-0300) Subject: Refactors autoscaling module to separate concerns and allow better control of db... X-Git-Tag: v6.0.0~10 X-Git-Url: https://osm.etsi.org/gitweb/?a=commitdiff_plain;h=refs%2Fchanges%2F78%2F7378%2F1;p=osm%2FPOL.git Refactors autoscaling module to separate concerns and allow better control of db connections Change-Id: I57034d34ae2fbb3c4aeda784f9542e643d348460 Signed-off-by: Benjamin Diaz --- diff --git a/osm_policy_module/autoscaling/__init__.py b/osm_policy_module/autoscaling/__init__.py new file mode 100644 index 0000000..e69de29 diff --git a/osm_policy_module/autoscaling/agent.py b/osm_policy_module/autoscaling/agent.py new file mode 100644 index 0000000..bc12dc7 --- /dev/null +++ b/osm_policy_module/autoscaling/agent.py @@ -0,0 +1,148 @@ +# -*- coding: utf-8 -*- + +# Copyright 2018 Whitestack, LLC +# ************************************************************* + +# This file is part of OSM Monitoring module +# All Rights Reserved to Whitestack, LLC + +# Licensed under the Apache License, Version 2.0 (the "License"); you may +# not use this file except in compliance with the License. You may obtain +# a copy of the License at + +# http://www.apache.org/licenses/LICENSE-2.0 + +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT +# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the +# License for the specific language governing permissions and limitations +# under the License. + +# For those usages not covered by the Apache License, Version 2.0 please +# contact: bdiaz@whitestack.com or glavado@whitestack.com +## +import asyncio +import logging + +import peewee + +from osm_policy_module.autoscaling.service import Service +from osm_policy_module.common.message_bus_client import MessageBusClient +from osm_policy_module.core.config import Config +from osm_policy_module.core.database import ScalingAlarm + +log = logging.getLogger(__name__) + +ALLOWED_KAFKA_KEYS = ['instantiated', 'scaled', 'terminated', 'notify_alarm'] + + +class PolicyModuleAgent: + def __init__(self, config: Config, loop=None): + self.conf = config + if not loop: + loop = asyncio.get_event_loop() + self.loop = loop + self.msg_bus = MessageBusClient(config) + self.service = Service(config, loop) + + def run(self): + self.loop.run_until_complete(self.start()) + + async def start(self): + topics = [ + "ns", + "alarm_response" + ] + await self.msg_bus.aioread(topics, self._process_msg) + log.critical("Exiting...") + + async def _process_msg(self, topic, key, msg): + log.debug("_process_msg topic=%s key=%s msg=%s", topic, key, msg) + log.info("Message arrived: %s", msg) + try: + if key in ALLOWED_KAFKA_KEYS: + + if key == 'instantiated': + await self._handle_instantiated(msg) + + if key == 'scaled': + await self._handle_scaled(msg) + + if key == 'terminated': + await self._handle_terminated(msg) + + if key == 'notify_alarm': + await self._handle_alarm_notification(msg) + else: + log.debug("Key %s is not in ALLOWED_KAFKA_KEYS", key) + except peewee.PeeweeException: + log.exception("Database error consuming message: ") + raise + except Exception: + log.exception("Error consuming message: ") + + async def _handle_alarm_notification(self, content): + log.debug("_handle_alarm_notification: %s", content) + alarm_uuid = content['notify_details']['alarm_uuid'] + metric_name = content['notify_details']['metric_name'] + operation = content['notify_details']['operation'] + threshold = content['notify_details']['threshold_value'] + vdu_name = content['notify_details']['vdu_name'] + vnf_member_index = content['notify_details']['vnf_member_index'] + nsr_id = content['notify_details']['ns_id'] + log.info( + "Received alarm notification for alarm %s, \ + metric %s, \ + operation %s, \ + threshold %s, \ + vdu_name %s, \ + vnf_member_index %s, \ + ns_id %s ", + alarm_uuid, metric_name, operation, threshold, vdu_name, vnf_member_index, nsr_id) + try: + alarm = self.service.get_alarm(alarm_uuid) + await self.service.scale(alarm) + except ScalingAlarm.DoesNotExist: + log.info("There is no action configured for alarm %s.", alarm_uuid) + + async def _handle_instantiated(self, content): + log.debug("_handle_instantiated: %s", content) + nslcmop_id = content['nslcmop_id'] + nslcmop = self.service.get_nslcmop(nslcmop_id) + if nslcmop['operationState'] == 'COMPLETED' or nslcmop['operationState'] == 'PARTIALLY_COMPLETED': + nsr_id = nslcmop['nsInstanceId'] + log.info("Configuring scaling groups for network service with nsr_id: %s", nsr_id) + await self.service.configure_scaling_groups(nsr_id) + else: + log.info( + "Network service is not in COMPLETED or PARTIALLY_COMPLETED state. " + "Current state is %s. Skipping...", + nslcmop['operationState']) + + async def _handle_scaled(self, content): + log.debug("_handle_scaled: %s", content) + nslcmop_id = content['nslcmop_id'] + nslcmop = self.service.get_nslcmop(nslcmop_id) + if nslcmop['operationState'] == 'COMPLETED' or nslcmop['operationState'] == 'PARTIALLY_COMPLETED': + nsr_id = nslcmop['nsInstanceId'] + log.info("Configuring scaling groups for network service with nsr_id: %s", nsr_id) + await self.service.configure_scaling_groups(nsr_id) + log.info("Checking for orphaned alarms to be deleted for network service with nsr_id: %s", nsr_id) + await self.service.delete_orphaned_alarms(nsr_id) + else: + log.info( + "Network service is not in COMPLETED or PARTIALLY_COMPLETED state. " + "Current state is %s. Skipping...", + nslcmop['operationState']) + + async def _handle_terminated(self, content): + log.debug("_handle_deleted: %s", content) + nsr_id = content['nsr_id'] + if content['operationState'] == 'COMPLETED' or content['operationState'] == 'PARTIALLY_COMPLETED': + log.info("Deleting scaling groups and alarms for network service with nsr_id: %s", nsr_id) + await self.service.delete_scaling_groups(nsr_id) + else: + log.info( + "Network service is not in COMPLETED or PARTIALLY_COMPLETED state. " + "Current state is %s. Skipping...", + content['operationState']) diff --git a/osm_policy_module/autoscaling/service.py b/osm_policy_module/autoscaling/service.py new file mode 100644 index 0000000..e1f0704 --- /dev/null +++ b/osm_policy_module/autoscaling/service.py @@ -0,0 +1,311 @@ +# -*- coding: utf-8 -*- + +# Copyright 2018 Whitestack, LLC +# ************************************************************* + +# This file is part of OSM Monitoring module +# All Rights Reserved to Whitestack, LLC + +# Licensed under the Apache License, Version 2.0 (the "License"); you may +# not use this file except in compliance with the License. You may obtain +# a copy of the License at + +# http://www.apache.org/licenses/LICENSE-2.0 + +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT +# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the +# License for the specific language governing permissions and limitations +# under the License. + +# For those usages not covered by the Apache License, Version 2.0 please +# contact: bdiaz@whitestack.com or glavado@whitestack.com +## +import asyncio +import datetime +import json +import logging + +from osm_policy_module.common.common_db_client import CommonDbClient +from osm_policy_module.common.lcm_client import LcmClient +from osm_policy_module.common.mon_client import MonClient +from osm_policy_module.core import database +from osm_policy_module.core.config import Config +from osm_policy_module.core.database import ScalingGroup, ScalingAlarm, ScalingPolicy, ScalingCriteria +from osm_policy_module.core.exceptions import VdurNotFound +from osm_policy_module.utils.vnfd import VnfdUtils + +log = logging.getLogger(__name__) + + +class Service: + + def __init__(self, config: Config, loop=None): + self.conf = config + if not loop: + loop = asyncio.get_event_loop() + self.loop = loop + self.db_client = CommonDbClient(config) + self.mon_client = MonClient(config, loop=self.loop) + self.lcm_client = LcmClient(config, loop=self.loop) + + async def configure_scaling_groups(self, nsr_id: str): + log.debug("_configure_scaling_groups: %s", nsr_id) + alarms_created = [] + database.db.connect() + with database.db.atomic() as tx: + try: + vnfrs = self.db_client.get_vnfrs(nsr_id) + for vnfr in vnfrs: + log.info("Processing vnfr: %s", vnfr) + vnfd = self.db_client.get_vnfd(vnfr['vnfd-id']) + log.info("Looking for vnfd %s", vnfr['vnfd-id']) + if 'scaling-group-descriptor' not in vnfd: + continue + scaling_groups = vnfd['scaling-group-descriptor'] + vnf_monitoring_params = vnfd['monitoring-param'] + for scaling_group in scaling_groups: + try: + scaling_group_record = ScalingGroup.select().where( + ScalingGroup.nsr_id == nsr_id, + ScalingGroup.vnf_member_index == int(vnfr['member-vnf-index-ref']), + ScalingGroup.name == scaling_group['name'] + ).get() + log.info("Found existing scaling group record in DB...") + except ScalingGroup.DoesNotExist: + log.info("Creating scaling group record in DB...") + scaling_group_record = ScalingGroup.create( + nsr_id=nsr_id, + vnf_member_index=vnfr['member-vnf-index-ref'], + name=scaling_group['name'], + content=json.dumps(scaling_group) + ) + log.info( + "Created scaling group record in DB : nsr_id=%s, vnf_member_index=%s, name=%s", + scaling_group_record.nsr_id, + scaling_group_record.vnf_member_index, + scaling_group_record.name) + for scaling_policy in scaling_group['scaling-policy']: + if scaling_policy['scaling-type'] != 'automatic': + continue + try: + scaling_policy_record = ScalingPolicy.select().join(ScalingGroup).where( + ScalingPolicy.name == scaling_policy['name'], + ScalingGroup.id == scaling_group_record.id + ).get() + log.info("Found existing scaling policy record in DB...") + except ScalingPolicy.DoesNotExist: + log.info("Creating scaling policy record in DB...") + scaling_policy_record = ScalingPolicy.create( + nsr_id=nsr_id, + name=scaling_policy['name'], + cooldown_time=scaling_policy['cooldown-time'], + scaling_group=scaling_group_record + ) + log.info("Created scaling policy record in DB : name=%s, scaling_group.name=%s", + scaling_policy_record.name, + scaling_policy_record.scaling_group.name) + + for scaling_criteria in scaling_policy['scaling-criteria']: + try: + scaling_criteria_record = ScalingCriteria.select().join(ScalingPolicy).where( + ScalingPolicy.id == scaling_policy_record.id, + ScalingCriteria.name == scaling_criteria['name'] + ).get() + log.info("Found existing scaling criteria record in DB...") + except ScalingCriteria.DoesNotExist: + log.info("Creating scaling criteria record in DB...") + scaling_criteria_record = ScalingCriteria.create( + nsr_id=nsr_id, + name=scaling_criteria['name'], + scaling_policy=scaling_policy_record + ) + log.info( + "Created scaling criteria record in DB : name=%s, scaling_policy.name=%s", + scaling_criteria_record.name, + scaling_criteria_record.scaling_policy.name) + + vnf_monitoring_param = next( + filter( + lambda param: param['id'] == scaling_criteria[ + 'vnf-monitoring-param-ref' + ], + vnf_monitoring_params) + ) + if 'vdu-monitoring-param' in vnf_monitoring_param: + vdurs = list( + filter( + lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param + ['vdu-monitoring-param'] + ['vdu-ref'], + vnfr['vdur'] + ) + ) + elif 'vdu-metric' in vnf_monitoring_param: + vdurs = list( + filter( + lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param + ['vdu-metric'] + ['vdu-ref'], + vnfr['vdur'] + ) + ) + elif 'vnf-metric' in vnf_monitoring_param: + vdu = VnfdUtils.get_mgmt_vdu(vnfd) + vdurs = list( + filter( + lambda vdur: vdur['vdu-id-ref'] == vdu['id'], + vnfr['vdur'] + ) + ) + else: + log.warning( + "Scaling criteria is referring to a vnf-monitoring-param that does not " + "contain a reference to a vdu or vnf metric.") + continue + for vdur in vdurs: + log.info("Creating alarm for vdur %s ", vdur) + try: + (ScalingAlarm.select() + .join(ScalingCriteria) + .join(ScalingPolicy) + .join(ScalingGroup) + .where( + ScalingAlarm.vdu_name == vdur['name'], + ScalingCriteria.name == scaling_criteria['name'], + ScalingPolicy.name == scaling_policy['name'], + ScalingGroup.nsr_id == nsr_id + ).get()) + log.debug("vdu %s already has an alarm configured", vdur['name']) + continue + except ScalingAlarm.DoesNotExist: + pass + alarm_uuid = await self.mon_client.create_alarm( + metric_name=vnf_monitoring_param['id'], + ns_id=nsr_id, + vdu_name=vdur['name'], + vnf_member_index=vnfr['member-vnf-index-ref'], + threshold=scaling_criteria['scale-in-threshold'], + operation=scaling_criteria['scale-in-relational-operation'], + statistic=vnf_monitoring_param['aggregation-type'] + ) + alarm = ScalingAlarm.create( + alarm_uuid=alarm_uuid, + action='scale_in', + vnf_member_index=int(vnfr['member-vnf-index-ref']), + vdu_name=vdur['name'], + scaling_criteria=scaling_criteria_record + ) + alarms_created.append(alarm) + alarm_uuid = await self.mon_client.create_alarm( + metric_name=vnf_monitoring_param['id'], + ns_id=nsr_id, + vdu_name=vdur['name'], + vnf_member_index=vnfr['member-vnf-index-ref'], + threshold=scaling_criteria['scale-out-threshold'], + operation=scaling_criteria['scale-out-relational-operation'], + statistic=vnf_monitoring_param['aggregation-type'] + ) + alarm = ScalingAlarm.create( + alarm_uuid=alarm_uuid, + action='scale_out', + vnf_member_index=int(vnfr['member-vnf-index-ref']), + vdu_name=vdur['name'], + scaling_criteria=scaling_criteria_record + ) + alarms_created.append(alarm) + + except Exception as e: + log.exception("Error configuring scaling groups:") + tx.rollback() + if len(alarms_created) > 0: + log.info("Cleaning alarm resources in MON") + for alarm in alarms_created: + await self.mon_client.delete_alarm(alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, + alarm.vnf_member_index, + alarm.vdu_name, + alarm.alarm_uuid) + raise e + database.db.close() + + async def delete_scaling_groups(self, nsr_id: str): + database.db.connect() + with database.db.atomic() as tx: + try: + for scaling_group in ScalingGroup.select().where(ScalingGroup.nsr_id == nsr_id): + for scaling_policy in scaling_group.scaling_policies: + for scaling_criteria in scaling_policy.scaling_criterias: + for alarm in scaling_criteria.scaling_alarms: + try: + await self.mon_client.delete_alarm( + alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, + alarm.vnf_member_index, + alarm.vdu_name, + alarm.alarm_uuid) + except ValueError: + log.exception("Error deleting alarm in MON %s", alarm.alarm_uuid) + alarm.delete_instance() + scaling_criteria.delete_instance() + scaling_policy.delete_instance() + scaling_group.delete_instance() + + except Exception as e: + log.exception("Error deleting scaling groups and alarms:") + tx.rollback() + raise e + database.db.close() + + async def delete_orphaned_alarms(self, nsr_id): + database.db.connect() + with database.db.atomic() as tx: + try: + for scaling_group in ScalingGroup.select().where(ScalingGroup.nsr_id == nsr_id): + for scaling_policy in scaling_group.scaling_policies: + for scaling_criteria in scaling_policy.scaling_criterias: + for alarm in scaling_criteria.scaling_alarms: + try: + self.db_client.get_vdur(nsr_id, alarm.vnf_member_index, alarm.vdu_name) + except VdurNotFound: + log.info("Deleting orphaned alarm %s", alarm.alarm_uuid) + try: + await self.mon_client.delete_alarm( + alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, + alarm.vnf_member_index, + alarm.vdu_name, + alarm.alarm_uuid) + except ValueError: + log.exception("Error deleting alarm in MON %s", alarm.alarm_uuid) + alarm.delete_instance() + + except Exception as e: + log.exception("Error deleting orphaned alarms:") + tx.rollback() + raise e + database.db.close() + + async def scale(self, alarm): + database.db.connect() + with database.db.atomic(): + delta = datetime.datetime.now() - alarm.scaling_criteria.scaling_policy.last_scale + if delta.total_seconds() > alarm.scaling_criteria.scaling_policy.cooldown_time: + log.info("Sending scaling action message for ns: %s", + alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id) + await self.lcm_client.scale(alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, + alarm.scaling_criteria.scaling_policy.scaling_group.name, + alarm.vnf_member_index, + alarm.action) + alarm.scaling_criteria.scaling_policy.last_scale = datetime.datetime.now() + alarm.scaling_criteria.scaling_policy.save() + else: + log.info("Time between last scale and now is less than cooldown time. Skipping.") + database.db.close() + + def get_alarm(self, alarm_uuid: str): + database.db.connect() + with database.db.atomic(): + alarm = ScalingAlarm.select().where(ScalingAlarm.alarm_uuid == alarm_uuid).get() + database.db.close() + return alarm + + def get_nslcmop(self, nslcmop_id): + return self.db_client.get_nslcmop(nslcmop_id) diff --git a/osm_policy_module/cmd/policy_module_agent.py b/osm_policy_module/cmd/policy_module_agent.py index af2f602..19e8054 100644 --- a/osm_policy_module/cmd/policy_module_agent.py +++ b/osm_policy_module/cmd/policy_module_agent.py @@ -26,7 +26,7 @@ import asyncio import logging import sys -from osm_policy_module.core.agent import PolicyModuleAgent +from osm_policy_module.autoscaling.agent import PolicyModuleAgent from osm_policy_module.core.config import Config from osm_policy_module.core.database import DatabaseManager diff --git a/osm_policy_module/core/agent.py b/osm_policy_module/core/agent.py deleted file mode 100644 index 65c741f..0000000 --- a/osm_policy_module/core/agent.py +++ /dev/null @@ -1,399 +0,0 @@ -# -*- coding: utf-8 -*- - -# Copyright 2018 Whitestack, LLC -# ************************************************************* - -# This file is part of OSM Monitoring module -# All Rights Reserved to Whitestack, LLC - -# Licensed under the Apache License, Version 2.0 (the "License"); you may -# not use this file except in compliance with the License. You may obtain -# a copy of the License at - -# http://www.apache.org/licenses/LICENSE-2.0 - -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT -# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the -# License for the specific language governing permissions and limitations -# under the License. - -# For those usages not covered by the Apache License, Version 2.0 please -# contact: bdiaz@whitestack.com or glavado@whitestack.com -## -import asyncio -import datetime -import json -import logging - -import peewee - -from osm_policy_module.common.common_db_client import CommonDbClient -from osm_policy_module.common.lcm_client import LcmClient -from osm_policy_module.common.message_bus_client import MessageBusClient -from osm_policy_module.common.mon_client import MonClient -from osm_policy_module.core import database -from osm_policy_module.core.config import Config -from osm_policy_module.core.database import ScalingGroup, ScalingAlarm, ScalingPolicy, ScalingCriteria, DatabaseManager -from osm_policy_module.core.exceptions import VdurNotFound -from osm_policy_module.utils.vnfd import VnfdUtils - -log = logging.getLogger(__name__) - -ALLOWED_KAFKA_KEYS = ['instantiated', 'scaled', 'terminated', 'notify_alarm'] - - -class PolicyModuleAgent: - def __init__(self, config: Config, loop=None): - self.conf = config - if not loop: - loop = asyncio.get_event_loop() - self.loop = loop - self.db_client = CommonDbClient(config) - self.mon_client = MonClient(config, loop=self.loop) - self.lcm_client = LcmClient(config, loop=self.loop) - self.database_manager = DatabaseManager(config) - self.msg_bus = MessageBusClient(config) - - def run(self): - self.loop.run_until_complete(self.start()) - - async def start(self): - topics = [ - "ns", - "alarm_response" - ] - await self.msg_bus.aioread(topics, self._process_msg) - log.critical("Exiting...") - - async def _process_msg(self, topic, key, msg): - log.debug("_process_msg topic=%s key=%s msg=%s", topic, key, msg) - log.info("Message arrived: %s", msg) - try: - if key in ALLOWED_KAFKA_KEYS: - - if key == 'instantiated': - await self._handle_instantiated(msg) - - if key == 'scaled': - await self._handle_scaled(msg) - - if key == 'terminated': - await self._handle_terminated(msg) - - if key == 'notify_alarm': - await self._handle_alarm_notification(msg) - else: - log.debug("Key %s is not in ALLOWED_KAFKA_KEYS", key) - except peewee.PeeweeException: - log.exception("Database error consuming message: ") - raise - except Exception: - log.exception("Error consuming message: ") - - async def _handle_alarm_notification(self, content): - log.debug("_handle_alarm_notification: %s", content) - alarm_uuid = content['notify_details']['alarm_uuid'] - metric_name = content['notify_details']['metric_name'] - operation = content['notify_details']['operation'] - threshold = content['notify_details']['threshold_value'] - vdu_name = content['notify_details']['vdu_name'] - vnf_member_index = content['notify_details']['vnf_member_index'] - nsr_id = content['notify_details']['ns_id'] - log.info( - "Received alarm notification for alarm %s, \ - metric %s, \ - operation %s, \ - threshold %s, \ - vdu_name %s, \ - vnf_member_index %s, \ - ns_id %s ", - alarm_uuid, metric_name, operation, threshold, vdu_name, vnf_member_index, nsr_id) - try: - alarm = self.database_manager.get_alarm(alarm_uuid) - delta = datetime.datetime.now() - alarm.scaling_criteria.scaling_policy.last_scale - log.debug("last_scale: %s", alarm.scaling_criteria.scaling_policy.last_scale) - log.debug("now: %s", datetime.datetime.now()) - log.debug("delta: %s", delta) - if delta.total_seconds() < alarm.scaling_criteria.scaling_policy.cooldown_time: - log.info("Time between last scale and now is less than cooldown time. Skipping.") - return - log.info("Sending scaling action message for ns: %s", nsr_id) - await self.lcm_client.scale(nsr_id, - alarm.scaling_criteria.scaling_policy.scaling_group.name, - alarm.vnf_member_index, - alarm.action) - alarm.scaling_criteria.scaling_policy.last_scale = datetime.datetime.now() - alarm.scaling_criteria.scaling_policy.save() - except ScalingAlarm.DoesNotExist: - log.info("There is no action configured for alarm %s.", alarm_uuid) - - async def _handle_instantiated(self, content): - log.debug("_handle_instantiated: %s", content) - nslcmop_id = content['nslcmop_id'] - nslcmop = self.db_client.get_nslcmop(nslcmop_id) - if nslcmop['operationState'] == 'COMPLETED' or nslcmop['operationState'] == 'PARTIALLY_COMPLETED': - nsr_id = nslcmop['nsInstanceId'] - log.info("Configuring scaling groups for network service with nsr_id: %s", nsr_id) - await self._configure_scaling_groups(nsr_id) - else: - log.info( - "Network service is not in COMPLETED or PARTIALLY_COMPLETED state. " - "Current state is %s. Skipping...", - nslcmop['operationState']) - - async def _handle_scaled(self, content): - log.debug("_handle_scaled: %s", content) - nslcmop_id = content['nslcmop_id'] - nslcmop = self.db_client.get_nslcmop(nslcmop_id) - if nslcmop['operationState'] == 'COMPLETED' or nslcmop['operationState'] == 'PARTIALLY_COMPLETED': - nsr_id = nslcmop['nsInstanceId'] - log.info("Configuring scaling groups for network service with nsr_id: %s", nsr_id) - await self._configure_scaling_groups(nsr_id) - log.info("Checking for orphaned alarms to be deleted for network service with nsr_id: %s", nsr_id) - await self._delete_orphaned_alarms(nsr_id) - else: - log.info( - "Network service is not in COMPLETED or PARTIALLY_COMPLETED state. " - "Current state is %s. Skipping...", - nslcmop['operationState']) - - async def _handle_terminated(self, content): - log.debug("_handle_deleted: %s", content) - nsr_id = content['nsr_id'] - if content['operationState'] == 'COMPLETED' or content['operationState'] == 'PARTIALLY_COMPLETED': - log.info("Deleting scaling groups and alarms for network service with nsr_id: %s", nsr_id) - await self._delete_scaling_groups(nsr_id) - else: - log.info( - "Network service is not in COMPLETED or PARTIALLY_COMPLETED state. " - "Current state is %s. Skipping...", - content['operationState']) - - async def _configure_scaling_groups(self, nsr_id: str): - log.debug("_configure_scaling_groups: %s", nsr_id) - alarms_created = [] - with database.db.atomic() as tx: - try: - vnfrs = self.db_client.get_vnfrs(nsr_id) - for vnfr in vnfrs: - log.info("Processing vnfr: %s", vnfr) - vnfd = self.db_client.get_vnfd(vnfr['vnfd-id']) - log.info("Looking for vnfd %s", vnfr['vnfd-id']) - if 'scaling-group-descriptor' not in vnfd: - continue - scaling_groups = vnfd['scaling-group-descriptor'] - vnf_monitoring_params = vnfd['monitoring-param'] - for scaling_group in scaling_groups: - try: - scaling_group_record = ScalingGroup.select().where( - ScalingGroup.nsr_id == nsr_id, - ScalingGroup.vnf_member_index == int(vnfr['member-vnf-index-ref']), - ScalingGroup.name == scaling_group['name'] - ).get() - log.info("Found existing scaling group record in DB...") - except ScalingGroup.DoesNotExist: - log.info("Creating scaling group record in DB...") - scaling_group_record = ScalingGroup.create( - nsr_id=nsr_id, - vnf_member_index=vnfr['member-vnf-index-ref'], - name=scaling_group['name'], - content=json.dumps(scaling_group) - ) - log.info( - "Created scaling group record in DB : nsr_id=%s, vnf_member_index=%s, name=%s", - scaling_group_record.nsr_id, - scaling_group_record.vnf_member_index, - scaling_group_record.name) - for scaling_policy in scaling_group['scaling-policy']: - if scaling_policy['scaling-type'] != 'automatic': - continue - try: - scaling_policy_record = ScalingPolicy.select().join(ScalingGroup).where( - ScalingPolicy.name == scaling_policy['name'], - ScalingGroup.id == scaling_group_record.id - ).get() - log.info("Found existing scaling policy record in DB...") - except ScalingPolicy.DoesNotExist: - log.info("Creating scaling policy record in DB...") - scaling_policy_record = ScalingPolicy.create( - nsr_id=nsr_id, - name=scaling_policy['name'], - cooldown_time=scaling_policy['cooldown-time'], - scaling_group=scaling_group_record - ) - log.info("Created scaling policy record in DB : name=%s, scaling_group.name=%s", - scaling_policy_record.name, - scaling_policy_record.scaling_group.name) - - for scaling_criteria in scaling_policy['scaling-criteria']: - try: - scaling_criteria_record = ScalingCriteria.select().join(ScalingPolicy).where( - ScalingPolicy.id == scaling_policy_record.id, - ScalingCriteria.name == scaling_criteria['name'] - ).get() - log.info("Found existing scaling criteria record in DB...") - except ScalingCriteria.DoesNotExist: - log.info("Creating scaling criteria record in DB...") - scaling_criteria_record = ScalingCriteria.create( - nsr_id=nsr_id, - name=scaling_criteria['name'], - scaling_policy=scaling_policy_record - ) - log.info( - "Created scaling criteria record in DB : name=%s, scaling_policy.name=%s", - scaling_criteria_record.name, - scaling_criteria_record.scaling_policy.name) - - vnf_monitoring_param = next( - filter( - lambda param: param['id'] == scaling_criteria[ - 'vnf-monitoring-param-ref' - ], - vnf_monitoring_params) - ) - if 'vdu-monitoring-param' in vnf_monitoring_param: - vdurs = list( - filter( - lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param - ['vdu-monitoring-param'] - ['vdu-ref'], - vnfr['vdur'] - ) - ) - elif 'vdu-metric' in vnf_monitoring_param: - vdurs = list( - filter( - lambda vdur: vdur['vdu-id-ref'] == vnf_monitoring_param - ['vdu-metric'] - ['vdu-ref'], - vnfr['vdur'] - ) - ) - elif 'vnf-metric' in vnf_monitoring_param: - vdu = VnfdUtils.get_mgmt_vdu(vnfd) - vdurs = list( - filter( - lambda vdur: vdur['vdu-id-ref'] == vdu['id'], - vnfr['vdur'] - ) - ) - else: - log.warning( - "Scaling criteria is referring to a vnf-monitoring-param that does not " - "contain a reference to a vdu or vnf metric.") - continue - for vdur in vdurs: - log.info("Creating alarm for vdur %s ", vdur) - try: - (ScalingAlarm.select() - .join(ScalingCriteria) - .join(ScalingPolicy) - .join(ScalingGroup) - .where( - ScalingAlarm.vdu_name == vdur['name'], - ScalingCriteria.name == scaling_criteria['name'], - ScalingPolicy.name == scaling_policy['name'], - ScalingGroup.nsr_id == nsr_id - ).get()) - log.debug("vdu %s already has an alarm configured", vdur['name']) - continue - except ScalingAlarm.DoesNotExist: - pass - alarm_uuid = await self.mon_client.create_alarm( - metric_name=vnf_monitoring_param['id'], - ns_id=nsr_id, - vdu_name=vdur['name'], - vnf_member_index=vnfr['member-vnf-index-ref'], - threshold=scaling_criteria['scale-in-threshold'], - operation=scaling_criteria['scale-in-relational-operation'], - statistic=vnf_monitoring_param['aggregation-type'] - ) - alarm = ScalingAlarm.create( - alarm_uuid=alarm_uuid, - action='scale_in', - vnf_member_index=int(vnfr['member-vnf-index-ref']), - vdu_name=vdur['name'], - scaling_criteria=scaling_criteria_record - ) - alarms_created.append(alarm) - alarm_uuid = await self.mon_client.create_alarm( - metric_name=vnf_monitoring_param['id'], - ns_id=nsr_id, - vdu_name=vdur['name'], - vnf_member_index=vnfr['member-vnf-index-ref'], - threshold=scaling_criteria['scale-out-threshold'], - operation=scaling_criteria['scale-out-relational-operation'], - statistic=vnf_monitoring_param['aggregation-type'] - ) - alarm = ScalingAlarm.create( - alarm_uuid=alarm_uuid, - action='scale_out', - vnf_member_index=int(vnfr['member-vnf-index-ref']), - vdu_name=vdur['name'], - scaling_criteria=scaling_criteria_record - ) - alarms_created.append(alarm) - - except Exception as e: - log.exception("Error configuring scaling groups:") - tx.rollback() - if len(alarms_created) > 0: - log.info("Cleaning alarm resources in MON") - for alarm in alarms_created: - await self.mon_client.delete_alarm(alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, - alarm.vnf_member_index, - alarm.vdu_name, - alarm.alarm_uuid) - raise e - - async def _delete_scaling_groups(self, nsr_id: str): - with database.db.atomic() as tx: - try: - for scaling_group in ScalingGroup.select().where(ScalingGroup.nsr_id == nsr_id): - for scaling_policy in scaling_group.scaling_policies: - for scaling_criteria in scaling_policy.scaling_criterias: - for alarm in scaling_criteria.scaling_alarms: - try: - await self.mon_client.delete_alarm( - alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, - alarm.vnf_member_index, - alarm.vdu_name, - alarm.alarm_uuid) - except ValueError: - log.exception("Error deleting alarm in MON %s", alarm.alarm_uuid) - alarm.delete_instance() - scaling_criteria.delete_instance() - scaling_policy.delete_instance() - scaling_group.delete_instance() - - except Exception as e: - log.exception("Error deleting scaling groups and alarms:") - tx.rollback() - raise e - - async def _delete_orphaned_alarms(self, nsr_id): - with database.db.atomic() as tx: - try: - for scaling_group in ScalingGroup.select().where(ScalingGroup.nsr_id == nsr_id): - for scaling_policy in scaling_group.scaling_policies: - for scaling_criteria in scaling_policy.scaling_criterias: - for alarm in scaling_criteria.scaling_alarms: - try: - self.db_client.get_vdur(nsr_id, alarm.vnf_member_index, alarm.vdu_name) - except VdurNotFound: - log.info("Deleting orphaned alarm %s", alarm.alarm_uuid) - try: - await self.mon_client.delete_alarm( - alarm.scaling_criteria.scaling_policy.scaling_group.nsr_id, - alarm.vnf_member_index, - alarm.vdu_name, - alarm.alarm_uuid) - except ValueError: - log.exception("Error deleting alarm in MON %s", alarm.alarm_uuid) - alarm.delete_instance() - - except Exception as e: - log.exception("Error deleting orphaned alarms:") - tx.rollback() - raise e diff --git a/osm_policy_module/core/database.py b/osm_policy_module/core/database.py index 3ca2e33..ecaa315 100644 --- a/osm_policy_module/core/database.py +++ b/osm_policy_module/core/database.py @@ -76,10 +76,8 @@ class DatabaseManager: db.initialize(connect(config.get('sql', 'database_uri'))) def create_tables(self) -> None: + db.connect() with db.atomic(): router = Router(db, os.path.dirname(migrations.__file__)) router.run() - - def get_alarm(self, alarm_uuid: str): - with db.atomic(): - return ScalingAlarm.select().where(ScalingAlarm.alarm_uuid == alarm_uuid).get() + db.close() diff --git a/osm_policy_module/tests/integration/test_policy_agent.py b/osm_policy_module/tests/integration/test_policy_agent.py index 7ad9944..4a7efbb 100644 --- a/osm_policy_module/tests/integration/test_policy_agent.py +++ b/osm_policy_module/tests/integration/test_policy_agent.py @@ -23,6 +23,7 @@ ## import asyncio import logging +import os import sys import unittest import uuid @@ -35,7 +36,7 @@ from playhouse.db_url import connect from osm_policy_module.common.common_db_client import CommonDbClient from osm_policy_module.common.mon_client import MonClient from osm_policy_module.core import database -from osm_policy_module.core.agent import PolicyModuleAgent +from osm_policy_module.autoscaling.agent import PolicyModuleAgent from osm_policy_module.core.config import Config from osm_policy_module.core.database import ScalingGroup, ScalingAlarm, ScalingPolicy, ScalingCriteria @@ -426,15 +427,17 @@ MODELS = [ScalingGroup, ScalingPolicy, ScalingCriteria, ScalingAlarm] class PolicyModuleAgentTest(unittest.TestCase): def setUp(self): super() - database.db.initialize(connect('sqlite://')) + database.db.initialize(connect('sqlite:///test_db.sqlite')) database.db.bind(MODELS) database.db.connect() database.db.drop_tables(MODELS) database.db.create_tables(MODELS) + database.db.close() self.loop = asyncio.new_event_loop() def tearDown(self): super() + os.remove('test_db.sqlite') @patch.object(DbMongo, 'db_connect', Mock()) @patch.object(KafkaProducer, '__init__') @@ -459,7 +462,7 @@ class PolicyModuleAgentTest(unittest.TestCase): create_alarm.side_effect = _test_configure_scaling_groups_create_alarm config = Config() agent = PolicyModuleAgent(config, self.loop) - self.loop.run_until_complete(agent._configure_scaling_groups("test_nsr_id")) + self.loop.run_until_complete(agent.service.configure_scaling_groups("test_nsr_id")) create_alarm.assert_any_call(metric_name='cirros_vnf_memory_util', ns_id='test_nsr_id', operation='GT', diff --git a/osm_policy_module/tests/unit/core/test_policy_agent.py b/osm_policy_module/tests/unit/core/test_policy_agent.py index 7fc2dc9..57b0db1 100644 --- a/osm_policy_module/tests/unit/core/test_policy_agent.py +++ b/osm_policy_module/tests/unit/core/test_policy_agent.py @@ -27,20 +27,20 @@ import unittest from unittest import mock from unittest.mock import Mock -from osm_policy_module.core.agent import PolicyModuleAgent +from osm_policy_module.autoscaling.agent import PolicyModuleAgent +from osm_policy_module.autoscaling.service import Service from osm_policy_module.core.config import Config -from osm_policy_module.core.database import DatabaseManager class PolicyAgentTest(unittest.TestCase): def setUp(self): self.loop = asyncio.new_event_loop() - @mock.patch('osm_policy_module.core.agent.CommonDbClient') - @mock.patch('osm_policy_module.core.agent.MonClient') - @mock.patch('osm_policy_module.core.agent.LcmClient') - @mock.patch.object(PolicyModuleAgent, '_configure_scaling_groups') - @mock.patch.object(PolicyModuleAgent, '_delete_orphaned_alarms') + @mock.patch('osm_policy_module.autoscaling.service.CommonDbClient') + @mock.patch('osm_policy_module.autoscaling.service.MonClient') + @mock.patch('osm_policy_module.autoscaling.service.LcmClient') + @mock.patch.object(Service, 'configure_scaling_groups') + @mock.patch.object(Service, 'delete_orphaned_alarms') def test_handle_instantiated(self, delete_orphaned_alarms, configure_scaling_groups, lcm_client, mon_client, db_client): async def mock_configure_scaling_groups(nsr_id): @@ -77,11 +77,12 @@ class PolicyAgentTest(unittest.TestCase): self.loop.run_until_complete(agent._handle_instantiated(content)) configure_scaling_groups.assert_not_called() - @mock.patch('osm_policy_module.core.agent.CommonDbClient') - @mock.patch('osm_policy_module.core.agent.MonClient') - @mock.patch('osm_policy_module.core.agent.LcmClient') - @mock.patch.object(DatabaseManager, 'get_alarm') - def test_handle_alarm_notification(self, get_alarm, lcm_client, mon_client, db_client): + @mock.patch('osm_policy_module.autoscaling.service.CommonDbClient') + @mock.patch('osm_policy_module.autoscaling.service.MonClient') + @mock.patch('osm_policy_module.autoscaling.service.LcmClient') + @mock.patch('osm_policy_module.core.database.db') + @mock.patch.object(Service, 'get_alarm') + def test_handle_alarm_notification(self, get_alarm, db, lcm_client, mon_client, db_client): async def mock_scale(nsr_id, scaling_group_name, vnf_member_index, action): pass