X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=osm_mon%2Fevaluator%2Fservice.py;h=9049d953a48529f062f0e0f91828a30e6d53aab0;hb=refs%2Fchanges%2F67%2F10967%2F1;hp=de3798bd7fd86e110164829c4d1849c1ca013e49;hpb=d0058e879923ad938e19fd0361f1e6e126a2322e;p=osm%2FMON.git diff --git a/osm_mon/evaluator/service.py b/osm_mon/evaluator/service.py index de3798b..9049d95 100644 --- a/osm_mon/evaluator/service.py +++ b/osm_mon/evaluator/service.py @@ -25,78 +25,70 @@ import multiprocessing from enum import Enum from typing import Tuple, List -from osm_mon.core import database from osm_mon.core.common_db import CommonDbClient from osm_mon.core.config import Config -from osm_mon.core.database import Alarm, AlarmRepository +from osm_mon.core.models import Alarm from osm_mon.evaluator.backends.prometheus import PrometheusBackend log = logging.getLogger(__name__) -BACKENDS = { - 'prometheus': PrometheusBackend -} +BACKENDS = {"prometheus": PrometheusBackend} class AlarmStatus(Enum): - ALARM = 'alarm' - OK = 'ok' - INSUFFICIENT = 'insufficient-data' + ALARM = "alarm" + OK = "ok" + INSUFFICIENT = "insufficient-data" + DISABLED = "disabled" class EvaluatorService: - def __init__(self, config: Config): self.conf = config self.common_db = CommonDbClient(self.conf) self.queue = multiprocessing.Queue() - def _get_metric_value(self, - metric_name: str, - tags: dict): - return BACKENDS[self.conf.get('evaluator', 'backend')](self.conf).get_metric_value(metric_name, tags) + def _get_metric_value(self, metric_name: str, tags: dict): + return BACKENDS[self.conf.get("evaluator", "backend")]( + self.conf + ).get_metric_value(metric_name, tags) - def _evaluate_metric(self, - alarm: Alarm, tags: dict): + def _evaluate_metric(self, alarm: Alarm): log.debug("_evaluate_metric") - metric_value = self._get_metric_value(alarm.metric, tags) - if metric_value is None: - log.warning("No metric result for alarm %s", alarm.id) - self.queue.put((alarm, AlarmStatus.INSUFFICIENT)) - else: - if alarm.operation.upper() == 'GT': - if metric_value > alarm.threshold: - self.queue.put((alarm, AlarmStatus.ALARM)) - else: - self.queue.put((alarm, AlarmStatus.OK)) - elif alarm.operation.upper() == 'LT': - if metric_value < alarm.threshold: - self.queue.put((alarm, AlarmStatus.ALARM)) - else: - self.queue.put((alarm, AlarmStatus.OK)) + metric_value = self._get_metric_value(alarm.metric, alarm.tags) + if alarm.alarm_status.upper() != AlarmStatus.DISABLED.value.upper(): + if metric_value is None: + log.warning("No metric result for alarm %s", alarm.uuid) + self.queue.put((alarm, AlarmStatus.INSUFFICIENT)) + else: + if alarm.operation.upper() == "GT": + if metric_value > alarm.threshold: + self.queue.put((alarm, AlarmStatus.ALARM)) + else: + self.queue.put((alarm, AlarmStatus.OK)) + elif alarm.operation.upper() == "LT": + if metric_value < alarm.threshold: + self.queue.put((alarm, AlarmStatus.ALARM)) + else: + self.queue.put((alarm, AlarmStatus.OK)) + + def update_alarm_status(self, alarm_state, uuid): + alarm_data = self.common_db.get_alarm_by_uuid(uuid) + if alarm_data.get("alarm_status").upper() != AlarmStatus.DISABLED.value.upper(): + self.common_db.update_alarm_status(alarm_state, uuid) + return def evaluate_alarms(self) -> List[Tuple[Alarm, AlarmStatus]]: - log.debug('evaluate_alarms') + log.debug("evaluate_alarms") processes = [] - database.db.connect() - try: - with database.db.atomic(): - for alarm in AlarmRepository.list(): - # Tags need to be passed inside a dict to avoid database locking issues related to process forking - tags = {} - for tag in alarm.tags: - tags[tag.name] = tag.value - p = multiprocessing.Process(target=self._evaluate_metric, - args=(alarm, tags)) - processes.append(p) - p.start() - - for process in processes: - process.join(timeout=10) - alarms_tuples = [] - log.info("Appending alarms to queue") - while not self.queue.empty(): - alarms_tuples.append(self.queue.get()) - return alarms_tuples - finally: - database.db.close() + for alarm in self.common_db.get_alarms(): + p = multiprocessing.Process(target=self._evaluate_metric, args=(alarm,)) + processes.append(p) + p.start() + + for process in processes: + process.join(timeout=10) + alarms_tuples = [] + while not self.queue.empty(): + alarms_tuples.append(self.queue.get()) + return alarms_tuples