Pin pylint version to 3.1.1 in tox.ini
[osm/LCM.git] / osm_lcm / lcm.py
index 3b8da7d..9b62d82 100644 (file)
 # under the License.
 ##
 
+
+# DEBUG WITH PDB
+import pdb
+
 import asyncio
 import yaml
 import logging
 import logging.handlers
 import getopt
 import sys
+from random import SystemRandom
 
-from osm_lcm import ROclient, ns, vim_sdn, netslice
-from time import time, sleep
+from osm_lcm import ns, vim_sdn, netslice
+from osm_lcm.ng_ro import NgRoException, NgRoClient
+from osm_lcm.ROclient import ROClient, ROClientException
+
+from time import time
 from osm_lcm.lcm_utils import versiontuple, LcmException, TaskRegistry, LcmExceptionExit
 from osm_lcm import version as lcm_version, version_date as lcm_version_date
 
-from osm_common import dbmemory, dbmongo, fslocal, msglocal, msgkafka
+from osm_common import msglocal, msgkafka
 from osm_common import version as common_version
 from osm_common.dbbase import DbException
 from osm_common.fsbase import FsException
 from osm_common.msgbase import MsgException
-from os import environ, path
-from random import choice as random_choice
+from osm_lcm.data_utils.database.database import Database
+from osm_lcm.data_utils.filesystem.filesystem import Filesystem
+from osm_lcm.data_utils.lcm_config import LcmCfg
+from osm_lcm.lcm_hc import get_health_check_file
+from os import path, getenv
 from n2vc import version as n2vc_version
+import traceback
+
+if getenv("OSMLCM_PDB_DEBUG", None) is not None:
+    pdb.set_trace()
 
 
 __author__ = "Alfonso Tierno"
-min_RO_version = "0.6.3"
+min_RO_version = "6.0.2"
 min_n2vc_version = "0.0.2"
+
 min_common_version = "0.1.19"
-# uncomment if LCM is installed as library and installed, and get them from __init__.py
-# lcm_version = '0.1.41'
-# lcm_version_date = '2019-06-19'
-health_check_file = path.expanduser("~") + "/time_last_ping"   # TODO find better location for this file
 
 
 class Lcm:
+    ping_interval_pace = (
+        120  # how many time ping is send once is confirmed all is running
+    )
+    ping_interval_boot = 5  # how many time ping is sent when booting
 
-    ping_interval_pace = 120  # how many time ping is send once is confirmed all is running
-    ping_interval_boot = 5    # how many time ping is sent when booting
+    main_config = LcmCfg()
 
-    def __init__(self, config_file, loop=None):
+    def __init__(self, config_file):
         """
         Init, Connect to database, filesystem storage, and messaging
         :param config: two level dictionary with configuration. Top level should contain 'database', 'storage',
         :return: None
         """
-
         self.db = None
         self.msg = None
         self.msg_admin = None
@@ -70,87 +84,89 @@ class Lcm:
         self.first_start = False
 
         # logging
-        self.logger = logging.getLogger('lcm')
+        self.logger = logging.getLogger("lcm")
         # get id
         self.worker_id = self.get_process_id()
         # load configuration
         config = self.read_config_file(config_file)
-        self.config = config
-        self.ro_config = {
-            "endpoint_url": "http://{}:{}/openmano".format(config["RO"]["host"], config["RO"]["port"]),
-            "tenant": config.get("tenant", "osm"),
-            "logger_name": "lcm.ROclient",
-            "loglevel": "ERROR",
-        }
-
-        self.vca_config = config["VCA"]
-
-        self.loop = loop or asyncio.get_event_loop()
+        self.main_config.set_from_dict(config)
+        self.main_config.transform()
+        self.main_config.load_from_env()
+        self.logger.critical("Loaded configuration:" + str(self.main_config.to_dict()))
+        # TODO: check if lcm_hc.py is necessary
+        self.health_check_file = get_health_check_file(self.main_config.to_dict())
+        self.ns = (
+            self.netslice
+        ) = (
+            self.vim
+        ) = self.wim = self.sdn = self.k8scluster = self.vca = self.k8srepo = None
 
         # logging
-        log_format_simple = "%(asctime)s %(levelname)s %(name)s %(filename)s:%(lineno)s %(message)s"
-        log_formatter_simple = logging.Formatter(log_format_simple, datefmt='%Y-%m-%dT%H:%M:%S')
-        config["database"]["logger_name"] = "lcm.db"
-        config["storage"]["logger_name"] = "lcm.fs"
-        config["message"]["logger_name"] = "lcm.msg"
-        if config["global"].get("logfile"):
-            file_handler = logging.handlers.RotatingFileHandler(config["global"]["logfile"],
-                                                                maxBytes=100e6, backupCount=9, delay=0)
+        log_format_simple = (
+            "%(asctime)s %(levelname)s %(name)s %(filename)s:%(lineno)s %(message)s"
+        )
+        log_formatter_simple = logging.Formatter(
+            log_format_simple, datefmt="%Y-%m-%dT%H:%M:%S"
+        )
+        if self.main_config.globalConfig.logfile:
+            file_handler = logging.handlers.RotatingFileHandler(
+                self.main_config.globalConfig.logfile,
+                maxBytes=100e6,
+                backupCount=9,
+                delay=0,
+            )
             file_handler.setFormatter(log_formatter_simple)
             self.logger.addHandler(file_handler)
-        if not config["global"].get("nologging"):
+        if not self.main_config.globalConfig.to_dict()["nologging"]:
             str_handler = logging.StreamHandler()
             str_handler.setFormatter(log_formatter_simple)
             self.logger.addHandler(str_handler)
 
-        if config["global"].get("loglevel"):
-            self.logger.setLevel(config["global"]["loglevel"])
+        if self.main_config.globalConfig.to_dict()["loglevel"]:
+            self.logger.setLevel(self.main_config.globalConfig.loglevel)
 
         # logging other modules
-        for k1, logname in {"message": "lcm.msg", "database": "lcm.db", "storage": "lcm.fs"}.items():
-            config[k1]["logger_name"] = logname
-            logger_module = logging.getLogger(logname)
-            if config[k1].get("logfile"):
-                file_handler = logging.handlers.RotatingFileHandler(config[k1]["logfile"],
-                                                                    maxBytes=100e6, backupCount=9, delay=0)
+        for logger in ("message", "database", "storage", "tsdb"):
+            logger_config = self.main_config.to_dict()[logger]
+            logger_module = logging.getLogger(logger_config["logger_name"])
+            if logger_config["logfile"]:
+                file_handler = logging.handlers.RotatingFileHandler(
+                    logger_config["logfile"], maxBytes=100e6, backupCount=9, delay=0
+                )
                 file_handler.setFormatter(log_formatter_simple)
                 logger_module.addHandler(file_handler)
-            if config[k1].get("loglevel"):
-                logger_module.setLevel(config[k1]["loglevel"])
-        self.logger.critical("starting osm/lcm version {} {}".format(lcm_version, lcm_version_date))
+            if logger_config["loglevel"]:
+                logger_module.setLevel(logger_config["loglevel"])
+        self.logger.critical(
+            "starting osm/lcm version {} {}".format(lcm_version, lcm_version_date)
+        )
 
         # check version of N2VC
         # TODO enhance with int conversion or from distutils.version import LooseVersion
         # or with list(map(int, version.split(".")))
         if versiontuple(n2vc_version) < versiontuple(min_n2vc_version):
-            raise LcmException("Not compatible osm/N2VC version '{}'. Needed '{}' or higher".format(
-                n2vc_version, min_n2vc_version))
+            raise LcmException(
+                "Not compatible osm/N2VC version '{}'. Needed '{}' or higher".format(
+                    n2vc_version, min_n2vc_version
+                )
+            )
         # check version of common
         if versiontuple(common_version) < versiontuple(min_common_version):
-            raise LcmException("Not compatible osm/common version '{}'. Needed '{}' or higher".format(
-                common_version, min_common_version))
+            raise LcmException(
+                "Not compatible osm/common version '{}'. Needed '{}' or higher".format(
+                    common_version, min_common_version
+                )
+            )
 
         try:
-            # TODO check database version
-            if config["database"]["driver"] == "mongo":
-                self.db = dbmongo.DbMongo()
-                self.db.db_connect(config["database"])
-            elif config["database"]["driver"] == "memory":
-                self.db = dbmemory.DbMemory()
-                self.db.db_connect(config["database"])
-            else:
-                raise LcmException("Invalid configuration param '{}' at '[database]':'driver'".format(
-                    config["database"]["driver"]))
+            self.db = Database(self.main_config.to_dict()).instance.db
 
-            if config["storage"]["driver"] == "local":
-                self.fs = fslocal.FsLocal()
-                self.fs.fs_connect(config["storage"])
-            else:
-                raise LcmException("Invalid configuration param '{}' at '[storage]':'driver'".format(
-                    config["storage"]["driver"]))
+            self.fs = Filesystem(self.main_config.to_dict()).instance.fs
+            self.fs.sync()
 
-            config_message = config["message"].copy()
-            config_message["loop"] = self.loop
+            # copy message configuration in order to remove 'group_id' for msg_admin
+            config_message = self.main_config.message.to_dict()
+            config_message["loop"] = asyncio.get_event_loop()
             if config_message["driver"] == "local":
                 self.msg = msglocal.MsgLocal()
                 self.msg.connect(config_message)
@@ -164,34 +180,65 @@ class Lcm:
                 config_message.pop("group_id", None)
                 self.msg_admin.connect(config_message)
             else:
-                raise LcmException("Invalid configuration param '{}' at '[message]':'driver'".format(
-                    config["message"]["driver"]))
+                raise LcmException(
+                    "Invalid configuration param '{}' at '[message]':'driver'".format(
+                        self.main_config.message.driver
+                    )
+                )
         except (DbException, FsException, MsgException) as e:
             self.logger.critical(str(e), exc_info=True)
             raise LcmException(str(e))
 
         # contains created tasks/futures to be able to cancel
-        self.lcm_tasks = TaskRegistry(self.worker_id, self.db, self.logger)
-
-        self.ns = ns.NsLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.ro_config, self.vca_config, self.loop)
-        self.netslice = netslice.NetsliceLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.ro_config, 
-                                             self.vca_config, self.loop)
-        self.vim = vim_sdn.VimLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.ro_config, self.loop)
-        self.wim = vim_sdn.WimLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.ro_config, self.loop)
-        self.sdn = vim_sdn.SdnLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.ro_config, self.loop)
+        self.lcm_tasks = TaskRegistry(self.worker_id, self.logger)
 
     async def check_RO_version(self):
-        try:
-            RO = ROclient.ROClient(self.loop, **self.ro_config)
-            RO_version = await RO.get_version()
-            if versiontuple(RO_version) < versiontuple(min_RO_version):
-                raise LcmException("Not compatible osm/RO version '{}.{}.{}'. Needed '{}.{}.{}' or higher".format(
-                    *RO_version, *min_RO_version
-                ))
-        except ROclient.ROClientException as e:
-            error_text = "Error while conneting to osm/RO " + str(e)
-            self.logger.critical(error_text, exc_info=True)
-            raise LcmException(error_text)
+        tries = 14
+        last_error = None
+        while True:
+            ro_uri = self.main_config.RO.uri
+            if not ro_uri:
+                ro_uri = ""
+            try:
+                # try new  RO, if fail old RO
+                try:
+                    self.main_config.RO.uri = ro_uri + "ro"
+                    ro_server = NgRoClient(**self.main_config.RO.to_dict())
+                    ro_version = await ro_server.get_version()
+                    self.main_config.RO.ng = True
+                except Exception:
+                    self.main_config.RO.uri = ro_uri + "openmano"
+                    ro_server = ROClient(**self.main_config.RO.to_dict())
+                    ro_version = await ro_server.get_version()
+                    self.main_config.RO.ng = False
+                if versiontuple(ro_version) < versiontuple(min_RO_version):
+                    raise LcmException(
+                        "Not compatible osm/RO version '{}'. Needed '{}' or higher".format(
+                            ro_version, min_RO_version
+                        )
+                    )
+                self.logger.info(
+                    "Connected to RO version {} new-generation version {}".format(
+                        ro_version, self.main_config.RO.ng
+                    )
+                )
+                return
+            except (ROClientException, NgRoException) as e:
+                self.main_config.RO.uri = ro_uri
+                tries -= 1
+                traceback.print_tb(e.__traceback__)
+                error_text = "Error while connecting to RO on {}: {}".format(
+                    self.main_config.RO.uri, e
+                )
+                if tries <= 0:
+                    self.logger.critical(error_text)
+                    raise LcmException(error_text)
+                if last_error != error_text:
+                    last_error = error_text
+                    self.logger.error(
+                        error_text + ". Waiting until {} seconds".format(5 * tries)
+                    )
+                await asyncio.sleep(5)
 
     async def test(self, param=None):
         self.logger.debug("Starting/Ending test task: {}".format(param))
@@ -205,15 +252,25 @@ class Lcm:
         while True:
             try:
                 await self.msg_admin.aiowrite(
-                    "admin", "ping",
-                    {"from": "lcm", "to": "lcm", "worker_id": self.worker_id, "version": lcm_version},
-                    self.loop)
+                    "admin",
+                    "ping",
+                    {
+                        "from": "lcm",
+                        "to": "lcm",
+                        "worker_id": self.worker_id,
+                        "version": lcm_version,
+                    },
+                )
                 # time between pings are low when it is not received and at starting
-                wait_time = self.ping_interval_boot if not kafka_has_received else self.ping_interval_pace
+                wait_time = (
+                    self.ping_interval_boot
+                    if not kafka_has_received
+                    else self.ping_interval_pace
+                )
                 if not self.pings_not_received:
                     kafka_has_received = True
                 self.pings_not_received += 1
-                await asyncio.sleep(wait_time, loop=self.loop)
+                await asyncio.sleep(wait_time)
                 if self.pings_not_received > 10:
                     raise LcmException("It is not receiving pings from Kafka bus")
                 consecutive_errors = 0
@@ -224,18 +281,26 @@ class Lcm:
                 # if not first_start is the first time after starting. So leave more time and wait
                 # to allow kafka starts
                 if consecutive_errors == 8 if not first_start else 30:
-                    self.logger.error("Task kafka_read task exit error too many errors. Exception: {}".format(e))
+                    self.logger.error(
+                        "Task kafka_read task exit error too many errors. Exception: {}".format(
+                            e
+                        )
+                    )
                     raise
                 consecutive_errors += 1
-                self.logger.error("Task kafka_read retrying after Exception {}".format(e))
+                self.logger.error(
+                    "Task kafka_read retrying after Exception {}".format(e)
+                )
                 wait_time = 2 if not first_start else 5
-                await asyncio.sleep(wait_time, loop=self.loop)
+                await asyncio.sleep(wait_time)
 
-    def kafka_read_callback(self, topic, command, params):
+    async def kafka_read_callback(self, topic, command, params):
         order_id = 1
 
         if topic != "admin" and command != "ping":
-            self.logger.debug("Task kafka_read receives {} {}: {}".format(topic, command, params))
+            self.logger.debug(
+                "Task kafka_read receives {} {}: {}".format(topic, command, params)
+            )
         self.consecutive_errors = 0
         self.first_start = False
         order_id += 1
@@ -249,7 +314,7 @@ class Lcm:
             sys.stdout.flush()
             return
         elif command == "test":
-            asyncio.Task(self.test(params), loop=self.loop)
+            asyncio.Task(self.test(params))
             return
 
         if topic == "admin":
@@ -258,11 +323,111 @@ class Lcm:
                     return
                 self.pings_not_received = 0
                 try:
-                    with open(health_check_file, "w") as f:
+                    with open(self.health_check_file, "w") as f:
                         f.write(str(time()))
                 except Exception as e:
-                    self.logger.error("Cannot write into '{}' for healthcheck: {}".format(health_check_file, e))
+                    self.logger.error(
+                        "Cannot write into '{}' for healthcheck: {}".format(
+                            self.health_check_file, e
+                        )
+                    )
             return
+        elif topic == "nslcmops":
+            if command == "cancel":
+                nslcmop_id = params["_id"]
+                self.logger.debug("Cancelling nslcmop {}".format(nslcmop_id))
+                nsr_id = params["nsInstanceId"]
+                # cancel the tasks and wait
+                for task in self.lcm_tasks.cancel("ns", nsr_id, nslcmop_id):
+                    try:
+                        await task
+                        self.logger.debug(
+                            "Cancelled task ended {},{},{}".format(
+                                nsr_id, nslcmop_id, task
+                            )
+                        )
+                    except asyncio.CancelledError:
+                        self.logger.debug(
+                            "Task already cancelled and finished {},{},{}".format(
+                                nsr_id, nslcmop_id, task
+                            )
+                        )
+                # update DB
+                q_filter = {"_id": nslcmop_id}
+                update_dict = {
+                    "operationState": "FAILED_TEMP",
+                    "isCancelPending": False,
+                }
+                unset_dict = {
+                    "cancelMode": None,
+                }
+                self.db.set_one(
+                    "nslcmops",
+                    q_filter=q_filter,
+                    update_dict=update_dict,
+                    fail_on_empty=False,
+                    unset=unset_dict,
+                )
+                self.logger.debug("LCM task cancelled {},{}".format(nsr_id, nslcmop_id))
+            return
+        elif topic == "pla":
+            if command == "placement":
+                self.ns.update_nsrs_with_pla_result(params)
+            return
+        elif topic == "k8scluster":
+            if command == "create" or command == "created":
+                k8scluster_id = params.get("_id")
+                task = asyncio.ensure_future(self.k8scluster.create(params, order_id))
+                self.lcm_tasks.register(
+                    "k8scluster", k8scluster_id, order_id, "k8scluster_create", task
+                )
+                return
+            elif command == "edit" or command == "edited":
+                k8scluster_id = params.get("_id")
+                task = asyncio.ensure_future(self.k8scluster.edit(params, order_id))
+                self.lcm_tasks.register(
+                    "k8scluster", k8scluster_id, order_id, "k8scluster_edit", task
+                )
+                return
+            elif command == "delete" or command == "deleted":
+                k8scluster_id = params.get("_id")
+                task = asyncio.ensure_future(self.k8scluster.delete(params, order_id))
+                self.lcm_tasks.register(
+                    "k8scluster", k8scluster_id, order_id, "k8scluster_delete", task
+                )
+                return
+        elif topic == "vca":
+            if command == "create" or command == "created":
+                vca_id = params.get("_id")
+                task = asyncio.ensure_future(self.vca.create(params, order_id))
+                self.lcm_tasks.register("vca", vca_id, order_id, "vca_create", task)
+                return
+            elif command == "edit" or command == "edited":
+                vca_id = params.get("_id")
+                task = asyncio.ensure_future(self.vca.edit(params, order_id))
+                self.lcm_tasks.register("vca", vca_id, order_id, "vca_edit", task)
+                return
+            elif command == "delete" or command == "deleted":
+                vca_id = params.get("_id")
+                task = asyncio.ensure_future(self.vca.delete(params, order_id))
+                self.lcm_tasks.register("vca", vca_id, order_id, "vca_delete", task)
+                return
+        elif topic == "k8srepo":
+            if command == "create" or command == "created":
+                k8srepo_id = params.get("_id")
+                self.logger.debug("k8srepo_id = {}".format(k8srepo_id))
+                task = asyncio.ensure_future(self.k8srepo.create(params, order_id))
+                self.lcm_tasks.register(
+                    "k8srepo", k8srepo_id, order_id, "k8srepo_create", task
+                )
+                return
+            elif command == "delete" or command == "deleted":
+                k8srepo_id = params.get("_id")
+                task = asyncio.ensure_future(self.k8srepo.delete(params, order_id))
+                self.lcm_tasks.register(
+                    "k8srepo", k8srepo_id, order_id, "k8srepo_delete", task
+                )
+                return
         elif topic == "ns":
             if command == "instantiate":
                 # self.logger.debug("Deploying NS {}".format(nsr_id))
@@ -270,7 +435,9 @@ class Lcm:
                 nslcmop_id = nslcmop["_id"]
                 nsr_id = nslcmop["nsInstanceId"]
                 task = asyncio.ensure_future(self.ns.instantiate(nsr_id, nslcmop_id))
-                self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_instantiate", task)
+                self.lcm_tasks.register(
+                    "ns", nsr_id, nslcmop_id, "ns_instantiate", task
+                )
                 return
             elif command == "terminate":
                 # self.logger.debug("Deleting NS {}".format(nsr_id))
@@ -281,6 +448,17 @@ class Lcm:
                 task = asyncio.ensure_future(self.ns.terminate(nsr_id, nslcmop_id))
                 self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_terminate", task)
                 return
+            elif command == "vca_status_refresh":
+                nslcmop = params
+                nslcmop_id = nslcmop["_id"]
+                nsr_id = nslcmop["nsInstanceId"]
+                task = asyncio.ensure_future(
+                    self.ns.vca_status_refresh(nsr_id, nslcmop_id)
+                )
+                self.lcm_tasks.register(
+                    "ns", nsr_id, nslcmop_id, "ns_vca_status_refresh", task
+                )
+                return
             elif command == "action":
                 # self.logger.debug("Update NS {}".format(nsr_id))
                 nslcmop = params
@@ -289,6 +467,14 @@ class Lcm:
                 task = asyncio.ensure_future(self.ns.action(nsr_id, nslcmop_id))
                 self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_action", task)
                 return
+            elif command == "update":
+                # self.logger.debug("Update NS {}".format(nsr_id))
+                nslcmop = params
+                nslcmop_id = nslcmop["_id"]
+                nsr_id = nslcmop["nsInstanceId"]
+                task = asyncio.ensure_future(self.ns.update(nsr_id, nslcmop_id))
+                self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_update", task)
+                return
             elif command == "scale":
                 # self.logger.debug("Update NS {}".format(nsr_id))
                 nslcmop = params
@@ -297,31 +483,84 @@ class Lcm:
                 task = asyncio.ensure_future(self.ns.scale(nsr_id, nslcmop_id))
                 self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_scale", task)
                 return
+            elif command == "heal":
+                # self.logger.debug("Healing NS {}".format(nsr_id))
+                nslcmop = params
+                nslcmop_id = nslcmop["_id"]
+                nsr_id = nslcmop["nsInstanceId"]
+                task = asyncio.ensure_future(self.ns.heal(nsr_id, nslcmop_id))
+                self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_heal", task)
+                return
+            elif command == "migrate":
+                nslcmop = params
+                nslcmop_id = nslcmop["_id"]
+                nsr_id = nslcmop["nsInstanceId"]
+                task = asyncio.ensure_future(self.ns.migrate(nsr_id, nslcmop_id))
+                self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_migrate", task)
+                return
+            elif command == "verticalscale":
+                nslcmop = params
+                nslcmop_id = nslcmop["_id"]
+                nsr_id = nslcmop["nsInstanceId"]
+                task = asyncio.ensure_future(self.ns.vertical_scale(nsr_id, nslcmop_id))
+                self.logger.debug(
+                    "nsr_id,nslcmop_id,task {},{},{}".format(nsr_id, nslcmop_id, task)
+                )
+                self.lcm_tasks.register(
+                    "ns", nsr_id, nslcmop_id, "ns_verticalscale", task
+                )
+                self.logger.debug(
+                    "LCM task registered {},{},{} ".format(nsr_id, nslcmop_id, task)
+                )
+                return
             elif command == "show":
                 nsr_id = params
                 try:
                     db_nsr = self.db.get_one("nsrs", {"_id": nsr_id})
-                    print("nsr:\n    _id={}\n    operational-status: {}\n    config-status: {}"
-                          "\n    detailed-status: {}\n    deploy: {}\n    tasks: {}"
-                          "".format(nsr_id, db_nsr["operational-status"], db_nsr["config-status"],
-                                    db_nsr["detailed-status"],
-                                    db_nsr["_admin"]["deployed"], self.lcm_ns_tasks.get(nsr_id)))
+                    print(
+                        "nsr:\n    _id={}\n    operational-status: {}\n    config-status: {}"
+                        "\n    detailed-status: {}\n    deploy: {}\n    tasks: {}"
+                        "".format(
+                            nsr_id,
+                            db_nsr["operational-status"],
+                            db_nsr["config-status"],
+                            db_nsr["detailed-status"],
+                            db_nsr["_admin"]["deployed"],
+                            self.lcm_tasks.task_registry["ns"].get(nsr_id, ""),
+                        )
+                    )
                 except Exception as e:
                     print("nsr {} not found: {}".format(nsr_id, e))
                 sys.stdout.flush()
                 return
             elif command == "deleted":
                 return  # TODO cleaning of task just in case should be done
-            elif command in ("terminated", "instantiated", "scaled", "actioned"):  # "scaled-cooldown-time"
+            elif command in (
+                "vnf_terminated",
+                "policy_updated",
+                "terminated",
+                "instantiated",
+                "scaled",
+                "healed",
+                "actioned",
+                "updated",
+                "migrated",
+                "verticalscaled",
+            ):  # "scaled-cooldown-time"
                 return
+
         elif topic == "nsi":  # netslice LCM processes (instantiate, terminate, etc)
             if command == "instantiate":
                 # self.logger.debug("Instantiating Network Slice {}".format(nsilcmop["netsliceInstanceId"]))
                 nsilcmop = params
                 nsilcmop_id = nsilcmop["_id"]  # slice operation id
                 nsir_id = nsilcmop["netsliceInstanceId"]  # slice record id
-                task = asyncio.ensure_future(self.netslice.instantiate(nsir_id, nsilcmop_id))
-                self.lcm_tasks.register("nsi", nsir_id, nsilcmop_id, "nsi_instantiate", task)
+                task = asyncio.ensure_future(
+                    self.netslice.instantiate(nsir_id, nsilcmop_id)
+                )
+                self.lcm_tasks.register(
+                    "nsi", nsir_id, nsilcmop_id, "nsi_instantiate", task
+                )
                 return
             elif command == "terminate":
                 # self.logger.debug("Terminating Network Slice NS {}".format(nsilcmop["netsliceInstanceId"]))
@@ -329,93 +568,154 @@ class Lcm:
                 nsilcmop_id = nsilcmop["_id"]  # slice operation id
                 nsir_id = nsilcmop["netsliceInstanceId"]  # slice record id
                 self.lcm_tasks.cancel(topic, nsir_id)
-                task = asyncio.ensure_future(self.netslice.terminate(nsir_id, nsilcmop_id))
-                self.lcm_tasks.register("nsi", nsir_id, nsilcmop_id, "nsi_terminate", task)
+                task = asyncio.ensure_future(
+                    self.netslice.terminate(nsir_id, nsilcmop_id)
+                )
+                self.lcm_tasks.register(
+                    "nsi", nsir_id, nsilcmop_id, "nsi_terminate", task
+                )
                 return
             elif command == "show":
                 nsir_id = params
                 try:
                     db_nsir = self.db.get_one("nsirs", {"_id": nsir_id})
-                    print("nsir:\n    _id={}\n    operational-status: {}\n    config-status: {}"
-                          "\n    detailed-status: {}\n    deploy: {}\n    tasks: {}"
-                          "".format(nsir_id, db_nsir["operational-status"], db_nsir["config-status"],
-                                    db_nsir["detailed-status"],
-                                    db_nsir["_admin"]["deployed"], self.lcm_netslice_tasks.get(nsir_id)))
+                    print(
+                        "nsir:\n    _id={}\n    operational-status: {}\n    config-status: {}"
+                        "\n    detailed-status: {}\n    deploy: {}\n    tasks: {}"
+                        "".format(
+                            nsir_id,
+                            db_nsir["operational-status"],
+                            db_nsir["config-status"],
+                            db_nsir["detailed-status"],
+                            db_nsir["_admin"]["deployed"],
+                            self.lcm_tasks.task_registry["nsi"].get(nsir_id, ""),
+                        )
+                    )
                 except Exception as e:
                     print("nsir {} not found: {}".format(nsir_id, e))
                 sys.stdout.flush()
                 return
             elif command == "deleted":
                 return  # TODO cleaning of task just in case should be done
-            elif command in ("terminated", "instantiated", "scaled", "actioned"):  # "scaled-cooldown-time"
+            elif command in (
+                "terminated",
+                "instantiated",
+                "scaled",
+                "healed",
+                "actioned",
+            ):  # "scaled-cooldown-time"
                 return
         elif topic == "vim_account":
             vim_id = params["_id"]
-            if command == "create":
-                task = asyncio.ensure_future(self.vim.create(params, order_id))
-                self.lcm_tasks.register("vim_account", vim_id, order_id, "vim_create", task)
+            if command in ("create", "created"):
+                if not self.main_config.RO.ng:
+                    task = asyncio.ensure_future(self.vim.create(params, order_id))
+                    self.lcm_tasks.register(
+                        "vim_account", vim_id, order_id, "vim_create", task
+                    )
                 return
-            elif command == "delete":
+            elif command == "delete" or command == "deleted":
                 self.lcm_tasks.cancel(topic, vim_id)
                 task = asyncio.ensure_future(self.vim.delete(params, order_id))
-                self.lcm_tasks.register("vim_account", vim_id, order_id, "vim_delete", task)
+                self.lcm_tasks.register(
+                    "vim_account", vim_id, order_id, "vim_delete", task
+                )
                 return
             elif command == "show":
                 print("not implemented show with vim_account")
                 sys.stdout.flush()
                 return
-            elif command == "edit":
-                task = asyncio.ensure_future(self.vim.edit(params, order_id))
-                self.lcm_tasks.register("vim_account", vim_id, order_id, "vim_edit", task)
+            elif command in ("edit", "edited"):
+                if not self.main_config.RO.ng:
+                    task = asyncio.ensure_future(self.vim.edit(params, order_id))
+                    self.lcm_tasks.register(
+                        "vim_account", vim_id, order_id, "vim_edit", task
+                    )
                 return
+            elif command == "deleted":
+                return  # TODO cleaning of task just in case should be done
         elif topic == "wim_account":
             wim_id = params["_id"]
-            if command == "create":
-                task = asyncio.ensure_future(self.wim.create(params, order_id))
-                self.lcm_tasks.register("wim_account", wim_id, order_id, "wim_create", task)
+            if command in ("create", "created"):
+                if not self.main_config.RO.ng:
+                    task = asyncio.ensure_future(self.wim.create(params, order_id))
+                    self.lcm_tasks.register(
+                        "wim_account", wim_id, order_id, "wim_create", task
+                    )
                 return
-            elif command == "delete":
+            elif command == "delete" or command == "deleted":
                 self.lcm_tasks.cancel(topic, wim_id)
                 task = asyncio.ensure_future(self.wim.delete(params, order_id))
-                self.lcm_tasks.register("wim_account", wim_id, order_id, "wim_delete", task)
+                self.lcm_tasks.register(
+                    "wim_account", wim_id, order_id, "wim_delete", task
+                )
                 return
             elif command == "show":
                 print("not implemented show with wim_account")
                 sys.stdout.flush()
                 return
-            elif command == "edit":
+            elif command in ("edit", "edited"):
                 task = asyncio.ensure_future(self.wim.edit(params, order_id))
-                self.lcm_tasks.register("wim_account", wim_id, order_id, "wim_edit", task)
+                self.lcm_tasks.register(
+                    "wim_account", wim_id, order_id, "wim_edit", task
+                )
                 return
+            elif command == "deleted":
+                return  # TODO cleaning of task just in case should be done
         elif topic == "sdn":
             _sdn_id = params["_id"]
-            if command == "create":
-                task = asyncio.ensure_future(self.sdn.create(params, order_id))
-                self.lcm_tasks.register("sdn", _sdn_id, order_id, "sdn_create", task)
+            if command in ("create", "created"):
+                if not self.main_config.RO.ng:
+                    task = asyncio.ensure_future(self.sdn.create(params, order_id))
+                    self.lcm_tasks.register(
+                        "sdn", _sdn_id, order_id, "sdn_create", task
+                    )
                 return
-            elif command == "delete":
+            elif command == "delete" or command == "deleted":
                 self.lcm_tasks.cancel(topic, _sdn_id)
                 task = asyncio.ensure_future(self.sdn.delete(params, order_id))
                 self.lcm_tasks.register("sdn", _sdn_id, order_id, "sdn_delete", task)
                 return
-            elif command == "edit":
+            elif command in ("edit", "edited"):
                 task = asyncio.ensure_future(self.sdn.edit(params, order_id))
                 self.lcm_tasks.register("sdn", _sdn_id, order_id, "sdn_edit", task)
                 return
+            elif command == "deleted":
+                return  # TODO cleaning of task just in case should be done
         self.logger.critical("unknown topic {} and command '{}'".format(topic, command))
 
     async def kafka_read(self):
-        self.logger.debug("Task kafka_read Enter with worker_id={}".format(self.worker_id))
-        # future = asyncio.Future()
+        self.logger.debug(
+            "Task kafka_read Enter with worker_id={}".format(self.worker_id)
+        )
         self.consecutive_errors = 0
         self.first_start = True
         while self.consecutive_errors < 10:
             try:
-                topics = ("ns", "vim_account", "wim_account", "sdn", "nsi")
-                topics_admin = ("admin", )
+                topics = (
+                    "ns",
+                    "vim_account",
+                    "wim_account",
+                    "sdn",
+                    "nsi",
+                    "k8scluster",
+                    "vca",
+                    "k8srepo",
+                    "pla",
+                    "nslcmops",
+                )
+                topics_admin = ("admin",)
                 await asyncio.gather(
-                    self.msg.aioread(topics, self.loop, self.kafka_read_callback),
-                    self.msg_admin.aioread(topics_admin, self.loop, self.kafka_read_callback, group_id=False)
+                    self.msg.aioread(
+                        topics,
+                        aiocallback=self.kafka_read_callback,
+                        from_beginning=True,
+                    ),
+                    self.msg_admin.aioread(
+                        topics_admin,
+                        aiocallback=self.kafka_read_callback,
+                        group_id=False,
+                    ),
                 )
 
             except LcmExceptionExit:
@@ -425,37 +725,56 @@ class Lcm:
                 # if not first_start is the first time after starting. So leave more time and wait
                 # to allow kafka starts
                 if self.consecutive_errors == 8 if not self.first_start else 30:
-                    self.logger.error("Task kafka_read task exit error too many errors. Exception: {}".format(e))
+                    self.logger.error(
+                        "Task kafka_read task exit error too many errors. Exception: {}".format(
+                            e
+                        )
+                    )
                     raise
                 self.consecutive_errors += 1
-                self.logger.error("Task kafka_read retrying after Exception {}".format(e))
+                self.logger.error(
+                    "Task kafka_read retrying after Exception {}".format(e)
+                )
                 wait_time = 2 if not self.first_start else 5
-                await asyncio.sleep(wait_time, loop=self.loop)
+                await asyncio.sleep(wait_time)
 
-        # self.logger.debug("Task kafka_read terminating")
         self.logger.debug("Task kafka_read exit")
 
-    def start(self):
+    async def kafka_read_ping(self):
+        await asyncio.gather(self.kafka_read(), self.kafka_ping())
 
+    async def start(self):
         # check RO version
-        self.loop.run_until_complete(self.check_RO_version())
+        await self.check_RO_version()
+
+        self.ns = ns.NsLcm(self.msg, self.lcm_tasks, self.main_config)
+        # TODO: modify the rest of classes to use the LcmCfg object instead of dicts
+        self.netslice = netslice.NetsliceLcm(
+            self.msg, self.lcm_tasks, self.main_config.to_dict(), self.ns
+        )
+        self.vim = vim_sdn.VimLcm(self.msg, self.lcm_tasks, self.main_config.to_dict())
+        self.wim = vim_sdn.WimLcm(self.msg, self.lcm_tasks, self.main_config.to_dict())
+        self.sdn = vim_sdn.SdnLcm(self.msg, self.lcm_tasks, self.main_config.to_dict())
+        self.k8scluster = vim_sdn.K8sClusterLcm(
+            self.msg, self.lcm_tasks, self.main_config.to_dict()
+        )
+        self.vca = vim_sdn.VcaLcm(self.msg, self.lcm_tasks, self.main_config.to_dict())
+        self.k8srepo = vim_sdn.K8sRepoLcm(
+            self.msg, self.lcm_tasks, self.main_config.to_dict()
+        )
+
+        await self.kafka_read_ping()
 
-        self.loop.run_until_complete(asyncio.gather(
-            self.kafka_read(),
-            self.kafka_ping()
-        ))
         # TODO
         # self.logger.debug("Terminating cancelling creation tasks")
         # self.lcm_tasks.cancel("ALL", "create")
         # timeout = 200
         # while self.is_pending_tasks():
         #     self.logger.debug("Task kafka_read terminating. Waiting for tasks termination")
-        #     await asyncio.sleep(2, loop=self.loop)
+        #     await asyncio.sleep(2)
         #     timeout -= 2
         #     if not timeout:
         #         self.lcm_tasks.cancel("ALL", "ALL")
-        self.loop.close()
-        self.loop = None
         if self.db:
             self.db.db_disconnect()
         if self.msg:
@@ -466,33 +785,9 @@ class Lcm:
             self.fs.fs_disconnect()
 
     def read_config_file(self, config_file):
-        # TODO make a [ini] + yaml inside parser
-        # the configparser library is not suitable, because it does not admit comments at the end of line,
-        # and not parse integer or boolean
         try:
             with open(config_file) as f:
-                conf = yaml.load(f)
-            for k, v in environ.items():
-                if not k.startswith("OSMLCM_"):
-                    continue
-                k_items = k.lower().split("_")
-                if len(k_items) < 3:
-                    continue
-                if k_items[1] in ("ro", "vca"):
-                    # put in capital letter
-                    k_items[1] = k_items[1].upper()
-                c = conf
-                try:
-                    for k_item in k_items[1:-1]:
-                        c = c[k_item]
-                    if k_items[-1] == "port":
-                        c[k_items[-1]] = int(v)
-                    else:
-                        c[k_items[-1]] = v
-                except Exception as e:
-                    self.logger.warn("skipping environ '{}' on exception '{}'".format(k, e))
-
-            return conf
+                return yaml.safe_load(f)
         except Exception as e:
             self.logger.critical("At config file '{}': {}".format(config_file, e))
             exit(1)
@@ -504,51 +799,50 @@ class Lcm:
         will provide a random one
         :return: Obtained ID
         """
-        # Try getting docker id. If fails, get pid
-        try:
-            with open("/proc/self/cgroup", "r") as f:
-                text_id_ = f.readline()
-                _, _, text_id = text_id_.rpartition("/")
-                text_id = text_id.replace('\n', '')[:12]
-                if text_id:
-                    return text_id
-        except Exception:
-            pass
-        # Return a random id
-        return ''.join(random_choice("0123456789abcdef") for _ in range(12))
+
+        def get_docker_id():
+            try:
+                with open("/proc/self/cgroup", "r") as f:
+                    text_id_ = f.readline()
+                    _, _, text_id = text_id_.rpartition("/")
+                    return text_id.replace("\n", "")[:12]
+            except Exception:
+                return None
+
+        def generate_random_id():
+            return "".join(SystemRandom().choice("0123456789abcdef") for _ in range(12))
+
+        # Try getting docker id. If it fails, generate a random id
+        docker_id = get_docker_id()
+        return docker_id if docker_id else generate_random_id()
 
 
 def usage():
-    print("""Usage: {} [options]
-        -c|--config [configuration_file]: loads the configuration file (default: ./nbi.cfg)
+    print(
+        """Usage: {} [options]
+        -c|--config [configuration_file]: loads the configuration file (default: ./lcm.cfg)
         --health-check: do not run lcm, but inspect kafka bus to determine if lcm is healthy
         -h|--help: shows this help
-        """.format(sys.argv[0]))
+        """.format(
+            sys.argv[0]
+        )
+    )
     # --log-socket-host HOST: send logs to this host")
     # --log-socket-port PORT: send logs using this port (default: 9022)")
 
 
-def health_check():
-    retry = 2
-    while retry:
-        retry -= 1
-        try:
-            with open(health_check_file, "r") as f:
-                last_received_ping = f.read()
-
-            if time() - float(last_received_ping) < Lcm.ping_interval_pace + 10:
-                exit(0)
-        except Exception:
-            pass
-        if retry:
-            sleep(6)
-    exit(1)
-
-
-if __name__ == '__main__':
+if __name__ == "__main__":
     try:
+        # print("SYS.PATH='{}'".format(sys.path))
         # load parameters and configuration
-        opts, args = getopt.getopt(sys.argv[1:], "hc:", ["config=", "help", "health-check"])
+        # -h
+        # -c value
+        # --config value
+        # --help
+        # --health-check
+        opts, args = getopt.getopt(
+            sys.argv[1:], "hc:", ["config=", "help", "health-check"]
+        )
         # TODO add  "log-socket-host=", "log-socket-port=", "log-file="
         config_file = None
         for o, a in opts:
@@ -558,28 +852,36 @@ if __name__ == '__main__':
             elif o in ("-c", "--config"):
                 config_file = a
             elif o == "--health-check":
-                health_check()
-            # elif o == "--log-socket-port":
-            #     log_socket_port = a
-            # elif o == "--log-socket-host":
-            #     log_socket_host = a
-            # elif o == "--log-file":
-            #     log_file = a
+                from osm_lcm.lcm_hc import health_check
+
+                health_check(config_file, Lcm.ping_interval_pace)
             else:
-                assert False, "Unhandled option"
+                print(f"Unhandled option: {o}")
+                exit(1)
+
         if config_file:
             if not path.isfile(config_file):
-                print("configuration file '{}' not exist".format(config_file), file=sys.stderr)
+                print(
+                    "configuration file '{}' does not exist".format(config_file),
+                    file=sys.stderr,
+                )
                 exit(1)
         else:
-            for config_file in (__file__[:__file__.rfind(".")] + ".cfg", "./lcm.cfg", "/etc/osm/lcm.cfg"):
+            for config_file in (
+                __file__[: __file__.rfind(".")] + ".cfg",
+                "./lcm.cfg",
+                "/etc/osm/lcm.cfg",
+            ):
                 if path.isfile(config_file):
                     break
             else:
-                print("No configuration file 'lcm.cfg' found neither at local folder nor at /etc/osm/", file=sys.stderr)
+                print(
+                    "No configuration file 'lcm.cfg' found neither at local folder nor at /etc/osm/",
+                    file=sys.stderr,
+                )
                 exit(1)
         lcm = Lcm(config_file)
-        lcm.start()
+        asyncio.run(lcm.start())
     except (LcmException, getopt.GetoptError) as e:
         print(str(e), file=sys.stderr)
         # usage()