Reformat LCM to standardized format
[osm/LCM.git] / osm_lcm / lcm.py
index dd7c4ec..851ba09 100644 (file)
@@ -37,16 +37,19 @@ from time import time
 from osm_lcm.lcm_utils import versiontuple, LcmException, TaskRegistry, LcmExceptionExit
 from osm_lcm import version as lcm_version, version_date as lcm_version_date
 
-from osm_common import dbmemory, dbmongo, fslocal, fsmongo, msglocal, msgkafka
+from osm_common import msglocal, msgkafka
 from osm_common import version as common_version
 from osm_common.dbbase import DbException
 from osm_common.fsbase import FsException
 from osm_common.msgbase import MsgException
+from osm_lcm.data_utils.database.database import Database
+from osm_lcm.data_utils.filesystem.filesystem import Filesystem
 from os import environ, path
 from random import choice as random_choice
 from n2vc import version as n2vc_version
+import traceback
 
-if os.getenv('OSMLCM_PDB_DEBUG', None) is not None:
+if os.getenv("OSMLCM_PDB_DEBUG", None) is not None:
     pdb.set_trace()
 
 
@@ -55,13 +58,24 @@ min_RO_version = "6.0.2"
 min_n2vc_version = "0.0.2"
 
 min_common_version = "0.1.19"
-health_check_file = path.expanduser("~") + "/time_last_ping"   # TODO find better location for this file
+health_check_file = (
+    path.expanduser("~") + "/time_last_ping"
+)  # TODO find better location for this file
 
 
 class Lcm:
 
-    ping_interval_pace = 120  # how many time ping is send once is confirmed all is running
-    ping_interval_boot = 5    # how many time ping is sent when booting
+    ping_interval_pace = (
+        120  # how many time ping is send once is confirmed all is running
+    )
+    ping_interval_boot = 5  # how many time ping is sent when booting
+    cfg_logger_name = {
+        "message": "lcm.msg",
+        "database": "lcm.db",
+        "storage": "lcm.fs",
+        "tsdb": "lcm.prometheus",
+    }
+    # ^ contains for each section at lcm.cfg the used logger name
 
     def __init__(self, config_file, loop=None):
         """
@@ -78,7 +92,7 @@ class Lcm:
         self.first_start = False
 
         # logging
-        self.logger = logging.getLogger('lcm')
+        self.logger = logging.getLogger("lcm")
         # get id
         self.worker_id = self.get_process_id()
         # load configuration
@@ -92,23 +106,38 @@ class Lcm:
             "loglevel": config["RO"].get("loglevel", "ERROR"),
         }
         if not self.config["ro_config"]["uri"]:
-            if not self.config["ro_config"]["ng"]:
-                self.config["ro_config"]["uri"] = "http://{}:{}/openmano".format(config["RO"]["host"],
-                                                                                 config["RO"]["port"])
-            else:
-                self.config["ro_config"]["uri"] = "http://{}:{}/ro".format(config["RO"]["host"], config["RO"]["port"])
+            self.config["ro_config"]["uri"] = "http://{}:{}/".format(
+                config["RO"]["host"], config["RO"]["port"]
+            )
+        elif (
+            "/ro" in self.config["ro_config"]["uri"][-4:]
+            or "/openmano" in self.config["ro_config"]["uri"][-10:]
+        ):
+            # uri ends with '/ro', '/ro/', '/openmano', '/openmano/'
+            index = self.config["ro_config"]["uri"][-1].rfind("/")
+            self.config["ro_config"]["uri"] = self.config["ro_config"]["uri"][index + 1]
 
         self.loop = loop or asyncio.get_event_loop()
+        self.ns = (
+            self.netslice
+        ) = (
+            self.vim
+        ) = self.wim = self.sdn = self.k8scluster = self.vca = self.k8srepo = None
 
         # logging
-        log_format_simple = "%(asctime)s %(levelname)s %(name)s %(filename)s:%(lineno)s %(message)s"
-        log_formatter_simple = logging.Formatter(log_format_simple, datefmt='%Y-%m-%dT%H:%M:%S')
+        log_format_simple = (
+            "%(asctime)s %(levelname)s %(name)s %(filename)s:%(lineno)s %(message)s"
+        )
+        log_formatter_simple = logging.Formatter(
+            log_format_simple, datefmt="%Y-%m-%dT%H:%M:%S"
+        )
         config["database"]["logger_name"] = "lcm.db"
         config["storage"]["logger_name"] = "lcm.fs"
         config["message"]["logger_name"] = "lcm.msg"
         if config["global"].get("logfile"):
-            file_handler = logging.handlers.RotatingFileHandler(config["global"]["logfile"],
-                                                                maxBytes=100e6, backupCount=9, delay=0)
+            file_handler = logging.handlers.RotatingFileHandler(
+                config["global"]["logfile"], maxBytes=100e6, backupCount=9, delay=0
+            )
             file_handler.setFormatter(log_formatter_simple)
             self.logger.addHandler(file_handler)
         if not config["global"].get("nologging"):
@@ -120,50 +149,42 @@ class Lcm:
             self.logger.setLevel(config["global"]["loglevel"])
 
         # logging other modules
-        for k1, logname in {"message": "lcm.msg", "database": "lcm.db", "storage": "lcm.fs"}.items():
+        for k1, logname in self.cfg_logger_name.items():
             config[k1]["logger_name"] = logname
             logger_module = logging.getLogger(logname)
             if config[k1].get("logfile"):
-                file_handler = logging.handlers.RotatingFileHandler(config[k1]["logfile"],
-                                                                    maxBytes=100e6, backupCount=9, delay=0)
+                file_handler = logging.handlers.RotatingFileHandler(
+                    config[k1]["logfile"], maxBytes=100e6, backupCount=9, delay=0
+                )
                 file_handler.setFormatter(log_formatter_simple)
                 logger_module.addHandler(file_handler)
             if config[k1].get("loglevel"):
                 logger_module.setLevel(config[k1]["loglevel"])
-        self.logger.critical("starting osm/lcm version {} {}".format(lcm_version, lcm_version_date))
+        self.logger.critical(
+            "starting osm/lcm version {} {}".format(lcm_version, lcm_version_date)
+        )
 
         # check version of N2VC
         # TODO enhance with int conversion or from distutils.version import LooseVersion
         # or with list(map(int, version.split(".")))
         if versiontuple(n2vc_version) < versiontuple(min_n2vc_version):
-            raise LcmException("Not compatible osm/N2VC version '{}'. Needed '{}' or higher".format(
-                n2vc_version, min_n2vc_version))
+            raise LcmException(
+                "Not compatible osm/N2VC version '{}'. Needed '{}' or higher".format(
+                    n2vc_version, min_n2vc_version
+                )
+            )
         # check version of common
         if versiontuple(common_version) < versiontuple(min_common_version):
-            raise LcmException("Not compatible osm/common version '{}'. Needed '{}' or higher".format(
-                common_version, min_common_version))
+            raise LcmException(
+                "Not compatible osm/common version '{}'. Needed '{}' or higher".format(
+                    common_version, min_common_version
+                )
+            )
 
         try:
-            # TODO check database version
-            if config["database"]["driver"] == "mongo":
-                self.db = dbmongo.DbMongo()
-                self.db.db_connect(config["database"])
-            elif config["database"]["driver"] == "memory":
-                self.db = dbmemory.DbMemory()
-                self.db.db_connect(config["database"])
-            else:
-                raise LcmException("Invalid configuration param '{}' at '[database]':'driver'".format(
-                    config["database"]["driver"]))
-
-            if config["storage"]["driver"] == "local":
-                self.fs = fslocal.FsLocal()
-                self.fs.fs_connect(config["storage"])
-            elif config["storage"]["driver"] == "mongo":
-                self.fs = fsmongo.FsMongo()
-                self.fs.fs_connect(config["storage"])
-            else:
-                raise LcmException("Invalid configuration param '{}' at '[storage]':'driver'".format(
-                    config["storage"]["driver"]))
+            self.db = Database(config).instance.db
+
+            self.fs = Filesystem(config).instance.fs
 
             # copy message configuration in order to remove 'group_id' for msg_admin
             config_message = config["message"].copy()
@@ -181,52 +202,76 @@ class Lcm:
                 config_message.pop("group_id", None)
                 self.msg_admin.connect(config_message)
             else:
-                raise LcmException("Invalid configuration param '{}' at '[message]':'driver'".format(
-                    config["message"]["driver"]))
+                raise LcmException(
+                    "Invalid configuration param '{}' at '[message]':'driver'".format(
+                        config["message"]["driver"]
+                    )
+                )
         except (DbException, FsException, MsgException) as e:
             self.logger.critical(str(e), exc_info=True)
             raise LcmException(str(e))
 
         # contains created tasks/futures to be able to cancel
-        self.lcm_tasks = TaskRegistry(self.worker_id, self.db, self.logger)
+        self.lcm_tasks = TaskRegistry(self.worker_id, self.logger)
 
-        if self.config.get("prometheus"):
-            self.prometheus = prometheus.Prometheus(self.config["prometheus"], self.worker_id, self.db, self.loop)
+        if self.config.get("tsdb") and self.config["tsdb"].get("driver"):
+            if self.config["tsdb"]["driver"] == "prometheus":
+                self.prometheus = prometheus.Prometheus(
+                    self.config["tsdb"], self.worker_id, self.loop
+                )
+            else:
+                raise LcmException(
+                    "Invalid configuration param '{}' at '[tsdb]':'driver'".format(
+                        config["tsdb"]["driver"]
+                    )
+                )
         else:
             self.prometheus = None
-        self.ns = ns.NsLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.config, self.loop, self.prometheus)
-        self.netslice = netslice.NetsliceLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.config, self.loop,
-                                             self.ns)
-        self.vim = vim_sdn.VimLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.config, self.loop)
-        self.wim = vim_sdn.WimLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.config, self.loop)
-        self.sdn = vim_sdn.SdnLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.config, self.loop)
-        self.k8scluster = vim_sdn.K8sClusterLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.config, self.loop)
-        self.k8srepo = vim_sdn.K8sRepoLcm(self.db, self.msg, self.fs, self.lcm_tasks, self.config, self.loop)
 
     async def check_RO_version(self):
         tries = 14
         last_error = None
         while True:
+            ro_uri = self.config["ro_config"]["uri"]
             try:
-                if self.config["ro_config"].get("ng"):
+                # try new  RO, if fail old RO
+                try:
+                    self.config["ro_config"]["uri"] = ro_uri + "ro"
                     ro_server = NgRoClient(self.loop, **self.config["ro_config"])
-                else:
+                    ro_version = await ro_server.get_version()
+                    self.config["ro_config"]["ng"] = True
+                except Exception:
+                    self.config["ro_config"]["uri"] = ro_uri + "openmano"
                     ro_server = ROClient(self.loop, **self.config["ro_config"])
-                ro_version = await ro_server.get_version()
+                    ro_version = await ro_server.get_version()
+                    self.config["ro_config"]["ng"] = False
                 if versiontuple(ro_version) < versiontuple(min_RO_version):
-                    raise LcmException("Not compatible osm/RO version '{}'. Needed '{}' or higher".format(
-                        ro_version, min_RO_version))
-                self.logger.info("Connected to RO version {}".format(ro_version))
+                    raise LcmException(
+                        "Not compatible osm/RO version '{}'. Needed '{}' or higher".format(
+                            ro_version, min_RO_version
+                        )
+                    )
+                self.logger.info(
+                    "Connected to RO version {} new-generation version {}".format(
+                        ro_version, self.config["ro_config"]["ng"]
+                    )
+                )
                 return
             except (ROClientException, NgRoException) as e:
+                self.config["ro_config"]["uri"] = ro_uri
                 tries -= 1
-                error_text = "Error while connecting to RO on {}: {}".format(self.config["ro_config"]["uri"], e)
+                traceback.print_tb(e.__traceback__)
+                error_text = "Error while connecting to RO on {}: {}".format(
+                    self.config["ro_config"]["uri"], e
+                )
                 if tries <= 0:
                     self.logger.critical(error_text)
                     raise LcmException(error_text)
                 if last_error != error_text:
                     last_error = error_text
-                    self.logger.error(error_text + ". Waiting until {} seconds".format(5*tries))
+                    self.logger.error(
+                        error_text + ". Waiting until {} seconds".format(5 * tries)
+                    )
                 await asyncio.sleep(5)
 
     async def test(self, param=None):
@@ -241,11 +286,22 @@ class Lcm:
         while True:
             try:
                 await self.msg_admin.aiowrite(
-                    "admin", "ping",
-                    {"from": "lcm", "to": "lcm", "worker_id": self.worker_id, "version": lcm_version},
-                    self.loop)
+                    "admin",
+                    "ping",
+                    {
+                        "from": "lcm",
+                        "to": "lcm",
+                        "worker_id": self.worker_id,
+                        "version": lcm_version,
+                    },
+                    self.loop,
+                )
                 # time between pings are low when it is not received and at starting
-                wait_time = self.ping_interval_boot if not kafka_has_received else self.ping_interval_pace
+                wait_time = (
+                    self.ping_interval_boot
+                    if not kafka_has_received
+                    else self.ping_interval_pace
+                )
                 if not self.pings_not_received:
                     kafka_has_received = True
                 self.pings_not_received += 1
@@ -260,10 +316,16 @@ class Lcm:
                 # if not first_start is the first time after starting. So leave more time and wait
                 # to allow kafka starts
                 if consecutive_errors == 8 if not first_start else 30:
-                    self.logger.error("Task kafka_read task exit error too many errors. Exception: {}".format(e))
+                    self.logger.error(
+                        "Task kafka_read task exit error too many errors. Exception: {}".format(
+                            e
+                        )
+                    )
                     raise
                 consecutive_errors += 1
-                self.logger.error("Task kafka_read retrying after Exception {}".format(e))
+                self.logger.error(
+                    "Task kafka_read retrying after Exception {}".format(e)
+                )
                 wait_time = 2 if not first_start else 5
                 await asyncio.sleep(wait_time, loop=self.loop)
 
@@ -271,7 +333,9 @@ class Lcm:
         order_id = 1
 
         if topic != "admin" and command != "ping":
-            self.logger.debug("Task kafka_read receives {} {}: {}".format(topic, command, params))
+            self.logger.debug(
+                "Task kafka_read receives {} {}: {}".format(topic, command, params)
+            )
         self.consecutive_errors = 0
         self.first_start = False
         order_id += 1
@@ -297,7 +361,11 @@ class Lcm:
                     with open(health_check_file, "w") as f:
                         f.write(str(time()))
                 except Exception as e:
-                    self.logger.error("Cannot write into '{}' for healthcheck: {}".format(health_check_file, e))
+                    self.logger.error(
+                        "Cannot write into '{}' for healthcheck: {}".format(
+                            health_check_file, e
+                        )
+                    )
             return
         elif topic == "pla":
             if command == "placement":
@@ -307,24 +375,43 @@ class Lcm:
             if command == "create" or command == "created":
                 k8scluster_id = params.get("_id")
                 task = asyncio.ensure_future(self.k8scluster.create(params, order_id))
-                self.lcm_tasks.register("k8scluster", k8scluster_id, order_id, "k8scluster_create", task)
+                self.lcm_tasks.register(
+                    "k8scluster", k8scluster_id, order_id, "k8scluster_create", task
+                )
                 return
             elif command == "delete" or command == "deleted":
                 k8scluster_id = params.get("_id")
                 task = asyncio.ensure_future(self.k8scluster.delete(params, order_id))
-                self.lcm_tasks.register("k8scluster", k8scluster_id, order_id, "k8scluster_delete", task)
+                self.lcm_tasks.register(
+                    "k8scluster", k8scluster_id, order_id, "k8scluster_delete", task
+                )
+                return
+        elif topic == "vca":
+            if command == "create" or command == "created":
+                vca_id = params.get("_id")
+                task = asyncio.ensure_future(self.vca.create(params, order_id))
+                self.lcm_tasks.register("vca", vca_id, order_id, "vca_create", task)
+                return
+            elif command == "delete" or command == "deleted":
+                vca_id = params.get("_id")
+                task = asyncio.ensure_future(self.vca.delete(params, order_id))
+                self.lcm_tasks.register("vca", vca_id, order_id, "vca_delete", task)
                 return
         elif topic == "k8srepo":
             if command == "create" or command == "created":
                 k8srepo_id = params.get("_id")
                 self.logger.debug("k8srepo_id = {}".format(k8srepo_id))
                 task = asyncio.ensure_future(self.k8srepo.create(params, order_id))
-                self.lcm_tasks.register("k8srepo", k8srepo_id, order_id, "k8srepo_create", task)
+                self.lcm_tasks.register(
+                    "k8srepo", k8srepo_id, order_id, "k8srepo_create", task
+                )
                 return
             elif command == "delete" or command == "deleted":
                 k8srepo_id = params.get("_id")
                 task = asyncio.ensure_future(self.k8srepo.delete(params, order_id))
-                self.lcm_tasks.register("k8srepo", k8srepo_id, order_id, "k8srepo_delete", task)
+                self.lcm_tasks.register(
+                    "k8srepo", k8srepo_id, order_id, "k8srepo_delete", task
+                )
                 return
         elif topic == "ns":
             if command == "instantiate":
@@ -333,7 +420,9 @@ class Lcm:
                 nslcmop_id = nslcmop["_id"]
                 nsr_id = nslcmop["nsInstanceId"]
                 task = asyncio.ensure_future(self.ns.instantiate(nsr_id, nslcmop_id))
-                self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_instantiate", task)
+                self.lcm_tasks.register(
+                    "ns", nsr_id, nslcmop_id, "ns_instantiate", task
+                )
                 return
             elif command == "terminate":
                 # self.logger.debug("Deleting NS {}".format(nsr_id))
@@ -344,6 +433,17 @@ class Lcm:
                 task = asyncio.ensure_future(self.ns.terminate(nsr_id, nslcmop_id))
                 self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "ns_terminate", task)
                 return
+            elif command == "vca_status_refresh":
+                nslcmop = params
+                nslcmop_id = nslcmop["_id"]
+                nsr_id = nslcmop["nsInstanceId"]
+                task = asyncio.ensure_future(
+                    self.ns.vca_status_refresh(nsr_id, nslcmop_id)
+                )
+                self.lcm_tasks.register(
+                    "ns", nsr_id, nslcmop_id, "ns_vca_status_refresh", task
+                )
+                return
             elif command == "action":
                 # self.logger.debug("Update NS {}".format(nsr_id))
                 nslcmop = params
@@ -364,18 +464,30 @@ class Lcm:
                 nsr_id = params
                 try:
                     db_nsr = self.db.get_one("nsrs", {"_id": nsr_id})
-                    print("nsr:\n    _id={}\n    operational-status: {}\n    config-status: {}"
-                          "\n    detailed-status: {}\n    deploy: {}\n    tasks: {}"
-                          "".format(nsr_id, db_nsr["operational-status"], db_nsr["config-status"],
-                                    db_nsr["detailed-status"],
-                                    db_nsr["_admin"]["deployed"], self.lcm_ns_tasks.get(nsr_id)))
+                    print(
+                        "nsr:\n    _id={}\n    operational-status: {}\n    config-status: {}"
+                        "\n    detailed-status: {}\n    deploy: {}\n    tasks: {}"
+                        "".format(
+                            nsr_id,
+                            db_nsr["operational-status"],
+                            db_nsr["config-status"],
+                            db_nsr["detailed-status"],
+                            db_nsr["_admin"]["deployed"],
+                            self.lcm_ns_tasks.get(nsr_id),
+                        )
+                    )
                 except Exception as e:
                     print("nsr {} not found: {}".format(nsr_id, e))
                 sys.stdout.flush()
                 return
             elif command == "deleted":
                 return  # TODO cleaning of task just in case should be done
-            elif command in ("terminated", "instantiated", "scaled", "actioned"):  # "scaled-cooldown-time"
+            elif command in (
+                "terminated",
+                "instantiated",
+                "scaled",
+                "actioned",
+            ):  # "scaled-cooldown-time"
                 return
         elif topic == "nsi":  # netslice LCM processes (instantiate, terminate, etc)
             if command == "instantiate":
@@ -383,8 +495,12 @@ class Lcm:
                 nsilcmop = params
                 nsilcmop_id = nsilcmop["_id"]  # slice operation id
                 nsir_id = nsilcmop["netsliceInstanceId"]  # slice record id
-                task = asyncio.ensure_future(self.netslice.instantiate(nsir_id, nsilcmop_id))
-                self.lcm_tasks.register("nsi", nsir_id, nsilcmop_id, "nsi_instantiate", task)
+                task = asyncio.ensure_future(
+                    self.netslice.instantiate(nsir_id, nsilcmop_id)
+                )
+                self.lcm_tasks.register(
+                    "nsi", nsir_id, nsilcmop_id, "nsi_instantiate", task
+                )
                 return
             elif command == "terminate":
                 # self.logger.debug("Terminating Network Slice NS {}".format(nsilcmop["netsliceInstanceId"]))
@@ -392,57 +508,86 @@ class Lcm:
                 nsilcmop_id = nsilcmop["_id"]  # slice operation id
                 nsir_id = nsilcmop["netsliceInstanceId"]  # slice record id
                 self.lcm_tasks.cancel(topic, nsir_id)
-                task = asyncio.ensure_future(self.netslice.terminate(nsir_id, nsilcmop_id))
-                self.lcm_tasks.register("nsi", nsir_id, nsilcmop_id, "nsi_terminate", task)
+                task = asyncio.ensure_future(
+                    self.netslice.terminate(nsir_id, nsilcmop_id)
+                )
+                self.lcm_tasks.register(
+                    "nsi", nsir_id, nsilcmop_id, "nsi_terminate", task
+                )
                 return
             elif command == "show":
                 nsir_id = params
                 try:
                     db_nsir = self.db.get_one("nsirs", {"_id": nsir_id})
-                    print("nsir:\n    _id={}\n    operational-status: {}\n    config-status: {}"
-                          "\n    detailed-status: {}\n    deploy: {}\n    tasks: {}"
-                          "".format(nsir_id, db_nsir["operational-status"], db_nsir["config-status"],
-                                    db_nsir["detailed-status"],
-                                    db_nsir["_admin"]["deployed"], self.lcm_netslice_tasks.get(nsir_id)))
+                    print(
+                        "nsir:\n    _id={}\n    operational-status: {}\n    config-status: {}"
+                        "\n    detailed-status: {}\n    deploy: {}\n    tasks: {}"
+                        "".format(
+                            nsir_id,
+                            db_nsir["operational-status"],
+                            db_nsir["config-status"],
+                            db_nsir["detailed-status"],
+                            db_nsir["_admin"]["deployed"],
+                            self.lcm_netslice_tasks.get(nsir_id),
+                        )
+                    )
                 except Exception as e:
                     print("nsir {} not found: {}".format(nsir_id, e))
                 sys.stdout.flush()
                 return
             elif command == "deleted":
                 return  # TODO cleaning of task just in case should be done
-            elif command in ("terminated", "instantiated", "scaled", "actioned"):  # "scaled-cooldown-time"
+            elif command in (
+                "terminated",
+                "instantiated",
+                "scaled",
+                "actioned",
+            ):  # "scaled-cooldown-time"
                 return
         elif topic == "vim_account":
             vim_id = params["_id"]
             if command in ("create", "created"):
-                task = asyncio.ensure_future(self.vim.create(params, order_id))
-                self.lcm_tasks.register("vim_account", vim_id, order_id, "vim_create", task)
+                if not self.config["ro_config"].get("ng"):
+                    task = asyncio.ensure_future(self.vim.create(params, order_id))
+                    self.lcm_tasks.register(
+                        "vim_account", vim_id, order_id, "vim_create", task
+                    )
                 return
             elif command == "delete" or command == "deleted":
                 self.lcm_tasks.cancel(topic, vim_id)
                 task = asyncio.ensure_future(self.vim.delete(params, order_id))
-                self.lcm_tasks.register("vim_account", vim_id, order_id, "vim_delete", task)
+                self.lcm_tasks.register(
+                    "vim_account", vim_id, order_id, "vim_delete", task
+                )
                 return
             elif command == "show":
                 print("not implemented show with vim_account")
                 sys.stdout.flush()
                 return
             elif command in ("edit", "edited"):
-                task = asyncio.ensure_future(self.vim.edit(params, order_id))
-                self.lcm_tasks.register("vim_account", vim_id, order_id, "vim_edit", task)
+                if not self.config["ro_config"].get("ng"):
+                    task = asyncio.ensure_future(self.vim.edit(params, order_id))
+                    self.lcm_tasks.register(
+                        "vim_account", vim_id, order_id, "vim_edit", task
+                    )
                 return
             elif command == "deleted":
                 return  # TODO cleaning of task just in case should be done
         elif topic == "wim_account":
             wim_id = params["_id"]
             if command in ("create", "created"):
-                task = asyncio.ensure_future(self.wim.create(params, order_id))
-                self.lcm_tasks.register("wim_account", wim_id, order_id, "wim_create", task)
+                if not self.config["ro_config"].get("ng"):
+                    task = asyncio.ensure_future(self.wim.create(params, order_id))
+                    self.lcm_tasks.register(
+                        "wim_account", wim_id, order_id, "wim_create", task
+                    )
                 return
             elif command == "delete" or command == "deleted":
                 self.lcm_tasks.cancel(topic, wim_id)
                 task = asyncio.ensure_future(self.wim.delete(params, order_id))
-                self.lcm_tasks.register("wim_account", wim_id, order_id, "wim_delete", task)
+                self.lcm_tasks.register(
+                    "wim_account", wim_id, order_id, "wim_delete", task
+                )
                 return
             elif command == "show":
                 print("not implemented show with wim_account")
@@ -450,15 +595,20 @@ class Lcm:
                 return
             elif command in ("edit", "edited"):
                 task = asyncio.ensure_future(self.wim.edit(params, order_id))
-                self.lcm_tasks.register("wim_account", wim_id, order_id, "wim_edit", task)
+                self.lcm_tasks.register(
+                    "wim_account", wim_id, order_id, "wim_edit", task
+                )
                 return
             elif command == "deleted":
                 return  # TODO cleaning of task just in case should be done
         elif topic == "sdn":
             _sdn_id = params["_id"]
             if command in ("create", "created"):
-                task = asyncio.ensure_future(self.sdn.create(params, order_id))
-                self.lcm_tasks.register("sdn", _sdn_id, order_id, "sdn_create", task)
+                if not self.config["ro_config"].get("ng"):
+                    task = asyncio.ensure_future(self.sdn.create(params, order_id))
+                    self.lcm_tasks.register(
+                        "sdn", _sdn_id, order_id, "sdn_create", task
+                    )
                 return
             elif command == "delete" or command == "deleted":
                 self.lcm_tasks.cancel(topic, _sdn_id)
@@ -474,17 +624,36 @@ class Lcm:
         self.logger.critical("unknown topic {} and command '{}'".format(topic, command))
 
     async def kafka_read(self):
-        self.logger.debug("Task kafka_read Enter with worker_id={}".format(self.worker_id))
+        self.logger.debug(
+            "Task kafka_read Enter with worker_id={}".format(self.worker_id)
+        )
         # future = asyncio.Future()
         self.consecutive_errors = 0
         self.first_start = True
         while self.consecutive_errors < 10:
             try:
-                topics = ("ns", "vim_account", "wim_account", "sdn", "nsi", "k8scluster", "k8srepo", "pla")
-                topics_admin = ("admin", )
+                topics = (
+                    "ns",
+                    "vim_account",
+                    "wim_account",
+                    "sdn",
+                    "nsi",
+                    "k8scluster",
+                    "vca",
+                    "k8srepo",
+                    "pla",
+                )
+                topics_admin = ("admin",)
                 await asyncio.gather(
-                    self.msg.aioread(topics, self.loop, self.kafka_read_callback, from_beginning=True),
-                    self.msg_admin.aioread(topics_admin, self.loop, self.kafka_read_callback, group_id=False)
+                    self.msg.aioread(
+                        topics, self.loop, self.kafka_read_callback, from_beginning=True
+                    ),
+                    self.msg_admin.aioread(
+                        topics_admin,
+                        self.loop,
+                        self.kafka_read_callback,
+                        group_id=False,
+                    ),
                 )
 
             except LcmExceptionExit:
@@ -494,10 +663,16 @@ class Lcm:
                 # if not first_start is the first time after starting. So leave more time and wait
                 # to allow kafka starts
                 if self.consecutive_errors == 8 if not self.first_start else 30:
-                    self.logger.error("Task kafka_read task exit error too many errors. Exception: {}".format(e))
+                    self.logger.error(
+                        "Task kafka_read task exit error too many errors. Exception: {}".format(
+                            e
+                        )
+                    )
                     raise
                 self.consecutive_errors += 1
-                self.logger.error("Task kafka_read retrying after Exception {}".format(e))
+                self.logger.error(
+                    "Task kafka_read retrying after Exception {}".format(e)
+                )
                 wait_time = 2 if not self.first_start else 5
                 await asyncio.sleep(wait_time, loop=self.loop)
 
@@ -509,14 +684,30 @@ class Lcm:
         # check RO version
         self.loop.run_until_complete(self.check_RO_version())
 
-        # configure Prometheus
+        self.ns = ns.NsLcm(
+            self.msg, self.lcm_tasks, self.config, self.loop, self.prometheus
+        )
+        self.netslice = netslice.NetsliceLcm(
+            self.msg, self.lcm_tasks, self.config, self.loop, self.ns
+        )
+        self.vim = vim_sdn.VimLcm(self.msg, self.lcm_tasks, self.config, self.loop)
+        self.wim = vim_sdn.WimLcm(self.msg, self.lcm_tasks, self.config, self.loop)
+        self.sdn = vim_sdn.SdnLcm(self.msg, self.lcm_tasks, self.config, self.loop)
+        self.k8scluster = vim_sdn.K8sClusterLcm(
+            self.msg, self.lcm_tasks, self.config, self.loop
+        )
+        self.vca = vim_sdn.VcaLcm(self.msg, self.lcm_tasks, self.config, self.loop)
+        self.k8srepo = vim_sdn.K8sRepoLcm(
+            self.msg, self.lcm_tasks, self.config, self.loop
+        )
+
+        # configure tsdb prometheus
         if self.prometheus:
             self.loop.run_until_complete(self.prometheus.start())
 
-        self.loop.run_until_complete(asyncio.gather(
-            self.kafka_read(),
-            self.kafka_ping()
-        ))
+        self.loop.run_until_complete(
+            asyncio.gather(self.kafka_read(), self.kafka_ping())
+        )
         # TODO
         # self.logger.debug("Terminating cancelling creation tasks")
         # self.lcm_tasks.cancel("ALL", "create")
@@ -547,7 +738,15 @@ class Lcm:
             with open(config_file) as f:
                 conf = yaml.load(f, Loader=yaml.Loader)
             # Ensure all sections are not empty
-            for k in ("global", "timeout", "RO", "VCA", "database", "storage", "message"):
+            for k in (
+                "global",
+                "timeout",
+                "RO",
+                "VCA",
+                "database",
+                "storage",
+                "message",
+            ):
                 if not conf.get(k):
                     conf[k] = {}
 
@@ -567,27 +766,29 @@ class Lcm:
                     else:
                         conf[subject][item] = v
                 except Exception as e:
-                    self.logger.warning("skipping environ '{}' on exception '{}'".format(k, e))
+                    self.logger.warning(
+                        "skipping environ '{}' on exception '{}'".format(k, e)
+                    )
 
             # backward compatibility of VCA parameters
 
-            if 'pubkey' in conf["VCA"]:
-                conf["VCA"]['public_key'] = conf["VCA"].pop('pubkey')
-            if 'cacert' in conf["VCA"]:
-                conf["VCA"]['ca_cert'] = conf["VCA"].pop('cacert')
-            if 'apiproxy' in conf["VCA"]:
-                conf["VCA"]['api_proxy'] = conf["VCA"].pop('apiproxy')
+            if "pubkey" in conf["VCA"]:
+                conf["VCA"]["public_key"] = conf["VCA"].pop("pubkey")
+            if "cacert" in conf["VCA"]:
+                conf["VCA"]["ca_cert"] = conf["VCA"].pop("cacert")
+            if "apiproxy" in conf["VCA"]:
+                conf["VCA"]["api_proxy"] = conf["VCA"].pop("apiproxy")
 
-            if 'enableosupgrade' in conf["VCA"]:
-                conf["VCA"]['enable_os_upgrade'] = conf["VCA"].pop('enableosupgrade')
-            if isinstance(conf["VCA"].get('enable_os_upgrade'), str):
-                if conf["VCA"]['enable_os_upgrade'].lower() == 'false':
-                    conf["VCA"]['enable_os_upgrade'] = False
-                elif conf["VCA"]['enable_os_upgrade'].lower() == 'true':
-                    conf["VCA"]['enable_os_upgrade'] = True
+            if "enableosupgrade" in conf["VCA"]:
+                conf["VCA"]["enable_os_upgrade"] = conf["VCA"].pop("enableosupgrade")
+            if isinstance(conf["VCA"].get("enable_os_upgrade"), str):
+                if conf["VCA"]["enable_os_upgrade"].lower() == "false":
+                    conf["VCA"]["enable_os_upgrade"] = False
+                elif conf["VCA"]["enable_os_upgrade"].lower() == "true":
+                    conf["VCA"]["enable_os_upgrade"] = True
 
-            if 'aptmirror' in conf["VCA"]:
-                conf["VCA"]['apt_mirror'] = conf["VCA"].pop('aptmirror')
+            if "aptmirror" in conf["VCA"]:
+                conf["VCA"]["apt_mirror"] = conf["VCA"].pop("aptmirror")
 
             return conf
         except Exception as e:
@@ -606,26 +807,30 @@ class Lcm:
             with open("/proc/self/cgroup", "r") as f:
                 text_id_ = f.readline()
                 _, _, text_id = text_id_.rpartition("/")
-                text_id = text_id.replace('\n', '')[:12]
+                text_id = text_id.replace("\n", "")[:12]
                 if text_id:
                     return text_id
         except Exception:
             pass
         # Return a random id
-        return ''.join(random_choice("0123456789abcdef") for _ in range(12))
+        return "".join(random_choice("0123456789abcdef") for _ in range(12))
 
 
 def usage():
-    print("""Usage: {} [options]
+    print(
+        """Usage: {} [options]
         -c|--config [configuration_file]: loads the configuration file (default: ./lcm.cfg)
         --health-check: do not run lcm, but inspect kafka bus to determine if lcm is healthy
         -h|--help: shows this help
-        """.format(sys.argv[0]))
+        """.format(
+            sys.argv[0]
+        )
+    )
     # --log-socket-host HOST: send logs to this host")
     # --log-socket-port PORT: send logs using this port (default: 9022)")
 
 
-if __name__ == '__main__':
+if __name__ == "__main__":
 
     try:
         # print("SYS.PATH='{}'".format(sys.path))
@@ -635,7 +840,9 @@ if __name__ == '__main__':
         # --config value
         # --help
         # --health-check
-        opts, args = getopt.getopt(sys.argv[1:], "hc:", ["config=", "help", "health-check"])
+        opts, args = getopt.getopt(
+            sys.argv[1:], "hc:", ["config=", "help", "health-check"]
+        )
         # TODO add  "log-socket-host=", "log-socket-port=", "log-file="
         config_file = None
         for o, a in opts:
@@ -646,6 +853,7 @@ if __name__ == '__main__':
                 config_file = a
             elif o == "--health-check":
                 from osm_lcm.lcm_hc import health_check
+
                 health_check(health_check_file, Lcm.ping_interval_pace)
             # elif o == "--log-socket-port":
             #     log_socket_port = a
@@ -658,14 +866,24 @@ if __name__ == '__main__':
 
         if config_file:
             if not path.isfile(config_file):
-                print("configuration file '{}' does not exist".format(config_file), file=sys.stderr)
+                print(
+                    "configuration file '{}' does not exist".format(config_file),
+                    file=sys.stderr,
+                )
                 exit(1)
         else:
-            for config_file in (__file__[:__file__.rfind(".")] + ".cfg", "./lcm.cfg", "/etc/osm/lcm.cfg"):
+            for config_file in (
+                __file__[: __file__.rfind(".")] + ".cfg",
+                "./lcm.cfg",
+                "/etc/osm/lcm.cfg",
+            ):
                 if path.isfile(config_file):
                     break
             else:
-                print("No configuration file 'lcm.cfg' found neither at local folder nor at /etc/osm/", file=sys.stderr)
+                print(
+                    "No configuration file 'lcm.cfg' found neither at local folder nor at /etc/osm/",
+                    file=sys.stderr,
+                )
                 exit(1)
         lcm = Lcm(config_file)
         lcm.start()