Revert "Bug 1964 fixed: removed the variable cluster_uuid from init_env method"

This reverts commit 25a1392579da2e8e4789e0b8f35abbaa372fde08.

Change-Id: If90fe74a1d7d4b915bbb839ace19999403aa5548
Signed-off-by: David Garcia <david.garcia@canonical.com>
diff --git a/n2vc/k8s_helm_base_conn.py b/n2vc/k8s_helm_base_conn.py
index 176d9fc..b6aa7e9 100644
--- a/n2vc/k8s_helm_base_conn.py
+++ b/n2vc/k8s_helm_base_conn.py
@@ -90,18 +90,14 @@
         if self._stable_repo_url == "None":
             self._stable_repo_url = None
 
-    def _get_namespace(self, cluster_uuid: str) -> str:
+    @staticmethod
+    def _get_namespace_cluster_id(cluster_uuid: str) -> (str, str):
         """
-        Obtains the namespace used by the cluster with the uuid passed by argument
-
-        param: cluster_uuid: cluster's uuid
+        Parses cluster_uuid stored at database that can be either 'namespace:cluster_id' or only
+        cluster_id for backward compatibility
         """
-
-        # first, obtain the cluster corresponding to the uuid passed by argument
-        k8scluster = self.db.get_one(
-            "k8sclusters", q_filter={"_id": cluster_uuid}, fail_on_empty=False
-        )
-        return k8scluster.get("namespace")
+        namespace, _, cluster_id = cluster_uuid.rpartition(":")
+        return namespace, cluster_id
 
     async def init_env(
         self,
@@ -125,9 +121,11 @@
         """
 
         if reuse_cluster_uuid:
-            cluster_id = reuse_cluster_uuid
+            namespace_, cluster_id = self._get_namespace_cluster_id(reuse_cluster_uuid)
+            namespace = namespace_ or namespace
         else:
             cluster_id = str(uuid4())
+        cluster_uuid = "{}:{}".format(namespace, cluster_id)
 
         self.log.debug(
             "Initializing K8S Cluster {}. namespace: {}".format(cluster_id, namespace)
@@ -149,24 +147,25 @@
 
         self.log.info("Cluster {} initialized".format(cluster_id))
 
-        return cluster_id, n2vc_installed_sw
+        return cluster_uuid, n2vc_installed_sw
 
     async def repo_add(
         self, cluster_uuid: str, name: str, url: str, repo_type: str = "chart"
     ):
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
         self.log.debug(
             "Cluster {}, adding {} repository {}. URL: {}".format(
-                cluster_uuid, repo_type, name, url
+                cluster_id, repo_type, name, url
             )
         )
 
         # init_env
         paths, env = self._init_paths_env(
-            cluster_name=cluster_uuid, create_if_not_exist=True
+            cluster_name=cluster_id, create_if_not_exist=True
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # helm repo add name url
         command = "env KUBECONFIG={} {} repo add {} {}".format(
@@ -221,15 +220,16 @@
         :return: list of registered repositories: [ (name, url) .... ]
         """
 
-        self.log.debug("list repositories for cluster {}".format(cluster_uuid))
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
+        self.log.debug("list repositories for cluster {}".format(cluster_id))
 
         # config filename
         paths, env = self._init_paths_env(
-            cluster_name=cluster_uuid, create_if_not_exist=True
+            cluster_name=cluster_id, create_if_not_exist=True
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         command = "env KUBECONFIG={} {} repo list --output yaml".format(
             paths["kube_config"], self._helm_command
@@ -241,7 +241,7 @@
         )
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
         if _rc == 0:
             if output and len(output) > 0:
@@ -254,17 +254,17 @@
             return []
 
     async def repo_remove(self, cluster_uuid: str, name: str):
-        self.log.debug(
-            "remove {} repositories for cluster {}".format(name, cluster_uuid)
-        )
+
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
+        self.log.debug("remove {} repositories for cluster {}".format(name, cluster_id))
 
         # init env, paths
         paths, env = self._init_paths_env(
-            cluster_name=cluster_uuid, create_if_not_exist=True
+            cluster_name=cluster_id, create_if_not_exist=True
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         command = "env KUBECONFIG={} {} repo remove {}".format(
             paths["kube_config"], self._helm_command, name
@@ -274,7 +274,7 @@
         )
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
     async def reset(
         self,
@@ -293,15 +293,15 @@
         :param kwargs: Additional parameters (None yet)
         :return: Returns True if successful or raises an exception.
         """
-        namespace = self._get_namespace(cluster_uuid=cluster_uuid)
+        namespace, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
         self.log.debug(
             "Resetting K8s environment. cluster uuid: {} uninstall={}".format(
-                cluster_uuid, uninstall_sw
+                cluster_id, uninstall_sw
             )
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # uninstall releases if needed.
         if uninstall_sw:
@@ -330,20 +330,20 @@
                 else:
                     msg = (
                         "Cluster uuid: {} has releases and not force. Leaving K8s helm environment"
-                    ).format(cluster_uuid)
+                    ).format(cluster_id)
                     self.log.warn(msg)
                     uninstall_sw = (
                         False  # Allow to remove k8s cluster without removing Tiller
                     )
 
         if uninstall_sw:
-            await self._uninstall_sw(cluster_id=cluster_uuid, namespace=namespace)
+            await self._uninstall_sw(cluster_id, namespace)
 
         # delete cluster directory
-        self.log.debug("Removing directory {}".format(cluster_uuid))
-        self.fs.file_delete(cluster_uuid, ignore_non_exist=True)
+        self.log.debug("Removing directory {}".format(cluster_id))
+        self.fs.file_delete(cluster_id, ignore_non_exist=True)
         # Remove also local directorio if still exist
-        direct = self.fs.path + "/" + cluster_uuid
+        direct = self.fs.path + "/" + cluster_id
         shutil.rmtree(direct, ignore_errors=True)
 
         return True
@@ -461,10 +461,11 @@
         params: dict = None,
         db_dict: dict = None,
     ):
-        self.log.debug("upgrading {} in cluster {}".format(kdu_model, cluster_uuid))
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
+        self.log.debug("upgrading {} in cluster {}".format(kdu_model, cluster_id))
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # look for instance to obtain namespace
         instance_info = await self.get_instance_info(cluster_uuid, kdu_instance)
@@ -473,15 +474,15 @@
 
         # init env, paths
         paths, env = self._init_paths_env(
-            cluster_name=cluster_uuid, create_if_not_exist=True
+            cluster_name=cluster_id, create_if_not_exist=True
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # params to str
         params_str, file_to_delete = self._params_to_file_option(
-            cluster_id=cluster_uuid, params=params
+            cluster_id=cluster_id, params=params
         )
 
         # version
@@ -520,7 +521,7 @@
             # write status in another task
             status_task = asyncio.ensure_future(
                 coro_or_future=self._store_status(
-                    cluster_id=cluster_uuid,
+                    cluster_id=cluster_id,
                     kdu_instance=kdu_instance,
                     namespace=instance_info["namespace"],
                     db_dict=db_dict,
@@ -548,7 +549,7 @@
 
         # write final status
         await self._store_status(
-            cluster_id=cluster_uuid,
+            cluster_id=cluster_id,
             kdu_instance=kdu_instance,
             namespace=instance_info["namespace"],
             db_dict=db_dict,
@@ -563,7 +564,7 @@
             raise K8sException(msg)
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
         # return new revision number
         instance = await self.get_instance_info(
@@ -597,14 +598,16 @@
     async def rollback(
         self, cluster_uuid: str, kdu_instance: str, revision=0, db_dict: dict = None
     ):
+
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
         self.log.debug(
             "rollback kdu_instance {} to revision {} from cluster {}".format(
-                kdu_instance, revision, cluster_uuid
+                kdu_instance, revision, cluster_id
             )
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # look for instance to obtain namespace
         instance_info = await self.get_instance_info(cluster_uuid, kdu_instance)
@@ -613,11 +616,11 @@
 
         # init env, paths
         paths, env = self._init_paths_env(
-            cluster_name=cluster_uuid, create_if_not_exist=True
+            cluster_name=cluster_id, create_if_not_exist=True
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         command = self._get_rollback_command(
             kdu_instance, instance_info["namespace"], revision, paths["kube_config"]
@@ -634,7 +637,7 @@
         # write status in another task
         status_task = asyncio.ensure_future(
             coro_or_future=self._store_status(
-                cluster_id=cluster_uuid,
+                cluster_id=cluster_id,
                 kdu_instance=kdu_instance,
                 namespace=instance_info["namespace"],
                 db_dict=db_dict,
@@ -653,7 +656,7 @@
 
         # write final status
         await self._store_status(
-            cluster_id=cluster_uuid,
+            cluster_id=cluster_id,
             kdu_instance=kdu_instance,
             namespace=instance_info["namespace"],
             db_dict=db_dict,
@@ -668,7 +671,7 @@
             raise K8sException(msg)
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
         # return new revision number
         instance = await self.get_instance_info(
@@ -693,14 +696,13 @@
         :return: True if successful
         """
 
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
         self.log.debug(
-            "uninstall kdu_instance {} from cluster {}".format(
-                kdu_instance, cluster_uuid
-            )
+            "uninstall kdu_instance {} from cluster {}".format(kdu_instance, cluster_id)
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # look for instance to obtain namespace
         instance_info = await self.get_instance_info(cluster_uuid, kdu_instance)
@@ -709,11 +711,11 @@
             return True
         # init env, paths
         paths, env = self._init_paths_env(
-            cluster_name=cluster_uuid, create_if_not_exist=True
+            cluster_name=cluster_id, create_if_not_exist=True
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         command = self._get_uninstall_command(
             kdu_instance, instance_info["namespace"], paths["kube_config"]
@@ -723,7 +725,7 @@
         )
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
         return self._output_to_table(output)
 
@@ -735,16 +737,17 @@
         :return:
         """
 
-        self.log.debug("list releases for cluster {}".format(cluster_uuid))
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
+        self.log.debug("list releases for cluster {}".format(cluster_id))
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # execute internal command
-        result = await self._instances_list(cluster_uuid)
+        result = await self._instances_list(cluster_id)
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
         return result
 
@@ -802,6 +805,7 @@
         - `external_ip` List of external ips (in case they are available)
         """
 
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
         self.log.debug(
             "get_services: cluster_uuid: {}, kdu_instance: {}".format(
                 cluster_uuid, kdu_instance
@@ -810,24 +814,24 @@
 
         # init env, paths
         paths, env = self._init_paths_env(
-            cluster_name=cluster_uuid, create_if_not_exist=True
+            cluster_name=cluster_id, create_if_not_exist=True
         )
 
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # get list of services names for kdu
         service_names = await self._get_services(
-            cluster_uuid, kdu_instance, namespace, paths["kube_config"]
+            cluster_id, kdu_instance, namespace, paths["kube_config"]
         )
 
         service_list = []
         for service in service_names:
-            service = await self._get_service(cluster_uuid, service, namespace)
+            service = await self._get_service(cluster_id, service, namespace)
             service_list.append(service)
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
         return service_list
 
@@ -841,13 +845,15 @@
             )
         )
 
-        # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
 
-        service = await self._get_service(cluster_uuid, service_name, namespace)
+        # sync local dir
+        self.fs.sync(from_path=cluster_id)
+
+        service = await self._get_service(cluster_id, service_name, namespace)
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
         return service
 
@@ -885,11 +891,13 @@
             )
         )
 
+        _, cluster_id = self._get_namespace_cluster_id(cluster_uuid)
+
         # sync local dir
-        self.fs.sync(from_path=cluster_uuid)
+        self.fs.sync(from_path=cluster_id)
 
         # get instance: needed to obtain namespace
-        instances = await self._instances_list(cluster_id=cluster_uuid)
+        instances = await self._instances_list(cluster_id=cluster_id)
         for instance in instances:
             if instance.get("name") == kdu_instance:
                 break
@@ -897,12 +905,12 @@
             # instance does not exist
             raise K8sException(
                 "Instance name: {} not found in cluster: {}".format(
-                    kdu_instance, cluster_uuid
+                    kdu_instance, cluster_id
                 )
             )
 
         status = await self._status_kdu(
-            cluster_id=cluster_uuid,
+            cluster_id=cluster_id,
             kdu_instance=kdu_instance,
             namespace=instance["namespace"],
             yaml_format=yaml_format,
@@ -910,7 +918,7 @@
         )
 
         # sync fs
-        self.fs.reverse_sync(from_path=cluster_uuid)
+        self.fs.reverse_sync(from_path=cluster_id)
 
         return status