X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;ds=sidebyside;f=osm_lcm%2Fns.py;h=02dde158149d8bcaaf9c6f114f5e37861da72421;hb=18ebc3ad0e546dfa7ce649d336bd1c838b38d94c;hp=2a0e8b40c0a3311c4deed90937e5a9a619330982;hpb=fa40e69c55578e4abc363891fd6709cbcc96ed6a;p=osm%2FLCM.git diff --git a/osm_lcm/ns.py b/osm_lcm/ns.py index 2a0e8b4..02dde15 100644 --- a/osm_lcm/ns.py +++ b/osm_lcm/ns.py @@ -28,6 +28,7 @@ from osm_lcm import ROclient from osm_lcm.ng_ro import NgRoClient, NgRoException from osm_lcm.lcm_utils import LcmException, LcmExceptionNoMgmtIP, LcmBase, deep_get, get_iterable, populate_dict from n2vc.k8s_helm_conn import K8sHelmConnector +from n2vc.k8s_helm3_conn import K8sHelm3Connector from n2vc.k8s_juju_conn import K8sJujuConnector from osm_common.dbbase import DbException @@ -133,7 +134,7 @@ class NsLcm(LcmBase): on_update_db=self._on_update_n2vc_db ) - self.k8sclusterhelm = K8sHelmConnector( + self.k8sclusterhelm2 = K8sHelmConnector( kubectl_command=self.vca_config.get("kubectlpath"), helm_command=self.vca_config.get("helmpath"), fs=self.fs, @@ -142,18 +143,30 @@ class NsLcm(LcmBase): on_update_db=None, ) + self.k8sclusterhelm3 = K8sHelm3Connector( + kubectl_command=self.vca_config.get("kubectlpath"), + helm_command=self.vca_config.get("helm3path"), + fs=self.fs, + log=self.logger, + db=self.db, + on_update_db=None, + ) + self.k8sclusterjuju = K8sJujuConnector( kubectl_command=self.vca_config.get("kubectlpath"), juju_command=self.vca_config.get("jujupath"), fs=self.fs, log=self.logger, db=self.db, + loop=self.loop, on_update_db=None, + vca_config=self.vca_config, ) self.k8scluster_map = { - "helm-chart": self.k8sclusterhelm, - "chart": self.k8sclusterhelm, + "helm-chart": self.k8sclusterhelm2, + "helm-chart-v3": self.k8sclusterhelm3, + "chart": self.k8sclusterhelm3, "juju-bundle": self.k8sclusterjuju, "juju": self.k8sclusterjuju, } @@ -162,7 +175,8 @@ class NsLcm(LcmBase): "lxc_proxy_charm": self.n2vc, "native_charm": self.n2vc, "k8s_proxy_charm": self.n2vc, - "helm": self.conn_helm_ee + "helm": self.conn_helm_ee, + "helm-v3": self.conn_helm_ee } self.prometheus = prometheus @@ -1532,7 +1546,7 @@ class NsLcm(LcmBase): ee_id = vca_deployed.get("ee_id") # create or register execution environment in VCA - if vca_type in ("lxc_proxy_charm", "k8s_proxy_charm", "helm"): + if vca_type in ("lxc_proxy_charm", "k8s_proxy_charm", "helm", "helm-v3"): self._write_configuration_status( nsr_id=nsr_id, @@ -1650,7 +1664,7 @@ class NsLcm(LcmBase): # if SSH access is required, then get execution environment SSH public # if native charm we have waited already to VM be UP - if vca_type in ("k8s_proxy_charm", "lxc_proxy_charm", "helm"): + if vca_type in ("k8s_proxy_charm", "lxc_proxy_charm", "helm", "helm-v3"): pub_key = None user = None # self.logger.debug("get ssh key block") @@ -1738,7 +1752,7 @@ class NsLcm(LcmBase): # TODO register in database that primitive is done # STEP 7 Configure metrics - if vca_type == "helm": + if vca_type == "helm" or vca_type == "helm-v3": prometheus_jobs = await self.add_prometheus_metrics( ee_id=ee_id, artifact_path=artifact_path, @@ -2567,7 +2581,7 @@ class NsLcm(LcmBase): async def deploy_kdus(self, logging_text, nsr_id, nslcmop_id, db_vnfrs, db_vnfds, task_instantiation_info): # Launch kdus if present in the descriptor - k8scluster_id_2_uuic = {"helm-chart": {}, "juju-bundle": {}} + k8scluster_id_2_uuic = {"helm-chart-v3": {}, "helm-chart": {}, "juju-bundle": {}} async def _get_cluster_id(cluster_id, cluster_type): nonlocal k8scluster_id_2_uuic @@ -2587,8 +2601,25 @@ class NsLcm(LcmBase): k8s_id = deep_get(db_k8scluster, ("_admin", cluster_type, "id")) if not k8s_id: - raise LcmException("K8s cluster '{}' has not been initialized for '{}'".format(cluster_id, - cluster_type)) + if cluster_type == "helm-chart-v3": + try: + # backward compatibility for existing clusters that have not been initialized for helm v3 + k8s_credentials = yaml.safe_dump(db_k8scluster.get("credentials")) + k8s_id, uninstall_sw = await self.k8sclusterhelm3.init_env(k8s_credentials, + reuse_cluster_uuid=cluster_id) + db_k8scluster_update = {} + db_k8scluster_update["_admin.helm-chart-v3.error_msg"] = None + db_k8scluster_update["_admin.helm-chart-v3.id"] = k8s_id + db_k8scluster_update["_admin.helm-chart-v3.created"] = uninstall_sw + db_k8scluster_update["_admin.helm-chart-v3.operationalState"] = "ENABLED" + self.update_db_2("k8sclusters", cluster_id, db_k8scluster_update) + except Exception as e: + self.logger.error(logging_text + "error initializing helm-v3 cluster: {}".format(str(e))) + raise LcmException("K8s cluster '{}' has not been initialized for '{}'".format(cluster_id, + cluster_type)) + else: + raise LcmException("K8s cluster '{}' has not been initialized for '{}'". + format(cluster_id, cluster_type)) k8scluster_id_2_uuic[cluster_type][cluster_id] = k8s_id return k8s_id @@ -2600,6 +2631,7 @@ class NsLcm(LcmBase): index = 0 updated_cluster_list = [] + updated_v3_cluster_list = [] for vnfr_data in db_vnfrs.values(): for kdu_index, kdur in enumerate(get_iterable(vnfr_data, "kdur")): @@ -2610,7 +2642,11 @@ class NsLcm(LcmBase): namespace = kdur.get("k8s-namespace") if kdur.get("helm-chart"): kdumodel = kdur["helm-chart"] - k8sclustertype = "helm-chart" + # Default version: helm3, if helm-version is v2 assign v2 + k8sclustertype = "helm-chart-v3" + self.logger.debug("kdur: {}".format(kdur)) + if kdur.get("helm-version") and kdur.get("helm-version") == "v2": + k8sclustertype = "helm-chart" elif kdur.get("juju-bundle"): kdumodel = kdur["juju-bundle"] k8sclustertype = "juju-bundle" @@ -2637,18 +2673,25 @@ class NsLcm(LcmBase): cluster_uuid = await _get_cluster_id(k8s_cluster_id, k8sclustertype) # Synchronize repos - if k8sclustertype == "helm-chart" and cluster_uuid not in updated_cluster_list: + if (k8sclustertype == "helm-chart" and cluster_uuid not in updated_cluster_list)\ + or (k8sclustertype == "helm-chart-v3" and cluster_uuid not in updated_v3_cluster_list): del_repo_list, added_repo_dict = await asyncio.ensure_future( - self.k8sclusterhelm.synchronize_repos(cluster_uuid=cluster_uuid)) + self.k8scluster_map[k8sclustertype].synchronize_repos(cluster_uuid=cluster_uuid)) if del_repo_list or added_repo_dict: - unset = {'_admin.helm_charts_added.' + item: None for item in del_repo_list} - updated = {'_admin.helm_charts_added.' + - item: name for item, name in added_repo_dict.items()} - self.logger.debug(logging_text + "repos synchronized on k8s cluster '{}' to_delete: {}, " - "to_add: {}".format(k8s_cluster_id, del_repo_list, - added_repo_dict)) + if k8sclustertype == "helm-chart": + unset = {'_admin.helm_charts_added.' + item: None for item in del_repo_list} + updated = {'_admin.helm_charts_added.' + + item: name for item, name in added_repo_dict.items()} + updated_cluster_list.append(cluster_uuid) + elif k8sclustertype == "helm-chart-v3": + unset = {'_admin.helm_charts_v3_added.' + item: None for item in del_repo_list} + updated = {'_admin.helm_charts_v3_added.' + + item: name for item, name in added_repo_dict.items()} + updated_v3_cluster_list.append(cluster_uuid) + self.logger.debug(logging_text + "repos synchronized on k8s cluster " + "'{}' to_delete: {}, to_add: {}". + format(k8s_cluster_id, del_repo_list, added_repo_dict)) self.db.set_one("k8sclusters", {"_id": k8s_cluster_id}, updated, unset=unset) - updated_cluster_list.append(cluster_uuid) # Instantiate kdu step = "Instantiating KDU {}.{} in k8s cluster {}".format(vnfr_data["member-vnf-index-ref"], @@ -2714,7 +2757,10 @@ class NsLcm(LcmBase): vca_type = "native_charm" elif ee_item.get("helm-chart"): vca_name = ee_item['helm-chart'] - vca_type = "helm" + if ee_item.get("helm-version") and ee_item.get("helm-version") == "v2": + vca_type = "helm" + else: + vca_type = "helm-v3" else: self.logger.debug(logging_text + "skipping non juju neither charm configuration") continue @@ -3328,7 +3374,7 @@ class NsLcm(LcmBase): vca.get("needed_terminate")) # For helm we must destroy_ee. Also for native_charm, as juju_model cannot be deleted if there are # pending native charms - destroy_ee = True if vca_type in ("helm", "native_charm") else False + destroy_ee = True if vca_type in ("helm", "helm-v3", "native_charm") else False # self.logger.debug(logging_text + "vca_index: {}, ee_id: {}, vca_type: {} destroy_ee: {}".format( # vca_index, vca.get("ee_id"), vca_type, destroy_ee)) task = asyncio.ensure_future(