X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=osm_lcm%2Fns.py;h=3b2dc2ed58a3360faee68a9639cfb344c3ae80e8;hb=c95569774f9119c7ad9d53405ca59b836f675134;hp=b172c2f0186da4f7af020da38a2387eff74d419e;hpb=16fedf5d361e74b049007da76ed29c57642e33a0;p=osm%2FLCM.git diff --git a/osm_lcm/ns.py b/osm_lcm/ns.py index b172c2f..3b2dc2e 100644 --- a/osm_lcm/ns.py +++ b/osm_lcm/ns.py @@ -29,7 +29,7 @@ from lcm_utils import LcmException, LcmExceptionNoMgmtIP, LcmBase from osm_common.dbbase import DbException from osm_common.fsbase import FsException -from n2vc.vnf import N2VC, N2VCPrimitiveExecutionFailed, NetworkServiceDoesNotExist +from n2vc.vnf import N2VC, N2VCPrimitiveExecutionFailed, NetworkServiceDoesNotExist, PrimitiveDoesNotExist from copy import copy, deepcopy from http import HTTPStatus @@ -628,6 +628,12 @@ class NsLcm(LcmBase): raise LcmException("ns_update_vnfr: Not found member_vnf_index={} at RO info".format(vnf_index)) async def instantiate(self, nsr_id, nslcmop_id): + + # Try to lock HA task here + task_is_locked_by_me = self.lcm_tasks.lock_HA('ns', 'nslcmops', nslcmop_id) + if not task_is_locked_by_me: + return + logging_text = "Task ns={} instantiate={} ".format(nsr_id, nslcmop_id) self.logger.debug(logging_text + "Enter") # get all needed from database @@ -641,8 +647,12 @@ class NsLcm(LcmBase): RO_descriptor_number = 0 # number of descriptors created at RO vnf_index_2_RO_id = {} # map between vnfd/nsd id to the id used at RO n2vc_info = {} + n2vc_key_list = [] # list of public keys to be injected as authorized to VMs exc = None try: + # wait for any previous tasks in process + await self.lcm_tasks.waitfor_related_HA('ns', 'nslcmops', nslcmop_id) + step = "Getting nslcmop={} from db".format(nslcmop_id) db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) step = "Getting nsr={} from db".format(nsr_id) @@ -651,21 +661,11 @@ class NsLcm(LcmBase): nsd = db_nsr["nsd"] nsr_name = db_nsr["name"] # TODO short-name?? - # look if previous tasks in process - task_name, task_dependency = self.lcm_tasks.lookfor_related("ns", nsr_id, nslcmop_id) - if task_dependency: - step = db_nslcmop_update["detailed-status"] = \ - "Waiting for related tasks to be completed: {}".format(task_name) - self.logger.debug(logging_text + step) - self.update_db_2("nslcmops", nslcmop_id, db_nslcmop_update) - _, pending = await asyncio.wait(task_dependency, timeout=3600) - if pending: - raise LcmException("Timeout waiting related tasks to be completed") - step = "Getting vnfrs from db" db_vnfrs_list = self.db.get_list("vnfrs", {"nsr-id-ref": nsr_id}) db_vnfds_ref = {} db_vnfds = {} + db_vnfds_index = {} for vnfr in db_vnfrs_list: db_vnfrs[vnfr["member-vnf-index-ref"]] = vnfr vnfd_id = vnfr["vnfd-id"] @@ -675,6 +675,7 @@ class NsLcm(LcmBase): vnfd = self.db.get_one("vnfds", {"_id": vnfd_id}) db_vnfds_ref[vnfd_ref] = vnfd db_vnfds[vnfd_id] = vnfd + db_vnfds_index[vnfr["member-vnf-index-ref"]] = db_vnfds[vnfd_id] # Get or generates the _admin.deployed,VCA list vca_deployed_list = None @@ -685,10 +686,12 @@ class NsLcm(LcmBase): if vca_deployed_list is None: vca_deployed_list = [] db_nsr_update["_admin.deployed.VCA"] = vca_deployed_list + populate_dict(db_nsr, ("_admin", "deployed", "VCA"), vca_deployed_list) elif isinstance(vca_deployed_list, dict): # maintain backward compatibility. Change a dict to list at database vca_deployed_list = list(vca_deployed_list.values()) db_nsr_update["_admin.deployed.VCA"] = vca_deployed_list + populate_dict(db_nsr, ("_admin", "deployed", "VCA"), vca_deployed_list) db_nsr_update["detailed-status"] = "creating" db_nsr_update["operational-status"] = "init" @@ -697,18 +700,347 @@ class NsLcm(LcmBase): populate_dict(db_nsr, ("_admin", "deployed", "RO", "vnfd"), []) db_nsr_update["_admin.deployed.RO.vnfd"] = [] - RO = ROclient.ROClient(self.loop, **self.ro_config) - # set state to INSTANTIATED. When instantiated NBI will not delete directly db_nsr_update["_admin.nsState"] = "INSTANTIATED" self.update_db_2("nsrs", nsr_id, db_nsr_update) + # Deploy charms + # The parameters we'll need to deploy a charm + number_to_configure = 0 + + def deploy_charm(vnf_index, vdu_id, vdu_name, vdu_count_index, charm_params, n2vc_info, native_charm=False): + """An inner function to deploy the charm from either ns, vnf or vdu + For ns both vnf_index and vdu_id are None. + For vnf only vdu_id is None + For vdu both vnf_index and vdu_id contain a value + """ + # if not charm_params.get("rw_mgmt_ip") and vnf_index: # if NS skip mgmt_ip checking + # raise LcmException("ns/vnfd/vdu has not management ip address to configure it") + + machine_spec = {} + if native_charm: + machine_spec["username"] = charm_params.get("username"), + machine_spec["hostname"] = charm_params.get("rw_mgmt_ip") + + # Note: The charm needs to exist on disk at the location + # specified by charm_path. + descriptor = vnfd if vnf_index else nsd + base_folder = descriptor["_admin"]["storage"] + storage_params = self.fs.get_params() + charm_path = "{}{}/{}/charms/{}".format( + storage_params["path"], + base_folder["folder"], + base_folder["pkg-dir"], + proxy_charm + ) + + # ns_name will be ignored in the current version of N2VC + # but will be implemented for the next point release. + model_name = nsr_id + vdu_id_text = (str(vdu_id) if vdu_id else "") + "-" + vnf_index_text = (str(vnf_index) if vnf_index else "") + "-" + application_name = self.n2vc.FormatApplicationName(nsr_name, vnf_index_text, vdu_id_text) + + vca_index = len(vca_deployed_list) + # trunk name and add two char index at the end to ensure that it is unique. It is assumed no more than + # 26*26 charm in the same NS + application_name = application_name[0:48] + application_name += chr(97 + vca_index // 26) + chr(97 + vca_index % 26) + vca_deployed_ = { + "member-vnf-index": vnf_index, + "vdu_id": vdu_id, + "model": model_name, + "application": application_name, + "operational-status": "init", + "detailed-status": "", + "step": "initial-deploy", + "vnfd_id": vnfd_id, + "vdu_name": vdu_name, + "vdu_count_index": vdu_count_index, + } + vca_deployed_list.append(vca_deployed_) + db_nsr_update["_admin.deployed.VCA.{}".format(vca_index)] = vca_deployed_ + self.update_db_2("nsrs", nsr_id, db_nsr_update) + + self.logger.debug("Task create_ns={} Passing artifacts path '{}' for {}".format(nsr_id, charm_path, + proxy_charm)) + if not n2vc_info: + n2vc_info["nsr_id"] = nsr_id + n2vc_info["nslcmop_id"] = nslcmop_id + n2vc_info["n2vc_event"] = asyncio.Event(loop=self.loop) + n2vc_info["lcmOperationType"] = "instantiate" + n2vc_info["deployed"] = vca_deployed_list + n2vc_info["db_update"] = db_nsr_update + task = asyncio.ensure_future( + self.n2vc.DeployCharms( + model_name, # The network service name + application_name, # The application name + descriptor, # The vnf/nsd descriptor + charm_path, # Path to charm + charm_params, # Runtime params, like mgmt ip + machine_spec, # for native charms only + self.n2vc_callback, # Callback for status changes + n2vc_info, # Callback parameter + None, # Callback parameter (task) + ) + ) + task.add_done_callback(functools.partial(self.n2vc_callback, model_name, application_name, None, None, + n2vc_info)) + self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "create_charm:" + application_name, task) + + step = "Looking for needed vnfd to configure with proxy charm" + self.logger.debug(logging_text + step) + + for c_vnf in get_iterable(nsd, "constituent-vnfd"): + vnfd_id = c_vnf["vnfd-id-ref"] + vnf_index = str(c_vnf["member-vnf-index"]) + vnfd = db_vnfds_ref[vnfd_id] + + # Get additional parameters + vnfr_params = {} + if db_vnfrs[vnf_index].get("additionalParamsForVnf"): + vnfr_params = db_vnfrs[vnf_index]["additionalParamsForVnf"].copy() + for k, v in vnfr_params.items(): + if isinstance(v, str) and v.startswith("!!yaml "): + vnfr_params[k] = yaml.safe_load(v[7:]) + + step = "deploying proxy charms for configuration" + # Check if this VNF has a charm configuration + vnf_config = vnfd.get("vnf-configuration") + if vnf_config and vnf_config.get("juju"): + proxy_charm = vnf_config["juju"]["charm"] + if vnf_config["juju"].get("proxy") is False: + # native_charm, will be deployed after VM. Skip + proxy_charm = None + + if proxy_charm: + if not vca_model_name: + step = "creating VCA model name '{}'".format(nsr_id) + self.logger.debug(logging_text + step) + await self.n2vc.CreateNetworkService(nsr_id) + vca_model_name = nsr_id + db_nsr_update["_admin.deployed.VCA-model-name"] = nsr_id + self.update_db_2("nsrs", nsr_id, db_nsr_update) + step = "deploying proxy charm to configure vnf {}".format(vnf_index) + vnfr_params["rw_mgmt_ip"] = db_vnfrs[vnf_index]["ip-address"] + charm_params = { + "user_values": vnfr_params, + "rw_mgmt_ip": db_vnfrs[vnf_index]["ip-address"], + "initial-config-primitive": {} # vnf_config.get('initial-config-primitive') or {} + } + + # Login to the VCA. If there are multiple calls to login(), + # subsequent calls will be a nop and return immediately. + await self.n2vc.login() + + deploy_charm(vnf_index, None, None, None, charm_params, n2vc_info) + number_to_configure += 1 + + # Deploy charms for each VDU that supports one. + for vdu_index, vdu in enumerate(get_iterable(vnfd, 'vdu')): + vdu_config = vdu.get('vdu-configuration') + proxy_charm = None + + if vdu_config and vdu_config.get("juju"): + proxy_charm = vdu_config["juju"]["charm"] + if vdu_config["juju"].get("proxy") is False: + # native_charm, will be deployed after VM. Skip + proxy_charm = None + + if proxy_charm: + if not vca_model_name: + step = "creating VCA model name" + await self.n2vc.CreateNetworkService(nsr_id) + vca_model_name = nsr_id + db_nsr_update["_admin.deployed.VCA-model-name"] = nsr_id + self.update_db_2("nsrs", nsr_id, db_nsr_update) + step = "deploying proxy charm to configure member_vnf_index={} vdu={}".format(vnf_index, + vdu["id"]) + await self.n2vc.login() + vdur = db_vnfrs[vnf_index]["vdur"][vdu_index] + # TODO for the moment only first vdu_id contains a charm deployed + if vdur["vdu-id-ref"] != vdu["id"]: + raise LcmException("Mismatch vdur {}, vdu {} at index {} for member_vnf_index={}" + .format(vdur["vdu-id-ref"], vdu["id"], vdu_index, vnf_index)) + vnfr_params["rw_mgmt_ip"] = vdur["ip-address"] + charm_params = { + "user_values": vnfr_params, + "rw_mgmt_ip": vdur["ip-address"], + "initial-config-primitive": {} # vdu_config.get('initial-config-primitive') or {} + } + deploy_charm(vnf_index, vdu["id"], vdur.get("name"), vdur["count-index"], + charm_params, n2vc_info) + number_to_configure += 1 + + # Check if this NS has a charm configuration + + ns_config = nsd.get("ns-configuration") + if ns_config and ns_config.get("juju"): + proxy_charm = ns_config["juju"]["charm"] + if ns_config["juju"].get("proxy") is False: + # native_charm, will be deployed after VM. Skip + proxy_charm = None + + if proxy_charm: + step = "deploying proxy charm to configure ns" + # TODO is NS magmt IP address needed? + + # Get additional parameters + additional_params = {} + if db_nsr.get("additionalParamsForNs"): + additional_params = db_nsr["additionalParamsForNs"].copy() + for k, v in additional_params.items(): + if isinstance(v, str) and v.startswith("!!yaml "): + additional_params[k] = yaml.safe_load(v[7:]) + + # additional_params["rw_mgmt_ip"] = db_nsr["ip-address"] + charm_params = { + "user_values": additional_params, + # "rw_mgmt_ip": db_nsr["ip-address"], + "initial-config-primitive": {} # ns_config.get('initial-config-primitive') or {} + } + + # Login to the VCA. If there are multiple calls to login(), + # subsequent calls will be a nop and return immediately. + await self.n2vc.login() + deploy_charm(None, None, None, None, charm_params, n2vc_info) + number_to_configure += 1 + + db_nsr_update["operational-status"] = "running" + + # Wait until all charms has reached blocked or active status + step = "waiting proxy charms to be ready" + if number_to_configure: + # wait until all charms are configured. + # steps are: + # initial-deploy + # get-ssh-public-key + # generate-ssh-key + # retry-get-ssh-public-key + # ssh-public-key-obtained + while time() <= start_deploy + self.total_deploy_timeout: + if db_nsr_update: + self.update_db_2("nsrs", nsr_id, db_nsr_update) + if db_nslcmop_update: + self.update_db_2("nslcmops", nslcmop_id, db_nslcmop_update) + + all_active = True + for vca_index, vca_deployed in enumerate(vca_deployed_list): + database_entry = "_admin.deployed.VCA.{}.".format(vca_index) + if vca_deployed["step"] == "initial-deploy": + if vca_deployed["operational-status"] in ("active", "blocked"): + step = "execute charm primitive get-ssh-public-key for member_vnf_index={} vdu_id={}" \ + .format(vca_deployed["member-vnf-index"], + vca_deployed["vdu_id"]) + self.logger.debug(logging_text + step) + try: + primitive_id = await self.n2vc.ExecutePrimitive( + vca_deployed["model"], + vca_deployed["application"], + "get-ssh-public-key", + None, + ) + vca_deployed["step"] = db_nsr_update[database_entry + "step"] = "get-ssh-public-key" + vca_deployed["primitive_id"] = db_nsr_update[database_entry + "primitive_id"] =\ + primitive_id + db_nsr_update[database_entry + "operational-status"] =\ + vca_deployed["operational-status"] + except PrimitiveDoesNotExist: + ssh_public_key = None + vca_deployed["step"] = db_nsr_update[database_entry + "step"] =\ + "ssh-public-key-obtained" + vca_deployed["ssh-public-key"] = db_nsr_update[database_entry + "ssh-public-key"] =\ + ssh_public_key + step = "charm ssh-public-key for member_vnf_index={} vdu_id={} not needed".format( + vca_deployed["member-vnf-index"], vca_deployed["vdu_id"]) + self.logger.debug(logging_text + step) + + elif vca_deployed["step"] in ("get-ssh-public-key", "retry-get-ssh-public-key"): + primitive_id = vca_deployed["primitive_id"] + primitive_status = await self.n2vc.GetPrimitiveStatus(vca_deployed["model"], + primitive_id) + if primitive_status in ("completed", "failed"): + primitive_result = await self.n2vc.GetPrimitiveOutput(vca_deployed["model"], + primitive_id) + vca_deployed["primitive_id"] = db_nsr_update[database_entry + "primitive_id"] = None + if primitive_status == "completed" and isinstance(primitive_result, dict) and \ + primitive_result.get("pubkey"): + ssh_public_key = primitive_result.get("pubkey") + vca_deployed["step"] = db_nsr_update[database_entry + "step"] =\ + "ssh-public-key-obtained" + vca_deployed["ssh-public-key"] = db_nsr_update[database_entry + "ssh-public-key"] =\ + ssh_public_key + n2vc_key_list.append(ssh_public_key) + step = "charm ssh-public-key for member_vnf_index={} vdu_id={} is '{}'".format( + vca_deployed["member-vnf-index"], vca_deployed["vdu_id"], ssh_public_key) + self.logger.debug(logging_text + step) + else: # primitive_status == "failed": + if vca_deployed["step"] == "get-ssh-public-key": + step = "execute charm primitive generate-ssh-public-key for member_vnf_index="\ + "{} vdu_id={}".format(vca_deployed["member-vnf-index"], + vca_deployed["vdu_id"]) + self.logger.debug(logging_text + step) + vca_deployed["step"] = db_nsr_update[database_entry + "step"] =\ + "generate-ssh-key" + primitive_id = await self.n2vc.ExecutePrimitive( + vca_deployed["model"], + vca_deployed["application"], + "generate-ssh-key", + None, + ) + vca_deployed["primitive_id"] = db_nsr_update[database_entry + "primitive_id"] =\ + primitive_id + else: # failed for second time + raise LcmException( + "error executing primitive get-ssh-public-key: {}".format(primitive_result)) + + elif vca_deployed["step"] == "generate-ssh-key": + primitive_id = vca_deployed["primitive_id"] + primitive_status = await self.n2vc.GetPrimitiveStatus(vca_deployed["model"], + primitive_id) + if primitive_status in ("completed", "failed"): + primitive_result = await self.n2vc.GetPrimitiveOutput(vca_deployed["model"], + primitive_id) + vca_deployed["primitive_id"] = db_nsr_update[ + database_entry + "primitive_id"] = None + if primitive_status == "completed": + step = "execute primitive get-ssh-public-key again for member_vnf_index={} "\ + "vdu_id={}".format(vca_deployed["member-vnf-index"], + vca_deployed["vdu_id"]) + self.logger.debug(logging_text + step) + vca_deployed["step"] = db_nsr_update[database_entry + "step"] = \ + "retry-get-ssh-public-key" + primitive_id = await self.n2vc.ExecutePrimitive( + vca_deployed["model"], + vca_deployed["application"], + "get-ssh-public-key", + None, + ) + vca_deployed["primitive_id"] = db_nsr_update[database_entry + "primitive_id"] =\ + primitive_id + + else: # primitive_status == "failed": + raise LcmException("error executing primitive generate-ssh-key: {}" + .format(primitive_result)) + + if vca_deployed["step"] != "ssh-public-key-obtained": + all_active = False + + if all_active: + break + await asyncio.sleep(5) + else: # total_deploy_timeout + raise LcmException("Timeout waiting charm to be initialized for member_vnf_index={} vdu_id={}" + .format(vca_deployed["member-vnf-index"], vca_deployed["vdu_id"])) + + # deploy RO + RO = ROclient.ROClient(self.loop, **self.ro_config) # get vnfds, instantiate at RO for c_vnf in nsd.get("constituent-vnfd", ()): member_vnf_index = c_vnf["member-vnf-index"] vnfd = db_vnfds_ref[c_vnf['vnfd-id-ref']] vnfd_ref = vnfd["id"] - step = db_nsr_update["detailed-status"] = "Creating vnfd='{}' member-vnf-index='{}' at RO".format( + step = db_nsr_update["detailed-status"] = "Creating vnfd='{}' member_vnf_index='{}' at RO".format( vnfd_ref, member_vnf_index) # self.logger.debug(logging_text + step) vnfd_id_RO = "{}.{}.{}".format(nsr_id, RO_descriptor_number, member_vnf_index[:23]) @@ -728,14 +1060,14 @@ class NsLcm(LcmBase): vnfd_list = await RO.get_list("vnfd", filter_by={"osm_id": vnfd_id_RO}) if vnfd_list: RO_update["id"] = vnfd_list[0]["uuid"] - self.logger.debug(logging_text + "vnfd='{}' member-vnf-index='{}' exists at RO. Using RO_id={}". + self.logger.debug(logging_text + "vnfd='{}' member_vnf_index='{}' exists at RO. Using RO_id={}". format(vnfd_ref, member_vnf_index, vnfd_list[0]["uuid"])) else: vnfd_RO = self.vnfd2RO(vnfd, vnfd_id_RO, db_vnfrs[c_vnf["member-vnf-index"]]. get("additionalParamsForVnf"), nsr_id) desc = await RO.create("vnfd", descriptor=vnfd_RO) RO_update["id"] = desc["uuid"] - self.logger.debug(logging_text + "vnfd='{}' member-vnf-index='{}' created at RO. RO_id={}".format( + self.logger.debug(logging_text + "vnfd='{}' member_vnf_index='{}' created at RO. RO_id={}".format( vnfd_ref, member_vnf_index, desc["uuid"])) db_nsr_update["_admin.deployed.RO.vnfd.{}".format(index)] = RO_update db_nsr["_admin"]["deployed"]["RO"]["vnfd"][index] = RO_update @@ -816,7 +1148,8 @@ class NsLcm(LcmBase): # feature 1429. Add n2vc public key to needed VMs n2vc_key = await self.n2vc.GetPublicKey() - RO_ns_params = self.ns_params_2_RO(ns_params, nsd, db_vnfds_ref, [n2vc_key]) + n2vc_key_list.append(n2vc_key) + RO_ns_params = self.ns_params_2_RO(ns_params, nsd, db_vnfds_ref, n2vc_key_list) step = db_nsr_update["detailed-status"] = "Creating ns at RO" desc = await RO.create("ns", descriptor=RO_ns_params, @@ -860,95 +1193,74 @@ class NsLcm(LcmBase): step = "Updating NSR" self.ns_update_nsr(db_nsr_update, db_nsr, desc) + db_nsr_update["operational-status"] = "running" db_nsr["detailed-status"] = "Configuring vnfr" self.update_db_2("nsrs", nsr_id, db_nsr_update) - # The parameters we'll need to deploy a charm - number_to_configure = 0 - - def deploy_charm(vnf_index, vdu_id, vdu_name, vdu_count_index, charm_params, n2vc_info): - """An inner function to deploy the charm from either ns, vnf or vdu - For ns both vnf_index and vdu_id are None. - For vnf only vdu_id is None - For vdu both vnf_index and vdu_id contain a value - """ - if not charm_params.get("rw_mgmt_ip") and vnf_index: # if NS skip mgmt_ip checking - raise LcmException("ns/vnfd/vdu has not management ip address to configure it") - # Login to the VCA. - # if number_to_configure == 0: - # self.logger.debug("Logging into N2VC...") - # task = asyncio.ensure_future(self.n2vc.login()) - # yield from asyncio.wait_for(task, 30.0) - # self.logger.debug("Logged into N2VC!") - - # # await self.n2vc.login() - - # Note: The charm needs to exist on disk at the location - # specified by charm_path. - descriptor = vnfd if vnf_index else nsd - base_folder = descriptor["_admin"]["storage"] - storage_params = self.fs.get_params() - charm_path = "{}{}/{}/charms/{}".format( - storage_params["path"], - base_folder["folder"], - base_folder["pkg-dir"], - proxy_charm - ) - - # ns_name will be ignored in the current version of N2VC - # but will be implemented for the next point release. - model_name = nsr_id - vdu_id_text = (str(vdu_id) if vdu_id else "") + "-" - vnf_index_text = (str(vnf_index) if vnf_index else "") + "-" - application_name = self.n2vc.FormatApplicationName(nsr_name, vnf_index_text, vdu_id_text) - - vca_index = len(vca_deployed_list) - # trunk name and add two char index at the end to ensure that it is unique. It is assumed no more than - # 26*26 charm in the same NS - application_name = application_name[0:48] - application_name += chr(97 + vca_index // 26) + chr(97 + vca_index % 26) - vca_deployed_ = { - "member-vnf-index": vnf_index, - "vdu_id": vdu_id, - "model": model_name, - "application": application_name, - "operational-status": "init", - "detailed-status": "", - "vnfd_id": vnfd_id, - "vdu_name": vdu_name, - "vdu_count_index": vdu_count_index, - } - vca_deployed_list.append(vca_deployed_) - db_nsr_update["_admin.deployed.VCA.{}".format(vca_index)] = vca_deployed_ - self.update_db_2("nsrs", nsr_id, db_nsr_update) - - self.logger.debug("Task create_ns={} Passing artifacts path '{}' for {}".format(nsr_id, charm_path, - proxy_charm)) - if not n2vc_info: - n2vc_info["nsr_id"] = nsr_id - n2vc_info["nslcmop_id"] = nslcmop_id - n2vc_info["n2vc_event"] = asyncio.Event(loop=self.loop) - n2vc_info["lcmOperationType"] = "instantiate" - n2vc_info["deployed"] = vca_deployed_list - n2vc_info["db_update"] = db_nsr_update - task = asyncio.ensure_future( - self.n2vc.DeployCharms( - model_name, # The network service name - application_name, # The application name - descriptor, # The vnf/nsd descriptor - charm_path, # Path to charm - charm_params, # Runtime params, like mgmt ip - {}, # for native charms only - self.n2vc_callback, # Callback for status changes - n2vc_info, # Callback parameter - None, # Callback parameter (task) - ) - ) - task.add_done_callback(functools.partial(self.n2vc_callback, model_name, application_name, None, None, - n2vc_info)) - self.lcm_tasks.register("ns", nsr_id, nslcmop_id, "create_charm:" + application_name, task) - - step = "Looking for needed vnfd to configure" + # Configure proxy charms once VMs are up + for vca_index, vca_deployed in enumerate(vca_deployed_list): + vnf_index = vca_deployed.get("member-vnf-index") + vdu_id = vca_deployed.get("vdu_id") + vdu_name = None + vdu_count_index = None + + step = "executing proxy charm initial primitives for member_vnf_index={} vdu_id={}".format(vnf_index, + vdu_id) + add_params = {} + initial_config_primitive_list = [] + if vnf_index: + if db_vnfrs[vnf_index].get("additionalParamsForVnf"): + add_params = db_vnfrs[vnf_index]["additionalParamsForVnf"].copy() + vnfd = db_vnfds_index[vnf_index] + + if vdu_id: + for vdu_index, vdu in enumerate(get_iterable(vnfd, 'vdu')): + if vdu["id"] == vdu_id: + initial_config_primitive_list = vdu['vdu-configuration'].get( + 'initial-config-primitive', []) + break + else: + raise LcmException("Not found vdu_id={} at vnfd:vdu".format(vdu_id)) + vdur = db_vnfrs[vnf_index]["vdur"][vdu_index] + # TODO for the moment only first vdu_id contains a charm deployed + if vdur["vdu-id-ref"] != vdu["id"]: + raise LcmException("Mismatch vdur {}, vdu {} at index {} for vnf {}" + .format(vdur["vdu-id-ref"], vdu["id"], vdu_index, vnf_index)) + add_params["rw_mgmt_ip"] = vdur["ip-address"] + else: + add_params["rw_mgmt_ip"] = db_vnfrs[vnf_index]["ip-address"] + initial_config_primitive_list = vnfd["vnf-configuration"].get('initial-config-primitive', []) + else: + if db_nsr.get("additionalParamsForNs"): + add_params = db_nsr["additionalParamsForNs"].copy() + for k, v in add_params.items(): + if isinstance(v, str) and v.startswith("!!yaml "): + add_params[k] = yaml.safe_load(v[7:]) + add_params["rw_mgmt_ip"] = None + initial_config_primitive_list = nsd["ns-configuration"].get('initial-config-primitive', []) + + # add primitive verify-ssh-credentials to the list after config only when is a vnf or vdu charm + initial_config_primitive_list = initial_config_primitive_list.copy() + if initial_config_primitive_list and vnf_index and vca_deployed.get("ssh-public-key"): + initial_config_primitive_list.insert(1, {"name": "verify-ssh-credentials", "parameter": []}) + + for initial_config_primitive in initial_config_primitive_list: + primitive_params_ = self._map_primitive_params(initial_config_primitive, {}, add_params) + self.logger.debug(logging_text + step + " primitive '{}' params '{}'" + .format(initial_config_primitive["name"], primitive_params_)) + primitive_result, primitive_detail = await self._ns_execute_primitive( + db_nsr["_admin"]["deployed"], vnf_index, vdu_id, vdu_name, vdu_count_index, + initial_config_primitive["name"], + primitive_params_, + retries=10 if initial_config_primitive["name"] == "verify-ssh-credentials" else 0, + retries_interval=30) + if primitive_result != "COMPLETED": + raise LcmException("charm error executing primitive {} for member_vnf_index={} vdu_id={}: '{}'" + .format(initial_config_primitive["name"], vca_deployed["member-vnf-index"], + vca_deployed["vdu_id"], primitive_detail)) + + # Deploy native charms + step = "Looking for needed vnfd to configure with native charm" self.logger.debug(logging_text + step) for c_vnf in get_iterable(nsd, "constituent-vnfd"): @@ -967,9 +1279,9 @@ class NsLcm(LcmBase): # Check if this VNF has a charm configuration vnf_config = vnfd.get("vnf-configuration") if vnf_config and vnf_config.get("juju"): - proxy_charm = vnf_config["juju"]["charm"] + native_charm = vnf_config["juju"].get("proxy") is False - if proxy_charm: + if native_charm: if not vca_model_name: step = "creating VCA model name '{}'".format(nsr_id) self.logger.debug(logging_text + step) @@ -977,37 +1289,49 @@ class NsLcm(LcmBase): vca_model_name = nsr_id db_nsr_update["_admin.deployed.VCA-model-name"] = nsr_id self.update_db_2("nsrs", nsr_id, db_nsr_update) - step = "connecting to N2VC to configure vnf {}".format(vnf_index) + step = "deploying native charm for vnf_member_index={}".format(vnf_index) vnfr_params["rw_mgmt_ip"] = db_vnfrs[vnf_index]["ip-address"] charm_params = { "user_values": vnfr_params, "rw_mgmt_ip": db_vnfrs[vnf_index]["ip-address"], - "initial-config-primitive": vnf_config.get('initial-config-primitive') or {} + "initial-config-primitive": vnf_config.get('initial-config-primitive') or {}, } + # get username + # TODO remove this when changes on IM regarding config-access:ssh-access:default-user were + # merged. Meanwhile let's get username from initial-config-primitive + if vnf_config.get("initial-config-primitive"): + for param in vnf_config["initial-config-primitive"][0].get("parameter", ()): + if param["name"] == "ssh-username": + charm_params["username"] = param["value"] + if vnf_config.get("config-access") and vnf_config["config-access"].get("ssh-access"): + if vnf_config["config-access"]["ssh-access"].get("required"): + charm_params["username"] = vnf_config["config-access"]["ssh-access"].get("default-user") + # Login to the VCA. If there are multiple calls to login(), # subsequent calls will be a nop and return immediately. await self.n2vc.login() - deploy_charm(vnf_index, None, None, None, charm_params, n2vc_info) + deploy_charm(vnf_index, None, None, None, charm_params, n2vc_info, native_charm) number_to_configure += 1 # Deploy charms for each VDU that supports one. for vdu_index, vdu in enumerate(get_iterable(vnfd, 'vdu')): vdu_config = vdu.get('vdu-configuration') - proxy_charm = None + native_charm = False if vdu_config and vdu_config.get("juju"): - proxy_charm = vdu_config["juju"]["charm"] + native_charm = vdu_config["juju"].get("proxy") is False - if proxy_charm: + if native_charm: if not vca_model_name: step = "creating VCA model name" await self.n2vc.CreateNetworkService(nsr_id) vca_model_name = nsr_id db_nsr_update["_admin.deployed.VCA-model-name"] = nsr_id self.update_db_2("nsrs", nsr_id, db_nsr_update) - step = "connecting to N2VC to configure vdu {} from vnf {}".format(vdu["id"], vnf_index) + step = "deploying native charm for vnf_member_index={} vdu_id={}".format(vnf_index, + vdu["id"]) await self.n2vc.login() vdur = db_vnfrs[vnf_index]["vdur"][vdu_index] # TODO for the moment only first vdu_id contains a charm deployed @@ -1020,18 +1344,31 @@ class NsLcm(LcmBase): "rw_mgmt_ip": vdur["ip-address"], "initial-config-primitive": vdu_config.get('initial-config-primitive') or {} } + + # get username + # TODO remove this when changes on IM regarding config-access:ssh-access:default-user were + # merged. Meanwhile let's get username from initial-config-primitive + if vdu_config.get("initial-config-primitive"): + for param in vdu_config["initial-config-primitive"][0].get("parameter", ()): + if param["name"] == "ssh-username": + charm_params["username"] = param["value"] + if vdu_config.get("config-access") and vdu_config["config-access"].get("ssh-access"): + if vdu_config["config-access"]["ssh-access"].get("required"): + charm_params["username"] = vdu_config["config-access"]["ssh-access"].get( + "default-user") + deploy_charm(vnf_index, vdu["id"], vdur.get("name"), vdur["count-index"], - charm_params, n2vc_info) + charm_params, n2vc_info, native_charm) number_to_configure += 1 # Check if this NS has a charm configuration ns_config = nsd.get("ns-configuration") if ns_config and ns_config.get("juju"): - proxy_charm = ns_config["juju"]["charm"] + native_charm = ns_config["juju"].get("proxy") is False - if proxy_charm: - step = "connecting to N2VC to configure ns" + if native_charm: + step = "deploying native charm to configure ns" # TODO is NS magmt IP address needed? # Get additional parameters @@ -1045,19 +1382,31 @@ class NsLcm(LcmBase): # additional_params["rw_mgmt_ip"] = db_nsr["ip-address"] charm_params = { "user_values": additional_params, - # "rw_mgmt_ip": db_nsr["ip-address"], + "rw_mgmt_ip": db_nsr.get("ip-address"), "initial-config-primitive": ns_config.get('initial-config-primitive') or {} } + # get username + # TODO remove this when changes on IM regarding config-access:ssh-access:default-user were + # merged. Meanwhile let's get username from initial-config-primitive + if ns_config.get("initial-config-primitive"): + for param in ns_config["initial-config-primitive"][0].get("parameter", ()): + if param["name"] == "ssh-username": + charm_params["username"] = param["value"] + if ns_config.get("config-access") and ns_config["config-access"].get("ssh-access"): + if ns_config["config-access"]["ssh-access"].get("required"): + charm_params["username"] = ns_config["config-access"]["ssh-access"].get("default-user") + # Login to the VCA. If there are multiple calls to login(), # subsequent calls will be a nop and return immediately. await self.n2vc.login() - deploy_charm(None, None, None, None, charm_params, n2vc_info) + deploy_charm(None, None, None, None, charm_params, n2vc_info, native_charm) number_to_configure += 1 - db_nsr_update["operational-status"] = "running" + # waiting all charms are ok configuration_failed = False if number_to_configure: + step = "Waiting all charms are active" old_status = "configuring: init: {}".format(number_to_configure) db_nsr_update["config-status"] = old_status db_nsr_update["detailed-status"] = old_status @@ -1211,8 +1560,9 @@ class NsLcm(LcmBase): else: return False - # Get a numerically sorted list of the sequences for this VNFD's terminate action - def _get_terminate_config_primitive_seq_list(self, vnfd): + @staticmethod + def _get_terminate_config_primitive_seq_list(vnfd): + """ Get a numerically sorted list of the sequences for this VNFD's terminate action """ # No need to check for existing primitive twice, already done before vnf_config = vnfd.get("vnf-configuration") seq_list = vnf_config.get("terminate-config-primitive") @@ -1254,10 +1604,10 @@ class NsLcm(LcmBase): } return nslcmop - # Create a primitive with params from VNFD - # - Called from terminate() before deleting instance - # - Calls action() to execute the primitive async def _terminate_action(self, db_nslcmop, nslcmop_id, nsr_id): + """ Create a primitive with params from VNFD + Called from terminate() before deleting instance + Calls action() to execute the primitive """ logging_text = "Task ns={} _terminate_action={} ".format(nsr_id, nslcmop_id) db_vnfds = {} db_vnfrs_list = self.db.get_list("vnfrs", {"nsr-id-ref": nsr_id}) @@ -1303,13 +1653,19 @@ class NsLcm(LcmBase): nsr_id, nslcmop_terminate_action_id) # Launch Exception if action() returns other than ['COMPLETED', 'PARTIALLY_COMPLETED'] nslcmop_operation_states_ok = ['COMPLETED', 'PARTIALLY_COMPLETED'] - if (nslcmop_operation_state not in nslcmop_operation_states_ok): + if nslcmop_operation_state not in nslcmop_operation_states_ok: raise LcmException( "terminate_primitive_action for vnf_member_index={}", " primitive={} fails with error {}".format( vnf_index, seq.get("name"), nslcmop_operation_state_detail)) async def terminate(self, nsr_id, nslcmop_id): + + # Try to lock HA task here + task_is_locked_by_me = self.lcm_tasks.lock_HA('ns', 'nslcmops', nslcmop_id) + if not task_is_locked_by_me: + return + logging_text = "Task ns={} terminate={} ".format(nsr_id, nslcmop_id) self.logger.debug(logging_text + "Enter") db_nsr = None @@ -1322,6 +1678,9 @@ class NsLcm(LcmBase): nslcmop_operation_state = None autoremove = False # autoremove after terminated try: + # wait for any previous tasks in process + await self.lcm_tasks.waitfor_related_HA("ns", 'nslcmops', nslcmop_id) + step = "Getting nslcmop={} from db".format(nslcmop_id) db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) step = "Getting nsr={} from db".format(nsr_id) @@ -1470,7 +1829,7 @@ class NsLcm(LcmBase): try: RO_vnfd_id = vnf_deployed["id"] step = db_nsr_update["detailed-status"] = db_nslcmop_update["detailed-status"] =\ - "Deleting member-vnf-index={} RO_vnfd_id={} from RO".format( + "Deleting member_vnf_index={} RO_vnfd_id={} from RO".format( vnf_deployed["member-vnf-index"], RO_vnfd_id) await RO.delete("vnfd", RO_vnfd_id) self.logger.debug(logging_text + "RO_vnfd_id={} deleted".format(RO_vnfd_id)) @@ -1569,12 +1928,15 @@ class NsLcm(LcmBase): param_name = parameter["name"] if param_name in params: calculated_params[param_name] = params[param_name] - elif "default-value" in parameter: - calculated_params[param_name] = parameter["default-value"] - if isinstance(parameter["default-value"], str) and parameter["default-value"].startswith("<") and \ - parameter["default-value"].endswith(">"): - if parameter["default-value"][1:-1] in instantiation_params: - calculated_params[param_name] = instantiation_params[parameter["default-value"][1:-1]] + elif "default-value" in parameter or "value" in parameter: + if "value" in parameter: + calculated_params[param_name] = parameter["value"] + else: + calculated_params[param_name] = parameter["default-value"] + if isinstance(calculated_params[param_name], str) and calculated_params[param_name].startswith("<") \ + and calculated_params[param_name].endswith(">"): + if calculated_params[param_name][1:-1] in instantiation_params: + calculated_params[param_name] = instantiation_params[calculated_params[param_name][1:-1]] else: raise LcmException("Parameter {} needed to execute primitive {} not provided". format(parameter["default-value"], primitive_desc["name"])) @@ -1590,7 +1952,7 @@ class NsLcm(LcmBase): return calculated_params async def _ns_execute_primitive(self, db_deployed, member_vnf_index, vdu_id, vdu_name, vdu_count_index, - primitive, primitive_params): + primitive, primitive_params, retries=0, retries_interval=30): start_primitive_time = time() try: for vca_deployed in db_deployed["VCA"]: @@ -1612,40 +1974,55 @@ class NsLcm(LcmBase): raise LcmException("charm for member_vnf_index={} vdu_id={} vdu_name={} vdu_count_index={} has not " "model or application name" .format(member_vnf_index, vdu_id, vdu_name, vdu_count_index)) - if vca_deployed["operational-status"] != "active": - raise LcmException("charm for member_vnf_index={} vdu_id={} operational_status={} not 'active'".format( - member_vnf_index, vdu_id, vca_deployed["operational-status"])) + # if vca_deployed["operational-status"] != "active": + # raise LcmException("charm for member_vnf_index={} vdu_id={} operational_status={} not 'active'".format( + # member_vnf_index, vdu_id, vca_deployed["operational-status"])) callback = None # self.n2vc_callback callback_args = () # [db_nsr, db_nslcmop, member_vnf_index, None] await self.n2vc.login() - primitive_id = await self.n2vc.ExecutePrimitive( - model_name, - application_name, - primitive, - callback, - *callback_args, - **primitive_params - ) - while time() - start_primitive_time < self.timeout_primitive: - primitive_result_ = await self.n2vc.GetPrimitiveStatus(model_name, primitive_id) - if primitive_result_ in ("running", "pending"): - pass - elif primitive_result_ in ("completed", "failed"): - primitive_result = "COMPLETED" if primitive_result_ == "completed" else "FAILED" - detailed_result = await self.n2vc.GetPrimitiveOutput(model_name, primitive_id) - break + if primitive == "config": + primitive_params = {"params": primitive_params} + while retries >= 0: + primitive_id = await self.n2vc.ExecutePrimitive( + model_name, + application_name, + primitive, + callback, + *callback_args, + **primitive_params + ) + while time() - start_primitive_time < self.timeout_primitive: + primitive_result_ = await self.n2vc.GetPrimitiveStatus(model_name, primitive_id) + if primitive_result_ in ("completed", "failed"): + primitive_result = "COMPLETED" if primitive_result_ == "completed" else "FAILED" + detailed_result = await self.n2vc.GetPrimitiveOutput(model_name, primitive_id) + break + elif primitive_result_ is None and primitive == "config": + primitive_result = "COMPLETED" + detailed_result = None + break + else: # ("running", "pending", None): + pass + await asyncio.sleep(5) else: - detailed_result = "Invalid N2VC.GetPrimitiveStatus = {} obtained".format(primitive_result_) - primitive_result = "FAILED" + raise LcmException("timeout after {} seconds".format(self.timeout_primitive)) + if primitive_result == "COMPLETED": break - await asyncio.sleep(5) - else: - raise LcmException("timeout after {} seconds".format(self.timeout_primitive)) + retries -= 1 + if retries >= 0: + await asyncio.sleep(retries_interval) + return primitive_result, detailed_result except (N2VCPrimitiveExecutionFailed, LcmException) as e: return "FAILED", str(e) async def action(self, nsr_id, nslcmop_id): + + # Try to lock HA task here + task_is_locked_by_me = self.lcm_tasks.lock_HA('ns', 'nslcmops', nslcmop_id) + if not task_is_locked_by_me: + return + logging_text = "Task ns={} action={} ".format(nsr_id, nslcmop_id) self.logger.debug(logging_text + "Enter") # get all needed from database @@ -1657,6 +2034,9 @@ class NsLcm(LcmBase): nslcmop_operation_state_detail = None exc = None try: + # wait for any previous tasks in process + await self.lcm_tasks.waitfor_related_HA('ns', 'nslcmops', nslcmop_id) + step = "Getting information from database" db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) db_nsr = self.db.get_one("nsrs", {"_id": nsr_id}) @@ -1679,17 +2059,6 @@ class NsLcm(LcmBase): step = "Getting nsd from database" db_nsd = self.db.get_one("nsds", {"_id": db_nsr["nsd-id"]}) - # look if previous tasks in process - task_name, task_dependency = self.lcm_tasks.lookfor_related("ns", nsr_id, nslcmop_id) - if task_dependency: - step = db_nslcmop_update["detailed-status"] = \ - "Waiting for related tasks to be completed: {}".format(task_name) - self.logger.debug(logging_text + step) - self.update_db_2("nslcmops", nslcmop_id, db_nslcmop_update) - _, pending = await asyncio.wait(task_dependency, timeout=3600) - if pending: - raise LcmException("Timeout waiting related tasks to be completed") - # for backward compatibility if nsr_deployed and isinstance(nsr_deployed.get("VCA"), dict): nsr_deployed["VCA"] = list(nsr_deployed["VCA"].values()) @@ -1777,6 +2146,12 @@ class NsLcm(LcmBase): return nslcmop_operation_state, nslcmop_operation_state_detail async def scale(self, nsr_id, nslcmop_id): + + # Try to lock HA task here + task_is_locked_by_me = self.lcm_tasks.lock_HA('ns', 'nslcmops', nslcmop_id) + if not task_is_locked_by_me: + return + logging_text = "Task ns={} scale={} ".format(nsr_id, nslcmop_id) self.logger.debug(logging_text + "Enter") # get all needed from database @@ -1792,26 +2167,19 @@ class NsLcm(LcmBase): old_config_status = "" vnfr_scaled = False try: + # wait for any previous tasks in process + await self.lcm_tasks.waitfor_related_HA('ns', 'nslcmops', nslcmop_id) + step = "Getting nslcmop from database" + self.logger.debug(step + " after having waited for previous tasks to be completed") db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) step = "Getting nsr from database" db_nsr = self.db.get_one("nsrs", {"_id": nsr_id}) old_operational_status = db_nsr["operational-status"] old_config_status = db_nsr["config-status"] - - # look if previous tasks in process - task_name, task_dependency = self.lcm_tasks.lookfor_related("ns", nsr_id, nslcmop_id) - if task_dependency: - step = db_nslcmop_update["detailed-status"] = \ - "Waiting for related tasks to be completed: {}".format(task_name) - self.logger.debug(logging_text + step) - self.update_db_2("nslcmops", nslcmop_id, db_nslcmop_update) - _, pending = await asyncio.wait(task_dependency, timeout=3600) - if pending: - raise LcmException("Timeout waiting related tasks to be completed") - step = "Parsing scaling parameters" + # self.logger.debug(step) db_nsr_update["operational-status"] = "scaling" self.update_db_2("nsrs", nsr_id, db_nsr_update) nsr_deployed = db_nsr["_admin"].get("deployed") @@ -1831,6 +2199,7 @@ class NsLcm(LcmBase): db_vnfr = self.db.get_one("vnfrs", {"member-vnf-index-ref": vnf_index, "nsr-id-ref": nsr_id}) step = "Getting vnfd from database" db_vnfd = self.db.get_one("vnfds", {"_id": db_vnfr["vnfd-id"]}) + step = "Getting scaling-group-descriptor" for scaling_descriptor in db_vnfd["scaling-group-descriptor"]: if scaling_descriptor["name"] == scaling_group: @@ -1838,6 +2207,7 @@ class NsLcm(LcmBase): else: raise LcmException("input parameter 'scaleByStepData':'scaling-group-descriptor':'{}' is not present " "at vnfd:scaling-group-descriptor".format(scaling_group)) + # cooldown_time = 0 # for scaling_policy_descriptor in scaling_descriptor.get("scaling-policy", ()): # cooldown_time = scaling_policy_descriptor.get("cooldown-time", 0) @@ -1864,16 +2234,21 @@ class NsLcm(LcmBase): # count if max-instance-count is reached if "max-instance-count" in scaling_descriptor and scaling_descriptor["max-instance-count"] is not None: max_instance_count = int(scaling_descriptor["max-instance-count"]) + + # self.logger.debug("MAX_INSTANCE_COUNT is {}".format(scaling_descriptor["max-instance-count"])) if nb_scale_op >= max_instance_count: - raise LcmException("reached the limit of {} (max-instance-count) scaling-out operations for the" - " scaling-group-descriptor '{}'".format(nb_scale_op, scaling_group)) - nb_scale_op = nb_scale_op + 1 + raise LcmException("reached the limit of {} (max-instance-count) " + "scaling-out operations for the " + "scaling-group-descriptor '{}'".format(nb_scale_op, scaling_group)) + + nb_scale_op += 1 vdu_scaling_info["scaling_direction"] = "OUT" vdu_scaling_info["vdu-create"] = {} for vdu_scale_info in scaling_descriptor["vdu"]: RO_scaling_info.append({"osm_vdu_id": vdu_scale_info["vdu-id-ref"], "member-vnf-index": vnf_index, "type": "create", "count": vdu_scale_info.get("count", 1)}) vdu_scaling_info["vdu-create"][vdu_scale_info["vdu-id-ref"]] = vdu_scale_info.get("count", 1) + elif scaling_type == "SCALE_IN": # count if min-instance-count is reached min_instance_count = 0 @@ -1882,7 +2257,7 @@ class NsLcm(LcmBase): if nb_scale_op <= min_instance_count: raise LcmException("reached the limit of {} (min-instance-count) scaling-in operations for the " "scaling-group-descriptor '{}'".format(nb_scale_op, scaling_group)) - nb_scale_op = nb_scale_op - 1 + nb_scale_op -= 1 vdu_scaling_info["scaling_direction"] = "IN" vdu_scaling_info["vdu-delete"] = {} for vdu_scale_info in scaling_descriptor["vdu"]: @@ -2040,7 +2415,7 @@ class NsLcm(LcmBase): step = db_nslcmop_update["detailed-status"] = \ "executing post-scale scaling-config-action '{}'".format(vnf_config_primitive) - vnfr_params = {"": vdu_scaling_info} + vnfr_params = {"VDU_SCALE_INFO": vdu_scaling_info} if db_vnfr.get("additionalParamsForVnf"): vnfr_params.update(db_vnfr["additionalParamsForVnf"]) @@ -2070,7 +2445,8 @@ class NsLcm(LcmBase): db_nslcmop_update["statusEnteredTime"] = time() db_nslcmop_update["detailed-status"] = "done" db_nsr_update["detailed-status"] = "" # "scaled {} {}".format(scaling_group, scaling_type) - db_nsr_update["operational-status"] = old_operational_status + db_nsr_update["operational-status"] = "running" if old_operational_status == "failed" \ + else old_operational_status db_nsr_update["config-status"] = old_config_status return except (ROclient.ROClientException, DbException, LcmException) as e: