X-Git-Url: https://osm.etsi.org/gitweb/?p=osm%2FNBI.git;a=blobdiff_plain;f=osm_nbi%2Fdescriptor_topics.py;h=6bf437d653fceb96f57dd48bcfbe4350e6fbfb9b;hp=3bdaa03381bec85df94ddfe3d508776209790ff5;hb=4568a372eb5a204e04d917213de03ec51f9110c1;hpb=960531ab7417bf6705399dae790899463a659da9 diff --git a/osm_nbi/descriptor_topics.py b/osm_nbi/descriptor_topics.py index 3bdaa03..6bf437d 100644 --- a/osm_nbi/descriptor_topics.py +++ b/osm_nbi/descriptor_topics.py @@ -16,6 +16,9 @@ import tarfile import yaml import json +import importlib +import copy + # import logging from hashlib import md5 from osm_common.dbbase import DbException, deep_update_rfc7396 @@ -23,25 +26,33 @@ from http import HTTPStatus from time import time from uuid import uuid4 from re import fullmatch -from osm_nbi.validation import ValidationError, pdu_new_schema, pdu_edit_schema, \ - validate_input, vnfpkgop_new_schema +from osm_nbi.validation import ( + ValidationError, + pdu_new_schema, + pdu_edit_schema, + validate_input, + vnfpkgop_new_schema, +) from osm_nbi.base_topic import BaseTopic, EngineException, get_iterable -from osm_im.vnfd import vnfd as vnfd_im -from osm_im.nsd import nsd as nsd_im + +etsi_nfv_vnfd = importlib.import_module("osm_im.etsi-nfv-vnfd") +etsi_nfv_nsd = importlib.import_module("osm_im.etsi-nfv-nsd") from osm_im.nst import nst as nst_im from pyangbind.lib.serialise import pybindJSONDecoder import pyangbind.lib.pybindJSON as pybindJSON +from osm_nbi import utils __author__ = "Alfonso Tierno " class DescriptorTopic(BaseTopic): - def __init__(self, db, fs, msg, auth): BaseTopic.__init__(self, db, fs, msg, auth) def check_conflict_on_edit(self, session, final_content, edit_content, _id): - super().check_conflict_on_edit(session, final_content, edit_content, _id) + final_content = super().check_conflict_on_edit( + session, final_content, edit_content, _id + ) def _check_unique_id_name(descriptor, position=""): for desc_key, desc_item in descriptor.items(): @@ -50,17 +61,27 @@ class DescriptorTopic(BaseTopic): desc_item_id = None for index, list_item in enumerate(desc_item): if isinstance(list_item, dict): - _check_unique_id_name(list_item, "{}.{}[{}]" - .format(position, desc_key, index)) + _check_unique_id_name( + list_item, "{}.{}[{}]".format(position, desc_key, index) + ) # Base case - if index == 0 and (list_item.get("id") or list_item.get("name")): + if index == 0 and ( + list_item.get("id") or list_item.get("name") + ): desc_item_id = "id" if list_item.get("id") else "name" if desc_item_id and list_item.get(desc_item_id): if list_item[desc_item_id] in used_ids: - position = "{}.{}[{}]".format(position, desc_key, index) - raise EngineException("Error: identifier {} '{}' is not unique and repeats at '{}'" - .format(desc_item_id, list_item[desc_item_id], - position), HTTPStatus.UNPROCESSABLE_ENTITY) + position = "{}.{}[{}]".format( + position, desc_key, index + ) + raise EngineException( + "Error: identifier {} '{}' is not unique and repeats at '{}'".format( + desc_item_id, + list_item[desc_item_id], + position, + ), + HTTPStatus.UNPROCESSABLE_ENTITY, + ) used_ids.append(list_item[desc_item_id]) _check_unique_id_name(final_content) @@ -71,25 +92,34 @@ class DescriptorTopic(BaseTopic): if k in final_content: internal_keys[k] = final_content.pop(k) storage_params = internal_keys["_admin"].get("storage") - serialized = self._validate_input_new(final_content, storage_params, session["force"]) + serialized = self._validate_input_new( + final_content, storage_params, session["force"] + ) + # 1.2. modify final_content with a serialized version - final_content.clear() - final_content.update(serialized) + final_content = copy.deepcopy(serialized) # 1.3. restore internal keys for k, v in internal_keys.items(): final_content[k] = v - if session["force"]: - return + return final_content + # 2. check that this id is not present if "id" in edit_content: _filter = self._get_project_filter(session) + _filter["id"] = final_content["id"] _filter["_id.neq"] = _id + if self.db.get_one(self.topic, _filter, fail_on_empty=False): - raise EngineException("{} with id '{}' already exists for this project".format(self.topic[:-1], - final_content["id"]), - HTTPStatus.CONFLICT) + raise EngineException( + "{} with id '{}' already exists for this project".format( + self.topic[:-1], final_content["id"] + ), + HTTPStatus.CONFLICT, + ) + + return final_content @staticmethod def format_on_new(content, project_id=None, make_public=False): @@ -119,20 +149,31 @@ class DescriptorTopic(BaseTopic): if len(desc_list) == 1: return desc_list[0] elif len(desc_list) > 1: - raise DbException("Found more than one {} with id='{}' belonging to this project".format(topic[:-1], id), - HTTPStatus.CONFLICT) + raise DbException( + "Found more than one {} with id='{}' belonging to this project".format( + topic[:-1], id + ), + HTTPStatus.CONFLICT, + ) # not found any: try to find public _filter = BaseTopic._get_project_filter(session) _filter["id"] = id desc_list = db.get_list(topic, _filter) if not desc_list: - raise DbException("Not found any {} with id='{}'".format(topic[:-1], id), HTTPStatus.NOT_FOUND) + raise DbException( + "Not found any {} with id='{}'".format(topic[:-1], id), + HTTPStatus.NOT_FOUND, + ) elif len(desc_list) == 1: return desc_list[0] else: - raise DbException("Found more than one public {} with id='{}'; and no one belonging to this project".format( - topic[:-1], id), HTTPStatus.CONFLICT) + raise DbException( + "Found more than one public {} with id='{}'; and no one belonging to this project".format( + topic[:-1], id + ), + HTTPStatus.CONFLICT, + ) def new(self, rollback, session, indata=None, kwargs=None, headers=None): """ @@ -154,7 +195,7 @@ class DescriptorTopic(BaseTopic): # _remove_envelop if indata: if "userDefinedData" in indata: - indata = indata['userDefinedData'] + indata = indata["userDefinedData"] # Override descriptor with query string kwargs self._update_input_with_kwargs(indata, kwargs) @@ -163,7 +204,9 @@ class DescriptorTopic(BaseTopic): # indata = DescriptorTopic._validate_input_new(self, indata, project_id=session["force"]) content = {"_admin": {"userDefinedData": indata}} - self.format_on_new(content, session["project_id"], make_public=session["public"]) + self.format_on_new( + content, session["project_id"], make_public=session["public"] + ) _id = self.db.create(self.topic, content) rollback.append({"topic": self.topic, "_id": _id}) self._send_msg("created", {"_id": _id}) @@ -187,8 +230,12 @@ class DescriptorTopic(BaseTopic): expected_md5 = headers.get("Content-File-MD5") compressed = None content_type = headers.get("Content-Type") - if content_type and "application/gzip" in content_type or "application/x-gzip" in content_type or \ - "application/zip" in content_type: + if ( + content_type + and "application/gzip" in content_type + or "application/x-gzip" in content_type + or "application/zip" in content_type + ): compressed = "gzip" filename = headers.get("Content-Filename") if not filename: @@ -198,19 +245,27 @@ class DescriptorTopic(BaseTopic): error_text = "" try: if content_range_text: - content_range = content_range_text.replace("-", " ").replace("/", " ").split() - if content_range[0] != "bytes": # TODO check x result # manyfiles yes X -> zip @@ -379,50 +479,58 @@ class DescriptorTopic(BaseTopic): # onefile yes no -> zip # X yes -> text contain_many_files = False - if storage.get('pkg-dir'): + if storage.get("pkg-dir"): # check if there are more than one file in the package, ignoring checksums.txt. - pkg_files = self.fs.dir_ls((storage['folder'], storage['pkg-dir'])) - if len(pkg_files) >= 3 or (len(pkg_files) == 2 and 'checksums.txt' not in pkg_files): + pkg_files = self.fs.dir_ls((storage["folder"], storage["pkg-dir"])) + if len(pkg_files) >= 3 or ( + len(pkg_files) == 2 and "checksums.txt" not in pkg_files + ): contain_many_files = True if accept_text and (not contain_many_files or path == "$DESCRIPTOR"): - return self.fs.file_open((storage['folder'], storage['descriptor']), "r"), "text/plain" + return ( + self.fs.file_open((storage["folder"], storage["descriptor"]), "r"), + "text/plain", + ) elif contain_many_files and not accept_zip: - raise EngineException("Packages that contains several files need to be retrieved with 'application/zip'" - "Accept header", http_code=HTTPStatus.NOT_ACCEPTABLE) + raise EngineException( + "Packages that contains several files need to be retrieved with 'application/zip'" + "Accept header", + http_code=HTTPStatus.NOT_ACCEPTABLE, + ) else: - if not storage.get('zipfile'): + if not storage.get("zipfile"): # TODO generate zipfile if not present - raise EngineException("Only allowed 'text/plain' Accept header for this descriptor. To be solved in " - "future versions", http_code=HTTPStatus.NOT_ACCEPTABLE) - return self.fs.file_open((storage['folder'], storage['zipfile']), "rb"), accept_zip + raise EngineException( + "Only allowed 'text/plain' Accept header for this descriptor. To be solved in " + "future versions", + http_code=HTTPStatus.NOT_ACCEPTABLE, + ) + return ( + self.fs.file_open((storage["folder"], storage["zipfile"]), "rb"), + accept_zip, + ) + + def _remove_yang_prefixes_from_descriptor(self, descriptor): + new_descriptor = {} + for k, v in descriptor.items(): + new_v = v + if isinstance(v, dict): + new_v = self._remove_yang_prefixes_from_descriptor(v) + elif isinstance(v, list): + new_v = list() + for x in v: + if isinstance(x, dict): + new_v.append(self._remove_yang_prefixes_from_descriptor(x)) + else: + new_v.append(x) + new_descriptor[k.split(":")[-1]] = new_v + return new_descriptor def pyangbind_validation(self, item, data, force=False): - try: - if item == "vnfds": - myvnfd = vnfd_im() - pybindJSONDecoder.load_ietf_json({'vnfd:vnfd-catalog': {'vnfd': [data]}}, None, None, obj=myvnfd, - path_helper=True, skip_unknown=force) - out = pybindJSON.dumps(myvnfd, mode="ietf") - elif item == "nsds": - mynsd = nsd_im() - pybindJSONDecoder.load_ietf_json({'nsd:nsd-catalog': {'nsd': [data]}}, None, None, obj=mynsd, - path_helper=True, skip_unknown=force) - out = pybindJSON.dumps(mynsd, mode="ietf") - elif item == "nsts": - mynst = nst_im() - pybindJSONDecoder.load_ietf_json({'nst': [data]}, None, None, obj=mynst, - path_helper=True, skip_unknown=force) - out = pybindJSON.dumps(mynst, mode="ietf") - else: - raise EngineException("Not possible to validate '{}' item".format(item), - http_code=HTTPStatus.INTERNAL_SERVER_ERROR) - - desc_out = self._remove_envelop(yaml.safe_load(out)) - return desc_out - - except Exception as e: - raise EngineException("Error in pyangbind validation: {}".format(str(e)), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + raise EngineException( + "Not possible to validate '{}' item".format(item), + http_code=HTTPStatus.INTERNAL_SERVER_ERROR, + ) def _validate_input_edit(self, indata, content, force=False): # not needed to validate with pyangbind becuase it will be validated at check_conflict_on_edit @@ -435,25 +543,38 @@ class DescriptorTopic(BaseTopic): if indata["operationalState"] in ("ENABLED", "DISABLED"): indata["_admin"]["operationalState"] = indata.pop("operationalState") else: - raise EngineException("State '{}' is not a valid operational state" - .format(indata["operationalState"]), - http_code=HTTPStatus.BAD_REQUEST) + raise EngineException( + "State '{}' is not a valid operational state".format( + indata["operationalState"] + ), + http_code=HTTPStatus.BAD_REQUEST, + ) - # In the case of user defined data, we need to put the data in the root of the object + # In the case of user defined data, we need to put the data in the root of the object # to preserve current expected behaviour if "userDefinedData" in indata: data = indata.pop("userDefinedData") if type(data) == dict: indata["_admin"]["userDefinedData"] = data else: - raise EngineException("userDefinedData should be an object, but is '{}' instead" - .format(type(data)), - http_code=HTTPStatus.BAD_REQUEST) - - if ("operationalState" in indata["_admin"] and - content["_admin"]["operationalState"] == indata["_admin"]["operationalState"]): - raise EngineException("operationalState already {}".format(content["_admin"]["operationalState"]), - http_code=HTTPStatus.CONFLICT) + raise EngineException( + "userDefinedData should be an object, but is '{}' instead".format( + type(data) + ), + http_code=HTTPStatus.BAD_REQUEST, + ) + + if ( + "operationalState" in indata["_admin"] + and content["_admin"]["operationalState"] + == indata["_admin"]["operationalState"] + ): + raise EngineException( + "operationalState already {}".format( + content["_admin"]["operationalState"] + ), + http_code=HTTPStatus.CONFLICT, + ) return indata @@ -465,27 +586,57 @@ class VnfdTopic(DescriptorTopic): def __init__(self, db, fs, msg, auth): DescriptorTopic.__init__(self, db, fs, msg, auth) + def pyangbind_validation(self, item, data, force=False): + if self._descriptor_data_is_in_old_format(data): + raise EngineException( + "ERROR: Unsupported descriptor format. Please, use an ETSI SOL006 descriptor.", + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + try: + myvnfd = etsi_nfv_vnfd.etsi_nfv_vnfd() + pybindJSONDecoder.load_ietf_json( + {"etsi-nfv-vnfd:vnfd": data}, + None, + None, + obj=myvnfd, + path_helper=True, + skip_unknown=force, + ) + out = pybindJSON.dumps(myvnfd, mode="ietf") + desc_out = self._remove_envelop(yaml.safe_load(out)) + desc_out = self._remove_yang_prefixes_from_descriptor(desc_out) + return utils.deep_update_dict(data, desc_out) + except Exception as e: + raise EngineException( + "Error in pyangbind validation: {}".format(str(e)), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + + @staticmethod + def _descriptor_data_is_in_old_format(data): + return ("vnfd-catalog" in data) or ("vnfd:vnfd-catalog" in data) + @staticmethod def _remove_envelop(indata=None): if not indata: return {} clean_indata = indata - if clean_indata.get('vnfd:vnfd-catalog'): - clean_indata = clean_indata['vnfd:vnfd-catalog'] - elif clean_indata.get('vnfd-catalog'): - clean_indata = clean_indata['vnfd-catalog'] - if clean_indata.get('vnfd'): - if not isinstance(clean_indata['vnfd'], list) or len(clean_indata['vnfd']) != 1: - raise EngineException("'vnfd' must be a list of only one element") - clean_indata = clean_indata['vnfd'][0] - elif clean_indata.get('vnfd:vnfd'): - if not isinstance(clean_indata['vnfd:vnfd'], list) or len(clean_indata['vnfd:vnfd']) != 1: - raise EngineException("'vnfd:vnfd' must be a list of only one element") - clean_indata = clean_indata['vnfd:vnfd'][0] + + if clean_indata.get("etsi-nfv-vnfd:vnfd"): + if not isinstance(clean_indata["etsi-nfv-vnfd:vnfd"], dict): + raise EngineException("'etsi-nfv-vnfd:vnfd' must be a dict") + clean_indata = clean_indata["etsi-nfv-vnfd:vnfd"] + elif clean_indata.get("vnfd"): + if not isinstance(clean_indata["vnfd"], dict): + raise EngineException("'vnfd' must be dict") + clean_indata = clean_indata["vnfd"] + return clean_indata def check_conflict_on_edit(self, session, final_content, edit_content, _id): - super().check_conflict_on_edit(session, final_content, edit_content, _id) + final_content = super().check_conflict_on_edit( + session, final_content, edit_content, _id + ) # set type of vnfd contains_pdu = False @@ -500,6 +651,7 @@ class VnfdTopic(DescriptorTopic): elif contains_vdu: final_content["_admin"]["type"] = "vnfd" # if neither vud nor pdu do not fill type + return final_content def check_conflict_on_del(self, session, _id, db_content): """ @@ -523,245 +675,345 @@ class VnfdTopic(DescriptorTopic): # check vnfrs using this vnfd _filter["vnfd-id"] = _id if self.db.get_list("vnfrs", _filter): - raise EngineException("There is at least one VNF using this descriptor", http_code=HTTPStatus.CONFLICT) + raise EngineException( + "There is at least one VNF instance using this descriptor", + http_code=HTTPStatus.CONFLICT, + ) # check NSD referencing this VNFD del _filter["vnfd-id"] - _filter["constituent-vnfd.ANYINDEX.vnfd-id-ref"] = descriptor_id + _filter["vnfd-id"] = descriptor_id if self.db.get_list("nsds", _filter): - raise EngineException("There is at least one NSD referencing this descriptor", - http_code=HTTPStatus.CONFLICT) + raise EngineException( + "There is at least one NS package referencing this descriptor", + http_code=HTTPStatus.CONFLICT, + ) def _validate_input_new(self, indata, storage_params, force=False): indata.pop("onboardingState", None) indata.pop("operationalState", None) indata.pop("usageState", None) - indata.pop("links", None) indata = self.pyangbind_validation("vnfds", indata, force) # Cross references validation in the descriptor - self.validate_mgmt_interfaces_connection_points(indata) + + self.validate_mgmt_interface_connection_point(indata) for vdu in get_iterable(indata.get("vdu")): - self.validate_vdu_connection_point_refs(vdu, indata) - self._validate_vdu_charms_in_package(storage_params, vdu, indata) + self.validate_vdu_internal_connection_points(vdu) self._validate_vdu_cloud_init_in_package(storage_params, vdu, indata) + self._validate_vdu_charms_in_package(storage_params, indata) self._validate_vnf_charms_in_package(storage_params, indata) - self.validate_internal_vlds(indata) + self.validate_external_connection_points(indata) + self.validate_internal_virtual_links(indata) self.validate_monitoring_params(indata) self.validate_scaling_group_descriptor(indata) return indata @staticmethod - def validate_mgmt_interfaces_connection_points(indata): + def validate_mgmt_interface_connection_point(indata): if not indata.get("vdu"): return - if not indata.get("mgmt-interface"): - raise EngineException("'mgmt-interface' is a mandatory field and it is not defined", - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - if indata["mgmt-interface"].get("cp"): - for cp in get_iterable(indata.get("connection-point")): - if cp["name"] == indata["mgmt-interface"]["cp"]: - break - else: - raise EngineException("mgmt-interface:cp='{}' must match an existing connection-point" - .format(indata["mgmt-interface"]["cp"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + if not indata.get("mgmt-cp"): + raise EngineException( + "'mgmt-cp' is a mandatory field and it is not defined", + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + + for cp in get_iterable(indata.get("ext-cpd")): + if cp["id"] == indata["mgmt-cp"]: + break + else: + raise EngineException( + "mgmt-cp='{}' must match an existing ext-cpd".format(indata["mgmt-cp"]), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) @staticmethod - def validate_vdu_connection_point_refs(vdu, indata): - icp_refs = [] - ecp_refs = [] - for interface in get_iterable(vdu.get("interface")): - if interface.get("external-connection-point-ref"): - if interface.get("external-connection-point-ref") in ecp_refs: - raise EngineException("vdu[id='{}']:interface[name='{}']:external-connection-point-ref='{}' " - "is referenced by other interface" - .format(vdu["id"], interface["name"], - interface["external-connection-point-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - ecp_refs.append(interface.get("external-connection-point-ref")) - for cp in get_iterable(indata.get("connection-point")): - if cp["name"] == interface["external-connection-point-ref"]: - break - else: - raise EngineException("vdu[id='{}']:interface[name='{}']:external-connection-point-ref='{}' " - "must match an existing connection-point" - .format(vdu["id"], interface["name"], - interface["external-connection-point-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - elif interface.get("internal-connection-point-ref"): - if interface.get("internal-connection-point-ref") in icp_refs: - raise EngineException("vdu[id='{}']:interface[name='{}']:internal-connection-point-ref='{}' " - "is referenced by other interface" - .format(vdu["id"], interface["name"], - interface["internal-connection-point-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - icp_refs.append(interface.get("internal-connection-point-ref")) - for internal_cp in get_iterable(vdu.get("internal-connection-point")): - if interface["internal-connection-point-ref"] == internal_cp.get("id"): - break - else: - raise EngineException("vdu[id='{}']:interface[name='{}']:internal-connection-point-ref='{}' " - "must match an existing vdu:internal-connection-point" - .format(vdu["id"], interface["name"], - interface["internal-connection-point-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - - def _validate_vdu_charms_in_package(self, storage_params, vdu, indata): - if not vdu.get("vdu-configuration"): - return - if vdu["vdu-configuration"].get("juju"): - if not self._validate_package_folders(storage_params, 'charms'): - raise EngineException("Charm defined in vnf[id={}]:vdu[id={}] but not present in " - "package".format(indata["id"], vdu["id"])) + def validate_vdu_internal_connection_points(vdu): + int_cpds = set() + for cpd in get_iterable(vdu.get("int-cpd")): + cpd_id = cpd.get("id") + if cpd_id and cpd_id in int_cpds: + raise EngineException( + "vdu[id='{}']:int-cpd[id='{}'] is already used by other int-cpd".format( + vdu["id"], cpd_id + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + int_cpds.add(cpd_id) + + @staticmethod + def validate_external_connection_points(indata): + all_vdus_int_cpds = set() + for vdu in get_iterable(indata.get("vdu")): + for int_cpd in get_iterable(vdu.get("int-cpd")): + all_vdus_int_cpds.add((vdu.get("id"), int_cpd.get("id"))) + + ext_cpds = set() + for cpd in get_iterable(indata.get("ext-cpd")): + cpd_id = cpd.get("id") + if cpd_id and cpd_id in ext_cpds: + raise EngineException( + "ext-cpd[id='{}'] is already used by other ext-cpd".format(cpd_id), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + ext_cpds.add(cpd_id) + + int_cpd = cpd.get("int-cpd") + if int_cpd: + if (int_cpd.get("vdu-id"), int_cpd.get("cpd")) not in all_vdus_int_cpds: + raise EngineException( + "ext-cpd[id='{}']:int-cpd must match an existing vdu int-cpd".format( + cpd_id + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + # TODO: Validate k8s-cluster-net points to a valid k8s-cluster:nets ? + + def _validate_vdu_charms_in_package(self, storage_params, indata): + for df in indata["df"]: + if ( + "lcm-operations-configuration" in df + and "operate-vnf-op-config" in df["lcm-operations-configuration"] + ): + configs = df["lcm-operations-configuration"][ + "operate-vnf-op-config" + ].get("day1-2", []) + vdus = df.get("vdu-profile", []) + for vdu in vdus: + for config in configs: + if config["id"] == vdu["id"] and utils.find_in_list( + config.get("execution-environment-list", []), + lambda ee: "juju" in ee, + ): + if not self._validate_package_folders( + storage_params, "charms" + ): + raise EngineException( + "Charm defined in vnf[id={}] but not present in " + "package".format(indata["id"]) + ) def _validate_vdu_cloud_init_in_package(self, storage_params, vdu, indata): if not vdu.get("cloud-init-file"): return - if not self._validate_package_folders(storage_params, 'cloud_init', vdu["cloud-init-file"]): - raise EngineException("Cloud-init defined in vnf[id={}]:vdu[id={}] but not present in " - "package".format(indata["id"], vdu["id"])) + if not self._validate_package_folders( + storage_params, "cloud_init", vdu["cloud-init-file"] + ): + raise EngineException( + "Cloud-init defined in vnf[id={}]:vdu[id={}] but not present in " + "package".format(indata["id"], vdu["id"]) + ) def _validate_vnf_charms_in_package(self, storage_params, indata): - if not indata.get("vnf-configuration"): - return - if indata["vnf-configuration"].get("juju"): - if not self._validate_package_folders(storage_params, 'charms'): - raise EngineException("Charm defined in vnf[id={}] but not present in " - "package".format(indata["id"])) + # Get VNF configuration through new container + for deployment_flavor in indata.get("df", []): + if "lcm-operations-configuration" not in deployment_flavor: + return + if ( + "operate-vnf-op-config" + not in deployment_flavor["lcm-operations-configuration"] + ): + return + for day_1_2_config in deployment_flavor["lcm-operations-configuration"][ + "operate-vnf-op-config" + ]["day1-2"]: + if day_1_2_config["id"] == indata["id"]: + if utils.find_in_list( + day_1_2_config.get("execution-environment-list", []), + lambda ee: "juju" in ee, + ): + if not self._validate_package_folders(storage_params, "charms"): + raise EngineException( + "Charm defined in vnf[id={}] but not present in " + "package".format(indata["id"]) + ) def _validate_package_folders(self, storage_params, folder, file=None): if not storage_params or not storage_params.get("pkg-dir"): return False else: - if self.fs.file_exists("{}_".format(storage_params["folder"]), 'dir'): - f = "{}_/{}/{}".format(storage_params["folder"], storage_params["pkg-dir"], folder) + if self.fs.file_exists("{}_".format(storage_params["folder"]), "dir"): + f = "{}_/{}/{}".format( + storage_params["folder"], storage_params["pkg-dir"], folder + ) else: - f = "{}/{}/{}".format(storage_params["folder"], storage_params["pkg-dir"], folder) + f = "{}/{}/{}".format( + storage_params["folder"], storage_params["pkg-dir"], folder + ) if file: - return self.fs.file_exists("{}/{}".format(f, file), 'file') + return self.fs.file_exists("{}/{}".format(f, file), "file") else: - if self.fs.file_exists(f, 'dir'): + if self.fs.file_exists(f, "dir"): if self.fs.dir_ls(f): return True return False @staticmethod - def validate_internal_vlds(indata): - vld_names = [] # For detection of duplicated VLD names - for ivld in get_iterable(indata.get("internal-vld")): - ivld_name = ivld.get("name") - if ivld_name and ivld_name in vld_names: - raise EngineException("Duplicated VLD name '{}' in vnfd[id={}]:internal-vld[id={}]" - .format(ivld["name"], indata["id"], ivld["id"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + def validate_internal_virtual_links(indata): + all_ivld_ids = set() + for ivld in get_iterable(indata.get("int-virtual-link-desc")): + ivld_id = ivld.get("id") + if ivld_id and ivld_id in all_ivld_ids: + raise EngineException( + "Duplicated VLD id in int-virtual-link-desc[id={}]".format(ivld_id), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) else: - vld_names.append(ivld_name) - - for icp in get_iterable(ivld.get("internal-connection-point")): - icp_mark = False - for vdu in get_iterable(indata.get("vdu")): - for internal_cp in get_iterable(vdu.get("internal-connection-point")): - if icp["id-ref"] == internal_cp["id"]: - icp_mark = True - break - if icp_mark: - break - else: - raise EngineException("internal-vld[id='{}']:internal-connection-point='{}' must match an existing " - "vdu:internal-connection-point".format(ivld["id"], icp["id-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - if ivld.get("ip-profile-ref"): - for ip_prof in get_iterable(indata.get("ip-profiles")): - if ip_prof["name"] == get_iterable(ivld.get("ip-profile-ref")): - break - else: - raise EngineException("internal-vld[id='{}']:ip-profile-ref='{}' does not exist".format( - ivld["id"], ivld["ip-profile-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + all_ivld_ids.add(ivld_id) + + for vdu in get_iterable(indata.get("vdu")): + for int_cpd in get_iterable(vdu.get("int-cpd")): + int_cpd_ivld_id = int_cpd.get("int-virtual-link-desc") + if int_cpd_ivld_id and int_cpd_ivld_id not in all_ivld_ids: + raise EngineException( + "vdu[id='{}']:int-cpd[id='{}']:int-virtual-link-desc='{}' must match an existing " + "int-virtual-link-desc".format( + vdu["id"], int_cpd["id"], int_cpd_ivld_id + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + + for df in get_iterable(indata.get("df")): + for vlp in get_iterable(df.get("virtual-link-profile")): + vlp_ivld_id = vlp.get("id") + if vlp_ivld_id and vlp_ivld_id not in all_ivld_ids: + raise EngineException( + "df[id='{}']:virtual-link-profile='{}' must match an existing " + "int-virtual-link-desc".format(df["id"], vlp_ivld_id), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) @staticmethod def validate_monitoring_params(indata): - for mp in get_iterable(indata.get("monitoring-param")): - if mp.get("vdu-monitoring-param"): - mp_vmp_mark = False - for vdu in get_iterable(indata.get("vdu")): - for vmp in get_iterable(vdu.get("monitoring-param")): - if vmp["id"] == mp["vdu-monitoring-param"].get("vdu-monitoring-param-ref") and vdu["id"] == \ - mp["vdu-monitoring-param"]["vdu-ref"]: - mp_vmp_mark = True - break - if mp_vmp_mark: - break + all_monitoring_params = set() + for ivld in get_iterable(indata.get("int-virtual-link-desc")): + for mp in get_iterable(ivld.get("monitoring-parameters")): + mp_id = mp.get("id") + if mp_id and mp_id in all_monitoring_params: + raise EngineException( + "Duplicated monitoring-parameter id in " + "int-virtual-link-desc[id='{}']:monitoring-parameters[id='{}']".format( + ivld["id"], mp_id + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) else: - raise EngineException("monitoring-param:vdu-monitoring-param:vdu-monitoring-param-ref='{}' not " - "defined at vdu[id='{}'] or vdu does not exist" - .format(mp["vdu-monitoring-param"]["vdu-monitoring-param-ref"], - mp["vdu-monitoring-param"]["vdu-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - elif mp.get("vdu-metric"): - mp_vm_mark = False - for vdu in get_iterable(indata.get("vdu")): - if vdu.get("vdu-configuration"): - for metric in get_iterable(vdu["vdu-configuration"].get("metrics")): - if metric["name"] == mp["vdu-metric"]["vdu-metric-name-ref"] and vdu["id"] == \ - mp["vdu-metric"]["vdu-ref"]: - mp_vm_mark = True - break - if mp_vm_mark: - break + all_monitoring_params.add(mp_id) + + for vdu in get_iterable(indata.get("vdu")): + for mp in get_iterable(vdu.get("monitoring-parameter")): + mp_id = mp.get("id") + if mp_id and mp_id in all_monitoring_params: + raise EngineException( + "Duplicated monitoring-parameter id in " + "vdu[id='{}']:monitoring-parameter[id='{}']".format( + vdu["id"], mp_id + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + else: + all_monitoring_params.add(mp_id) + + for df in get_iterable(indata.get("df")): + for mp in get_iterable(df.get("monitoring-parameter")): + mp_id = mp.get("id") + if mp_id and mp_id in all_monitoring_params: + raise EngineException( + "Duplicated monitoring-parameter id in " + "df[id='{}']:monitoring-parameter[id='{}']".format( + df["id"], mp_id + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) else: - raise EngineException("monitoring-param:vdu-metric:vdu-metric-name-ref='{}' not defined at " - "vdu[id='{}'] or vdu does not exist" - .format(mp["vdu-metric"]["vdu-metric-name-ref"], - mp["vdu-metric"]["vdu-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + all_monitoring_params.add(mp_id) @staticmethod def validate_scaling_group_descriptor(indata): - for sgd in get_iterable(indata.get("scaling-group-descriptor")): - for sp in get_iterable(sgd.get("scaling-policy")): - for sc in get_iterable(sp.get("scaling-criteria")): - for mp in get_iterable(indata.get("monitoring-param")): - if mp["id"] == get_iterable(sc.get("vnf-monitoring-param-ref")): - break - else: - raise EngineException("scaling-group-descriptor[name='{}']:scaling-criteria[name='{}']:" - "vnf-monitoring-param-ref='{}' not defined in any monitoring-param" - .format(sgd["name"], sc["name"], sc["vnf-monitoring-param-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - for sgd_vdu in get_iterable(sgd.get("vdu")): - sgd_vdu_mark = False - for vdu in get_iterable(indata.get("vdu")): - if vdu["id"] == sgd_vdu["vdu-id-ref"]: - sgd_vdu_mark = True - break - if sgd_vdu_mark: - break - else: - raise EngineException("scaling-group-descriptor[name='{}']:vdu-id-ref={} does not match any vdu" - .format(sgd["name"], sgd_vdu["vdu-id-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - for sca in get_iterable(sgd.get("scaling-config-action")): - if not indata.get("vnf-configuration"): - raise EngineException("'vnf-configuration' not defined in the descriptor but it is referenced by " - "scaling-group-descriptor[name='{}']:scaling-config-action" - .format(sgd["name"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - for primitive in get_iterable(indata["vnf-configuration"].get("config-primitive")): - if primitive["name"] == sca["vnf-config-primitive-name-ref"]: - break - else: - raise EngineException("scaling-group-descriptor[name='{}']:scaling-config-action:vnf-config-" - "primitive-name-ref='{}' does not match any " - "vnf-configuration:config-primitive:name" - .format(sgd["name"], sca["vnf-config-primitive-name-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + all_monitoring_params = set() + for ivld in get_iterable(indata.get("int-virtual-link-desc")): + for mp in get_iterable(ivld.get("monitoring-parameters")): + all_monitoring_params.add(mp.get("id")) + + for vdu in get_iterable(indata.get("vdu")): + for mp in get_iterable(vdu.get("monitoring-parameter")): + all_monitoring_params.add(mp.get("id")) + + for df in get_iterable(indata.get("df")): + for mp in get_iterable(df.get("monitoring-parameter")): + all_monitoring_params.add(mp.get("id")) + + for df in get_iterable(indata.get("df")): + for sa in get_iterable(df.get("scaling-aspect")): + for sp in get_iterable(sa.get("scaling-policy")): + for sc in get_iterable(sp.get("scaling-criteria")): + sc_monitoring_param = sc.get("vnf-monitoring-param-ref") + if ( + sc_monitoring_param + and sc_monitoring_param not in all_monitoring_params + ): + raise EngineException( + "df[id='{}']:scaling-aspect[id='{}']:scaling-policy" + "[name='{}']:scaling-criteria[name='{}']: " + "vnf-monitoring-param-ref='{}' not defined in any monitoring-param".format( + df["id"], + sa["id"], + sp["name"], + sc["name"], + sc_monitoring_param, + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + + for sca in get_iterable(sa.get("scaling-config-action")): + if ( + "lcm-operations-configuration" not in df + or "operate-vnf-op-config" + not in df["lcm-operations-configuration"] + or not utils.find_in_list( + df["lcm-operations-configuration"][ + "operate-vnf-op-config" + ].get("day1-2", []), + lambda config: config["id"] == indata["id"], + ) + ): + raise EngineException( + "'day1-2 configuration' not defined in the descriptor but it is " + "referenced by df[id='{}']:scaling-aspect[id='{}']:scaling-config-action".format( + df["id"], sa["id"] + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + for configuration in get_iterable( + df["lcm-operations-configuration"]["operate-vnf-op-config"].get( + "day1-2", [] + ) + ): + for primitive in get_iterable( + configuration.get("config-primitive") + ): + if ( + primitive["name"] + == sca["vnf-config-primitive-name-ref"] + ): + break + else: + raise EngineException( + "df[id='{}']:scaling-aspect[id='{}']:scaling-config-action:vnf-" + "config-primitive-name-ref='{}' does not match any " + "day1-2 configuration:config-primitive:name".format( + df["id"], + sa["id"], + sca["vnf-config-primitive-name-ref"], + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) def delete_extra(self, session, _id, db_content, not_send_msg=None): """ @@ -784,7 +1036,9 @@ class VnfdTopic(DescriptorTopic): links = {} links["self"] = {"href": "/vnfpkgm/v1/vnf_packages/{}".format(data["_id"])} links["vnfd"] = {"href": "/vnfpkgm/v1/vnf_packages/{}/vnfd".format(data["_id"])} - links["packageContent"] = {"href": "/vnfpkgm/v1/vnf_packages/{}/package_content".format(data["_id"])} + links["packageContent"] = { + "href": "/vnfpkgm/v1/vnf_packages/{}/package_content".format(data["_id"]) + } data["_links"] = links return super().sol005_projection(data) @@ -797,24 +1051,56 @@ class NsdTopic(DescriptorTopic): def __init__(self, db, fs, msg, auth): DescriptorTopic.__init__(self, db, fs, msg, auth) + def pyangbind_validation(self, item, data, force=False): + if self._descriptor_data_is_in_old_format(data): + raise EngineException( + "ERROR: Unsupported descriptor format. Please, use an ETSI SOL006 descriptor.", + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + try: + nsd_vnf_profiles = data.get("df", [{}])[0].get("vnf-profile", []) + mynsd = etsi_nfv_nsd.etsi_nfv_nsd() + pybindJSONDecoder.load_ietf_json( + {"nsd": {"nsd": [data]}}, + None, + None, + obj=mynsd, + path_helper=True, + skip_unknown=force, + ) + out = pybindJSON.dumps(mynsd, mode="ietf") + desc_out = self._remove_envelop(yaml.safe_load(out)) + desc_out = self._remove_yang_prefixes_from_descriptor(desc_out) + if nsd_vnf_profiles: + desc_out["df"][0]["vnf-profile"] = nsd_vnf_profiles + return desc_out + except Exception as e: + raise EngineException( + "Error in pyangbind validation: {}".format(str(e)), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + + @staticmethod + def _descriptor_data_is_in_old_format(data): + return ("nsd-catalog" in data) or ("nsd:nsd-catalog" in data) + @staticmethod def _remove_envelop(indata=None): if not indata: return {} clean_indata = indata - if clean_indata.get('nsd:nsd-catalog'): - clean_indata = clean_indata['nsd:nsd-catalog'] - elif clean_indata.get('nsd-catalog'): - clean_indata = clean_indata['nsd-catalog'] - if clean_indata.get('nsd'): - if not isinstance(clean_indata['nsd'], list) or len(clean_indata['nsd']) != 1: + if clean_indata.get("nsd"): + clean_indata = clean_indata["nsd"] + elif clean_indata.get("etsi-nfv-nsd:nsd"): + clean_indata = clean_indata["etsi-nfv-nsd:nsd"] + if clean_indata.get("nsd"): + if ( + not isinstance(clean_indata["nsd"], list) + or len(clean_indata["nsd"]) != 1 + ): raise EngineException("'nsd' must be a list of only one element") - clean_indata = clean_indata['nsd'][0] - elif clean_indata.get('nsd:nsd'): - if not isinstance(clean_indata['nsd:nsd'], list) or len(clean_indata['nsd:nsd']) != 1: - raise EngineException("'nsd:nsd' must be a list of only one element") - clean_indata = clean_indata['nsd:nsd'][0] + clean_indata = clean_indata["nsd"][0] return clean_indata def _validate_input_new(self, indata, storage_params, force=False): @@ -827,61 +1113,50 @@ class NsdTopic(DescriptorTopic): indata = self.pyangbind_validation("nsds", indata, force) # Cross references validation in the descriptor # TODO validata that if contains cloud-init-file or charms, have artifacts _admin.storage."pkg-dir" is not none - for vld in get_iterable(indata.get("vld")): - self.validate_vld_mgmt_network_with_ip_profile_ref(vld) - self.validate_vld_connection_point_refs(vld, indata) + for vld in get_iterable(indata.get("virtual-link-desc")): + self.validate_vld_mgmt_network_with_virtual_link_protocol_data(vld, indata) - for fgd in get_iterable(indata.get("vnffgd")): - self.validate_fgd_classifiers(fgd) + self.validate_vnf_profiles_vnfd_id(indata) return indata @staticmethod - def validate_vld_mgmt_network_with_ip_profile_ref(vld): - if vld.get("mgmt-network") and vld.get("ip-profile-ref"): - raise EngineException("Error at vld[id='{}']:ip-profile-ref" - " You cannot set an ip-profile when mgmt-network is True" - .format(vld["id"]), http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - - @staticmethod - def validate_vld_connection_point_refs(vld, indata): - for vnfd_cp in get_iterable(vld.get("vnfd-connection-point-ref")): - for constituent_vnfd in get_iterable(indata.get("constituent-vnfd")): - if vnfd_cp["member-vnf-index-ref"] == constituent_vnfd["member-vnf-index"]: - if vnfd_cp.get("vnfd-id-ref") and vnfd_cp["vnfd-id-ref"] != constituent_vnfd["vnfd-id-ref"]: - raise EngineException("Error at vld[id='{}']:vnfd-connection-point-ref[vnfd-id-ref='{}'] " - "does not match constituent-vnfd[member-vnf-index='{}']:vnfd-id-ref" - " '{}'".format(vld["id"], vnfd_cp["vnfd-id-ref"], - constituent_vnfd["member-vnf-index"], - constituent_vnfd["vnfd-id-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) - break - else: - raise EngineException("Error at vld[id='{}']:vnfd-connection-point-ref[member-vnf-index-ref='{}'] " - "does not match any constituent-vnfd:member-vnf-index" - .format(vld["id"], vnfd_cp["member-vnf-index-ref"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + def validate_vld_mgmt_network_with_virtual_link_protocol_data(vld, indata): + if not vld.get("mgmt-network"): + return + vld_id = vld.get("id") + for df in get_iterable(indata.get("df")): + for vlp in get_iterable(df.get("virtual-link-profile")): + if vld_id and vld_id == vlp.get("virtual-link-desc-id"): + if vlp.get("virtual-link-protocol-data"): + raise EngineException( + "Error at df[id='{}']:virtual-link-profile[id='{}']:virtual-link-" + "protocol-data You cannot set a virtual-link-protocol-data " + "when mgmt-network is True".format(df["id"], vlp["id"]), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) @staticmethod - def validate_fgd_classifiers(fgd): - for cls in get_iterable(fgd.get("classifier")): - rspref = cls.get("rsp-id-ref") - for rsp in get_iterable(fgd.get("rsp")): - rspid = rsp.get("id") - if rspid and rspref and rspid == rspref: - break - else: - raise EngineException( - "Error at vnffgd[id='{}']:classifier[id='{}']:rsp-id-ref '{}' does not match any rsp:id" - .format(fgd["id"], cls["id"], rspref), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + def validate_vnf_profiles_vnfd_id(indata): + all_vnfd_ids = set(get_iterable(indata.get("vnfd-id"))) + for df in get_iterable(indata.get("df")): + for vnf_profile in get_iterable(df.get("vnf-profile")): + vnfd_id = vnf_profile.get("vnfd-id") + if vnfd_id and vnfd_id not in all_vnfd_ids: + raise EngineException( + "Error at df[id='{}']:vnf_profile[id='{}']:vnfd-id='{}' " + "does not match any vnfd-id".format( + df["id"], vnf_profile["id"], vnfd_id + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) def _validate_input_edit(self, indata, content, force=False): # not needed to validate with pyangbind becuase it will be validated at check_conflict_on_edit """ indata looks as follows: - - In the new case (conformant) - {'nsdOperationalState': 'DISABLED', 'userDefinedData': {'id': 'string23', + - In the new case (conformant) + {'nsdOperationalState': 'DISABLED', 'userDefinedData': {'id': 'string23', '_id': 'c6ddc544-cede-4b94-9ebe-be07b298a3c1', 'name': 'simon46'}} - In the old case (backwards-compatible) {'id': 'string23', '_id': 'c6ddc544-cede-4b94-9ebe-be07b298a3c1', 'name': 'simon46'} @@ -893,24 +1168,37 @@ class NsdTopic(DescriptorTopic): if indata["nsdOperationalState"] in ("ENABLED", "DISABLED"): indata["_admin"]["operationalState"] = indata.pop("nsdOperationalState") else: - raise EngineException("State '{}' is not a valid operational state" - .format(indata["nsdOperationalState"]), - http_code=HTTPStatus.BAD_REQUEST) + raise EngineException( + "State '{}' is not a valid operational state".format( + indata["nsdOperationalState"] + ), + http_code=HTTPStatus.BAD_REQUEST, + ) - # In the case of user defined data, we need to put the data in the root of the object + # In the case of user defined data, we need to put the data in the root of the object # to preserve current expected behaviour if "userDefinedData" in indata: data = indata.pop("userDefinedData") if type(data) == dict: indata["_admin"]["userDefinedData"] = data else: - raise EngineException("userDefinedData should be an object, but is '{}' instead" - .format(type(data)), - http_code=HTTPStatus.BAD_REQUEST) - if ("operationalState" in indata["_admin"] and - content["_admin"]["operationalState"] == indata["_admin"]["operationalState"]): - raise EngineException("nsdOperationalState already {}".format(content["_admin"]["operationalState"]), - http_code=HTTPStatus.CONFLICT) + raise EngineException( + "userDefinedData should be an object, but is '{}' instead".format( + type(data) + ), + http_code=HTTPStatus.BAD_REQUEST, + ) + if ( + "operationalState" in indata["_admin"] + and content["_admin"]["operationalState"] + == indata["_admin"]["operationalState"] + ): + raise EngineException( + "nsdOperationalState already {}".format( + content["_admin"]["operationalState"] + ), + http_code=HTTPStatus.CONFLICT, + ) return indata def _check_descriptor_dependencies(self, session, descriptor): @@ -923,48 +1211,65 @@ class NsdTopic(DescriptorTopic): """ if session["force"]: return - member_vnfd_index = self._get_descriptor_constituent_vnfds_by_member_vnfd_index(session, descriptor) + vnfds_index = self._get_descriptor_constituent_vnfds_index(session, descriptor) # Cross references validation in the descriptor and vnfd connection point validation - for vld in get_iterable(descriptor.get("vld")): - self.validate_vld_connection_point_refs_vnfd_connection_points(vld, member_vnfd_index) - - def _get_descriptor_constituent_vnfds_by_member_vnfd_index(self, session, descriptor): - member_vnfd_index = {} - if descriptor.get("constituent-vnfd") and not session["force"]: - for vnf in descriptor["constituent-vnfd"]: - vnfd_id = vnf["vnfd-id-ref"] + for df in get_iterable(descriptor.get("df")): + self.validate_df_vnf_profiles_constituent_connection_points(df, vnfds_index) + + def _get_descriptor_constituent_vnfds_index(self, session, descriptor): + vnfds_index = {} + if descriptor.get("vnfd-id") and not session["force"]: + for vnfd_id in get_iterable(descriptor.get("vnfd-id")): query_filter = self._get_project_filter(session) query_filter["id"] = vnfd_id vnf_list = self.db.get_list("vnfds", query_filter) if not vnf_list: - raise EngineException("Descriptor error at 'constituent-vnfd':'vnfd-id-ref'='{}' references a non " - "existing vnfd".format(vnfd_id), http_code=HTTPStatus.CONFLICT) - - member_vnfd_index[vnf["member-vnf-index"]] = vnf_list[0] - return member_vnfd_index + raise EngineException( + "Descriptor error at 'vnfd-id'='{}' references a non " + "existing vnfd".format(vnfd_id), + http_code=HTTPStatus.CONFLICT, + ) + vnfds_index[vnfd_id] = vnf_list[0] + return vnfds_index @staticmethod - def validate_vld_connection_point_refs_vnfd_connection_points(vld, member_vnfd_index): - for referenced_vnfd_cp in get_iterable(vld.get("vnfd-connection-point-ref")): - # look if this vnfd contains this connection point - vnfd = member_vnfd_index.get(referenced_vnfd_cp["member-vnf-index-ref"]) - for vnfd_cp in get_iterable(vnfd.get("connection-point")): - if referenced_vnfd_cp.get("vnfd-connection-point-ref") == vnfd_cp["name"]: - break - else: - raise EngineException( - "Error at vld[id='{}']:vnfd-connection-point-ref[member-vnf-index-ref='{}']:vnfd-" - "connection-point-ref='{}' references a non existing conection-point:name inside vnfd '{}'" - .format(vld["id"], referenced_vnfd_cp["member-vnf-index-ref"], - referenced_vnfd_cp["vnfd-connection-point-ref"], vnfd["id"]), - http_code=HTTPStatus.UNPROCESSABLE_ENTITY) + def validate_df_vnf_profiles_constituent_connection_points(df, vnfds_index): + for vnf_profile in get_iterable(df.get("vnf-profile")): + vnfd = vnfds_index.get(vnf_profile["vnfd-id"]) + all_vnfd_ext_cpds = set() + for ext_cpd in get_iterable(vnfd.get("ext-cpd")): + if ext_cpd.get("id"): + all_vnfd_ext_cpds.add(ext_cpd.get("id")) + + for virtual_link in get_iterable( + vnf_profile.get("virtual-link-connectivity") + ): + for vl_cpd in get_iterable(virtual_link.get("constituent-cpd-id")): + vl_cpd_id = vl_cpd.get("constituent-cpd-id") + if vl_cpd_id and vl_cpd_id not in all_vnfd_ext_cpds: + raise EngineException( + "Error at df[id='{}']:vnf-profile[id='{}']:virtual-link-connectivity" + "[virtual-link-profile-id='{}']:constituent-cpd-id='{}' references a " + "non existing ext-cpd:id inside vnfd '{}'".format( + df["id"], + vnf_profile["id"], + virtual_link["virtual-link-profile-id"], + vl_cpd_id, + vnfd["id"], + ), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) def check_conflict_on_edit(self, session, final_content, edit_content, _id): - super().check_conflict_on_edit(session, final_content, edit_content, _id) + final_content = super().check_conflict_on_edit( + session, final_content, edit_content, _id + ) self._check_descriptor_dependencies(session, final_content) + return final_content + def check_conflict_on_del(self, session, _id, db_content): """ Check that there is not any NSR that uses this NSD. Only NSRs belonging to this project are considered. Note @@ -985,14 +1290,19 @@ class NsdTopic(DescriptorTopic): _filter = self._get_project_filter(session) _filter["nsd-id"] = _id if self.db.get_list("nsrs", _filter): - raise EngineException("There is at least one NS using this descriptor", http_code=HTTPStatus.CONFLICT) + raise EngineException( + "There is at least one NS instance using this descriptor", + http_code=HTTPStatus.CONFLICT, + ) # check NSD referenced by NST del _filter["nsd-id"] _filter["netslice-subnet.ANYINDEX.nsd-ref"] = descriptor_id if self.db.get_list("nsts", _filter): - raise EngineException("There is at least one NetSlice Template referencing this descriptor", - http_code=HTTPStatus.CONFLICT) + raise EngineException( + "There is at least one NetSlice Template referencing this descriptor", + http_code=HTTPStatus.CONFLICT, + ) def sol005_projection(self, data): data["nsdOnboardingState"] = data["_admin"]["onboardingState"] @@ -1001,7 +1311,9 @@ class NsdTopic(DescriptorTopic): links = {} links["self"] = {"href": "/nsd/v1/ns_descriptors/{}".format(data["_id"])} - links["nsd_content"] = {"href": "/nsd/v1/ns_descriptors/{}/nsd_content".format(data["_id"])} + links["nsd_content"] = { + "href": "/nsd/v1/ns_descriptors/{}/nsd_content".format(data["_id"]) + } data["_links"] = links return super().sol005_projection(data) @@ -1015,20 +1327,46 @@ class NstTopic(DescriptorTopic): def __init__(self, db, fs, msg, auth): DescriptorTopic.__init__(self, db, fs, msg, auth) + def pyangbind_validation(self, item, data, force=False): + try: + mynst = nst_im() + pybindJSONDecoder.load_ietf_json( + {"nst": [data]}, + None, + None, + obj=mynst, + path_helper=True, + skip_unknown=force, + ) + out = pybindJSON.dumps(mynst, mode="ietf") + desc_out = self._remove_envelop(yaml.safe_load(out)) + return desc_out + except Exception as e: + raise EngineException( + "Error in pyangbind validation: {}".format(str(e)), + http_code=HTTPStatus.UNPROCESSABLE_ENTITY, + ) + @staticmethod def _remove_envelop(indata=None): if not indata: return {} clean_indata = indata - if clean_indata.get('nst'): - if not isinstance(clean_indata['nst'], list) or len(clean_indata['nst']) != 1: + if clean_indata.get("nst"): + if ( + not isinstance(clean_indata["nst"], list) + or len(clean_indata["nst"]) != 1 + ): raise EngineException("'nst' must be a list only one element") - clean_indata = clean_indata['nst'][0] - elif clean_indata.get('nst:nst'): - if not isinstance(clean_indata['nst:nst'], list) or len(clean_indata['nst:nst']) != 1: + clean_indata = clean_indata["nst"][0] + elif clean_indata.get("nst:nst"): + if ( + not isinstance(clean_indata["nst:nst"], list) + or len(clean_indata["nst:nst"]) != 1 + ): raise EngineException("'nst:nst' must be a list only one element") - clean_indata = clean_indata['nst:nst'][0] + clean_indata = clean_indata["nst:nst"][0] return clean_indata def _validate_input_new(self, indata, storage_params, force=False): @@ -1052,13 +1390,19 @@ class NstTopic(DescriptorTopic): filter_q = self._get_project_filter(session) filter_q["id"] = nsd_id if not self.db.get_list("nsds", filter_q): - raise EngineException("Descriptor error at 'netslice-subnet':'nsd-ref'='{}' references a non " - "existing nsd".format(nsd_id), http_code=HTTPStatus.CONFLICT) + raise EngineException( + "Descriptor error at 'netslice-subnet':'nsd-ref'='{}' references a non " + "existing nsd".format(nsd_id), + http_code=HTTPStatus.CONFLICT, + ) def check_conflict_on_edit(self, session, final_content, edit_content, _id): - super().check_conflict_on_edit(session, final_content, edit_content, _id) + final_content = super().check_conflict_on_edit( + session, final_content, edit_content, _id + ) self._check_descriptor_dependencies(session, final_content) + return final_content def check_conflict_on_del(self, session, _id, db_content): """ @@ -1076,8 +1420,10 @@ class NstTopic(DescriptorTopic): _filter = self._get_project_filter(session) _filter["_admin.nst-id"] = _id if self.db.get_list("nsis", _filter): - raise EngineException("there is at least one Netslice Instance using this descriptor", - http_code=HTTPStatus.CONFLICT) + raise EngineException( + "there is at least one Netslice Instance using this descriptor", + http_code=HTTPStatus.CONFLICT, + ) def sol005_projection(self, data): data["onboardingState"] = data["_admin"]["onboardingState"] @@ -1123,7 +1469,10 @@ class PduTopic(BaseTopic): _filter = self._get_project_filter(session) _filter["vdur.pdu-id"] = _id if self.db.get_list("vnfrs", _filter): - raise EngineException("There is at least one VNF using this PDU", http_code=HTTPStatus.CONFLICT) + raise EngineException( + "There is at least one VNF instance using this PDU", + http_code=HTTPStatus.CONFLICT, + ) class VnfPkgOpTopic(BaseTopic): @@ -1136,16 +1485,22 @@ class VnfPkgOpTopic(BaseTopic): BaseTopic.__init__(self, db, fs, msg, auth) def edit(self, session, _id, indata=None, kwargs=None, content=None): - raise EngineException("Method 'edit' not allowed for topic '{}'".format(self.topic), - HTTPStatus.METHOD_NOT_ALLOWED) + raise EngineException( + "Method 'edit' not allowed for topic '{}'".format(self.topic), + HTTPStatus.METHOD_NOT_ALLOWED, + ) def delete(self, session, _id, dry_run=False): - raise EngineException("Method 'delete' not allowed for topic '{}'".format(self.topic), - HTTPStatus.METHOD_NOT_ALLOWED) + raise EngineException( + "Method 'delete' not allowed for topic '{}'".format(self.topic), + HTTPStatus.METHOD_NOT_ALLOWED, + ) def delete_list(self, session, filter_q=None): - raise EngineException("Method 'delete_list' not allowed for topic '{}'".format(self.topic), - HTTPStatus.METHOD_NOT_ALLOWED) + raise EngineException( + "Method 'delete_list' not allowed for topic '{}'".format(self.topic), + HTTPStatus.METHOD_NOT_ALLOWED, + ) def new(self, rollback, session, indata=None, kwargs=None, headers=None): """ @@ -1173,7 +1528,9 @@ class VnfPkgOpTopic(BaseTopic): juju_bundle = kdu.get("juju-bundle") break else: - raise EngineException("Not found vnfd[id='{}']:kdu[name='{}']".format(vnfpkg_id, kdu_name)) + raise EngineException( + "Not found vnfd[id='{}']:kdu[name='{}']".format(vnfpkg_id, kdu_name) + ) if helm_chart: indata["helm-chart"] = helm_chart match = fullmatch(r"([^/]*)/([^/]*)", helm_chart) @@ -1183,8 +1540,11 @@ class VnfPkgOpTopic(BaseTopic): match = fullmatch(r"([^/]*)/([^/]*)", juju_bundle) repo_name = match.group(1) if match else None else: - raise EngineException("Found neither 'helm-chart' nor 'juju-bundle' in vnfd[id='{}']:kdu[name='{}']" - .format(vnfpkg_id, kdu_name)) + raise EngineException( + "Found neither 'helm-chart' nor 'juju-bundle' in vnfd[id='{}']:kdu[name='{}']".format( + vnfpkg_id, kdu_name + ) + ) if repo_name: del filter_q["_id"] filter_q["name"] = repo_name @@ -1208,9 +1568,11 @@ class VnfPkgOpTopic(BaseTopic): "links": { "self": "/osm/vnfpkgm/v1/vnfpkg_op_occs/" + vnfpkgop_id, "vnfpkg": "/osm/vnfpkgm/v1/vnf_packages/" + vnfpkg_id, - } + }, } - self.format_on_new(vnfpkgop_desc, session["project_id"], make_public=session["public"]) + self.format_on_new( + vnfpkgop_desc, session["project_id"], make_public=session["public"] + ) ctime = vnfpkgop_desc["_admin"]["created"] vnfpkgop_desc["statusEnteredTime"] = ctime vnfpkgop_desc["startTime"] = ctime