From: Mark Beierl Date: Wed, 10 May 2023 17:55:55 +0000 (-0400) Subject: Update to Python 3.10 and Ubuntu 22.04 X-Git-Tag: release-v14.0-start~14 X-Git-Url: https://osm.etsi.org/gitweb/?a=commitdiff_plain;h=refs%2Fchanges%2F66%2F13366%2F2;p=osm%2FNBI.git Update to Python 3.10 and Ubuntu 22.04 Removed stale test file that has linting errors Removed event loops Updated Python dependencies Change-Id: I9462b0d67ea6b5bd4c869b5f2bc8d6c57d78660c Signed-off-by: Mark Beierl --- diff --git a/Dockerfile b/Dockerfile index 2a64fdc..c4c9b6c 100644 --- a/Dockerfile +++ b/Dockerfile @@ -21,7 +21,7 @@ # devops-stages/stage-build.sh # -FROM ubuntu:20.04 +FROM ubuntu:22.04 ARG APT_PROXY RUN if [ ! -z $APT_PROXY ] ; then \ @@ -34,13 +34,15 @@ RUN DEBIAN_FRONTEND=noninteractive apt-get update && \ debhelper \ dh-python \ git \ - python3.8 \ + python3 \ python3-all \ - python3.8-dev \ - python3-setuptools + python3-dev \ + python3-setuptools \ + python3-pip \ + tox -RUN python3 -m easy_install pip==21.3.1 -RUN pip install tox==3.24.5 +ENV LC_ALL C.UTF-8 +ENV LANG C.UTF-8 RUN DEBIAN_FRONTEND=noninteractive apt-get -y install wget diff --git a/attic/run_test.py b/attic/run_test.py new file mode 100755 index 0000000..b7768ba --- /dev/null +++ b/attic/run_test.py @@ -0,0 +1,5889 @@ +#! /usr/bin/python3 +# -*- coding: utf-8 -*- + +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or +# implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import getopt +import sys +import requests +import json +import logging +import yaml + +# import json +# import tarfile +from time import sleep +from random import randint +import os +from sys import stderr +from uuid import uuid4 +import re + +__author__ = "Alfonso Tierno, alfonso.tiernosepulveda@telefonica.com" +__date__ = "$2018-03-01$" +__version__ = "0.3" +version_date = "Oct 2018" + + +def usage(): + print("Usage: ", sys.argv[0], "[options]") + print( + " Performs system tests over running NBI. It can be used for real OSM test using option '--test-osm'" + ) + print( + " If this is the case env variables 'OSMNBITEST_VIM_NAME' must be supplied to create a VIM if not exist " + "where deployment is done" + ) + print("OPTIONS") + print(" -h|--help: shows this help") + print(" --insecure: Allows non trusted https NBI server") + print(" --list: list available tests") + print( + " --manual-check: Deployment tests stop after deployed to allow manual inspection. Only make sense with " + "'--test-osm'" + ) + print(" -p|--password PASSWORD: NBI access password. 'admin' by default") + print(" ---project PROJECT: NBI access project. 'admin' by default") + print( + " --test TEST[,...]: Execute only a test or a comma separated list of tests" + ) + print( + " --params key=val: params to the previous test. key can be vnfd-files, nsd-file, ns-name, ns-config" + ) + print( + " --test-osm: If missing this test is intended for NBI only, no other OSM components are expected. Use " + "this flag to test the system. LCM and RO components are expected to be up and running" + ) + print( + " --timeout TIMEOUT: General NBI timeout, by default {}s".format(timeout) + ) + print( + " --timeout-deploy TIMEOUT: Timeout used for getting NS deployed, by default {}s".format( + timeout_deploy + ) + ) + print( + " --timeout-configure TIMEOUT: Timeout used for getting NS deployed and configured," + " by default {}s".format(timeout_configure) + ) + print(" -u|--user USERNAME: NBI access username. 'admin' by default") + print( + " --url URL: complete NBI server URL. 'https//localhost:9999/osm' by default" + ) + print(" -v|--verbose print debug information, can be used several times") + print(" --no-verbose remove verbosity") + print(" --version: prints current version") + print("ENV variables used for real deployment tests with option osm-test.") + print(" export OSMNBITEST_VIM_NAME=vim-name") + print(" export OSMNBITEST_VIM_URL=vim-url") + print(" export OSMNBITEST_VIM_TYPE=vim-type") + print(" export OSMNBITEST_VIM_TENANT=vim-tenant") + print(" export OSMNBITEST_VIM_USER=vim-user") + print(" export OSMNBITEST_VIM_PASSWORD=vim-password") + print(' export OSMNBITEST_VIM_CONFIG="vim-config"') + print(' export OSMNBITEST_NS_NAME="vim-config"') + return + + +r_header_json = {"Content-type": "application/json"} +headers_json = {"Content-type": "application/json", "Accept": "application/json"} +r_header_yaml = {"Content-type": "application/yaml"} +headers_yaml = {"Content-type": "application/yaml", "Accept": "application/yaml"} +r_header_text = {"Content-type": "text/plain"} +r_header_octect = {"Content-type": "application/octet-stream"} +headers_text = {"Accept": "text/plain,application/yaml"} +r_header_zip = {"Content-type": "application/zip"} +headers_zip = {"Accept": "application/zip,application/yaml"} +headers_zip_yaml = {"Accept": "application/yaml", "Content-type": "application/zip"} +headers_zip_json = {"Accept": "application/json", "Content-type": "application/zip"} +headers_txt_json = {"Accept": "application/json", "Content-type": "text/plain"} +r_headers_yaml_location_vnfd = { + "Location": "/vnfpkgm/v1/vnf_packages_content/", + "Content-Type": "application/yaml", +} +r_headers_yaml_location_nsd = { + "Location": "/nsd/v1/ns_descriptors_content/", + "Content-Type": "application/yaml", +} +r_headers_yaml_location_nst = { + "Location": "/nst/v1/netslice_templates_content", + "Content-Type": "application/yaml", +} +r_headers_yaml_location_nslcmop = { + "Location": "nslcm/v1/ns_lcm_op_occs/", + "Content-Type": "application/yaml", +} +r_headers_yaml_location_nsilcmop = { + "Location": "/osm/nsilcm/v1/nsi_lcm_op_occs/", + "Content-Type": "application/yaml", +} + +# test ones authorized +test_authorized_list = ( + ( + "AU1", + "Invalid vnfd id", + "GET", + "/vnfpkgm/v1/vnf_packages/non-existing-id", + headers_json, + None, + 404, + r_header_json, + "json", + ), + ( + "AU2", + "Invalid nsd id", + "GET", + "/nsd/v1/ns_descriptors/non-existing-id", + headers_yaml, + None, + 404, + r_header_yaml, + "yaml", + ), + ( + "AU3", + "Invalid nsd id", + "DELETE", + "/nsd/v1/ns_descriptors_content/non-existing-id", + headers_yaml, + None, + 404, + r_header_yaml, + "yaml", + ), +) +timeout = 120 # general timeout +timeout_deploy = 60 * 10 # timeout for NS deploying without charms +timeout_configure = 60 * 20 # timeout for NS deploying and configuring + + +class TestException(Exception): + pass + + +class TestRest: + def __init__( + self, + url_base, + header_base=None, + verify=False, + user="admin", + password="admin", + project="admin", + ): + self.url_base = url_base + if header_base is None: + self.header_base = {} + else: + self.header_base = header_base.copy() + self.s = requests.session() + self.s.headers = self.header_base + self.verify = verify + self.token = False + self.user = user + self.password = password + self.project = project + self.vim_id = None + # contains ID of tests obtained from Location response header. "" key contains last obtained id + self.last_id = "" + self.test_name = None + self.step = 0 # number of subtest under test + self.passed_tests = 0 + self.failed_tests = 0 + + def set_test_name(self, test_name): + self.test_name = test_name + self.step = 0 + self.last_id = "" + + def set_header(self, header): + self.s.headers.update(header) + + def set_tet_name(self, test_name): + self.test_name = test_name + + def unset_header(self, key): + if key in self.s.headers: + del self.s.headers[key] + + def test( + self, + description, + method, + url, + headers, + payload, + expected_codes, + expected_headers, + expected_payload, + store_file=None, + pooling=False, + ): + """ + Performs an http request and check http code response. Exit if different than allowed. It get the returned id + that can be used by following test in the URL with {name} where name is the name of the test + :param description: description of the test + :param method: HTTP method: GET,PUT,POST,DELETE,... + :param url: complete URL or relative URL + :param headers: request headers to add to the base headers + :param payload: Can be a dict, transformed to json, a text or a file if starts with '@' + :param expected_codes: expected response codes, can be int, int tuple or int range + :param expected_headers: expected response headers, dict with key values + :param expected_payload: expected payload, 0 if empty, 'yaml', 'json', 'text', 'zip', 'octet-stream' + :param store_file: filename to store content + :param pooling: if True do not count neither log this test. Because a pooling is done with many equal requests + :return: requests response + """ + r = None + try: + if not self.s: + self.s = requests.session() + # URL + if not url: + url = self.url_base + elif not url.startswith("http"): + url = self.url_base + url + + # replace url <> with the last ID + url = url.replace("<>", self.last_id) + if payload: + if isinstance(payload, str): + if payload.startswith("@"): + mode = "r" + file_name = payload[1:] + if payload.startswith("@b"): + mode = "rb" + file_name = payload[2:] + with open(file_name, mode) as f: + payload = f.read() + elif isinstance(payload, dict): + payload = json.dumps(payload) + + if not pooling: + test_description = "Test {}{} {} {} {}".format( + self.test_name, self.step, description, method, url + ) + logger.warning(test_description) + self.step += 1 + stream = False + if expected_payload in ("zip", "octet-string") or store_file: + stream = True + __retry = 0 + while True: + try: + r = getattr(self.s, method.lower())( + url, + data=payload, + headers=headers, + verify=self.verify, + stream=stream, + ) + break + except requests.exceptions.ConnectionError as e: + if __retry == 2: + raise + logger.error("Exception {}. Retrying".format(e)) + __retry += 1 + + if expected_payload in ("zip", "octet-string") or store_file: + logger.debug("RX {}".format(r.status_code)) + else: + logger.debug("RX {}: {}".format(r.status_code, r.text)) + + # check response + if expected_codes: + if isinstance(expected_codes, int): + expected_codes = (expected_codes,) + if r.status_code not in expected_codes: + raise TestException( + "Got status {}. Expected {}. {}".format( + r.status_code, expected_codes, r.text + ) + ) + + if expected_headers: + for header_key, header_val in expected_headers.items(): + if header_key.lower() not in r.headers: + raise TestException("Header {} not present".format(header_key)) + if header_val and header_val.lower() not in r.headers[header_key]: + raise TestException( + "Header {} does not contain {} but {}".format( + header_key, header_val, r.headers[header_key] + ) + ) + + if expected_payload is not None: + if expected_payload == 0 and len(r.content) > 0: + raise TestException("Expected empty payload") + elif expected_payload == "json": + try: + r.json() + except Exception as e: + raise TestException( + "Expected json response payload, but got Exception {}".format( + e + ) + ) + elif expected_payload == "yaml": + try: + yaml.safe_load(r.text) + except Exception as e: + raise TestException( + "Expected yaml response payload, but got Exception {}".format( + e + ) + ) + elif expected_payload in ("zip", "octet-string"): + if len(r.content) == 0: + raise TestException( + "Expected some response payload, but got empty" + ) + # try: + # tar = tarfile.open(None, 'r:gz', fileobj=r.raw) + # for tarinfo in tar: + # tarname = tarinfo.name + # print(tarname) + # except Exception as e: + # raise TestException("Expected zip response payload, but got Exception {}".format(e)) + elif expected_payload == "text": + if len(r.content) == 0: + raise TestException( + "Expected some response payload, but got empty" + ) + # r.text + if store_file: + with open(store_file, "wb") as fd: + for chunk in r.iter_content(chunk_size=128): + fd.write(chunk) + + location = r.headers.get("Location") + if location: + _id = location[location.rfind("/") + 1 :] + if _id: + self.last_id = str(_id) + if not pooling: + self.passed_tests += 1 + return r + except TestException as e: + self.failed_tests += 1 + r_status_code = None + r_text = None + if r: + r_status_code = r.status_code + r_text = r.text + logger.error("{} \nRX code{}: {}".format(e, r_status_code, r_text)) + return None + # exit(1) + except IOError as e: + if store_file: + logger.error("Cannot open file {}: {}".format(store_file, e)) + else: + logger.error("Exception: {}".format(e), exc_info=True) + self.failed_tests += 1 + return None + # exit(1) + except requests.exceptions.RequestException as e: + logger.error("Exception: {}".format(e)) + + def get_autorization(self): # user=None, password=None, project=None): + if ( + self.token + ): # and self.user == user and self.password == password and self.project == project: + return + # self.user = user + # self.password = password + # self.project = project + r = self.test( + "Obtain token", + "POST", + "/admin/v1/tokens", + headers_json, + { + "username": self.user, + "password": self.password, + "project_id": self.project, + }, + (200, 201), + r_header_json, + "json", + ) + if not r: + return + response = r.json() + self.token = response["id"] + self.set_header({"Authorization": "Bearer {}".format(self.token)}) + + def remove_authorization(self): + if self.token: + self.test( + "Delete token", + "DELETE", + "/admin/v1/tokens/{}".format(self.token), + headers_json, + None, + (200, 201, 204), + None, + None, + ) + self.token = None + self.unset_header("Authorization") + + def get_create_vim(self, test_osm): + if self.vim_id: + return self.vim_id + self.get_autorization() + if test_osm: + vim_name = os.environ.get("OSMNBITEST_VIM_NAME") + if not vim_name: + raise TestException( + "Needed to define OSMNBITEST_VIM_XXX variables to create a real VIM for deployment" + ) + else: + vim_name = "fakeVim" + # Get VIM + r = self.test( + "Get VIM ID", + "GET", + "/admin/v1/vim_accounts?name={}".format(vim_name), + headers_json, + None, + 200, + r_header_json, + "json", + ) + if not r: + return + vims = r.json() + if vims: + return vims[0]["_id"] + # Add VIM + if test_osm: + # check needed environ parameters: + if not os.environ.get("OSMNBITEST_VIM_URL") or not os.environ.get( + "OSMNBITEST_VIM_TENANT" + ): + raise TestException( + "Env OSMNBITEST_VIM_URL and OSMNBITEST_VIM_TENANT are needed for create a real VIM" + " to deploy on whit the --test-osm option" + ) + vim_data = ( + "{{schema_version: '1.0', name: '{}', vim_type: {}, vim_url: '{}'," + "vim_tenant_name: '{}', " + "vim_user: {}, vim_password: {}" + ).format( + vim_name, + os.environ.get("OSMNBITEST_VIM_TYPE", "openstack"), + os.environ.get("OSMNBITEST_VIM_URL"), + os.environ.get("OSMNBITEST_VIM_TENANT"), + os.environ.get("OSMNBITEST_VIM_USER"), + os.environ.get("OSMNBITEST_VIM_PASSWORD"), + ) + if os.environ.get("OSMNBITEST_VIM_CONFIG"): + vim_data += " ,config: {}".format( + os.environ.get("OSMNBITEST_VIM_CONFIG") + ) + vim_data += "}" + else: + vim_data = ( + "{schema_version: '1.0', name: fakeVim, vim_type: openstack, vim_url: 'http://10.11.12.13/fake'" + ", vim_tenant_name: 'vimtenant', vim_user: vimuser, vim_password: vimpassword}" + ) + self.test( + "Create VIM", + "POST", + "/admin/v1/vim_accounts", + headers_yaml, + vim_data, + (201, 202), + {"Location": "/admin/v1/vim_accounts/", "Content-Type": "application/yaml"}, + "yaml", + ) + return self.last_id + + def print_results(self): + print("\n\n\n--------------------------------------------") + print( + "TEST RESULTS: Total: {}, Passed: {}, Failed: {}".format( + self.passed_tests + self.failed_tests, + self.passed_tests, + self.failed_tests, + ) + ) + print("--------------------------------------------") + + def wait_until_delete(self, url_op, timeout_delete): + """ + Make a pooling until topic is not present, because of deleted + :param url_op: + :param timeout_delete: + :return: + """ + description = "Wait to topic being deleted" + test_description = "Test {}{} {} {} {}".format( + self.test_name, self.step, description, "GET", url_op + ) + logger.warning(test_description) + self.step += 1 + + wait = timeout_delete + while wait >= 0: + r = self.test( + description, + "GET", + url_op, + headers_yaml, + None, + (200, 404), + None, + r_header_yaml, + "yaml", + pooling=True, + ) + if not r: + return + if r.status_code == 404: + self.passed_tests += 1 + break + elif r.status_code == 200: + wait -= 5 + sleep(5) + else: + raise TestException( + "Topic is not deleted after {} seconds".format(timeout_delete) + ) + self.failed_tests += 1 + + def wait_operation_ready(self, ns_nsi, opp_id, timeout, expected_fail=False): + """ + Wait until nslcmop or nsilcmop finished + :param ns_nsi: "ns" o "nsi" + :param opp_id: Id o fthe operation + :param timeout: + :param expected_fail: + :return: None. Updates passed/failed_tests + """ + if ns_nsi == "ns": + url_op = "/nslcm/v1/ns_lcm_op_occs/{}".format(opp_id) + else: + url_op = "/nsilcm/v1/nsi_lcm_op_occs/{}".format(opp_id) + description = "Wait to {} lcm operation complete".format(ns_nsi) + test_description = "Test {}{} {} {} {}".format( + self.test_name, self.step, description, "GET", url_op + ) + logger.warning(test_description) + self.step += 1 + wait = timeout + while wait >= 0: + r = self.test( + description, + "GET", + url_op, + headers_json, + None, + 200, + r_header_json, + "json", + pooling=True, + ) + if not r: + return + nslcmop = r.json() + if "COMPLETED" in nslcmop["operationState"]: + if expected_fail: + logger.error( + "NS terminate has success, expecting failing: {}".format( + nslcmop["detailed-status"] + ) + ) + self.failed_tests += 1 + else: + self.passed_tests += 1 + break + elif "FAILED" in nslcmop["operationState"]: + if not expected_fail: + logger.error( + "NS terminate has failed: {}".format(nslcmop["detailed-status"]) + ) + self.failed_tests += 1 + else: + self.passed_tests += 1 + break + + print(".", end="", file=stderr) + wait -= 10 + sleep(10) + else: + self.failed_tests += 1 + logger.error( + "NS instantiate is not terminate after {} seconds".format(timeout) + ) + return + print("", file=stderr) + + +class TestNonAuthorized: + description = "Test invalid URLs. methods and no authorization" + + @staticmethod + def run(engine, test_osm, manual_check, test_params=None): + engine.set_test_name("NonAuth") + engine.remove_authorization() + test_not_authorized_list = ( + ( + "Invalid token", + "GET", + "/admin/v1/users", + headers_json, + None, + 401, + r_header_json, + "json", + ), + ( + "Invalid URL", + "POST", + "/admin/v1/nonexist", + headers_yaml, + None, + 405, + r_header_yaml, + "yaml", + ), + ( + "Invalid version", + "DELETE", + "/admin/v2/users", + headers_yaml, + None, + 405, + r_header_yaml, + "yaml", + ), + ) + for t in test_not_authorized_list: + engine.test(*t) + + +class TestUsersProjects: + description = "test project and user creation" + + @staticmethod + def run(engine, test_osm, manual_check, test_params=None): + engine.set_test_name("UserProject") + # backend = test_params.get("backend") if test_params else None # UNUSED + + # Initialisation + p1 = p2 = p3 = None + padmin = pbad = None + u1 = u2 = u3 = u4 = None + + engine.get_autorization() + + res = engine.test( + "Create project non admin 1", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "P1"}, + (201, 204), + {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, + "json", + ) + p1 = engine.last_id if res else None + + res = engine.test( + "Create project admin", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "Padmin", "admin": True}, + (201, 204), + {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, + "json", + ) + padmin = engine.last_id if res else None + + res = engine.test( + "Create project bad format", + "POST", + "/admin/v1/projects", + headers_json, + {"name": 1}, + (400, 422), + r_header_json, + "json", + ) + pbad = engine.last_id if res else None + + res = engine.test( + "Get project admin role", + "GET", + "/admin/v1/roles?name=project_admin", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + rpa = res.json()[0]["_id"] if res else None + res = engine.test( + "Get project user role", + "GET", + "/admin/v1/roles?name=project_user", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + rpu = res.json()[0]["_id"] if res else None + res = engine.test( + "Get system admin role", + "GET", + "/admin/v1/roles?name=system_admin", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + rsa = res.json()[0]["_id"] if res else None + + data = {"username": "U1", "password": "pw1"} + p2 = uuid4().hex + data["project_role_mappings"] = [ + {"project": p1, "role": rpa}, + {"project": p2, "role": rpa}, + {"project": padmin, "role": rpu}, + ] + rc = 201 + xhd = {"Location": "/admin/v1/users/", "Content-Type": "application/json"} + res = engine.test( + "Create user with bad project and force", + "POST", + "/admin/v1/users?FORCE=True", + headers_json, + data, + rc, + xhd, + "json", + ) + if res: + u1 = engine.last_id + else: + # User is created sometimes even though an exception is raised + res = engine.test( + "Get user U1", + "GET", + "/admin/v1/users?username=U1", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + u1 = res.json()[0]["_id"] if res else None + + data = {"username": "U2", "password": "pw2"} + data["project_role_mappings"] = [ + {"project": p1, "role": rpa}, + {"project": padmin, "role": rsa}, + ] + res = engine.test( + "Create user 2", + "POST", + "/admin/v1/users", + headers_json, + data, + 201, + {"Location": "/admin/v1/users/", "Content-Type": "application/json"}, + "json", + ) + u2 = engine.last_id if res else None + + if u1: + ftt = "project_role_mappings" + xpr = [{"project": p1, "role": rpa}, {"project": padmin, "role": rpu}] + data = {ftt: xpr} + engine.test( + "Edit user U1, delete P2 project", + "PATCH", + "/admin/v1/users/" + u1, + headers_json, + data, + 204, + None, + None, + ) + res = engine.test( + "Check user U1, contains the right projects", + "GET", + "/admin/v1/users/" + u1, + headers_json, + None, + 200, + None, + json, + ) + if res: + rj = res.json() + xpr[0]["project_name"] = "P1" + xpr[0]["role_name"] = "project_admin" + xpr[1]["project_name"] = "Padmin" + xpr[1]["role_name"] = "project_user" + ok = True + for pr in rj[ftt]: + if pr not in xpr: + ok = False + for pr in xpr: + if pr not in rj[ftt]: + ok = False + if not ok: + logger.error( + "User {} '{}' are different than expected '{}'. Edition was not done properly".format( + ftt, rj[ftt], xpr + ) + ) + engine.failed_tests += 1 + + p2 = None # To prevent deletion attempts + + # Add a test of 'default project' for Keystone? + + if u2: + engine.test( + "Edit user U2, change password", + "PUT", + "/admin/v1/users/" + u2, + headers_json, + {"password": "pw2_new"}, + 204, + None, + None, + ) + + if p1: + engine.test( + "Change to project P1 non existing", + "POST", + "/admin/v1/tokens/", + headers_json, + {"project_id": p1}, + 401, + r_header_json, + "json", + ) + + if u2 and p1: + res = engine.test( + "Change to user U2 project P1", + "POST", + "/admin/v1/tokens", + headers_json, + {"username": "U2", "password": "pw2_new", "project_id": "P1"}, + (200, 201), + r_header_json, + "json", + ) + if res: + rj = res.json() + engine.set_header({"Authorization": "Bearer {}".format(rj["id"])}) + + engine.test( + "Edit user projects non admin", + "PUT", + "/admin/v1/users/U1", + headers_json, + {"remove_project_role_mappings": [{"project": "P1", "role": None}]}, + 401, + r_header_json, + "json", + ) + + res = engine.test( + "Add new project non admin", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "P2"}, + 401, + r_header_json, + "json", + ) + if res is None or res.status_code == 201: + # The project has been created even though it shouldn't + res = engine.test( + "Get project P2", + "GET", + "/admin/v1/projects/P2", + headers_json, + None, + 200, + r_header_json, + "json", + ) + p2 = res.json()["_id"] if res else None + + if p1: + data = {"username": "U3", "password": "pw3"} + data["project_role_mappings"] = [{"project": p1, "role": rpu}] + res = engine.test( + "Add new user non admin", + "POST", + "/admin/v1/users", + headers_json, + data, + 401, + r_header_json, + "json", + ) + if res is None or res.status_code == 201: + # The user has been created even though it shouldn't + res = engine.test( + "Get user U3", + "GET", + "/admin/v1/users/U3", + headers_json, + None, + 200, + r_header_json, + "json", + ) + u3 = res.json()["_id"] if res else None + else: + u3 = None + + if padmin: + res = engine.test( + "Change to user U2 project Padmin", + "POST", + "/admin/v1/tokens", + headers_json, + { + "project_id": "Padmin" + }, # Caused a Keystone authentication error + # {"username": "U2", "password": "pw2_new", "project_id": "Padmin"}, + (200, 201), + r_header_json, + "json", + ) + if res: + rj = res.json() + engine.set_header( + {"Authorization": "Bearer {}".format(rj["id"])} + ) + + res = engine.test( + "Add new project admin", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "P3"}, + (201, 204), + { + "Location": "/admin/v1/projects/", + "Content-Type": "application/json", + }, + "json", + ) + p3 = engine.last_id if res else None + + if p1: + data = {"username": "U4", "password": "pw4"} + data["project_role_mappings"] = [ + {"project": p1, "role": rpa} + ] + res = engine.test( + "Add new user admin", + "POST", + "/admin/v1/users", + headers_json, + data, + (201, 204), + { + "Location": "/admin/v1/users/", + "Content-Type": "application/json", + }, + "json", + ) + u4 = engine.last_id if res else None + else: + u4 = None + + if u4 and p3: + data = { + "project_role_mappings": [{"project": p3, "role": rpa}] + } + engine.test( + "Edit user projects admin", + "PUT", + "/admin/v1/users/U4", + headers_json, + data, + 204, + None, + None, + ) + # Project is deleted even though it shouldn't - PROVISIONAL? + res = engine.test( + "Delete project P3 conflict", + "DELETE", + "/admin/v1/projects/" + p3, + headers_json, + None, + 409, + None, + None, + ) + if res and res.status_code in (200, 204): + p3 = None + if p3: + res = engine.test( + "Delete project P3 forcing", + "DELETE", + "/admin/v1/projects/" + p3 + "?FORCE=True", + headers_json, + None, + 204, + None, + None, + ) + if res and res.status_code in (200, 204): + p3 = None + + if u2: + res = engine.test( + "Delete user U2. Conflict deleting own user", + "DELETE", + "/admin/v1/users/" + u2, + headers_json, + None, + 409, + r_header_json, + "json", + ) + if res is None or res.status_code in (200, 204): + u2 = None + if u4: + res = engine.test( + "Delete user U4", + "DELETE", + "/admin/v1/users/" + u4, + headers_json, + None, + 204, + None, + None, + ) + if res and res.status_code in (200, 204): + u4 = None + if p3: + res = engine.test( + "Delete project P3", + "DELETE", + "/admin/v1/projects/" + p3, + headers_json, + None, + 204, + None, + None, + ) + if res and res.status_code in (200, 204): + p3 = None + + if u3: + res = engine.test( + "Delete user U3", + "DELETE", + "/admin/v1/users/" + u3, + headers_json, + None, + 204, + None, + None, + ) + if res: + u3 = None + + # change to admin + engine.remove_authorization() # To force get authorization + engine.get_autorization() + if u1: + engine.test( + "Delete user U1", + "DELETE", + "/admin/v1/users/" + u1, + headers_json, + None, + 204, + None, + None, + ) + if u2: + engine.test( + "Delete user U2", + "DELETE", + "/admin/v1/users/" + u2, + headers_json, + None, + 204, + None, + None, + ) + if u3: + engine.test( + "Delete user U3", + "DELETE", + "/admin/v1/users/" + u3, + headers_json, + None, + 204, + None, + None, + ) + if u4: + engine.test( + "Delete user U4", + "DELETE", + "/admin/v1/users/" + u4, + headers_json, + None, + 204, + None, + None, + ) + if p1: + engine.test( + "Delete project P1", + "DELETE", + "/admin/v1/projects/" + p1, + headers_json, + None, + 204, + None, + None, + ) + if p2: + engine.test( + "Delete project P2", + "DELETE", + "/admin/v1/projects/" + p2, + headers_json, + None, + 204, + None, + None, + ) + if p3: + engine.test( + "Delete project P3", + "DELETE", + "/admin/v1/projects/" + p3, + headers_json, + None, + 204, + None, + None, + ) + if padmin: + engine.test( + "Delete project Padmin", + "DELETE", + "/admin/v1/projects/" + padmin, + headers_json, + None, + 204, + None, + None, + ) + if pbad: + engine.test( + "Delete bad project", + "DELETE", + "/admin/v1/projects/" + pbad, + headers_json, + None, + 204, + None, + None, + ) + + # BEGIN New Tests - Addressing Projects/Users by Name/ID + pid1 = pid2 = None + uid1 = uid2 = None + res = engine.test( + "Create new project P1", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "P1"}, + 201, + {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, + "json", + ) + if res: + pid1 = res.json()["id"] + # print("# pid =", pid1) + res = engine.test( + "Create new project P2", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "P2"}, + 201, + {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, + "json", + ) + if res: + pid2 = res.json()["id"] + # print("# pid =", pid2) + data = {"username": "U1", "password": "pw1"} + data["project_role_mappings"] = [{"project": pid1, "role": rpu}] + res = engine.test( + "Create new user U1", + "POST", + "/admin/v1/users", + headers_json, + data, + 201, + {"Location": "/admin/v1/users/", "Content-Type": "application/json"}, + "json", + ) + if res: + uid1 = res.json()["id"] + # print("# uid =", uid1) + data = {"username": "U2", "password": "pw2"} + data["project_role_mappings"] = [{"project": pid2, "role": rpu}] + res = engine.test( + "Create new user U2", + "POST", + "/admin/v1/users", + headers_json, + data, + 201, + {"Location": "/admin/v1/users/", "Content-Type": "application/json"}, + "json", + ) + if res: + uid2 = res.json()["id"] + # print("# uid =", uid2) + if pid1: + engine.test( + "Get Project P1 by Name", + "GET", + "/admin/v1/projects/P1", + headers_json, + None, + 200, + None, + "json", + ) + engine.test( + "Get Project P1 by ID", + "GET", + "/admin/v1/projects/" + pid1, + headers_json, + None, + 200, + None, + "json", + ) + if uid1: + engine.test( + "Get User U1 by Name", + "GET", + "/admin/v1/users/U1", + headers_json, + None, + 200, + None, + "json", + ) + engine.test( + "Get User U1 by ID", + "GET", + "/admin/v1/users/" + uid1, + headers_json, + None, + 200, + None, + "json", + ) + if pid1: + res = engine.test( + "Rename Project P1 by Name", + "PUT", + "/admin/v1/projects/P1", + headers_json, + {"name": "P3"}, + 204, + None, + None, + ) + if res: + engine.test( + "Get Project P1 by new Name", + "GET", + "/admin/v1/projects/P3", + headers_json, + None, + 200, + None, + "json", + ) + if pid2: + res = engine.test( + "Rename Project P2 by ID", + "PUT", + "/admin/v1/projects/" + pid2, + headers_json, + {"name": "P4"}, + 204, + None, + None, + ) + if res: + engine.test( + "Get Project P2 by new Name", + "GET", + "/admin/v1/projects/P4", + headers_json, + None, + 200, + None, + "json", + ) + + if uid1: + res = engine.test( + "Rename User U1 by Name", + "PUT", + "/admin/v1/users/U1", + headers_json, + {"username": "U3"}, + 204, + None, + None, + ) + if res: + engine.test( + "Get User U1 by new Name", + "GET", + "/admin/v1/users/U3", + headers_json, + None, + 200, + None, + "json", + ) + + if uid2: + res = engine.test( + "Rename User U2 by ID", + "PUT", + "/admin/v1/users/" + uid2, + headers_json, + {"username": "U4"}, + 204, + None, + None, + ) + if res: + engine.test( + "Get User U2 by new Name", + "GET", + "/admin/v1/users/U4", + headers_json, + None, + 200, + None, + "json", + ) + if uid1: + res = engine.test( + "Delete User U1 by Name", + "DELETE", + "/admin/v1/users/U3", + headers_json, + None, + 204, + None, + None, + ) + if res: + uid1 = None + + if uid2: + res = engine.test( + "Delete User U2 by ID", + "DELETE", + "/admin/v1/users/" + uid2, + headers_json, + None, + 204, + None, + None, + ) + if res: + uid2 = None + + if pid1: + res = engine.test( + "Delete Project P1 by Name", + "DELETE", + "/admin/v1/projects/P3", + headers_json, + None, + 204, + None, + None, + ) + if res: + pid1 = None + + if pid2: + res = engine.test( + "Delete Project P2 by ID", + "DELETE", + "/admin/v1/projects/" + pid2, + headers_json, + None, + 204, + None, + None, + ) + if res: + pid2 = None + + # END New Tests - Addressing Projects/Users by Name + + # CLEANUP + if pid1: + engine.test( + "Delete Project P1", + "DELETE", + "/admin/v1/projects/" + pid1, + headers_json, + None, + 204, + None, + None, + ) + if pid2: + engine.test( + "Delete Project P2", + "DELETE", + "/admin/v1/projects/" + pid2, + headers_json, + None, + 204, + None, + None, + ) + if uid1: + engine.test( + "Delete User U1", + "DELETE", + "/admin/v1/users/" + uid1, + headers_json, + None, + 204, + None, + None, + ) + if uid2: + engine.test( + "Delete User U2", + "DELETE", + "/admin/v1/users/" + uid2, + headers_json, + None, + 204, + None, + None, + ) + + engine.remove_authorization() # To finish + + +class TestProjectsDescriptors: + description = "test descriptors visibility among projects" + + @staticmethod + def run(engine, test_osm, manual_check, test_params=None): + vnfd_ids = [] + engine.set_test_name("ProjectDescriptors") + engine.get_autorization() + + project_admin_id = None + res = engine.test( + "Get my project Padmin", + "GET", + "/admin/v1/projects/{}".format(engine.project), + headers_json, + None, + 200, + r_header_json, + "json", + ) + if res: + response = res.json() + project_admin_id = response["_id"] + engine.test( + "Create project Padmin", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "Padmin", "admin": True}, + (201, 204), + {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, + "json", + ) + engine.test( + "Create project P2", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "P2"}, + (201, 204), + {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, + "json", + ) + engine.test( + "Create project P3", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "P3"}, + (201, 204), + {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, + "json", + ) + + engine.test( + "Create user U1", + "POST", + "/admin/v1/users", + headers_json, + { + "username": "U1", + "password": "pw1", + "project_role_mappings": [ + {"project": "Padmin", "role": "system_admin"}, + {"project": "P2", "role": "project_admin"}, + {"project": "P3", "role": "project_admin"}, + ], + }, + 201, + {"Location": "/admin/v1/users/", "Content-Type": "application/json"}, + "json", + ) + + engine.test( + "Onboard VNFD id1", + "POST", + "/vnfpkgm/v1/vnf_packages_content?id=id1", + headers_yaml, + TestDescriptors.vnfd_empty, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + vnfd_ids.append(engine.last_id) + engine.test( + "Onboard VNFD id2 PUBLIC", + "POST", + "/vnfpkgm/v1/vnf_packages_content?id=id2&PUBLIC=TRUE", + headers_yaml, + TestDescriptors.vnfd_empty, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + vnfd_ids.append(engine.last_id) + engine.test( + "Onboard VNFD id3", + "POST", + "/vnfpkgm/v1/vnf_packages_content?id=id3&PUBLIC=FALSE", + headers_yaml, + TestDescriptors.vnfd_empty, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + vnfd_ids.append(engine.last_id) + + res = engine.test( + "Get VNFD descriptors", + "GET", + "/vnfpkgm/v1/vnf_packages?id=id1,id2,id3", + headers_json, + None, + 200, + r_header_json, + "json", + ) + response = res.json() + if len(response) != 3: + logger.error( + "Only 3 vnfds should be present for project admin. {} listed".format( + len(response) + ) + ) + engine.failed_tests += 1 + + # Change to other project Padmin + res = engine.test( + "Change to user U1 project Padmin", + "POST", + "/admin/v1/tokens", + headers_json, + {"username": "U1", "password": "pw1", "project_id": "Padmin"}, + (200, 201), + r_header_json, + "json", + ) + if res: + response = res.json() + engine.set_header({"Authorization": "Bearer {}".format(response["id"])}) + + # list vnfds + res = engine.test( + "List VNFD descriptors for Padmin", + "GET", + "/vnfpkgm/v1/vnf_packages", + headers_json, + None, + 200, + r_header_json, + "json", + ) + response = res.json() + if len(response) != 0: + logger.error( + "Only 0 vnfds should be present for project Padmin. {} listed".format( + len(response) + ) + ) + engine.failed_tests += 1 + + # list Public vnfds + res = engine.test( + "List VNFD public descriptors", + "GET", + "/vnfpkgm/v1/vnf_packages?PUBLIC=True", + headers_json, + None, + 200, + r_header_json, + "json", + ) + response = res.json() + if len(response) != 1: + logger.error( + "Only 1 vnfds should be present for project Padmin. {} listed".format( + len(response) + ) + ) + engine.failed_tests += 1 + + # list vnfds belonging to project "admin" + res = engine.test( + "List VNFD of admin project", + "GET", + "/vnfpkgm/v1/vnf_packages?ADMIN={}".format(project_admin_id), + headers_json, + None, + 200, + r_header_json, + "json", + ) + if res: + response = res.json() + if len(response) != 3: + logger.error( + "Only 3 vnfds should be present for project Padmin. {} listed".format( + len(response) + ) + ) + engine.failed_tests += 1 + + # Get Public vnfds + engine.test( + "Get VNFD public descriptors", + "GET", + "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[1]), + headers_json, + None, + 200, + r_header_json, + "json", + ) + # Edit not owned vnfd + engine.test( + "Edit VNFD ", + "PATCH", + "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[0]), + headers_yaml, + "{name: pepe}", + 404, + r_header_yaml, + "yaml", + ) + + # Add to my catalog + engine.test( + "Add VNFD id2 to my catalog", + "PATCH", + "/vnfpkgm/v1/vnf_packages/{}?SET_PROJECT".format(vnfd_ids[1]), + headers_json, + None, + 204, + None, + 0, + ) + + # Add a new vnfd + engine.test( + "Onboard VNFD id4", + "POST", + "/vnfpkgm/v1/vnf_packages_content?id=id4", + headers_yaml, + TestDescriptors.vnfd_empty, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + vnfd_ids.append(engine.last_id) + + # list vnfds + res = engine.test( + "List VNFD public descriptors", + "GET", + "/vnfpkgm/v1/vnf_packages", + headers_json, + None, + 200, + r_header_json, + "json", + ) + response = res.json() + if len(response) != 2: + logger.error( + "Only 2 vnfds should be present for project Padmin. {} listed".format( + len(response) + ) + ) + engine.failed_tests += 1 + + if manual_check: + input( + "VNFDs have been omboarded. Perform manual check and press enter to resume" + ) + + test_rest.test( + "Delete VNFD id2", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[1]), + headers_yaml, + None, + 204, + None, + 0, + ) + + # change to admin project + engine.remove_authorization() # To force get authorization + engine.get_autorization() + test_rest.test( + "Delete VNFD id1", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[0]), + headers_yaml, + None, + 204, + None, + 0, + ) + test_rest.test( + "Delete VNFD id2", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[1]), + headers_yaml, + None, + 204, + None, + 0, + ) + test_rest.test( + "Delete VNFD id3", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[2]), + headers_yaml, + None, + 204, + None, + 0, + ) + test_rest.test( + "Delete VNFD id4", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[3]), + headers_yaml, + None, + 404, + r_header_yaml, + "yaml", + ) + test_rest.test( + "Delete VNFD id4", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[3]), + headers_yaml, + None, + 204, + None, + 0, + ) + # Get Public vnfds + engine.test( + "Get VNFD deleted id1", + "GET", + "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[0]), + headers_json, + None, + 404, + r_header_json, + "json", + ) + engine.test( + "Get VNFD deleted id2", + "GET", + "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[1]), + headers_json, + None, + 404, + r_header_json, + "json", + ) + engine.test( + "Get VNFD deleted id3", + "GET", + "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[2]), + headers_json, + None, + 404, + r_header_json, + "json", + ) + engine.test( + "Get VNFD deleted id4", + "GET", + "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[3]), + headers_json, + None, + 404, + r_header_json, + "json", + ) + + engine.test( + "Delete user U1", + "DELETE", + "/admin/v1/users/U1", + headers_json, + None, + 204, + None, + None, + ) + engine.test( + "Delete project Padmin", + "DELETE", + "/admin/v1/projects/Padmin", + headers_json, + None, + 204, + None, + None, + ) + engine.test( + "Delete project P2", + "DELETE", + "/admin/v1/projects/P2", + headers_json, + None, + 204, + None, + None, + ) + engine.test( + "Delete project P3", + "DELETE", + "/admin/v1/projects/P3", + headers_json, + None, + 204, + None, + None, + ) + + +class TestFakeVim: + description = "Creates/edit/delete fake VIMs and SDN controllers" + + def __init__(self): + self.vim = { + "schema_version": "1.0", + "schema_type": "No idea", + "name": "myVim", + "description": "Descriptor name", + "vim_type": "openstack", + "vim_url": "http://localhost:/vim", + "vim_tenant_name": "vimTenant", + "vim_user": "user", + "vim_password": "password", + "config": {"config_param": 1}, + } + self.sdn = { + "name": "sdn-name", + "description": "sdn-description", + "dpid": "50:50:52:54:00:94:21:21", + "ip": "192.168.15.17", + "port": 8080, + "type": "opendaylight", + "version": "3.5.6", + "user": "user", + "password": "passwd", + } + self.port_mapping = [ + { + "compute_node": "compute node 1", + "ports": [ + { + "pci": "0000:81:00.0", + "switch_port": "port-2/1", + "switch_mac": "52:54:00:94:21:21", + }, + { + "pci": "0000:81:00.1", + "switch_port": "port-2/2", + "switch_mac": "52:54:00:94:21:22", + }, + ], + }, + { + "compute_node": "compute node 2", + "ports": [ + { + "pci": "0000:81:00.0", + "switch_port": "port-2/3", + "switch_mac": "52:54:00:94:21:23", + }, + { + "pci": "0000:81:00.1", + "switch_port": "port-2/4", + "switch_mac": "52:54:00:94:21:24", + }, + ], + }, + ] + + def run(self, engine, test_osm, manual_check, test_params=None): + vim_bad = self.vim.copy() + vim_bad.pop("name") + + engine.set_test_name("FakeVim") + engine.get_autorization() + engine.test( + "Create VIM", + "POST", + "/admin/v1/vim_accounts", + headers_json, + self.vim, + (201, 202), + {"Location": "/admin/v1/vim_accounts/", "Content-Type": "application/json"}, + "json", + ) + vim_id = engine.last_id + engine.test( + "Create VIM without name, bad schema", + "POST", + "/admin/v1/vim_accounts", + headers_json, + vim_bad, + 422, + None, + headers_json, + ) + engine.test( + "Create VIM name repeated", + "POST", + "/admin/v1/vim_accounts", + headers_json, + self.vim, + 409, + None, + headers_json, + ) + engine.test( + "Show VIMs", + "GET", + "/admin/v1/vim_accounts", + headers_yaml, + None, + 200, + r_header_yaml, + "yaml", + ) + engine.test( + "Show VIM", + "GET", + "/admin/v1/vim_accounts/{}".format(vim_id), + headers_yaml, + None, + 200, + r_header_yaml, + "yaml", + ) + if not test_osm: + # delete with FORCE + engine.test( + "Delete VIM", + "DELETE", + "/admin/v1/vim_accounts/{}?FORCE=True".format(vim_id), + headers_yaml, + None, + 202, + None, + 0, + ) + engine.test( + "Check VIM is deleted", + "GET", + "/admin/v1/vim_accounts/{}".format(vim_id), + headers_yaml, + None, + 404, + r_header_yaml, + "yaml", + ) + else: + # delete and wait until is really deleted + engine.test( + "Delete VIM", + "DELETE", + "/admin/v1/vim_accounts/{}".format(vim_id), + headers_yaml, + None, + 202, + None, + 0, + ) + engine.wait_until_delete( + "/admin/v1/vim_accounts/{}".format(vim_id), timeout + ) + + +class TestVIMSDN(TestFakeVim): + description = "Creates VIM with SDN editing SDN controllers and port_mapping" + + def __init__(self): + TestFakeVim.__init__(self) + self.wim = { + "schema_version": "1.0", + "schema_type": "No idea", + "name": "myWim", + "description": "Descriptor name", + "wim_type": "odl", + "wim_url": "http://localhost:/wim", + "user": "user", + "password": "password", + "config": {"config_param": 1}, + } + + def run(self, engine, test_osm, manual_check, test_params=None): + engine.set_test_name("VimSdn") + engine.get_autorization() + # Added SDN + engine.test( + "Create SDN", + "POST", + "/admin/v1/sdns", + headers_json, + self.sdn, + (201, 202), + {"Location": "/admin/v1/sdns/", "Content-Type": "application/json"}, + "json", + ) + sdnc_id = engine.last_id + # sleep(5) + # Edit SDN + engine.test( + "Edit SDN", + "PATCH", + "/admin/v1/sdns/{}".format(sdnc_id), + headers_json, + {"name": "new_sdn_name"}, + (202, 204), + None, + None, + ) + # sleep(5) + # VIM with SDN + self.vim["config"]["sdn-controller"] = sdnc_id + self.vim["config"]["sdn-port-mapping"] = self.port_mapping + engine.test( + "Create VIM", + "POST", + "/admin/v1/vim_accounts", + headers_json, + self.vim, + (200, 202, 201), + {"Location": "/admin/v1/vim_accounts/", "Content-Type": "application/json"}, + "json", + ), + + vim_id = engine.last_id + self.port_mapping[0]["compute_node"] = "compute node XX" + engine.test( + "Edit VIM change port-mapping", + "PUT", + "/admin/v1/vim_accounts/{}".format(vim_id), + headers_json, + {"config": {"sdn-port-mapping": self.port_mapping}}, + (202, 204), + None, + None, + ) + engine.test( + "Edit VIM remove port-mapping", + "PUT", + "/admin/v1/vim_accounts/{}".format(vim_id), + headers_json, + {"config": {"sdn-port-mapping": None}}, + (202, 204), + None, + None, + ) + + engine.test( + "Create WIM", + "POST", + "/admin/v1/wim_accounts", + headers_json, + self.wim, + (200, 202, 201), + {"Location": "/admin/v1/wim_accounts/", "Content-Type": "application/json"}, + "json", + ), + wim_id = engine.last_id + + if not test_osm: + # delete with FORCE + engine.test( + "Delete VIM remove port-mapping", + "DELETE", + "/admin/v1/vim_accounts/{}?FORCE=True".format(vim_id), + headers_json, + None, + 202, + None, + 0, + ) + engine.test( + "Delete SDNC", + "DELETE", + "/admin/v1/sdns/{}?FORCE=True".format(sdnc_id), + headers_json, + None, + 202, + None, + 0, + ) + + engine.test( + "Delete WIM", + "DELETE", + "/admin/v1/wim_accounts/{}?FORCE=True".format(wim_id), + headers_json, + None, + 202, + None, + 0, + ) + engine.test( + "Check VIM is deleted", + "GET", + "/admin/v1/vim_accounts/{}".format(vim_id), + headers_yaml, + None, + 404, + r_header_yaml, + "yaml", + ) + engine.test( + "Check SDN is deleted", + "GET", + "/admin/v1/sdns/{}".format(sdnc_id), + headers_yaml, + None, + 404, + r_header_yaml, + "yaml", + ) + engine.test( + "Check WIM is deleted", + "GET", + "/admin/v1/wim_accounts/{}".format(wim_id), + headers_yaml, + None, + 404, + r_header_yaml, + "yaml", + ) + else: + if manual_check: + input( + "VIM, SDN, WIM has been deployed. Perform manual check and press enter to resume" + ) + # delete and wait until is really deleted + engine.test( + "Delete VIM remove port-mapping", + "DELETE", + "/admin/v1/vim_accounts/{}".format(vim_id), + headers_json, + None, + (202, 201, 204), + None, + 0, + ) + engine.test( + "Delete SDN", + "DELETE", + "/admin/v1/sdns/{}".format(sdnc_id), + headers_json, + None, + (202, 201, 204), + None, + 0, + ) + engine.test( + "Delete VIM", + "DELETE", + "/admin/v1/wim_accounts/{}".format(wim_id), + headers_json, + None, + (202, 201, 204), + None, + 0, + ) + engine.wait_until_delete( + "/admin/v1/vim_accounts/{}".format(vim_id), timeout + ) + engine.wait_until_delete("/admin/v1/sdns/{}".format(sdnc_id), timeout) + engine.wait_until_delete( + "/admin/v1/wim_accounts/{}".format(wim_id), timeout + ) + + +class TestDeploy: + description = "Base class for downloading descriptors from ETSI, onboard and deploy in real VIM" + + def __init__(self): + self.test_name = "DEPLOY" + self.nsd_id = None + self.vim_id = None + self.ns_id = None + self.vnfds_id = [] + self.descriptor_url = ( + "https://osm-download.etsi.org/ftp/osm-3.0-three/2nd-hackfest/packages/" + ) + self.vnfd_filenames = ("cirros_vnf.tar.gz",) + self.nsd_filename = "cirros_2vnf_ns.tar.gz" + self.descriptor_edit = None + self.uses_configuration = False + self.users = {} + self.passwords = {} + self.commands = {} + self.keys = {} + self.timeout = 120 + self.qforce = "" + self.ns_params = None + self.vnfr_ip_list = {} + + def create_descriptors(self, engine): + temp_dir = os.path.dirname(os.path.abspath(__file__)) + "/temp/" + if not os.path.exists(temp_dir): + os.makedirs(temp_dir) + for vnfd_index, vnfd_filename in enumerate(self.vnfd_filenames): + if "/" in vnfd_filename: + vnfd_filename_path = vnfd_filename + if not os.path.exists(vnfd_filename_path): + raise TestException( + "File '{}' does not exist".format(vnfd_filename_path) + ) + else: + vnfd_filename_path = temp_dir + vnfd_filename + if not os.path.exists(vnfd_filename_path): + with open(vnfd_filename_path, "wb") as file: + response = requests.get(self.descriptor_url + vnfd_filename) + if response.status_code >= 300: + raise TestException( + "Error downloading descriptor from '{}': {}".format( + self.descriptor_url + vnfd_filename, + response.status_code, + ) + ) + file.write(response.content) + if vnfd_filename_path.endswith(".yaml"): + headers = headers_yaml + else: + headers = headers_zip_yaml + if randint(0, 1) == 0: + # vnfd CREATE AND UPLOAD in one step: + engine.test( + "Onboard VNFD in one step", + "POST", + "/vnfpkgm/v1/vnf_packages_content" + self.qforce, + headers, + "@b" + vnfd_filename_path, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + self.vnfds_id.append(engine.last_id) + else: + # vnfd CREATE AND UPLOAD ZIP + engine.test( + "Onboard VNFD step 1", + "POST", + "/vnfpkgm/v1/vnf_packages", + headers_json, + None, + 201, + { + "Location": "/vnfpkgm/v1/vnf_packages/", + "Content-Type": "application/json", + }, + "json", + ) + self.vnfds_id.append(engine.last_id) + engine.test( + "Onboard VNFD step 2 as ZIP", + "PUT", + "/vnfpkgm/v1/vnf_packages/<>/package_content" + self.qforce, + headers, + "@b" + vnfd_filename_path, + 204, + None, + 0, + ) + + if self.descriptor_edit: + if "vnfd{}".format(vnfd_index) in self.descriptor_edit: + # Modify VNFD + engine.test( + "Edit VNFD ", + "PATCH", + "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfds_id[-1]), + headers_yaml, + self.descriptor_edit["vnfd{}".format(vnfd_index)], + 204, + None, + None, + ) + + if "/" in self.nsd_filename: + nsd_filename_path = self.nsd_filename + if not os.path.exists(nsd_filename_path): + raise TestException( + "File '{}' does not exist".format(nsd_filename_path) + ) + else: + nsd_filename_path = temp_dir + self.nsd_filename + if not os.path.exists(nsd_filename_path): + with open(nsd_filename_path, "wb") as file: + response = requests.get(self.descriptor_url + self.nsd_filename) + if response.status_code >= 300: + raise TestException( + "Error downloading descriptor from '{}': {}".format( + self.descriptor_url + self.nsd_filename, + response.status_code, + ) + ) + file.write(response.content) + if nsd_filename_path.endswith(".yaml"): + headers = headers_yaml + else: + headers = headers_zip_yaml + + if randint(0, 1) == 0: + # nsd CREATE AND UPLOAD in one step: + engine.test( + "Onboard NSD in one step", + "POST", + "/nsd/v1/ns_descriptors_content" + self.qforce, + headers, + "@b" + nsd_filename_path, + 201, + r_headers_yaml_location_nsd, + yaml, + ) + self.nsd_id = engine.last_id + else: + # nsd CREATE AND UPLOAD ZIP + engine.test( + "Onboard NSD step 1", + "POST", + "/nsd/v1/ns_descriptors", + headers_json, + None, + 201, + { + "Location": "/nsd/v1/ns_descriptors/", + "Content-Type": "application/json", + }, + "json", + ) + self.nsd_id = engine.last_id + engine.test( + "Onboard NSD step 2 as ZIP", + "PUT", + "/nsd/v1/ns_descriptors/<>/nsd_content" + self.qforce, + headers, + "@b" + nsd_filename_path, + 204, + None, + 0, + ) + + if self.descriptor_edit and "nsd" in self.descriptor_edit: + # Modify NSD + engine.test( + "Edit NSD ", + "PATCH", + "/nsd/v1/ns_descriptors/{}".format(self.nsd_id), + headers_yaml, + self.descriptor_edit["nsd"], + 204, + None, + None, + ) + + def delete_descriptors(self, engine): + # delete descriptors + engine.test( + "Delete NSSD SOL005", + "DELETE", + "/nsd/v1/ns_descriptors/{}".format(self.nsd_id), + headers_yaml, + None, + 204, + None, + 0, + ) + for vnfd_id in self.vnfds_id: + engine.test( + "Delete VNFD SOL005", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + def instantiate(self, engine, ns_data): + ns_data_text = yaml.safe_dump(ns_data, default_flow_style=True, width=256) + # create NS Two steps + r = engine.test( + "Create NS step 1", + "POST", + "/nslcm/v1/ns_instances", + headers_yaml, + ns_data_text, + (201, 202), + {"Location": "nslcm/v1/ns_instances/", "Content-Type": "application/yaml"}, + "yaml", + ) + if not r: + return + self.ns_id = engine.last_id + engine.test( + "Instantiate NS step 2", + "POST", + "/nslcm/v1/ns_instances/{}/instantiate".format(self.ns_id), + headers_yaml, + ns_data_text, + (201, 202), + r_headers_yaml_location_nslcmop, + "yaml", + ) + nslcmop_id = engine.last_id + + if test_osm: + # Wait until status is Ok + timeout = timeout_configure if self.uses_configuration else timeout_deploy + engine.wait_operation_ready("ns", nslcmop_id, timeout) + + def terminate(self, engine): + # remove deployment + if test_osm: + engine.test( + "Terminate NS", + "POST", + "/nslcm/v1/ns_instances/{}/terminate".format(self.ns_id), + headers_yaml, + None, + (201, 202), + r_headers_yaml_location_nslcmop, + "yaml", + ) + nslcmop2_id = engine.last_id + # Wait until status is Ok + engine.wait_operation_ready("ns", nslcmop2_id, timeout_deploy) + + engine.test( + "Delete NS", + "DELETE", + "/nslcm/v1/ns_instances/{}".format(self.ns_id), + headers_yaml, + None, + 204, + None, + 0, + ) + else: + engine.test( + "Delete NS with FORCE", + "DELETE", + "/nslcm/v1/ns_instances/{}?FORCE=True".format(self.ns_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + # check all it is deleted + engine.test( + "Check NS is deleted", + "GET", + "/nslcm/v1/ns_instances/{}".format(self.ns_id), + headers_yaml, + None, + 404, + None, + "yaml", + ) + r = engine.test( + "Check NSLCMOPs are deleted", + "GET", + "/nslcm/v1/ns_lcm_op_occs?nsInstanceId={}".format(self.ns_id), + headers_json, + None, + 200, + None, + "json", + ) + if not r: + return + nslcmops = r.json() + if not isinstance(nslcmops, list) or nslcmops: + raise TestException( + "NS {} deleted but with ns_lcm_op_occ active: {}".format( + self.ns_id, nslcmops + ) + ) + + def test_ns( + self, + engine, + test_osm, + commands=None, + users=None, + passwds=None, + keys=None, + timeout=0, + ): + r = engine.test( + "GET VNFR IDs", + "GET", + "/nslcm/v1/ns_instances/{}".format(self.ns_id), + headers_json, + None, + 200, + r_header_json, + "json", + ) + if not r: + return + ns_data = r.json() + + vnfr_list = ns_data["constituent-vnfr-ref"] + time = 0 + _commands = commands if commands is not None else self.commands + _users = users if users is not None else self.users + _passwds = passwds if passwds is not None else self.passwords + _keys = keys if keys is not None else self.keys + _timeout = timeout if timeout != 0 else self.timeout + + # vnfr_list=[d8272263-6bd3-4680-84ca-6a4be23b3f2d, 88b22e2f-994a-4b61-94fd-4a3c90de3dc4] + for vnfr_id in vnfr_list: + r = engine.test( + "Get VNFR to get IP_ADDRESS", + "GET", + "/nslcm/v1/vnfrs/{}".format(vnfr_id), + headers_json, + None, + 200, + r_header_json, + "json", + ) + if not r: + continue + vnfr_data = r.json() + + vnf_index = str(vnfr_data["member-vnf-index-ref"]) + + ip_address = self.get_vnfr_ip(engine, vnf_index) + description = "Exec command='{}' at VNFR={} IP={}".format( + _commands.get(vnf_index)[0], vnf_index, ip_address + ) + engine.step += 1 + test_description = "{}{} {}".format( + engine.test_name, engine.step, description + ) + logger.warning(test_description) + while _timeout >= time: + result, message = self.do_checks( + [ip_address], + vnf_index=vnfr_data["member-vnf-index-ref"], + commands=_commands.get(vnf_index), + user=_users.get(vnf_index), + passwd=_passwds.get(vnf_index), + key=_keys.get(vnf_index), + ) + if result == 1: + engine.passed_tests += 1 + logger.debug(message) + break + elif result == 0: + time += 20 + sleep(20) + elif result == -1: + engine.failed_tests += 1 + logger.error(message) + break + else: + time -= 20 + engine.failed_tests += 1 + logger.error(message) + else: + engine.failed_tests += 1 + logger.error( + "VNFR {} has not mgmt address. Check failed".format(vnf_index) + ) + + def do_checks(self, ip, vnf_index, commands=[], user=None, passwd=None, key=None): + try: + import urllib3 + from pssh.clients import ParallelSSHClient + from pssh.utils import load_private_key + from ssh2 import exceptions as ssh2Exception + except ImportError as e: + logger.critical( + "Package or/and is not installed. Please add them with 'pip3 install " + "parallel-ssh urllib3': {}".format(e) + ) + return -1, "install needed packages 'pip3 install parallel-ssh urllib3'" + urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning) + try: + p_host = os.environ.get("PROXY_HOST") + p_user = os.environ.get("PROXY_USER") + p_password = os.environ.get("PROXY_PASSWD") + + if key: + pkey = load_private_key(key) + else: + pkey = None + + client = ParallelSSHClient( + ip, + user=user, + password=passwd, + pkey=pkey, + proxy_host=p_host, + proxy_user=p_user, + proxy_password=p_password, + timeout=10, + num_retries=0, + ) + for cmd in commands: + output = client.run_command(cmd) + client.join(output) + if output[ip[0]].exit_code: + return -1, "VNFR {} command '{}' returns error: '{}'".format( + ip[0], cmd, "\n".join(output[ip[0]].stderr) + ) + else: + return 1, "VNFR {} command '{}' successful".format(ip[0], cmd) + except ( + ssh2Exception.ChannelFailure, + ssh2Exception.SocketDisconnectError, + ssh2Exception.SocketTimeout, + ssh2Exception.SocketRecvError, + ) as e: + return 0, "Timeout accessing the VNFR {}: {}".format(ip[0], str(e)) + except Exception as e: + return -1, "ERROR checking the VNFR {}: {}".format(ip[0], str(e)) + + def additional_operations(self, engine, test_osm, manual_check): + pass + + def run(self, engine, test_osm, manual_check, test_params=None): + engine.set_test_name(self.test_name) + engine.get_autorization() + nsname = os.environ.get("OSMNBITEST_NS_NAME", "OSMNBITEST") + if test_params: + if "vnfd-files" in test_params: + self.vnfd_filenames = test_params["vnfd-files"].split(",") + if "nsd-file" in test_params: + self.nsd_filename = test_params["nsd-file"] + if test_params.get("ns-name"): + nsname = test_params["ns-name"] + self.create_descriptors(engine) + + # create real VIM if not exist + self.vim_id = engine.get_create_vim(test_osm) + ns_data = { + "nsDescription": "default description", + "nsName": nsname, + "nsdId": self.nsd_id, + "vimAccountId": self.vim_id, + } + if self.ns_params: + ns_data.update(self.ns_params) + if test_params and test_params.get("ns-config"): + if isinstance(test_params["ns-config"], str): + ns_data.update(yaml.safe_load(test_params["ns-config"])) + else: + ns_data.update(test_params["ns-config"]) + self.instantiate(engine, ns_data) + + if manual_check: + input( + "NS has been deployed. Perform manual check and press enter to resume" + ) + if test_osm and self.commands: + self.test_ns(engine, test_osm) + self.additional_operations(engine, test_osm, manual_check) + self.terminate(engine) + self.delete_descriptors(engine) + + def get_first_ip(self, ip_string): + # When using a floating IP, the vnfr_data['ip-address'] contains a semicolon-separated list of IP:s. + first_ip = ip_string.split(";")[0] if ip_string else "" + return first_ip + + def get_vnfr_ip(self, engine, vnfr_index_wanted): + # If the IP address list has been obtained before, it has been stored in 'vnfr_ip_list' + ip = self.vnfr_ip_list.get(vnfr_index_wanted, "") + if ip: + return self.get_first_ip(ip) + r = engine.test( + "Get VNFR to get IP_ADDRESS", + "GET", + "/nslcm/v1/vnfrs?member-vnf-index-ref={}&nsr-id-ref={}".format( + vnfr_index_wanted, self.ns_id + ), + headers_json, + None, + 200, + r_header_json, + "json", + ) + if not r: + return "" + vnfr_data = r.json() + if not (vnfr_data and vnfr_data[0]): + return "" + # Store the IP (or list of IPs) in 'vnfr_ip_list' + ip_list = vnfr_data[0].get("ip-address", "") + if ip_list: + self.vnfr_ip_list[vnfr_index_wanted] = ip_list + ip = self.get_first_ip(ip_list) + return ip + + +class TestDeployHackfestCirros(TestDeploy): + description = "Load and deploy Hackfest cirros_2vnf_ns example" + + def __init__(self): + super().__init__() + self.test_name = "CIRROS" + self.vnfd_filenames = ("cirros_vnf.tar.gz",) + self.nsd_filename = "cirros_2vnf_ns.tar.gz" + self.commands = { + "1": [ + "ls -lrt", + ], + "2": [ + "ls -lrt", + ], + } + self.users = {"1": "cirros", "2": "cirros"} + self.passwords = {"1": "cubswin:)", "2": "cubswin:)"} + + def terminate(self, engine): + # Make a delete in one step, overriding the normal two step of TestDeploy that launched terminate and delete + if test_osm: + engine.test( + "Terminate and delete NS in one step", + "DELETE", + "/nslcm/v1/ns_instances_content/{}".format(self.ns_id), + headers_yaml, + None, + 202, + None, + "yaml", + ) + + engine.wait_until_delete( + "/nslcm/v1/ns_instances/{}".format(self.ns_id), timeout_deploy + ) + else: + engine.test( + "Delete NS with FORCE", + "DELETE", + "/nslcm/v1/ns_instances/{}?FORCE=True".format(self.ns_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + # check all it is deleted + engine.test( + "Check NS is deleted", + "GET", + "/nslcm/v1/ns_instances/{}".format(self.ns_id), + headers_yaml, + None, + 404, + None, + "yaml", + ) + r = engine.test( + "Check NSLCMOPs are deleted", + "GET", + "/nslcm/v1/ns_lcm_op_occs?nsInstanceId={}".format(self.ns_id), + headers_json, + None, + 200, + None, + "json", + ) + if not r: + return + nslcmops = r.json() + if not isinstance(nslcmops, list) or nslcmops: + raise TestException( + "NS {} deleted but with ns_lcm_op_occ active: {}".format( + self.ns_id, nslcmops + ) + ) + + +class TestDeployHackfest1(TestDeploy): + description = "Load and deploy Hackfest_1_vnfd example" + + def __init__(self): + super().__init__() + self.test_name = "HACKFEST1-" + self.vnfd_filenames = ("hackfest_1_vnfd.tar.gz",) + self.nsd_filename = "hackfest_1_nsd.tar.gz" + # self.commands = {'1': ['ls -lrt', ], '2': ['ls -lrt', ]} + # self.users = {'1': "cirros", '2': "cirros"} + # self.passwords = {'1': "cubswin:)", '2': "cubswin:)"} + + +class TestDeployHackfestCirrosScaling(TestDeploy): + description = ( + "Load and deploy Hackfest cirros_2vnf_ns example with scaling modifications" + ) + + def __init__(self): + super().__init__() + self.test_name = "CIRROS-SCALE" + self.vnfd_filenames = ("cirros_vnf.tar.gz",) + self.nsd_filename = "cirros_2vnf_ns.tar.gz" + # Modify VNFD to add scaling and count=2 + self.descriptor_edit = { + "vnfd0": { + "vdu": {"$id: 'cirros_vnfd-VM'": {"count": 2}}, + "scaling-group-descriptor": [ + { + "name": "scale_cirros", + "max-instance-count": 2, + "vdu": [{"vdu-id-ref": "cirros_vnfd-VM", "count": 2}], + } + ], + } + } + + def additional_operations(self, engine, test_osm, manual_check): + if not test_osm: + return + # 2 perform scale out twice + payload = ( + "{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_OUT, scaleByStepData: " + '{scaling-group-descriptor: scale_cirros, member-vnf-index: "1"}}}' + ) + for i in range(0, 2): + engine.test( + "Execute scale action over NS", + "POST", + "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), + headers_yaml, + payload, + (201, 202), + r_headers_yaml_location_nslcmop, + "yaml", + ) + nslcmop2_scale_out = engine.last_id + engine.wait_operation_ready("ns", nslcmop2_scale_out, timeout_deploy) + if manual_check: + input("NS scale out done. Check that two more vdus are there") + # TODO check automatic + + # 2 perform scale in + payload = ( + "{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_IN, scaleByStepData: " + '{scaling-group-descriptor: scale_cirros, member-vnf-index: "1"}}}' + ) + for i in range(0, 2): + engine.test( + "Execute scale IN action over NS", + "POST", + "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), + headers_yaml, + payload, + (201, 202), + r_headers_yaml_location_nslcmop, + "yaml", + ) + nslcmop2_scale_in = engine.last_id + engine.wait_operation_ready("ns", nslcmop2_scale_in, timeout_deploy) + if manual_check: + input("NS scale in done. Check that two less vdus are there") + # TODO check automatic + + # perform scale in that must fail as reached limit + engine.test( + "Execute scale IN out of limit action over NS", + "POST", + "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), + headers_yaml, + payload, + (201, 202), + r_headers_yaml_location_nslcmop, + "yaml", + ) + nslcmop2_scale_in = engine.last_id + engine.wait_operation_ready( + "ns", nslcmop2_scale_in, timeout_deploy, expected_fail=True + ) + + +class TestDeployIpMac(TestDeploy): + description = "Load and deploy descriptor examples setting mac, ip address at descriptor and instantiate params" + + def __init__(self): + super().__init__() + self.test_name = "SetIpMac" + self.vnfd_filenames = ( + "vnfd_2vdu_set_ip_mac2.yaml", + "vnfd_2vdu_set_ip_mac.yaml", + ) + self.nsd_filename = "scenario_2vdu_set_ip_mac.yaml" + self.descriptor_url = "https://osm.etsi.org/gitweb/?p=osm/RO.git;a=blob_plain;f=test/RO_tests/v3_2vdu_set_ip_mac/" + self.commands = { + "1": [ + "ls -lrt", + ], + "2": [ + "ls -lrt", + ], + } + self.users = {"1": "osm", "2": "osm"} + self.passwords = {"1": "osm4u", "2": "osm4u"} + self.timeout = 360 + + def run(self, engine, test_osm, manual_check, test_params=None): + # super().run(engine, test_osm, manual_check, test_params) + # run again setting IPs with instantiate parameters + instantiation_params = { + "vnf": [ + { + "member-vnf-index": "1", + "internal-vld": [ + { + "name": "internal_vld1", # net_internal + "ip-profile": { + "ip-version": "ipv4", + "subnet-address": "10.9.8.0/24", + "dhcp-params": { + "count": 100, + "start-address": "10.9.8.100", + }, + }, + "internal-connection-point": [ + { + "id-ref": "eth2", + "ip-address": "10.9.8.2", + }, + { + "id-ref": "eth3", + "ip-address": "10.9.8.3", + }, + ], + }, + ], + "vdu": [ + { + "id": "VM1", + "interface": [ + # { + # "name": "iface11", + # "floating-ip-required": True, + # }, + {"name": "iface13", "mac-address": "52:33:44:55:66:13"}, + ], + }, + { + "id": "VM2", + "interface": [ + { + "name": "iface21", + "ip-address": "10.31.31.22", + "mac-address": "52:33:44:55:66:21", + }, + ], + }, + ], + }, + ] + } + + super().run( + engine, + test_osm, + manual_check, + test_params={"ns-config": instantiation_params}, + ) + + +class TestDeployHackfest4(TestDeploy): + description = "Load and deploy Hackfest 4 example." + + def __init__(self): + super().__init__() + self.test_name = "HACKFEST4-" + self.vnfd_filenames = ("hackfest_4_vnfd.tar.gz",) + self.nsd_filename = "hackfest_4_nsd.tar.gz" + self.uses_configuration = True + self.commands = { + "1": [ + "ls -lrt", + ], + "2": [ + "ls -lrt", + ], + } + self.users = {"1": "ubuntu", "2": "ubuntu"} + self.passwords = {"1": "osm4u", "2": "osm4u"} + # Modify VNFD to add scaling + # self.descriptor_edit = { + # "vnfd0": { + # 'vnf-configuration': { + # 'config-primitive': [{ + # 'name': 'touch', + # 'parameter': [{ + # 'name': 'filename', + # 'data-type': 'STRING', + # 'default-value': '/home/ubuntu/touched' + # }] + # }] + # }, + # 'scaling-group-descriptor': [{ + # 'name': 'scale_dataVM', + # 'scaling-policy': [{ + # 'threshold-time': 0, + # 'name': 'auto_cpu_util_above_threshold', + # 'scaling-type': 'automatic', + # 'scaling-criteria': [{ + # 'name': 'cpu_util_above_threshold', + # 'vnf-monitoring-param-ref': 'all_aaa_cpu_util', + # 'scale-out-relational-operation': 'GE', + # 'scale-in-threshold': 15, + # 'scale-out-threshold': 60, + # 'scale-in-relational-operation': 'LE' + # }], + # 'cooldown-time': 60 + # }], + # 'max-instance-count': 10, + # 'scaling-config-action': [ + # {'vnf-config-primitive-name-ref': 'touch', + # 'trigger': 'post-scale-out'}, + # {'vnf-config-primitive-name-ref': 'touch', + # 'trigger': 'pre-scale-in'} + # ], + # 'vdu': [{ + # 'vdu-id-ref': 'dataVM', + # 'count': 1 + # }] + # }] + # } + # } + + +class TestDeployHackfest3Charmed(TestDeploy): + description = "Load and deploy Hackfest 3charmed_ns example" + + def __init__(self): + super().__init__() + self.test_name = "HACKFEST3-" + self.vnfd_filenames = ("hackfest_3charmed_vnfd.tar.gz",) + self.nsd_filename = "hackfest_3charmed_nsd.tar.gz" + self.uses_configuration = True + self.commands = { + "1": ["ls -lrt /home/ubuntu/first-touch"], + "2": ["ls -lrt /home/ubuntu/first-touch"], + } + self.users = {"1": "ubuntu", "2": "ubuntu"} + self.passwords = {"1": "osm4u", "2": "osm4u"} + self.descriptor_edit = { + "vnfd0": yaml.safe_load( + """ + vnf-configuration: + terminate-config-primitive: + - seq: '1' + name: touch + parameter: + - name: filename + value: '/home/ubuntu/last-touch1' + - seq: '3' + name: touch + parameter: + - name: filename + value: '/home/ubuntu/last-touch3' + - seq: '2' + name: touch + parameter: + - name: filename + value: '/home/ubuntu/last-touch2' + """ + ) + } + + def additional_operations(self, engine, test_osm, manual_check): + if not test_osm: + return + # 1 perform action + vnfr_index_selected = "2" + payload = '{member_vnf_index: "2", primitive: touch, primitive_params: { filename: /home/ubuntu/OSMTESTNBI }}' + engine.test( + "Exec service primitive over NS", + "POST", + "/nslcm/v1/ns_instances/{}/action".format(self.ns_id), + headers_yaml, + payload, + (201, 202), + r_headers_yaml_location_nslcmop, + "yaml", + ) + nslcmop2_action = engine.last_id + # Wait until status is Ok + engine.wait_operation_ready("ns", nslcmop2_action, timeout_deploy) + vnfr_ip = self.get_vnfr_ip(engine, vnfr_index_selected) + if manual_check: + input( + "NS service primitive has been executed." + "Check that file /home/ubuntu/OSMTESTNBI is present at {}".format( + vnfr_ip + ) + ) + if test_osm: + commands = { + "1": [""], + "2": [ + "ls -lrt /home/ubuntu/OSMTESTNBI", + ], + } + self.test_ns(engine, test_osm, commands=commands) + + # # 2 perform scale out + # payload = '{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_OUT, scaleByStepData: ' \ + # '{scaling-group-descriptor: scale_dataVM, member-vnf-index: "1"}}}' + # engine.test("Execute scale action over NS", "POST", + # "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), headers_yaml, payload, + # (201, 202), r_headers_yaml_location_nslcmop, "yaml") + # nslcmop2_scale_out = engine.last_id + # engine.wait_operation_ready("ns", nslcmop2_scale_out, timeout_deploy) + # if manual_check: + # input('NS scale out done. Check that file /home/ubuntu/touched is present and new VM is created') + # # TODO check automatic + # + # # 2 perform scale in + # payload = '{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_IN, scaleByStepData: ' \ + # '{scaling-group-descriptor: scale_dataVM, member-vnf-index: "1"}}}' + # engine.test("Execute scale action over NS", "POST", + # "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), headers_yaml, payload, + # (201, 202), r_headers_yaml_location_nslcmop, "yaml") + # nslcmop2_scale_in = engine.last_id + # engine.wait_operation_ready("ns", nslcmop2_scale_in, timeout_deploy) + # if manual_check: + # input('NS scale in done. Check that file /home/ubuntu/touched is updated and new VM is deleted') + # # TODO check automatic + + +class TestDeployHackfest3Charmed2(TestDeployHackfest3Charmed): + description = ( + "Load and deploy Hackfest 3charmed_ns example modified version of descriptors to have dots in " + "ids and member-vnf-index." + ) + + def __init__(self): + super().__init__() + self.test_name = "HACKFEST3v2-" + self.qforce = "?FORCE=True" + self.descriptor_edit = { + "vnfd0": { + "vdu": { + "$[0]": { + "interface": { + "$[0]": {"external-connection-point-ref": "pdu-mgmt"} + } + }, + "$[1]": None, + }, + "vnf-configuration": None, + "connection-point": { + "$[0]": { + "id": "pdu-mgmt", + "name": "pdu-mgmt", + "short-name": "pdu-mgmt", + }, + "$[1]": None, + }, + "mgmt-interface": {"cp": "pdu-mgmt"}, + "description": "A vnf single vdu to be used as PDU", + "id": "vdu-as-pdu", + "internal-vld": { + "$[0]": { + "id": "pdu_internal", + "name": "pdu_internal", + "internal-connection-point": {"$[1]": None}, + "short-name": "pdu_internal", + "type": "ELAN", + } + }, + }, + # Modify NSD accordingly + "nsd": { + "constituent-vnfd": { + "$[0]": {"vnfd-id-ref": "vdu-as-pdu"}, + "$[1]": None, + }, + "description": "A nsd to deploy the vnf to act as as PDU", + "id": "nsd-as-pdu", + "name": "nsd-as-pdu", + "short-name": "nsd-as-pdu", + "vld": { + "$[0]": { + "id": "mgmt_pdu", + "name": "mgmt_pdu", + "short-name": "mgmt_pdu", + "vnfd-connection-point-ref": { + "$[0]": { + "vnfd-connection-point-ref": "pdu-mgmt", + "vnfd-id-ref": "vdu-as-pdu", + }, + "$[1]": None, + }, + "type": "ELAN", + }, + "$[1]": None, + }, + }, + } + + +class TestDeployHackfest3Charmed3(TestDeployHackfest3Charmed): + description = "Load and deploy Hackfest 3charmed_ns example modified version to test scaling and NS parameters" + + def __init__(self): + super().__init__() + self.test_name = "HACKFEST3v3-" + self.commands = { + "1": ["ls -lrt /home/ubuntu/first-touch-1"], + "2": ["ls -lrt /home/ubuntu/first-touch-2"], + } + self.descriptor_edit = { + "vnfd0": yaml.safe_load( + """ + scaling-group-descriptor: + - name: "scale_dataVM" + max-instance-count: 10 + scaling-policy: + - name: "auto_cpu_util_above_threshold" + scaling-type: "automatic" + threshold-time: 0 + cooldown-time: 60 + scaling-criteria: + - name: "cpu_util_above_threshold" + scale-in-threshold: 15 + scale-in-relational-operation: "LE" + scale-out-threshold: 60 + scale-out-relational-operation: "GE" + vnf-monitoring-param-ref: "monitor1" + vdu: + - vdu-id-ref: dataVM + count: 1 + scaling-config-action: + - trigger: post-scale-out + vnf-config-primitive-name-ref: touch + - trigger: pre-scale-in + vnf-config-primitive-name-ref: touch + vdu: + "$id: dataVM": + monitoring-param: + - id: "dataVM_cpu_util" + nfvi-metric: "cpu_utilization" + + monitoring-param: + - id: "monitor1" + name: "monitor1" + aggregation-type: AVERAGE + vdu-monitoring-param: + vdu-ref: "dataVM" + vdu-monitoring-param-ref: "dataVM_cpu_util" + vnf-configuration: + initial-config-primitive: + "$[1]": + parameter: + "$[0]": + value: "" # default-value: /home/ubuntu/first-touch + config-primitive: + "$[0]": + parameter: + "$[0]": + default-value: "" + """, + ) + } + self.ns_params = { + "additionalParamsForVnf": [ + { + "member-vnf-index": "1", + "additionalParams": { + "touch_filename": "/home/ubuntu/first-touch-1", + "touch_filename2": "/home/ubuntu/second-touch-1", + }, + }, + { + "member-vnf-index": "2", + "additionalParams": { + "touch_filename": "/home/ubuntu/first-touch-2", + "touch_filename2": "/home/ubuntu/second-touch-2", + }, + }, + ] + } + + def additional_operations(self, engine, test_osm, manual_check): + super().additional_operations(engine, test_osm, manual_check) + if not test_osm: + return + + # 2 perform scale out + payload = ( + "{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_OUT, scaleByStepData: " + '{scaling-group-descriptor: scale_dataVM, member-vnf-index: "1"}}}' + ) + engine.test( + "Execute scale action over NS", + "POST", + "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), + headers_yaml, + payload, + (201, 202), + r_headers_yaml_location_nslcmop, + "yaml", + ) + nslcmop2_scale_out = engine.last_id + engine.wait_operation_ready("ns", nslcmop2_scale_out, timeout_deploy) + if manual_check: + input( + "NS scale out done. Check that file /home/ubuntu/second-touch-1 is present and new VM is created" + ) + if test_osm: + commands = { + "1": [ + "ls -lrt /home/ubuntu/second-touch-1", + ] + } + self.test_ns(engine, test_osm, commands=commands) + # TODO check automatic connection to scaled VM + + # 2 perform scale in + payload = ( + "{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_IN, scaleByStepData: " + '{scaling-group-descriptor: scale_dataVM, member-vnf-index: "1"}}}' + ) + engine.test( + "Execute scale action over NS", + "POST", + "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), + headers_yaml, + payload, + (201, 202), + r_headers_yaml_location_nslcmop, + "yaml", + ) + nslcmop2_scale_in = engine.last_id + engine.wait_operation_ready("ns", nslcmop2_scale_in, timeout_deploy) + if manual_check: + input( + "NS scale in done. Check that file /home/ubuntu/second-touch-1 is updated and new VM is deleted" + ) + # TODO check automatic + + +class TestDeploySimpleCharm(TestDeploy): + description = "Deploy hackfest-4 hackfest_simplecharm example" + + def __init__(self): + super().__init__() + self.test_name = "HACKFEST-SIMPLE" + self.descriptor_url = ( + "https://osm-download.etsi.org/ftp/osm-4.0-four/4th-hackfest/packages/" + ) + self.vnfd_filenames = ("hackfest_simplecharm_vnf.tar.gz",) + self.nsd_filename = "hackfest_simplecharm_ns.tar.gz" + self.uses_configuration = True + self.commands = { + "1": [""], + "2": [ + "ls -lrt /home/ubuntu/first-touch", + ], + } + self.users = {"1": "ubuntu", "2": "ubuntu"} + self.passwords = {"1": "osm4u", "2": "osm4u"} + + +class TestDeploySimpleCharm2(TestDeploySimpleCharm): + description = ( + "Deploy hackfest-4 hackfest_simplecharm example changing naming to contain dots on ids and " + "vnf-member-index" + ) + + def __init__(self): + super().__init__() + self.test_name = "HACKFEST-SIMPLE2-" + self.qforce = "?FORCE=True" + self.descriptor_edit = { + "vnfd0": {"id": "hackfest.simplecharm.vnf"}, + "nsd": { + "id": "hackfest.simplecharm.ns", + "constituent-vnfd": { + "$[0]": { + "vnfd-id-ref": "hackfest.simplecharm.vnf", + "member-vnf-index": "$1", + }, + "$[1]": { + "vnfd-id-ref": "hackfest.simplecharm.vnf", + "member-vnf-index": "$2", + }, + }, + "vld": { + "$[0]": { + "vnfd-connection-point-ref": { + "$[0]": { + "member-vnf-index-ref": "$1", + "vnfd-id-ref": "hackfest.simplecharm.vnf", + }, + "$[1]": { + "member-vnf-index-ref": "$2", + "vnfd-id-ref": "hackfest.simplecharm.vnf", + }, + }, + }, + "$[1]": { + "vnfd-connection-point-ref": { + "$[0]": { + "member-vnf-index-ref": "$1", + "vnfd-id-ref": "hackfest.simplecharm.vnf", + }, + "$[1]": { + "member-vnf-index-ref": "$2", + "vnfd-id-ref": "hackfest.simplecharm.vnf", + }, + }, + }, + }, + }, + } + + +class TestDeploySingleVdu(TestDeployHackfest3Charmed): + description = ( + "Generate a single VDU base on editing Hackfest3Charmed descriptors and deploy" + ) + + def __init__(self): + super().__init__() + self.test_name = "SingleVDU" + self.qforce = "?FORCE=True" + self.descriptor_edit = { + # Modify VNFD to remove one VDU + "vnfd0": { + "vdu": { + "$[0]": { + "interface": { + "$[0]": {"external-connection-point-ref": "pdu-mgmt"} + } + }, + "$[1]": None, + }, + "vnf-configuration": None, + "connection-point": { + "$[0]": { + "id": "pdu-mgmt", + "name": "pdu-mgmt", + "short-name": "pdu-mgmt", + }, + "$[1]": None, + }, + "mgmt-interface": {"cp": "pdu-mgmt"}, + "description": "A vnf single vdu to be used as PDU", + "id": "vdu-as-pdu", + "internal-vld": { + "$[0]": { + "id": "pdu_internal", + "name": "pdu_internal", + "internal-connection-point": {"$[1]": None}, + "short-name": "pdu_internal", + "type": "ELAN", + } + }, + }, + # Modify NSD accordingly + "nsd": { + "constituent-vnfd": { + "$[0]": {"vnfd-id-ref": "vdu-as-pdu"}, + "$[1]": None, + }, + "description": "A nsd to deploy the vnf to act as as PDU", + "id": "nsd-as-pdu", + "name": "nsd-as-pdu", + "short-name": "nsd-as-pdu", + "vld": { + "$[0]": { + "id": "mgmt_pdu", + "name": "mgmt_pdu", + "short-name": "mgmt_pdu", + "vnfd-connection-point-ref": { + "$[0]": { + "vnfd-connection-point-ref": "pdu-mgmt", + "vnfd-id-ref": "vdu-as-pdu", + }, + "$[1]": None, + }, + "type": "ELAN", + }, + "$[1]": None, + }, + }, + } + + +class TestDeployHnfd(TestDeployHackfest3Charmed): + description = ( + "Generate a HNFD base on editing Hackfest3Charmed descriptors and deploy" + ) + + def __init__(self): + super().__init__() + self.test_name = "HNFD" + self.pduDeploy = TestDeploySingleVdu() + self.pdu_interface_0 = {} + self.pdu_interface_1 = {} + + self.pdu_id = None + # self.vnf_to_pdu = """ + # vdu: + # "$[0]": + # pdu-type: PDU-TYPE-1 + # interface: + # "$[0]": + # name: mgmt-iface + # "$[1]": + # name: pdu-iface-internal + # id: hfn1 + # description: HFND, one PDU + One VDU + # name: hfn1 + # short-name: hfn1 + # + # """ + + self.pdu_descriptor = { + "name": "my-PDU", + "type": "PDU-TYPE-1", + "vim_accounts": "to-override", + "interfaces": [ + { + "name": "mgmt-iface", + "mgmt": True, + "type": "overlay", + "ip-address": "to override", + "mac-address": "mac_address", + "vim-network-name": "mgmt", + }, + { + "name": "pdu-iface-internal", + "mgmt": False, + "type": "overlay", + "ip-address": "to override", + "mac-address": "mac_address", + "vim-network-name": "pdu_internal", # OSMNBITEST-PDU-pdu_internal + }, + ], + } + self.vnfd_filenames = ( + "hackfest_3charmed_vnfd.tar.gz", + "hackfest_3charmed_vnfd.tar.gz", + ) + + self.descriptor_edit = { + "vnfd0": { + "id": "hfnd1", + "name": "hfn1", + "short-name": "hfn1", + "vdu": { + "$[0]": { + "pdu-type": "PDU-TYPE-1", + "interface": { + "$[0]": {"name": "mgmt-iface"}, + "$[1]": {"name": "pdu-iface-internal"}, + }, + } + }, + }, + "nsd": { + "constituent-vnfd": {"$[1]": {"vnfd-id-ref": "hfnd1"}}, + "vld": { + "$[0]": { + "vnfd-connection-point-ref": {"$[1]": {"vnfd-id-ref": "hfnd1"}} + }, + "$[1]": { + "vnfd-connection-point-ref": {"$[1]": {"vnfd-id-ref": "hfnd1"}} + }, + }, + }, + } + + def create_descriptors(self, engine): + super().create_descriptors(engine) + + # Create PDU + self.pdu_descriptor["interfaces"][0].update(self.pdu_interface_0) + self.pdu_descriptor["interfaces"][1].update(self.pdu_interface_1) + self.pdu_descriptor["vim_accounts"] = [self.vim_id] + # TODO get vim-network-name from vnfr.vld.name + self.pdu_descriptor["interfaces"][1]["vim-network-name"] = "{}-{}-{}".format( + os.environ.get("OSMNBITEST_NS_NAME", "OSMNBITEST"), + "PDU", + self.pdu_descriptor["interfaces"][1]["vim-network-name"], + ) + engine.test( + "Onboard PDU descriptor", + "POST", + "/pdu/v1/pdu_descriptors", + { + "Location": "/pdu/v1/pdu_descriptors/", + "Content-Type": "application/yaml", + }, + self.pdu_descriptor, + 201, + r_header_yaml, + "yaml", + ) + self.pdu_id = engine.last_id + + def run(self, engine, test_osm, manual_check, test_params=None): + engine.get_autorization() + engine.set_test_name(self.test_name) + nsname = os.environ.get("OSMNBITEST_NS_NAME", "OSMNBITEST") + + # create real VIM if not exist + self.vim_id = engine.get_create_vim(test_osm) + # instantiate PDU + self.pduDeploy.create_descriptors(engine) + self.pduDeploy.instantiate( + engine, + { + "nsDescription": "to be used as PDU", + "nsName": nsname + "-PDU", + "nsdId": self.pduDeploy.nsd_id, + "vimAccountId": self.vim_id, + }, + ) + if manual_check: + input( + "VNF to be used as PDU has been deployed. Perform manual check and press enter to resume" + ) + if test_osm: + self.pduDeploy.test_ns(engine, test_osm) + + if test_osm: + r = engine.test( + "Get VNFR to obtain IP_ADDRESS", + "GET", + "/nslcm/v1/vnfrs?nsr-id-ref={}".format(self.pduDeploy.ns_id), + headers_json, + None, + 200, + r_header_json, + "json", + ) + if not r: + return + vnfr_data = r.json() + # print(vnfr_data) + + self.pdu_interface_0["ip-address"] = vnfr_data[0]["vdur"][0]["interfaces"][ + 0 + ].get("ip-address") + self.pdu_interface_1["ip-address"] = vnfr_data[0]["vdur"][0]["interfaces"][ + 1 + ].get("ip-address") + self.pdu_interface_0["mac-address"] = vnfr_data[0]["vdur"][0]["interfaces"][ + 0 + ].get("mac-address") + self.pdu_interface_1["mac-address"] = vnfr_data[0]["vdur"][0]["interfaces"][ + 1 + ].get("mac-address") + if not self.pdu_interface_0["ip-address"]: + raise TestException("Vnfr has not managment ip address") + else: + self.pdu_interface_0["ip-address"] = "192.168.10.10" + self.pdu_interface_1["ip-address"] = "192.168.11.10" + self.pdu_interface_0["mac-address"] = "52:33:44:55:66:13" + self.pdu_interface_1["mac-address"] = "52:33:44:55:66:14" + + self.create_descriptors(engine) + + ns_data = { + "nsDescription": "default description", + "nsName": nsname, + "nsdId": self.nsd_id, + "vimAccountId": self.vim_id, + } + if test_params and test_params.get("ns-config"): + if isinstance(test_params["ns-config"], str): + ns_data.update(yaml.safe_load(test_params["ns-config"])) + else: + ns_data.update(test_params["ns-config"]) + + self.instantiate(engine, ns_data) + if manual_check: + input( + "NS has been deployed. Perform manual check and press enter to resume" + ) + if test_osm: + self.test_ns(engine, test_osm) + self.additional_operations(engine, test_osm, manual_check) + self.terminate(engine) + self.pduDeploy.terminate(engine) + self.delete_descriptors(engine) + self.pduDeploy.delete_descriptors(engine) + + def delete_descriptors(self, engine): + super().delete_descriptors(engine) + # delete pdu + engine.test( + "Delete PDU SOL005", + "DELETE", + "/pdu/v1/pdu_descriptors/{}".format(self.pdu_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + +class TestDescriptors: + description = "Test VNFD, NSD, PDU descriptors CRUD and dependencies" + vnfd_empty = """vnfd:vnfd-catalog: + vnfd: + - name: prova + short-name: prova + id: prova + """ + vnfd_prova = """vnfd:vnfd-catalog: + vnfd: + - connection-point: + - name: cp_0h8m + type: VPORT + id: prova + name: prova + short-name: prova + vdu: + - id: vdu_z4bm + image: ubuntu + interface: + - external-connection-point-ref: cp_0h8m + name: eth0 + virtual-interface: + type: VIRTIO + name: vdu_z4bm + version: '1.0' + """ + + def __init__(self): + self.vnfd_filename = "hackfest_3charmed_vnfd.tar.gz" + self.nsd_filename = "hackfest_3charmed_nsd.tar.gz" + self.descriptor_url = ( + "https://osm-download.etsi.org/ftp/osm-3.0-three/2nd-hackfest/packages/" + ) + self.vnfd_id = None + self.nsd_id = None + + def run(self, engine, test_osm, manual_check, test_params=None): + engine.set_test_name("Descriptors") + engine.get_autorization() + temp_dir = os.path.dirname(os.path.abspath(__file__)) + "/temp/" + if not os.path.exists(temp_dir): + os.makedirs(temp_dir) + + # download files + for filename in (self.vnfd_filename, self.nsd_filename): + filename_path = temp_dir + filename + if not os.path.exists(filename_path): + with open(filename_path, "wb") as file: + response = requests.get(self.descriptor_url + filename) + if response.status_code >= 300: + raise TestException( + "Error downloading descriptor from '{}': {}".format( + self.descriptor_url + filename, response.status_code + ) + ) + file.write(response.content) + + vnfd_filename_path = temp_dir + self.vnfd_filename + nsd_filename_path = temp_dir + self.nsd_filename + + engine.test( + "Onboard empty VNFD in one step", + "POST", + "/vnfpkgm/v1/vnf_packages_content", + headers_yaml, + self.vnfd_empty, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + self.vnfd_id = engine.last_id + + # test bug 605 + engine.test( + "Upload invalid VNFD ", + "PUT", + "/vnfpkgm/v1/vnf_packages/{}/package_content".format(self.vnfd_id), + headers_yaml, + self.vnfd_prova, + 422, + r_header_yaml, + "yaml", + ) + + engine.test( + "Upload VNFD {}".format(self.vnfd_filename), + "PUT", + "/vnfpkgm/v1/vnf_packages/{}/package_content".format(self.vnfd_id), + headers_zip_yaml, + "@b" + vnfd_filename_path, + 204, + None, + 0, + ) + + queries = [ + "mgmt-interface.cp=mgmt", + "vdu.0.interface.0.external-connection-point-ref=mgmt", + "vdu.0.interface.1.internal-connection-point-ref=internal", + "internal-vld.0.internal-connection-point.0.id-ref=internal", + # Detection of duplicated VLD names in VNF Descriptors + # URL: internal-vld=[ + # {id: internal1, name: internal, type:ELAN, + # internal-connection-point: [{id-ref: mgmtVM-internal}, {id-ref: dataVM-internal}]}, + # {id: internal2, name: internal, type:ELAN, + # internal-connection-point: [{id-ref: mgmtVM-internal}, {id-ref: dataVM-internal}]} + # ] + "internal-vld=%5B%7Bid%3A%20internal1%2C%20name%3A%20internal%2C%20type%3A%20ELAN%2C%20" + "internal-connection-point%3A%20%5B%7Bid-ref%3A%20mgmtVM-internal%7D%2C%20%7Bid-ref%3A%20" + "dataVM-internal%7D%5D%7D%2C%20%7Bid%3A%20internal2%2C%20name%3A%20internal%2C%20type%3A%20" + "ELAN%2C%20internal-connection-point%3A%20%5B%7Bid-ref%3A%20mgmtVM-internal%7D%2C%20%7B" + "id-ref%3A%20dataVM-internal%7D%5D%7D%5D", + ] + for query in queries: + engine.test( + "Upload invalid VNFD ", + "PUT", + "/vnfpkgm/v1/vnf_packages/{}/package_content?{}".format( + self.vnfd_id, query + ), + headers_zip_yaml, + "@b" + vnfd_filename_path, + 422, + r_header_yaml, + "yaml", + ) + + # test bug 605 + engine.test( + "Upload invalid VNFD ", + "PUT", + "/vnfpkgm/v1/vnf_packages/{}/package_content".format(self.vnfd_id), + headers_yaml, + self.vnfd_prova, + 422, + r_header_yaml, + "yaml", + ) + + # get vnfd descriptor + engine.test( + "Get VNFD descriptor", + "GET", + "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_id), + headers_yaml, + None, + 200, + r_header_yaml, + "yaml", + ) + + # get vnfd file descriptor + engine.test( + "Get VNFD file descriptor", + "GET", + "/vnfpkgm/v1/vnf_packages/{}/vnfd".format(self.vnfd_id), + headers_text, + None, + 200, + r_header_text, + "text", + temp_dir + "vnfd-yaml", + ) + # TODO compare files: diff vnfd-yaml hackfest_3charmed_vnfd/hackfest_3charmed_vnfd.yaml + + # get vnfd zip file package + engine.test( + "Get VNFD zip package", + "GET", + "/vnfpkgm/v1/vnf_packages/{}/package_content".format(self.vnfd_id), + headers_zip, + None, + 200, + r_header_zip, + "zip", + temp_dir + "vnfd-zip", + ) + # TODO compare files: diff vnfd-zip hackfest_3charmed_vnfd.tar.gz + + # get vnfd artifact + engine.test( + "Get VNFD artifact package", + "GET", + "/vnfpkgm/v1/vnf_packages/{}/artifacts/icons/osm.png".format(self.vnfd_id), + headers_zip, + None, + 200, + r_header_octect, + "octet-string", + temp_dir + "vnfd-icon", + ) + # TODO compare files: diff vnfd-icon hackfest_3charmed_vnfd/icons/osm.png + + # nsd CREATE AND UPLOAD in one step: + engine.test( + "Onboard NSD in one step", + "POST", + "/nsd/v1/ns_descriptors_content", + headers_zip_yaml, + "@b" + nsd_filename_path, + 201, + r_headers_yaml_location_nsd, + "yaml", + ) + self.nsd_id = engine.last_id + + queries = ["vld.0.vnfd-connection-point-ref.0.vnfd-id-ref=hf"] + for query in queries: + engine.test( + "Upload invalid NSD ", + "PUT", + "/nsd/v1/ns_descriptors/{}/nsd_content?{}".format(self.nsd_id, query), + headers_zip_yaml, + "@b" + nsd_filename_path, + 422, + r_header_yaml, + "yaml", + ) + + # get nsd descriptor + engine.test( + "Get NSD descriptor", + "GET", + "/nsd/v1/ns_descriptors/{}".format(self.nsd_id), + headers_yaml, + None, + 200, + r_header_yaml, + "yaml", + ) + + # get nsd file descriptor + engine.test( + "Get NSD file descriptor", + "GET", + "/nsd/v1/ns_descriptors/{}/nsd".format(self.nsd_id), + headers_text, + None, + 200, + r_header_text, + "text", + temp_dir + "nsd-yaml", + ) + # TODO compare files: diff nsd-yaml hackfest_3charmed_nsd/hackfest_3charmed_nsd.yaml + + # get nsd zip file package + engine.test( + "Get NSD zip package", + "GET", + "/nsd/v1/ns_descriptors/{}/nsd_content".format(self.nsd_id), + headers_zip, + None, + 200, + r_header_zip, + "zip", + temp_dir + "nsd-zip", + ) + # TODO compare files: diff nsd-zip hackfest_3charmed_nsd.tar.gz + + # get nsd artifact + engine.test( + "Get NSD artifact package", + "GET", + "/nsd/v1/ns_descriptors/{}/artifacts/icons/osm.png".format(self.nsd_id), + headers_zip, + None, + 200, + r_header_octect, + "octet-string", + temp_dir + "nsd-icon", + ) + # TODO compare files: diff nsd-icon hackfest_3charmed_nsd/icons/osm.png + + # vnfd DELETE + test_rest.test( + "Delete VNFD conflict", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_id), + headers_yaml, + None, + 409, + None, + None, + ) + + test_rest.test( + "Delete VNFD force", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}?FORCE=TRUE".format(self.vnfd_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + # nsd DELETE + test_rest.test( + "Delete NSD", + "DELETE", + "/nsd/v1/ns_descriptors/{}".format(self.nsd_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + +class TestNetSliceTemplates: + description = "Upload a NST to OSM" + + def __init__(self): + self.vnfd_filename = "@./slice_shared/vnfd/slice_shared_vnfd.yaml" + self.vnfd_filename_middle = "@./slice_shared/vnfd/slice_shared_middle_vnfd.yaml" + self.nsd_filename = "@./slice_shared/nsd/slice_shared_nsd.yaml" + self.nsd_filename_middle = "@./slice_shared/nsd/slice_shared_middle_nsd.yaml" + self.nst_filenames = "@./slice_shared/slice_shared_nstd.yaml" + + def run(self, engine, test_osm, manual_check, test_params=None): + # nst CREATE + engine.set_test_name("NST step ") + engine.get_autorization() + temp_dir = os.path.dirname(os.path.abspath(__file__)) + "/temp/" + if not os.path.exists(temp_dir): + os.makedirs(temp_dir) + + # Onboard VNFDs + engine.test( + "Onboard edge VNFD", + "POST", + "/vnfpkgm/v1/vnf_packages_content", + headers_yaml, + self.vnfd_filename, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + self.vnfd_edge_id = engine.last_id + + engine.test( + "Onboard middle VNFD", + "POST", + "/vnfpkgm/v1/vnf_packages_content", + headers_yaml, + self.vnfd_filename_middle, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + self.vnfd_middle_id = engine.last_id + + # Onboard NSDs + engine.test( + "Onboard NSD edge", + "POST", + "/nsd/v1/ns_descriptors_content", + headers_yaml, + self.nsd_filename, + 201, + r_headers_yaml_location_nsd, + "yaml", + ) + self.nsd_edge_id = engine.last_id + + engine.test( + "Onboard NSD middle", + "POST", + "/nsd/v1/ns_descriptors_content", + headers_yaml, + self.nsd_filename_middle, + 201, + r_headers_yaml_location_nsd, + "yaml", + ) + self.nsd_middle_id = engine.last_id + + # Onboard NST + engine.test( + "Onboard NST", + "POST", + "/nst/v1/netslice_templates_content", + headers_yaml, + self.nst_filenames, + 201, + r_headers_yaml_location_nst, + "yaml", + ) + nst_id = engine.last_id + + # nstd SHOW OSM format + engine.test( + "Show NSTD OSM format", + "GET", + "/nst/v1/netslice_templates/{}".format(nst_id), + headers_json, + None, + 200, + r_header_json, + "json", + ) + + # nstd DELETE + engine.test( + "Delete NSTD", + "DELETE", + "/nst/v1/netslice_templates/{}".format(nst_id), + headers_json, + None, + 204, + None, + 0, + ) + + # NSDs DELETE + test_rest.test( + "Delete NSD middle", + "DELETE", + "/nsd/v1/ns_descriptors/{}".format(self.nsd_middle_id), + headers_json, + None, + 204, + None, + 0, + ) + + test_rest.test( + "Delete NSD edge", + "DELETE", + "/nsd/v1/ns_descriptors/{}".format(self.nsd_edge_id), + headers_json, + None, + 204, + None, + 0, + ) + + # VNFDs DELETE + test_rest.test( + "Delete VNFD edge", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_edge_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + test_rest.test( + "Delete VNFD middle", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_middle_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + +class TestNetSliceInstances: + """ + Test procedure: + 1. Populate databases with VNFD, NSD, NST with the following scenario + +-----------------management-----------------+ + | | | + +--+---+ +----+----+ +---+--+ + | | | | | | + | edge +---data1----+ middle +---data2-----+ edge | + | | | | | | + +------+ +---------+ +------+ + shared-nss + 2. Create NSI-1 + 3. Instantiate NSI-1 + 4. Create NSI-2 + 5. Instantiate NSI-2 + Manual check - Are 2 slices instantiated correctly? + NSI-1 3 nss (2 nss-edges + 1 nss-middle) + NSI-2 2 nss (2 nss-edge sharing nss-middle) + 6. Terminate NSI-1 + 7. Delete NSI-1 + Manual check - Is slice NSI-1 deleted correctly? + NSI-2 with 2 nss-edge + 1 nss-middle (The one from NSI-1) + 8. Create NSI-3 + 9. Instantiate NSI-3 + Manual check - Is slice NSI-3 instantiated correctly? + NSI-3 reuse nss-middle. NSI-3 only create 2 nss-edge + 10. Delete NSI-2 + 11. Terminate NSI-2 + 12. Delete NSI-3 + 13. Terminate NSI-3 + Manual check - All cleaned correctly? + NSI-2 and NSI-3 were terminated and deleted + 14. Cleanup database + """ + + description = "Upload a NST to OSM" + + def __init__(self): + self.vim_id = None + self.vnfd_filename = "@./slice_shared/vnfd/slice_shared_vnfd.yaml" + self.vnfd_filename_middle = "@./slice_shared/vnfd/slice_shared_middle_vnfd.yaml" + self.nsd_filename = "@./slice_shared/nsd/slice_shared_nsd.yaml" + self.nsd_filename_middle = "@./slice_shared/nsd/slice_shared_middle_nsd.yaml" + self.nst_filenames = "@./slice_shared/slice_shared_nstd.yaml" + + def create_slice(self, engine, nsi_data, name): + ns_data_text = yaml.safe_dump(nsi_data, default_flow_style=True, width=256) + r = engine.test( + name, + "POST", + "/nsilcm/v1/netslice_instances", + headers_yaml, + ns_data_text, + (201, 202), + { + "Location": "nsilcm/v1/netslice_instances/", + "Content-Type": "application/yaml", + }, + "yaml", + ) + return r + + def instantiate_slice(self, engine, nsi_data, nsi_id, name): + ns_data_text = yaml.safe_dump(nsi_data, default_flow_style=True, width=256) + engine.test( + name, + "POST", + "/nsilcm/v1/netslice_instances/{}/instantiate".format(nsi_id), + headers_yaml, + ns_data_text, + (201, 202), + r_headers_yaml_location_nsilcmop, + "yaml", + ) + + def terminate_slice(self, engine, nsi_id, name): + engine.test( + name, + "POST", + "/nsilcm/v1/netslice_instances/{}/terminate".format(nsi_id), + headers_yaml, + None, + (201, 202), + r_headers_yaml_location_nsilcmop, + "yaml", + ) + + def delete_slice(self, engine, nsi_id, name): + engine.test( + name, + "DELETE", + "/nsilcm/v1/netslice_instances/{}".format(nsi_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + def run(self, engine, test_osm, manual_check, test_params=None): + # nst CREATE + engine.set_test_name("NSI") + engine.get_autorization() + + # Onboard VNFDs + engine.test( + "Onboard edge VNFD", + "POST", + "/vnfpkgm/v1/vnf_packages_content", + headers_yaml, + self.vnfd_filename, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + self.vnfd_edge_id = engine.last_id + + engine.test( + "Onboard middle VNFD", + "POST", + "/vnfpkgm/v1/vnf_packages_content", + headers_yaml, + self.vnfd_filename_middle, + 201, + r_headers_yaml_location_vnfd, + "yaml", + ) + self.vnfd_middle_id = engine.last_id + + # Onboard NSDs + engine.test( + "Onboard NSD edge", + "POST", + "/nsd/v1/ns_descriptors_content", + headers_yaml, + self.nsd_filename, + 201, + r_headers_yaml_location_nsd, + "yaml", + ) + self.nsd_edge_id = engine.last_id + + engine.test( + "Onboard NSD middle", + "POST", + "/nsd/v1/ns_descriptors_content", + headers_yaml, + self.nsd_filename_middle, + 201, + r_headers_yaml_location_nsd, + "yaml", + ) + self.nsd_middle_id = engine.last_id + + # Onboard NST + engine.test( + "Onboard NST", + "POST", + "/nst/v1/netslice_templates_content", + headers_yaml, + self.nst_filenames, + 201, + r_headers_yaml_location_nst, + "yaml", + ) + nst_id = engine.last_id + + self.vim_id = engine.get_create_vim(test_osm) + + # CREATE NSI-1 + ns_data = { + "nsiName": "Deploy-NSI-1", + "vimAccountId": self.vim_id, + "nstId": nst_id, + "nsiDescription": "default", + } + r = self.create_slice(engine, ns_data, "Create NSI-1 step 1") + if not r: + return + self.nsi_id1 = engine.last_id + + # INSTANTIATE NSI-1 + self.instantiate_slice( + engine, ns_data, self.nsi_id1, "Instantiate NSI-1 step 2" + ) + nsilcmop_id1 = engine.last_id + + # Waiting for NSI-1 + if test_osm: + engine.wait_operation_ready("nsi", nsilcmop_id1, timeout_deploy) + + # CREATE NSI-2 + ns_data = { + "nsiName": "Deploy-NSI-2", + "vimAccountId": self.vim_id, + "nstId": nst_id, + "nsiDescription": "default", + } + r = self.create_slice(engine, ns_data, "Create NSI-2 step 1") + if not r: + return + self.nsi_id2 = engine.last_id + + # INSTANTIATE NSI-2 + self.instantiate_slice( + engine, ns_data, self.nsi_id2, "Instantiate NSI-2 step 2" + ) + nsilcmop_id2 = engine.last_id + + # Waiting for NSI-2 + if test_osm: + engine.wait_operation_ready("nsi", nsilcmop_id2, timeout_deploy) + + if manual_check: + input( + "NSI-1 AND NSI-2 has been deployed. Perform manual check and press enter to resume" + ) + + # TERMINATE NSI-1 + if test_osm: + self.terminate_slice(engine, self.nsi_id1, "Terminate NSI-1") + nsilcmop1_id = engine.last_id + + # Wait terminate NSI-1 + engine.wait_operation_ready("nsi", nsilcmop1_id, timeout_deploy) + + # DELETE NSI-1 + self.delete_slice(engine, self.nsi_id1, "Delete NS") + + if manual_check: + input( + "NSI-1 has been deleted. Perform manual check and press enter to resume" + ) + + # CREATE NSI-3 + ns_data = { + "nsiName": "Deploy-NSI-3", + "vimAccountId": self.vim_id, + "nstId": nst_id, + "nsiDescription": "default", + } + r = self.create_slice(engine, ns_data, "Create NSI-3 step 1") + + if not r: + return + self.nsi_id3 = engine.last_id + + # INSTANTIATE NSI-3 + self.instantiate_slice( + engine, ns_data, self.nsi_id3, "Instantiate NSI-3 step 2" + ) + nsilcmop_id3 = engine.last_id + + # Wait Instantiate NSI-3 + if test_osm: + engine.wait_operation_ready("nsi", nsilcmop_id3, timeout_deploy) + + if manual_check: + input( + "NSI-3 has been deployed. Perform manual check and press enter to resume" + ) + + # TERMINATE NSI-2 + if test_osm: + self.terminate_slice(engine, self.nsi_id2, "Terminate NSI-2") + nsilcmop2_id = engine.last_id + + # Wait terminate NSI-2 + engine.wait_operation_ready("nsi", nsilcmop2_id, timeout_deploy) + + # DELETE NSI-2 + self.delete_slice(engine, self.nsi_id2, "DELETE NSI-2") + + # TERMINATE NSI-3 + if test_osm: + self.terminate_slice(engine, self.nsi_id3, "Terminate NSI-3") + nsilcmop3_id = engine.last_id + + # Wait terminate NSI-3 + engine.wait_operation_ready("nsi", nsilcmop3_id, timeout_deploy) + + # DELETE NSI-3 + self.delete_slice(engine, self.nsi_id3, "DELETE NSI-3") + + if manual_check: + input( + "NSI-2 and NSI-3 has been deleted. Perform manual check and press enter to resume" + ) + + # nstd DELETE + engine.test( + "Delete NSTD", + "DELETE", + "/nst/v1/netslice_templates/{}".format(nst_id), + headers_json, + None, + 204, + None, + 0, + ) + + # NSDs DELETE + test_rest.test( + "Delete NSD middle", + "DELETE", + "/nsd/v1/ns_descriptors/{}".format(self.nsd_middle_id), + headers_json, + None, + 204, + None, + 0, + ) + + test_rest.test( + "Delete NSD edge", + "DELETE", + "/nsd/v1/ns_descriptors/{}".format(self.nsd_edge_id), + headers_json, + None, + 204, + None, + 0, + ) + + # VNFDs DELETE + test_rest.test( + "Delete VNFD edge", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_edge_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + test_rest.test( + "Delete VNFD middle", + "DELETE", + "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_middle_id), + headers_yaml, + None, + 204, + None, + 0, + ) + + +class TestAuthentication: + description = "Test Authentication" + + @staticmethod + def run(engine, test_osm, manual_check, test_params=None): + engine.set_test_name("Authentication") + # backend = test_params.get("backend") if test_params else None # UNUSED + + admin_project_id = test_project_id = None + project_admin_role_id = project_user_role_id = None + test_user_id = empty_user_id = None + default_role_id = empty_role_id = token_role_id = None + + engine.get_autorization() + + # GET + engine.test( + "Get tokens", + "GET", + "/admin/v1/tokens", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + engine.test( + "Get projects", + "GET", + "/admin/v1/projects", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + engine.test( + "Get users", + "GET", + "/admin/v1/users", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + engine.test( + "Get roles", + "GET", + "/admin/v1/roles", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + res = engine.test( + "Get admin project", + "GET", + "/admin/v1/projects?name=admin", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + admin_project_id = res.json()[0]["_id"] if res else None + res = engine.test( + "Get project admin role", + "GET", + "/admin/v1/roles?name=project_admin", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + project_admin_role_id = res.json()[0]["_id"] if res else None + res = engine.test( + "Get project user role", + "GET", + "/admin/v1/roles?name=project_user", + headers_json, + {}, + (200), + {"Content-Type": "application/json"}, + "json", + ) + project_user_role_id = res.json()[0]["_id"] if res else None + + # POST + res = engine.test( + "Create test project", + "POST", + "/admin/v1/projects", + headers_json, + {"name": "test"}, + (201), + {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, + "json", + ) + test_project_id = engine.last_id if res else None + res = engine.test( + "Create role without permissions", + "POST", + "/admin/v1/roles", + headers_json, + {"name": "empty"}, + (201), + {"Content-Type": "application/json"}, + "json", + ) + empty_role_id = engine.last_id if res else None + res = engine.test( + "Create role with default permissions", + "POST", + "/admin/v1/roles", + headers_json, + {"name": "default", "permissions": {"default": True}}, + (201), + {"Location": "/admin/v1/roles/", "Content-Type": "application/json"}, + "json", + ) + default_role_id = engine.last_id if res else None + res = engine.test( + "Create role with token permissions", + "POST", + "/admin/v1/roles", + headers_json, + { + "name": "tokens", + "permissions": {"tokens": True}, + }, # is default required ? + (201), + {"Location": "/admin/v1/roles/", "Content-Type": "application/json"}, + "json", + ) + token_role_id = engine.last_id if res else None + pr = "project-role mappings" + res = engine.test( + "Create user without " + pr, + "POST", + "/admin/v1/users", + headers_json, + {"username": "empty", "password": "empty"}, + 201, + {"Content-Type": "application/json"}, + "json", + ) + empty_user_id = engine.last_id if res else None + if ( + admin_project_id + and test_project_id + and project_admin_role_id + and project_user_role_id + ): + data = {"username": "test", "password": "test"} + data["project_role_mappings"] = [ + {"project": test_project_id, "role": project_admin_role_id}, + {"project": admin_project_id, "role": project_user_role_id}, + ] + res = engine.test( + "Create user with " + pr, + "POST", + "/admin/v1/users", + headers_json, + data, + (201), + {"Content-Type": "application/json"}, + "json", + ) + test_user_id = engine.last_id if res else None + + # PUT + if test_user_id: + engine.test( + "Modify test user's password", + "PUT", + "/admin/v1/users/" + test_user_id, + headers_json, + {"password": "password"}, + (204), + {}, + 0, + ) + if ( + empty_user_id + and admin_project_id + and test_project_id + and project_admin_role_id + and project_user_role_id + ): + data = { + "project_role_mappings": [ + {"project": test_project_id, "role": project_admin_role_id}, + {"project": admin_project_id, "role": project_user_role_id}, + ] + } + engine.test( + "Modify empty user's " + pr, + "PUT", + "/admin/v1/users/" + empty_user_id, + headers_json, + data, + (204), + {}, + 0, + ) + + # DELETE + if empty_user_id: + engine.test( + "Delete empty user", + "DELETE", + "/admin/v1/users/" + empty_user_id, + headers_json, + {}, + (204), + {}, + 0, + ) + if test_user_id: + engine.test( + "Delete test user", + "DELETE", + "/admin/v1/users/" + test_user_id, + headers_json, + {}, + (204), + {}, + 0, + ) + if empty_role_id: + engine.test( + "Delete empty role", + "DELETE", + "/admin/v1/roles/" + empty_role_id, + headers_json, + {}, + (204), + {}, + 0, + ) + if default_role_id: + engine.test( + "Delete default role", + "DELETE", + "/admin/v1/roles/" + default_role_id, + headers_json, + {}, + (204), + {}, + 0, + ) + if token_role_id: + engine.test( + "Delete token role", + "DELETE", + "/admin/v1/roles/" + token_role_id, + headers_json, + {}, + (204), + {}, + 0, + ) + if test_project_id: + engine.test( + "Delete test project", + "DELETE", + "/admin/v1/projects/" + test_project_id, + headers_json, + {}, + (204), + {}, + 0, + ) + + # END Tests + + engine.remove_authorization() # To finish + + +class TestNbiQuotas: + description = "Test NBI Quotas" + + @staticmethod + def run(engine, test_osm, manual_check, test_params=None): + engine.set_test_name("NBI-Quotas_") + # backend = test_params.get("backend") if test_params else None # UNUSED + + test_username = "test-nbi-quotas" + test_password = "test-nbi-quotas" + test_project = "test-nbi-quotas" + + test_vim = "test-nbi-quotas" + test_wim = "test-nbi-quotas" + test_sdn = "test-nbi-quotas" + + test_user_id = None + test_project_id = None + + test_vim_ids = [] + test_wim_ids = [] + test_sdn_ids = [] + test_vnfd_ids = [] + test_nsd_ids = [] + test_nst_ids = [] + test_pdu_ids = [] + test_nsr_ids = [] + test_nsi_ids = [] + + # Save admin access data + admin_username = engine.user + admin_password = engine.password + admin_project = engine.project + + # Get admin access + engine.get_autorization() + admin_token = engine.last_id + + # Check that test project,user do not exist + res1 = engine.test( + "Check that test project doesn't exist", + "GET", + "/admin/v1/projects/" + test_project, + headers_json, + {}, + (404), + {}, + True, + ) + res2 = engine.test( + "Check that test user doesn't exist", + "GET", + "/admin/v1/users/" + test_username, + headers_json, + {}, + (404), + {}, + True, + ) + if None in [res1, res2]: + engine.remove_authorization() + logger.error("Test project and/or user already exist") + return + + # Create test project&user + res = engine.test( + "Create test project", + "POST", + "/admin/v1/projects", + headers_json, + { + "name": test_username, + "quotas": { + "vnfds": 2, + "nsds": 2, + "nsts": 1, + "pdus": 1, + "nsrs": 2, + "nsis": 1, + "vim_accounts": 1, + "wim_accounts": 1, + "sdns": 1, + }, + }, + (201), + r_header_json, + "json", + ) + test_project_id = engine.last_id if res else None + res = engine.test( + "Create test user", + "POST", + "/admin/v1/users", + headers_json, + { + "username": test_username, + "password": test_password, + "project_role_mappings": [ + {"project": test_project, "role": "project_admin"} + ], + }, + (201), + r_header_json, + "json", + ) + test_user_id = engine.last_id if res else None + + if test_project_id and test_user_id: + # Get user access + engine.token = None + engine.user = test_username + engine.password = test_password + engine.project = test_project + engine.get_autorization() + user_token = engine.last_id + + # Create test VIM + res = engine.test( + "Create test VIM", + "POST", + "/admin/v1/vim_accounts", + headers_json, + { + "name": test_vim, + "vim_type": "openvim", + "vim_user": test_username, + "vim_password": test_password, + "vim_tenant_name": test_project, + "vim_url": "https://0.0.0.0:0/v0.0", + }, + (202), + r_header_json, + "json", + ) + test_vim_ids += [engine.last_id if res else None] + + res = engine.test( + "Try to create second test VIM", + "POST", + "/admin/v1/vim_accounts", + headers_json, + { + "name": test_vim + "_2", + "vim_type": "openvim", + "vim_user": test_username, + "vim_password": test_password, + "vim_tenant_name": test_project, + "vim_url": "https://0.0.0.0:0/v0.0", + }, + (422), + r_header_json, + "json", + ) + test_vim_ids += [engine.last_id if res is None else None] + + res = engine.test( + "Try to create second test VIM with FORCE", + "POST", + "/admin/v1/vim_accounts?FORCE", + headers_json, + { + "name": test_vim + "_3", + "vim_type": "openvim", + "vim_user": test_username, + "vim_password": test_password, + "vim_tenant_name": test_project, + "vim_url": "https://0.0.0.0:0/v0.0", + }, + (202), + r_header_json, + "json", + ) + test_vim_ids += [engine.last_id if res else None] + + if test_vim_ids[0]: + # Download descriptor files (if required) + test_dir = "/tmp/" + test_username + "/" + test_url = "https://osm-download.etsi.org/ftp/osm-6.0-six/7th-hackfest/packages/" + vnfd_filenames = [ + "slice_hackfest_vnfd.tar.gz", + "slice_hackfest_middle_vnfd.tar.gz", + ] + nsd_filenames = [ + "slice_hackfest_nsd.tar.gz", + "slice_hackfest_middle_nsd.tar.gz", + ] + nst_filenames = ["slice_hackfest_nstd.yaml"] + pdu_filenames = ["PDU_router.yaml"] + desc_filenames = ( + vnfd_filenames + nsd_filenames + nst_filenames + pdu_filenames + ) + if not os.path.exists(test_dir): + os.makedirs(test_dir) + for filename in desc_filenames: + if not os.path.exists(test_dir + filename): + res = requests.get(test_url + filename) + if res.status_code < 300: + with open(test_dir + filename, "wb") as file: + file.write(res.content) + + if all([os.path.exists(test_dir + p) for p in desc_filenames]): + # Test VNFD Quotas + res = engine.test( + "Create test VNFD #1", + "POST", + "/vnfpkgm/v1/vnf_packages_content", + headers_zip_json, + "@b" + test_dir + vnfd_filenames[0], + (201), + r_header_json, + "json", + ) + test_vnfd_ids += [engine.last_id if res else None] + res = engine.test( + "Create test VNFD #2", + "POST", + "/vnfpkgm/v1/vnf_packages_content", + headers_zip_json, + "@b" + test_dir + vnfd_filenames[1], + (201), + r_header_json, + "json", + ) + test_vnfd_ids += [engine.last_id if res else None] + res = engine.test( + "Try to create extra test VNFD", + "POST", + "/vnfpkgm/v1/vnf_packages_content", + headers_zip_json, + "@b" + test_dir + vnfd_filenames[0], + (422), + r_header_json, + "json", + ) + test_vnfd_ids += [engine.last_id if res is None else None] + res = engine.test( + "Try to create extra test VNFD with FORCE", + "POST", + "/vnfpkgm/v1/vnf_packages_content?FORCE", + headers_zip_json, + "@b" + test_dir + vnfd_filenames[0], + (201), + r_header_json, + "json", + ) + test_vnfd_ids += [engine.last_id if res else None] + + # Remove extra VNFDs to prevent further errors + for i in [2, 3]: + if test_vnfd_ids[i]: + res = engine.test( + "Delete test VNFD #" + str(i), + "DELETE", + "/vnfpkgm/v1/vnf_packages_content/" + + test_vnfd_ids[i] + + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + if res: + test_vnfd_ids[i] = None + + if test_vnfd_ids[0] and test_vnfd_ids[1]: + # Test NSD Quotas + res = engine.test( + "Create test NSD #1", + "POST", + "/nsd/v1/ns_descriptors_content", + headers_zip_json, + "@b" + test_dir + nsd_filenames[0], + (201), + r_header_json, + "json", + ) + test_nsd_ids += [engine.last_id if res else None] + res = engine.test( + "Create test NSD #2", + "POST", + "/nsd/v1/ns_descriptors_content", + headers_zip_json, + "@b" + test_dir + nsd_filenames[1], + (201), + r_header_json, + "json", + ) + test_nsd_ids += [engine.last_id if res else None] + res = engine.test( + "Try to create extra test NSD", + "POST", + "/nsd/v1/ns_descriptors_content", + headers_zip_json, + "@b" + test_dir + nsd_filenames[0], + (422), + r_header_json, + "json", + ) + test_nsd_ids += [engine.last_id if res is None else None] + res = engine.test( + "Try to create extra test NSD with FORCE", + "POST", + "/nsd/v1/ns_descriptors_content?FORCE", + headers_zip_json, + "@b" + test_dir + nsd_filenames[0], + (201), + r_header_json, + "json", + ) + test_nsd_ids += [engine.last_id if res else None] + + # Remove extra NSDs to prevent further errors + for i in [2, 3]: + if test_nsd_ids[i]: + res = engine.test( + "Delete test NSD #" + str(i), + "DELETE", + "/nsd/v1/ns_descriptors_content/" + + test_nsd_ids[i] + + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + if res: + test_nsd_ids[i] = None + + if test_nsd_ids[0] and test_nsd_ids[1]: + # Test NSR Quotas + res = engine.test( + "Create test NSR #1", + "POST", + "/nslcm/v1/ns_instances_content", + headers_json, + { + "nsName": test_username + "_1", + "nsdId": test_nsd_ids[0], + "vimAccountId": test_vim_ids[0], + }, + (201), + r_header_json, + "json", + ) + test_nsr_ids += [engine.last_id if res else None] + res = engine.test( + "Create test NSR #2", + "POST", + "/nslcm/v1/ns_instances_content", + headers_json, + { + "nsName": test_username + "_2", + "nsdId": test_nsd_ids[1], + "vimAccountId": test_vim_ids[0], + }, + (201), + r_header_json, + "json", + ) + test_nsr_ids += [engine.last_id if res else None] + res = engine.test( + "Try to create extra test NSR", + "POST", + "/nslcm/v1/ns_instances_content", + headers_json, + { + "nsName": test_username + "_3", + "nsdId": test_nsd_ids[0], + "vimAccountId": test_vim_ids[0], + }, + (422), + r_header_json, + "json", + ) + test_nsr_ids += [engine.last_id if res is None else None] + res = engine.test( + "Try to create test NSR with FORCE", + "POST", + "/nslcm/v1/ns_instances_content?FORCE", + headers_json, + { + "nsName": test_username + "_4", + "nsdId": test_nsd_ids[0], + "vimAccountId": test_vim_ids[0], + }, + (201), + r_header_json, + "json", + ) + test_nsr_ids += [engine.last_id if res else None] + + # Test NST Quotas + res = engine.test( + "Create test NST", + "POST", + "/nst/v1/netslice_templates_content", + headers_txt_json, + "@b" + test_dir + nst_filenames[0], + (201), + r_header_json, + "json", + ) + test_nst_ids += [engine.last_id if res else None] + res = engine.test( + "Try to create extra test NST", + "POST", + "/nst/v1/netslice_templates_content", + headers_txt_json, + "@b" + test_dir + nst_filenames[0], + (422), + r_header_json, + "json", + ) + test_nst_ids += [engine.last_id if res is None else None] + res = engine.test( + "Try to create extra test NST with FORCE", + "POST", + "/nst/v1/netslice_templates_content?FORCE", + headers_txt_json, + "@b" + test_dir + nst_filenames[0], + (201), + r_header_json, + "json", + ) + test_nst_ids += [engine.last_id if res else None] + + if test_nst_ids[0]: + # Remove NSR Quota + engine.set_header( + {"Authorization": "Bearer {}".format(admin_token)} + ) + res = engine.test( + "Remove NSR Quota", + "PUT", + "/admin/v1/projects/" + test_project_id, + headers_json, + {"quotas": {"nsrs": None}}, + (204), + {}, + 0, + ) + engine.set_header( + {"Authorization": "Bearer {}".format(user_token)} + ) + if res: + # Test NSI Quotas + res = engine.test( + "Create test NSI", + "POST", + "/nsilcm/v1/netslice_instances_content", + headers_json, + { + "nsiName": test_username, + "nstId": test_nst_ids[0], + "vimAccountId": test_vim_ids[0], + }, + (201), + r_header_json, + "json", + ) + test_nsi_ids += [engine.last_id if res else None] + res = engine.test( + "Try to create extra test NSI", + "POST", + "/nsilcm/v1/netslice_instances_content", + headers_json, + { + "nsiName": test_username, + "nstId": test_nst_ids[0], + "vimAccountId": test_vim_ids[0], + }, + (400), + r_header_json, + "json", + ) + test_nsi_ids += [ + engine.last_id if res is None else None + ] + res = engine.test( + "Try to create extra test NSI with FORCE", + "POST", + "/nsilcm/v1/netslice_instances_content?FORCE", + headers_json, + { + "nsiName": test_username, + "nstId": test_nst_ids[0], + "vimAccountId": test_vim_ids[0], + }, + (201), + r_header_json, + "json", + ) + test_nsi_ids += [engine.last_id if res else None] + + # Test PDU Quotas + with open(test_dir + pdu_filenames[0], "rb") as file: + pdu_text = re.sub( + r"ip-address: *\[[^\]]*\]", + "ip-address: '0.0.0.0'", + file.read().decode("utf-8"), + ) + with open(test_dir + pdu_filenames[0], "wb") as file: + file.write(pdu_text.encode("utf-8")) + res = engine.test( + "Create test PDU", + "POST", + "/pdu/v1/pdu_descriptors", + headers_yaml, + "@b" + test_dir + pdu_filenames[0], + (201), + r_header_yaml, + "yaml", + ) + test_pdu_ids += [engine.last_id if res else None] + res = engine.test( + "Try to create extra test PDU", + "POST", + "/pdu/v1/pdu_descriptors", + headers_yaml, + "@b" + test_dir + pdu_filenames[0], + (422), + r_header_yaml, + "yaml", + ) + test_pdu_ids += [engine.last_id if res is None else None] + res = engine.test( + "Try to create extra test PDU with FORCE", + "POST", + "/pdu/v1/pdu_descriptors?FORCE", + headers_yaml, + "@b" + test_dir + pdu_filenames[0], + (201), + r_header_yaml, + "yaml", + ) + test_pdu_ids += [engine.last_id if res else None] + + # Cleanup + for i, id in enumerate(test_nsi_ids): + if id: + engine.test( + "Delete test NSI #" + str(i), + "DELETE", + "/nsilcm/v1/netslice_instances_content/" + + id + + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + for i, id in enumerate(test_nsr_ids): + if id: + engine.test( + "Delete test NSR #" + str(i), + "DELETE", + "/nslcm/v1/ns_instances_content/" + id + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + for i, id in enumerate(test_nst_ids): + if id: + engine.test( + "Delete test NST #" + str(i), + "DELETE", + "/nst/v1/netslice_templates_content/" + id + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + for i, id in enumerate(test_nsd_ids): + if id: + engine.test( + "Delete test NSD #" + str(i), + "DELETE", + "/nsd/v1/ns_descriptors_content/" + id + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + for i, id in enumerate(test_vnfd_ids): + if id: + engine.test( + "Delete test VNFD #" + str(i), + "DELETE", + "/vnfpkgm/v1/vnf_packages_content/" + id + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + for i, id in enumerate(test_pdu_ids): + if id: + engine.test( + "Delete test PDU #" + str(i), + "DELETE", + "/pdu/v1/pdu_descriptors/" + id + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + + # END Test NBI Quotas + + # Test WIM Quotas + res = engine.test( + "Create test WIM", + "POST", + "/admin/v1/wim_accounts", + headers_json, + { + "name": test_wim, + "wim_type": "onos", + "wim_url": "https://0.0.0.0:0/v0.0", + }, + (202), + r_header_json, + "json", + ) + test_wim_ids += [engine.last_id if res else None] + res = engine.test( + "Try to create second test WIM", + "POST", + "/admin/v1/wim_accounts", + headers_json, + { + "name": test_wim + "_2", + "wim_type": "onos", + "wim_url": "https://0.0.0.0:0/v0.0", + }, + (422), + r_header_json, + "json", + ) + test_wim_ids += [engine.last_id if res is None else None] + res = engine.test( + "Try to create second test WIM with FORCE", + "POST", + "/admin/v1/wim_accounts?FORCE", + headers_json, + { + "name": test_wim + "_3", + "wim_type": "onos", + "wim_url": "https://0.0.0.0:0/v0.0", + }, + (202), + r_header_json, + "json", + ) + test_wim_ids += [engine.last_id if res else None] + + # Test SDN Quotas + res = engine.test( + "Create test SDN", + "POST", + "/admin/v1/sdns", + headers_json, + { + "name": test_sdn, + "type": "onos", + "ip": "0.0.0.0", + "port": 9999, + "dpid": "00:00:00:00:00:00:00:00", + }, + (202), + r_header_json, + "json", + ) + test_sdn_ids += [engine.last_id if res else None] + res = engine.test( + "Try to create second test SDN", + "POST", + "/admin/v1/sdns", + headers_json, + { + "name": test_sdn + "_2", + "type": "onos", + "ip": "0.0.0.0", + "port": 9999, + "dpid": "00:00:00:00:00:00:00:00", + }, + (422), + r_header_json, + "json", + ) + test_sdn_ids += [engine.last_id if res is None else None] + res = engine.test( + "Try to create second test SDN with FORCE", + "POST", + "/admin/v1/sdns?FORCE", + headers_json, + { + "name": test_sdn + "_3", + "type": "onos", + "ip": "0.0.0.0", + "port": 9999, + "dpid": "00:00:00:00:00:00:00:00", + }, + (202), + r_header_json, + "json", + ) + test_sdn_ids += [engine.last_id if res else None] + + # Cleanup + for i, id in enumerate(test_vim_ids): + if id: + engine.test( + "Delete test VIM #" + str(i), + "DELETE", + "/admin/v1/vim_accounts/" + id + "?FORCE", + headers_json, + {}, + (202), + {}, + 0, + ) + for i, id in enumerate(test_wim_ids): + if id: + engine.test( + "Delete test WIM #" + str(i), + "DELETE", + "/admin/v1/wim_accounts/" + id + "?FORCE", + headers_json, + {}, + (202), + {}, + 0, + ) + for i, id in enumerate(test_sdn_ids): + if id: + engine.test( + "Delete test SDN #" + str(i), + "DELETE", + "/admin/v1/sdns/" + id + "?FORCE", + headers_json, + {}, + (202), + {}, + 0, + ) + + # Release user access + engine.remove_authorization() + + # Cleanup + engine.user = admin_username + engine.password = admin_password + engine.project = admin_project + engine.get_autorization() + if test_user_id: + engine.test( + "Delete test user", + "DELETE", + "/admin/v1/users/" + test_user_id + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + if test_project_id: + engine.test( + "Delete test project", + "DELETE", + "/admin/v1/projects/" + test_project_id + "?FORCE", + headers_json, + {}, + (204), + {}, + 0, + ) + engine.remove_authorization() + + # END class TestNbiQuotas + + +if __name__ == "__main__": + global logger + test = "" + + # Disable warnings from self-signed certificates. + requests.packages.urllib3.disable_warnings() + try: + logging.basicConfig(format="%(levelname)s %(message)s", level=logging.ERROR) + logger = logging.getLogger("NBI") + # load parameters and configuration + opts, args = getopt.getopt( + sys.argv[1:], + "hvu:p:", + [ + "url=", + "user=", + "password=", + "help", + "version", + "verbose", + "no-verbose", + "project=", + "insecure", + "timeout", + "timeout-deploy", + "timeout-configure", + "test=", + "list", + "test-osm", + "manual-check", + "params=", + "fail-fast", + ], + ) + url = "https://localhost:9999/osm" + user = password = project = "admin" + test_osm = False + manual_check = False + verbose = 0 + verify = True + fail_fast = False + test_classes = { + "NonAuthorized": TestNonAuthorized, + "FakeVIM": TestFakeVim, + "Users-Projects": TestUsersProjects, + "Projects-Descriptors": TestProjectsDescriptors, + "VIM-SDN": TestVIMSDN, + "Deploy-Custom": TestDeploy, + "Deploy-Hackfest-Cirros": TestDeployHackfestCirros, + "Deploy-Hackfest-Cirros-Scaling": TestDeployHackfestCirrosScaling, + "Deploy-Hackfest-3Charmed": TestDeployHackfest3Charmed, + "Deploy-Hackfest-3Charmed2": TestDeployHackfest3Charmed2, + "Deploy-Hackfest-3Charmed3": TestDeployHackfest3Charmed3, + "Deploy-Hackfest-4": TestDeployHackfest4, + "Deploy-CirrosMacIp": TestDeployIpMac, + "Descriptors": TestDescriptors, + "Deploy-Hackfest1": TestDeployHackfest1, + # "Deploy-MultiVIM": TestDeployMultiVIM, + "Deploy-SingleVdu": TestDeploySingleVdu, + "Deploy-Hnfd": TestDeployHnfd, + "Upload-Slice-Template": TestNetSliceTemplates, + "Deploy-Slice-Instance": TestNetSliceInstances, + "Deploy-SimpleCharm": TestDeploySimpleCharm, + "Deploy-SimpleCharm2": TestDeploySimpleCharm2, + "Authentication": TestAuthentication, + "NBI-Quotas": TestNbiQuotas, + } + test_to_do = [] + test_params = {} + + for o, a in opts: + # print("parameter:", o, a) + if o == "--version": + print("test version " + __version__ + " " + version_date) + exit() + elif o == "--list": + for test, test_class in sorted(test_classes.items()): + print("{:32} {}".format(test + ":", test_class.description)) + exit() + elif o in ("-v", "--verbose"): + verbose += 1 + elif o == "no-verbose": + verbose = -1 + elif o in ("-h", "--help"): + usage() + sys.exit() + elif o == "--test-osm": + test_osm = True + elif o == "--manual-check": + manual_check = True + elif o == "--url": + url = a + elif o in ("-u", "--user"): + user = a + elif o in ("-p", "--password"): + password = a + elif o == "--project": + project = a + elif o == "--fail-fast": + fail_fast = True + elif o == "--test": + for _test in a.split(","): + if _test not in test_classes: + print( + "Invalid test name '{}'. Use option '--list' to show available tests".format( + _test + ), + file=sys.stderr, + ) + exit(1) + test_to_do.append(_test) + elif o == "--params": + param_key, _, param_value = a.partition("=") + text_index = len(test_to_do) + if text_index not in test_params: + test_params[text_index] = {} + test_params[text_index][param_key] = param_value + elif o == "--insecure": + verify = False + elif o == "--timeout": + timeout = int(a) + elif o == "--timeout-deploy": + timeout_deploy = int(a) + elif o == "--timeout-configure": + timeout_configure = int(a) + else: + assert False, "Unhandled option" + if verbose == 0: + logger.setLevel(logging.WARNING) + elif verbose > 1: + logger.setLevel(logging.DEBUG) + else: + logger.setLevel(logging.ERROR) + + test_rest = TestRest(url, user=user, password=password, project=project) + # print("tests to do:", test_to_do) + if test_to_do: + text_index = 0 + for test in test_to_do: + if fail_fast and test_rest.failed_tests: + break + text_index += 1 + test_class = test_classes[test] + test_class().run( + test_rest, test_osm, manual_check, test_params.get(text_index) + ) + else: + for test, test_class in sorted(test_classes.items()): + if fail_fast and test_rest.failed_tests: + break + test_class().run(test_rest, test_osm, manual_check, test_params.get(0)) + test_rest.print_results() + exit(1 if test_rest.failed_tests else 0) + + except TestException as e: + logger.error(test + "Test {} Exception: {}".format(test, str(e))) + exit(1) + except getopt.GetoptError as e: + logger.error(e) + print(e, file=sys.stderr) + exit(1) + except Exception as e: + logger.critical(test + " Exception: " + str(e), exc_info=True) diff --git a/osm_nbi/notifications.py b/osm_nbi/notifications.py index a62670b..63d4ce8 100644 --- a/osm_nbi/notifications.py +++ b/osm_nbi/notifications.py @@ -108,12 +108,12 @@ class NotificationBase: return payload async def send_notifications( - self, subscribers: list, loop: asyncio.AbstractEventLoop = None + self, + subscribers: list, ): """ Generate tasks for all notification for an event. :param subscribers: A list of subscribers who want to be notified for event. - :param loop: Event loop object. """ notifications = [] for subscriber in subscribers: @@ -154,21 +154,19 @@ class NotificationBase: if notifications: tasks = [] - async with aiohttp.ClientSession(loop=loop) as session: + async with aiohttp.ClientSession() as session: for notification in notifications: tasks.append( asyncio.ensure_future( - self.send_notification(session, notification, loop=loop), - loop=loop, + self.send_notification(session, notification), ) ) - await asyncio.gather(*tasks, loop=loop) + await asyncio.gather(*tasks) async def send_notification( self, session: aiohttp.ClientSession, notification: dict, - loop: asyncio.AbstractEventLoop = None, retry_count: int = 5, timeout: float = 5.0, ): @@ -177,7 +175,6 @@ class NotificationBase: after maximum number of reties, then notification is dropped. :param session: An aiohttp client session object to maintain http session. :param notification: A dictionary containing all necessary data to make POST request. - :param loop: Event loop object. :param retry_count: An integer specifying the maximum number of reties for a notification. :param timeout: A float representing client timeout of each HTTP request. """ @@ -226,7 +223,7 @@ class NotificationBase: notification["payload"]["subscriptionId"], backoff_delay ) ) - await asyncio.sleep(backoff_delay, loop=loop) + await asyncio.sleep(backoff_delay) # Dropping notification self.logger.debug( "Notification {} sent failed to subscriber:{}.".format( diff --git a/osm_nbi/subscriptions.py b/osm_nbi/subscriptions.py index b178e5b..846e7d3 100644 --- a/osm_nbi/subscriptions.py +++ b/osm_nbi/subscriptions.py @@ -53,7 +53,6 @@ class SubscriptionThread(threading.Thread): self.db = None self.msg = None self.engine = engine - self.loop = None self.logger = logging.getLogger("nbi.subscriptions") self.aiomain_task_admin = ( None # asyncio task for receiving admin actions from kafka bus @@ -81,41 +80,38 @@ class SubscriptionThread(threading.Thread): # created. # Before subscribe, send dummy messages await self.msg.aiowrite( - "admin", "echo", "dummy message", loop=self.loop + "admin", + "echo", + "dummy message", ) - await self.msg.aiowrite("ns", "echo", "dummy message", loop=self.loop) - await self.msg.aiowrite("nsi", "echo", "dummy message", loop=self.loop) - await self.msg.aiowrite("vnf", "echo", "dummy message", loop=self.loop) + await self.msg.aiowrite("ns", "echo", "dummy message") + await self.msg.aiowrite("nsi", "echo", "dummy message") + await self.msg.aiowrite("vnf", "echo", "dummy message") if not kafka_working: self.logger.critical("kafka is working again") kafka_working = True if not self.aiomain_task_admin: - await asyncio.sleep(10, loop=self.loop) + await asyncio.sleep(10) self.logger.debug("Starting admin subscription task") self.aiomain_task_admin = asyncio.ensure_future( self.msg.aioread( ("admin",), - loop=self.loop, group_id=False, aiocallback=self._msg_callback, ), - loop=self.loop, ) if not self.aiomain_task: - await asyncio.sleep(10, loop=self.loop) + await asyncio.sleep(10) self.logger.debug("Starting non-admin subscription task") self.aiomain_task = asyncio.ensure_future( self.msg.aioread( ("ns", "nsi", "vnf"), - loop=self.loop, aiocallback=self._msg_callback, ), - loop=self.loop, ) done, _ = await asyncio.wait( [self.aiomain_task, self.aiomain_task_admin], timeout=None, - loop=self.loop, return_when=asyncio.FIRST_COMPLETED, ) try: @@ -142,14 +138,13 @@ class SubscriptionThread(threading.Thread): "Error accessing kafka '{}'. Retrying ...".format(e) ) kafka_working = False - await asyncio.sleep(10, loop=self.loop) + await asyncio.sleep(10) def run(self): """ Start of the thread :return: None """ - self.loop = asyncio.new_event_loop() try: if not self.db: if self.config["database"]["driver"] == "mongo": @@ -166,7 +161,6 @@ class SubscriptionThread(threading.Thread): ) if not self.msg: config_msg = self.config["message"].copy() - config_msg["loop"] = self.loop if config_msg["driver"] == "local": self.msg = msglocal.MsgLocal() self.msg.connect(config_msg) @@ -187,11 +181,7 @@ class SubscriptionThread(threading.Thread): self.logger.debug("Starting") while not self.to_terminate: try: - self.loop.run_until_complete( - asyncio.ensure_future(self.start_kafka(), loop=self.loop) - ) - # except asyncio.CancelledError: - # break # if cancelled it should end, breaking loop + asyncio.run(self.start_kafka()) except Exception as e: if not self.to_terminate: self.logger.exception( @@ -200,7 +190,6 @@ class SubscriptionThread(threading.Thread): self.logger.debug("Finishing") self._stop() - self.loop.close() async def _msg_callback(self, topic, command, params): """ @@ -265,10 +254,7 @@ class SubscriptionThread(threading.Thread): # self.logger.debug(subscribers) if subscribers: asyncio.ensure_future( - self.nslcm.send_notifications( - subscribers, loop=self.loop - ), - loop=self.loop, + self.nslcm.send_notifications(subscribers), ) else: self.logger.debug( @@ -296,8 +282,7 @@ class SubscriptionThread(threading.Thread): ) if subscribers: asyncio.ensure_future( - self.vnflcm.send_notifications(subscribers, loop=self.loop), - loop=self.loop, + self.vnflcm.send_notifications(subscribers), ) elif topic == "nsi": if command == "terminated" and params["operationState"] in ( @@ -343,7 +328,7 @@ class SubscriptionThread(threading.Thread): # writing to kafka must be done with our own loop. For this reason it is not allowed Engine to do that, # but content to be written is stored at msg_to_send for msg in msg_to_send: - await self.msg.aiowrite(*msg, loop=self.loop) + await self.msg.aiowrite(*msg) except (EngineException, DbException, MsgException) as e: self.logger.error( "Error while processing topic={} command={}: {}".format( @@ -379,6 +364,8 @@ class SubscriptionThread(threading.Thread): """ self.to_terminate = True if self.aiomain_task: - self.loop.call_soon_threadsafe(self.aiomain_task.cancel) + asyncio.get_event_loop().call_soon_threadsafe(self.aiomain_task.cancel) if self.aiomain_task_admin: - self.loop.call_soon_threadsafe(self.aiomain_task_admin.cancel) + asyncio.get_event_loop().call_soon_threadsafe( + self.aiomain_task_admin.cancel + ) diff --git a/osm_nbi/tests/run_test.py b/osm_nbi/tests/run_test.py deleted file mode 100755 index b7768ba..0000000 --- a/osm_nbi/tests/run_test.py +++ /dev/null @@ -1,5889 +0,0 @@ -#! /usr/bin/python3 -# -*- coding: utf-8 -*- - -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or -# implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -import getopt -import sys -import requests -import json -import logging -import yaml - -# import json -# import tarfile -from time import sleep -from random import randint -import os -from sys import stderr -from uuid import uuid4 -import re - -__author__ = "Alfonso Tierno, alfonso.tiernosepulveda@telefonica.com" -__date__ = "$2018-03-01$" -__version__ = "0.3" -version_date = "Oct 2018" - - -def usage(): - print("Usage: ", sys.argv[0], "[options]") - print( - " Performs system tests over running NBI. It can be used for real OSM test using option '--test-osm'" - ) - print( - " If this is the case env variables 'OSMNBITEST_VIM_NAME' must be supplied to create a VIM if not exist " - "where deployment is done" - ) - print("OPTIONS") - print(" -h|--help: shows this help") - print(" --insecure: Allows non trusted https NBI server") - print(" --list: list available tests") - print( - " --manual-check: Deployment tests stop after deployed to allow manual inspection. Only make sense with " - "'--test-osm'" - ) - print(" -p|--password PASSWORD: NBI access password. 'admin' by default") - print(" ---project PROJECT: NBI access project. 'admin' by default") - print( - " --test TEST[,...]: Execute only a test or a comma separated list of tests" - ) - print( - " --params key=val: params to the previous test. key can be vnfd-files, nsd-file, ns-name, ns-config" - ) - print( - " --test-osm: If missing this test is intended for NBI only, no other OSM components are expected. Use " - "this flag to test the system. LCM and RO components are expected to be up and running" - ) - print( - " --timeout TIMEOUT: General NBI timeout, by default {}s".format(timeout) - ) - print( - " --timeout-deploy TIMEOUT: Timeout used for getting NS deployed, by default {}s".format( - timeout_deploy - ) - ) - print( - " --timeout-configure TIMEOUT: Timeout used for getting NS deployed and configured," - " by default {}s".format(timeout_configure) - ) - print(" -u|--user USERNAME: NBI access username. 'admin' by default") - print( - " --url URL: complete NBI server URL. 'https//localhost:9999/osm' by default" - ) - print(" -v|--verbose print debug information, can be used several times") - print(" --no-verbose remove verbosity") - print(" --version: prints current version") - print("ENV variables used for real deployment tests with option osm-test.") - print(" export OSMNBITEST_VIM_NAME=vim-name") - print(" export OSMNBITEST_VIM_URL=vim-url") - print(" export OSMNBITEST_VIM_TYPE=vim-type") - print(" export OSMNBITEST_VIM_TENANT=vim-tenant") - print(" export OSMNBITEST_VIM_USER=vim-user") - print(" export OSMNBITEST_VIM_PASSWORD=vim-password") - print(' export OSMNBITEST_VIM_CONFIG="vim-config"') - print(' export OSMNBITEST_NS_NAME="vim-config"') - return - - -r_header_json = {"Content-type": "application/json"} -headers_json = {"Content-type": "application/json", "Accept": "application/json"} -r_header_yaml = {"Content-type": "application/yaml"} -headers_yaml = {"Content-type": "application/yaml", "Accept": "application/yaml"} -r_header_text = {"Content-type": "text/plain"} -r_header_octect = {"Content-type": "application/octet-stream"} -headers_text = {"Accept": "text/plain,application/yaml"} -r_header_zip = {"Content-type": "application/zip"} -headers_zip = {"Accept": "application/zip,application/yaml"} -headers_zip_yaml = {"Accept": "application/yaml", "Content-type": "application/zip"} -headers_zip_json = {"Accept": "application/json", "Content-type": "application/zip"} -headers_txt_json = {"Accept": "application/json", "Content-type": "text/plain"} -r_headers_yaml_location_vnfd = { - "Location": "/vnfpkgm/v1/vnf_packages_content/", - "Content-Type": "application/yaml", -} -r_headers_yaml_location_nsd = { - "Location": "/nsd/v1/ns_descriptors_content/", - "Content-Type": "application/yaml", -} -r_headers_yaml_location_nst = { - "Location": "/nst/v1/netslice_templates_content", - "Content-Type": "application/yaml", -} -r_headers_yaml_location_nslcmop = { - "Location": "nslcm/v1/ns_lcm_op_occs/", - "Content-Type": "application/yaml", -} -r_headers_yaml_location_nsilcmop = { - "Location": "/osm/nsilcm/v1/nsi_lcm_op_occs/", - "Content-Type": "application/yaml", -} - -# test ones authorized -test_authorized_list = ( - ( - "AU1", - "Invalid vnfd id", - "GET", - "/vnfpkgm/v1/vnf_packages/non-existing-id", - headers_json, - None, - 404, - r_header_json, - "json", - ), - ( - "AU2", - "Invalid nsd id", - "GET", - "/nsd/v1/ns_descriptors/non-existing-id", - headers_yaml, - None, - 404, - r_header_yaml, - "yaml", - ), - ( - "AU3", - "Invalid nsd id", - "DELETE", - "/nsd/v1/ns_descriptors_content/non-existing-id", - headers_yaml, - None, - 404, - r_header_yaml, - "yaml", - ), -) -timeout = 120 # general timeout -timeout_deploy = 60 * 10 # timeout for NS deploying without charms -timeout_configure = 60 * 20 # timeout for NS deploying and configuring - - -class TestException(Exception): - pass - - -class TestRest: - def __init__( - self, - url_base, - header_base=None, - verify=False, - user="admin", - password="admin", - project="admin", - ): - self.url_base = url_base - if header_base is None: - self.header_base = {} - else: - self.header_base = header_base.copy() - self.s = requests.session() - self.s.headers = self.header_base - self.verify = verify - self.token = False - self.user = user - self.password = password - self.project = project - self.vim_id = None - # contains ID of tests obtained from Location response header. "" key contains last obtained id - self.last_id = "" - self.test_name = None - self.step = 0 # number of subtest under test - self.passed_tests = 0 - self.failed_tests = 0 - - def set_test_name(self, test_name): - self.test_name = test_name - self.step = 0 - self.last_id = "" - - def set_header(self, header): - self.s.headers.update(header) - - def set_tet_name(self, test_name): - self.test_name = test_name - - def unset_header(self, key): - if key in self.s.headers: - del self.s.headers[key] - - def test( - self, - description, - method, - url, - headers, - payload, - expected_codes, - expected_headers, - expected_payload, - store_file=None, - pooling=False, - ): - """ - Performs an http request and check http code response. Exit if different than allowed. It get the returned id - that can be used by following test in the URL with {name} where name is the name of the test - :param description: description of the test - :param method: HTTP method: GET,PUT,POST,DELETE,... - :param url: complete URL or relative URL - :param headers: request headers to add to the base headers - :param payload: Can be a dict, transformed to json, a text or a file if starts with '@' - :param expected_codes: expected response codes, can be int, int tuple or int range - :param expected_headers: expected response headers, dict with key values - :param expected_payload: expected payload, 0 if empty, 'yaml', 'json', 'text', 'zip', 'octet-stream' - :param store_file: filename to store content - :param pooling: if True do not count neither log this test. Because a pooling is done with many equal requests - :return: requests response - """ - r = None - try: - if not self.s: - self.s = requests.session() - # URL - if not url: - url = self.url_base - elif not url.startswith("http"): - url = self.url_base + url - - # replace url <> with the last ID - url = url.replace("<>", self.last_id) - if payload: - if isinstance(payload, str): - if payload.startswith("@"): - mode = "r" - file_name = payload[1:] - if payload.startswith("@b"): - mode = "rb" - file_name = payload[2:] - with open(file_name, mode) as f: - payload = f.read() - elif isinstance(payload, dict): - payload = json.dumps(payload) - - if not pooling: - test_description = "Test {}{} {} {} {}".format( - self.test_name, self.step, description, method, url - ) - logger.warning(test_description) - self.step += 1 - stream = False - if expected_payload in ("zip", "octet-string") or store_file: - stream = True - __retry = 0 - while True: - try: - r = getattr(self.s, method.lower())( - url, - data=payload, - headers=headers, - verify=self.verify, - stream=stream, - ) - break - except requests.exceptions.ConnectionError as e: - if __retry == 2: - raise - logger.error("Exception {}. Retrying".format(e)) - __retry += 1 - - if expected_payload in ("zip", "octet-string") or store_file: - logger.debug("RX {}".format(r.status_code)) - else: - logger.debug("RX {}: {}".format(r.status_code, r.text)) - - # check response - if expected_codes: - if isinstance(expected_codes, int): - expected_codes = (expected_codes,) - if r.status_code not in expected_codes: - raise TestException( - "Got status {}. Expected {}. {}".format( - r.status_code, expected_codes, r.text - ) - ) - - if expected_headers: - for header_key, header_val in expected_headers.items(): - if header_key.lower() not in r.headers: - raise TestException("Header {} not present".format(header_key)) - if header_val and header_val.lower() not in r.headers[header_key]: - raise TestException( - "Header {} does not contain {} but {}".format( - header_key, header_val, r.headers[header_key] - ) - ) - - if expected_payload is not None: - if expected_payload == 0 and len(r.content) > 0: - raise TestException("Expected empty payload") - elif expected_payload == "json": - try: - r.json() - except Exception as e: - raise TestException( - "Expected json response payload, but got Exception {}".format( - e - ) - ) - elif expected_payload == "yaml": - try: - yaml.safe_load(r.text) - except Exception as e: - raise TestException( - "Expected yaml response payload, but got Exception {}".format( - e - ) - ) - elif expected_payload in ("zip", "octet-string"): - if len(r.content) == 0: - raise TestException( - "Expected some response payload, but got empty" - ) - # try: - # tar = tarfile.open(None, 'r:gz', fileobj=r.raw) - # for tarinfo in tar: - # tarname = tarinfo.name - # print(tarname) - # except Exception as e: - # raise TestException("Expected zip response payload, but got Exception {}".format(e)) - elif expected_payload == "text": - if len(r.content) == 0: - raise TestException( - "Expected some response payload, but got empty" - ) - # r.text - if store_file: - with open(store_file, "wb") as fd: - for chunk in r.iter_content(chunk_size=128): - fd.write(chunk) - - location = r.headers.get("Location") - if location: - _id = location[location.rfind("/") + 1 :] - if _id: - self.last_id = str(_id) - if not pooling: - self.passed_tests += 1 - return r - except TestException as e: - self.failed_tests += 1 - r_status_code = None - r_text = None - if r: - r_status_code = r.status_code - r_text = r.text - logger.error("{} \nRX code{}: {}".format(e, r_status_code, r_text)) - return None - # exit(1) - except IOError as e: - if store_file: - logger.error("Cannot open file {}: {}".format(store_file, e)) - else: - logger.error("Exception: {}".format(e), exc_info=True) - self.failed_tests += 1 - return None - # exit(1) - except requests.exceptions.RequestException as e: - logger.error("Exception: {}".format(e)) - - def get_autorization(self): # user=None, password=None, project=None): - if ( - self.token - ): # and self.user == user and self.password == password and self.project == project: - return - # self.user = user - # self.password = password - # self.project = project - r = self.test( - "Obtain token", - "POST", - "/admin/v1/tokens", - headers_json, - { - "username": self.user, - "password": self.password, - "project_id": self.project, - }, - (200, 201), - r_header_json, - "json", - ) - if not r: - return - response = r.json() - self.token = response["id"] - self.set_header({"Authorization": "Bearer {}".format(self.token)}) - - def remove_authorization(self): - if self.token: - self.test( - "Delete token", - "DELETE", - "/admin/v1/tokens/{}".format(self.token), - headers_json, - None, - (200, 201, 204), - None, - None, - ) - self.token = None - self.unset_header("Authorization") - - def get_create_vim(self, test_osm): - if self.vim_id: - return self.vim_id - self.get_autorization() - if test_osm: - vim_name = os.environ.get("OSMNBITEST_VIM_NAME") - if not vim_name: - raise TestException( - "Needed to define OSMNBITEST_VIM_XXX variables to create a real VIM for deployment" - ) - else: - vim_name = "fakeVim" - # Get VIM - r = self.test( - "Get VIM ID", - "GET", - "/admin/v1/vim_accounts?name={}".format(vim_name), - headers_json, - None, - 200, - r_header_json, - "json", - ) - if not r: - return - vims = r.json() - if vims: - return vims[0]["_id"] - # Add VIM - if test_osm: - # check needed environ parameters: - if not os.environ.get("OSMNBITEST_VIM_URL") or not os.environ.get( - "OSMNBITEST_VIM_TENANT" - ): - raise TestException( - "Env OSMNBITEST_VIM_URL and OSMNBITEST_VIM_TENANT are needed for create a real VIM" - " to deploy on whit the --test-osm option" - ) - vim_data = ( - "{{schema_version: '1.0', name: '{}', vim_type: {}, vim_url: '{}'," - "vim_tenant_name: '{}', " - "vim_user: {}, vim_password: {}" - ).format( - vim_name, - os.environ.get("OSMNBITEST_VIM_TYPE", "openstack"), - os.environ.get("OSMNBITEST_VIM_URL"), - os.environ.get("OSMNBITEST_VIM_TENANT"), - os.environ.get("OSMNBITEST_VIM_USER"), - os.environ.get("OSMNBITEST_VIM_PASSWORD"), - ) - if os.environ.get("OSMNBITEST_VIM_CONFIG"): - vim_data += " ,config: {}".format( - os.environ.get("OSMNBITEST_VIM_CONFIG") - ) - vim_data += "}" - else: - vim_data = ( - "{schema_version: '1.0', name: fakeVim, vim_type: openstack, vim_url: 'http://10.11.12.13/fake'" - ", vim_tenant_name: 'vimtenant', vim_user: vimuser, vim_password: vimpassword}" - ) - self.test( - "Create VIM", - "POST", - "/admin/v1/vim_accounts", - headers_yaml, - vim_data, - (201, 202), - {"Location": "/admin/v1/vim_accounts/", "Content-Type": "application/yaml"}, - "yaml", - ) - return self.last_id - - def print_results(self): - print("\n\n\n--------------------------------------------") - print( - "TEST RESULTS: Total: {}, Passed: {}, Failed: {}".format( - self.passed_tests + self.failed_tests, - self.passed_tests, - self.failed_tests, - ) - ) - print("--------------------------------------------") - - def wait_until_delete(self, url_op, timeout_delete): - """ - Make a pooling until topic is not present, because of deleted - :param url_op: - :param timeout_delete: - :return: - """ - description = "Wait to topic being deleted" - test_description = "Test {}{} {} {} {}".format( - self.test_name, self.step, description, "GET", url_op - ) - logger.warning(test_description) - self.step += 1 - - wait = timeout_delete - while wait >= 0: - r = self.test( - description, - "GET", - url_op, - headers_yaml, - None, - (200, 404), - None, - r_header_yaml, - "yaml", - pooling=True, - ) - if not r: - return - if r.status_code == 404: - self.passed_tests += 1 - break - elif r.status_code == 200: - wait -= 5 - sleep(5) - else: - raise TestException( - "Topic is not deleted after {} seconds".format(timeout_delete) - ) - self.failed_tests += 1 - - def wait_operation_ready(self, ns_nsi, opp_id, timeout, expected_fail=False): - """ - Wait until nslcmop or nsilcmop finished - :param ns_nsi: "ns" o "nsi" - :param opp_id: Id o fthe operation - :param timeout: - :param expected_fail: - :return: None. Updates passed/failed_tests - """ - if ns_nsi == "ns": - url_op = "/nslcm/v1/ns_lcm_op_occs/{}".format(opp_id) - else: - url_op = "/nsilcm/v1/nsi_lcm_op_occs/{}".format(opp_id) - description = "Wait to {} lcm operation complete".format(ns_nsi) - test_description = "Test {}{} {} {} {}".format( - self.test_name, self.step, description, "GET", url_op - ) - logger.warning(test_description) - self.step += 1 - wait = timeout - while wait >= 0: - r = self.test( - description, - "GET", - url_op, - headers_json, - None, - 200, - r_header_json, - "json", - pooling=True, - ) - if not r: - return - nslcmop = r.json() - if "COMPLETED" in nslcmop["operationState"]: - if expected_fail: - logger.error( - "NS terminate has success, expecting failing: {}".format( - nslcmop["detailed-status"] - ) - ) - self.failed_tests += 1 - else: - self.passed_tests += 1 - break - elif "FAILED" in nslcmop["operationState"]: - if not expected_fail: - logger.error( - "NS terminate has failed: {}".format(nslcmop["detailed-status"]) - ) - self.failed_tests += 1 - else: - self.passed_tests += 1 - break - - print(".", end="", file=stderr) - wait -= 10 - sleep(10) - else: - self.failed_tests += 1 - logger.error( - "NS instantiate is not terminate after {} seconds".format(timeout) - ) - return - print("", file=stderr) - - -class TestNonAuthorized: - description = "Test invalid URLs. methods and no authorization" - - @staticmethod - def run(engine, test_osm, manual_check, test_params=None): - engine.set_test_name("NonAuth") - engine.remove_authorization() - test_not_authorized_list = ( - ( - "Invalid token", - "GET", - "/admin/v1/users", - headers_json, - None, - 401, - r_header_json, - "json", - ), - ( - "Invalid URL", - "POST", - "/admin/v1/nonexist", - headers_yaml, - None, - 405, - r_header_yaml, - "yaml", - ), - ( - "Invalid version", - "DELETE", - "/admin/v2/users", - headers_yaml, - None, - 405, - r_header_yaml, - "yaml", - ), - ) - for t in test_not_authorized_list: - engine.test(*t) - - -class TestUsersProjects: - description = "test project and user creation" - - @staticmethod - def run(engine, test_osm, manual_check, test_params=None): - engine.set_test_name("UserProject") - # backend = test_params.get("backend") if test_params else None # UNUSED - - # Initialisation - p1 = p2 = p3 = None - padmin = pbad = None - u1 = u2 = u3 = u4 = None - - engine.get_autorization() - - res = engine.test( - "Create project non admin 1", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "P1"}, - (201, 204), - {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, - "json", - ) - p1 = engine.last_id if res else None - - res = engine.test( - "Create project admin", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "Padmin", "admin": True}, - (201, 204), - {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, - "json", - ) - padmin = engine.last_id if res else None - - res = engine.test( - "Create project bad format", - "POST", - "/admin/v1/projects", - headers_json, - {"name": 1}, - (400, 422), - r_header_json, - "json", - ) - pbad = engine.last_id if res else None - - res = engine.test( - "Get project admin role", - "GET", - "/admin/v1/roles?name=project_admin", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - rpa = res.json()[0]["_id"] if res else None - res = engine.test( - "Get project user role", - "GET", - "/admin/v1/roles?name=project_user", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - rpu = res.json()[0]["_id"] if res else None - res = engine.test( - "Get system admin role", - "GET", - "/admin/v1/roles?name=system_admin", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - rsa = res.json()[0]["_id"] if res else None - - data = {"username": "U1", "password": "pw1"} - p2 = uuid4().hex - data["project_role_mappings"] = [ - {"project": p1, "role": rpa}, - {"project": p2, "role": rpa}, - {"project": padmin, "role": rpu}, - ] - rc = 201 - xhd = {"Location": "/admin/v1/users/", "Content-Type": "application/json"} - res = engine.test( - "Create user with bad project and force", - "POST", - "/admin/v1/users?FORCE=True", - headers_json, - data, - rc, - xhd, - "json", - ) - if res: - u1 = engine.last_id - else: - # User is created sometimes even though an exception is raised - res = engine.test( - "Get user U1", - "GET", - "/admin/v1/users?username=U1", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - u1 = res.json()[0]["_id"] if res else None - - data = {"username": "U2", "password": "pw2"} - data["project_role_mappings"] = [ - {"project": p1, "role": rpa}, - {"project": padmin, "role": rsa}, - ] - res = engine.test( - "Create user 2", - "POST", - "/admin/v1/users", - headers_json, - data, - 201, - {"Location": "/admin/v1/users/", "Content-Type": "application/json"}, - "json", - ) - u2 = engine.last_id if res else None - - if u1: - ftt = "project_role_mappings" - xpr = [{"project": p1, "role": rpa}, {"project": padmin, "role": rpu}] - data = {ftt: xpr} - engine.test( - "Edit user U1, delete P2 project", - "PATCH", - "/admin/v1/users/" + u1, - headers_json, - data, - 204, - None, - None, - ) - res = engine.test( - "Check user U1, contains the right projects", - "GET", - "/admin/v1/users/" + u1, - headers_json, - None, - 200, - None, - json, - ) - if res: - rj = res.json() - xpr[0]["project_name"] = "P1" - xpr[0]["role_name"] = "project_admin" - xpr[1]["project_name"] = "Padmin" - xpr[1]["role_name"] = "project_user" - ok = True - for pr in rj[ftt]: - if pr not in xpr: - ok = False - for pr in xpr: - if pr not in rj[ftt]: - ok = False - if not ok: - logger.error( - "User {} '{}' are different than expected '{}'. Edition was not done properly".format( - ftt, rj[ftt], xpr - ) - ) - engine.failed_tests += 1 - - p2 = None # To prevent deletion attempts - - # Add a test of 'default project' for Keystone? - - if u2: - engine.test( - "Edit user U2, change password", - "PUT", - "/admin/v1/users/" + u2, - headers_json, - {"password": "pw2_new"}, - 204, - None, - None, - ) - - if p1: - engine.test( - "Change to project P1 non existing", - "POST", - "/admin/v1/tokens/", - headers_json, - {"project_id": p1}, - 401, - r_header_json, - "json", - ) - - if u2 and p1: - res = engine.test( - "Change to user U2 project P1", - "POST", - "/admin/v1/tokens", - headers_json, - {"username": "U2", "password": "pw2_new", "project_id": "P1"}, - (200, 201), - r_header_json, - "json", - ) - if res: - rj = res.json() - engine.set_header({"Authorization": "Bearer {}".format(rj["id"])}) - - engine.test( - "Edit user projects non admin", - "PUT", - "/admin/v1/users/U1", - headers_json, - {"remove_project_role_mappings": [{"project": "P1", "role": None}]}, - 401, - r_header_json, - "json", - ) - - res = engine.test( - "Add new project non admin", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "P2"}, - 401, - r_header_json, - "json", - ) - if res is None or res.status_code == 201: - # The project has been created even though it shouldn't - res = engine.test( - "Get project P2", - "GET", - "/admin/v1/projects/P2", - headers_json, - None, - 200, - r_header_json, - "json", - ) - p2 = res.json()["_id"] if res else None - - if p1: - data = {"username": "U3", "password": "pw3"} - data["project_role_mappings"] = [{"project": p1, "role": rpu}] - res = engine.test( - "Add new user non admin", - "POST", - "/admin/v1/users", - headers_json, - data, - 401, - r_header_json, - "json", - ) - if res is None or res.status_code == 201: - # The user has been created even though it shouldn't - res = engine.test( - "Get user U3", - "GET", - "/admin/v1/users/U3", - headers_json, - None, - 200, - r_header_json, - "json", - ) - u3 = res.json()["_id"] if res else None - else: - u3 = None - - if padmin: - res = engine.test( - "Change to user U2 project Padmin", - "POST", - "/admin/v1/tokens", - headers_json, - { - "project_id": "Padmin" - }, # Caused a Keystone authentication error - # {"username": "U2", "password": "pw2_new", "project_id": "Padmin"}, - (200, 201), - r_header_json, - "json", - ) - if res: - rj = res.json() - engine.set_header( - {"Authorization": "Bearer {}".format(rj["id"])} - ) - - res = engine.test( - "Add new project admin", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "P3"}, - (201, 204), - { - "Location": "/admin/v1/projects/", - "Content-Type": "application/json", - }, - "json", - ) - p3 = engine.last_id if res else None - - if p1: - data = {"username": "U4", "password": "pw4"} - data["project_role_mappings"] = [ - {"project": p1, "role": rpa} - ] - res = engine.test( - "Add new user admin", - "POST", - "/admin/v1/users", - headers_json, - data, - (201, 204), - { - "Location": "/admin/v1/users/", - "Content-Type": "application/json", - }, - "json", - ) - u4 = engine.last_id if res else None - else: - u4 = None - - if u4 and p3: - data = { - "project_role_mappings": [{"project": p3, "role": rpa}] - } - engine.test( - "Edit user projects admin", - "PUT", - "/admin/v1/users/U4", - headers_json, - data, - 204, - None, - None, - ) - # Project is deleted even though it shouldn't - PROVISIONAL? - res = engine.test( - "Delete project P3 conflict", - "DELETE", - "/admin/v1/projects/" + p3, - headers_json, - None, - 409, - None, - None, - ) - if res and res.status_code in (200, 204): - p3 = None - if p3: - res = engine.test( - "Delete project P3 forcing", - "DELETE", - "/admin/v1/projects/" + p3 + "?FORCE=True", - headers_json, - None, - 204, - None, - None, - ) - if res and res.status_code in (200, 204): - p3 = None - - if u2: - res = engine.test( - "Delete user U2. Conflict deleting own user", - "DELETE", - "/admin/v1/users/" + u2, - headers_json, - None, - 409, - r_header_json, - "json", - ) - if res is None or res.status_code in (200, 204): - u2 = None - if u4: - res = engine.test( - "Delete user U4", - "DELETE", - "/admin/v1/users/" + u4, - headers_json, - None, - 204, - None, - None, - ) - if res and res.status_code in (200, 204): - u4 = None - if p3: - res = engine.test( - "Delete project P3", - "DELETE", - "/admin/v1/projects/" + p3, - headers_json, - None, - 204, - None, - None, - ) - if res and res.status_code in (200, 204): - p3 = None - - if u3: - res = engine.test( - "Delete user U3", - "DELETE", - "/admin/v1/users/" + u3, - headers_json, - None, - 204, - None, - None, - ) - if res: - u3 = None - - # change to admin - engine.remove_authorization() # To force get authorization - engine.get_autorization() - if u1: - engine.test( - "Delete user U1", - "DELETE", - "/admin/v1/users/" + u1, - headers_json, - None, - 204, - None, - None, - ) - if u2: - engine.test( - "Delete user U2", - "DELETE", - "/admin/v1/users/" + u2, - headers_json, - None, - 204, - None, - None, - ) - if u3: - engine.test( - "Delete user U3", - "DELETE", - "/admin/v1/users/" + u3, - headers_json, - None, - 204, - None, - None, - ) - if u4: - engine.test( - "Delete user U4", - "DELETE", - "/admin/v1/users/" + u4, - headers_json, - None, - 204, - None, - None, - ) - if p1: - engine.test( - "Delete project P1", - "DELETE", - "/admin/v1/projects/" + p1, - headers_json, - None, - 204, - None, - None, - ) - if p2: - engine.test( - "Delete project P2", - "DELETE", - "/admin/v1/projects/" + p2, - headers_json, - None, - 204, - None, - None, - ) - if p3: - engine.test( - "Delete project P3", - "DELETE", - "/admin/v1/projects/" + p3, - headers_json, - None, - 204, - None, - None, - ) - if padmin: - engine.test( - "Delete project Padmin", - "DELETE", - "/admin/v1/projects/" + padmin, - headers_json, - None, - 204, - None, - None, - ) - if pbad: - engine.test( - "Delete bad project", - "DELETE", - "/admin/v1/projects/" + pbad, - headers_json, - None, - 204, - None, - None, - ) - - # BEGIN New Tests - Addressing Projects/Users by Name/ID - pid1 = pid2 = None - uid1 = uid2 = None - res = engine.test( - "Create new project P1", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "P1"}, - 201, - {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, - "json", - ) - if res: - pid1 = res.json()["id"] - # print("# pid =", pid1) - res = engine.test( - "Create new project P2", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "P2"}, - 201, - {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, - "json", - ) - if res: - pid2 = res.json()["id"] - # print("# pid =", pid2) - data = {"username": "U1", "password": "pw1"} - data["project_role_mappings"] = [{"project": pid1, "role": rpu}] - res = engine.test( - "Create new user U1", - "POST", - "/admin/v1/users", - headers_json, - data, - 201, - {"Location": "/admin/v1/users/", "Content-Type": "application/json"}, - "json", - ) - if res: - uid1 = res.json()["id"] - # print("# uid =", uid1) - data = {"username": "U2", "password": "pw2"} - data["project_role_mappings"] = [{"project": pid2, "role": rpu}] - res = engine.test( - "Create new user U2", - "POST", - "/admin/v1/users", - headers_json, - data, - 201, - {"Location": "/admin/v1/users/", "Content-Type": "application/json"}, - "json", - ) - if res: - uid2 = res.json()["id"] - # print("# uid =", uid2) - if pid1: - engine.test( - "Get Project P1 by Name", - "GET", - "/admin/v1/projects/P1", - headers_json, - None, - 200, - None, - "json", - ) - engine.test( - "Get Project P1 by ID", - "GET", - "/admin/v1/projects/" + pid1, - headers_json, - None, - 200, - None, - "json", - ) - if uid1: - engine.test( - "Get User U1 by Name", - "GET", - "/admin/v1/users/U1", - headers_json, - None, - 200, - None, - "json", - ) - engine.test( - "Get User U1 by ID", - "GET", - "/admin/v1/users/" + uid1, - headers_json, - None, - 200, - None, - "json", - ) - if pid1: - res = engine.test( - "Rename Project P1 by Name", - "PUT", - "/admin/v1/projects/P1", - headers_json, - {"name": "P3"}, - 204, - None, - None, - ) - if res: - engine.test( - "Get Project P1 by new Name", - "GET", - "/admin/v1/projects/P3", - headers_json, - None, - 200, - None, - "json", - ) - if pid2: - res = engine.test( - "Rename Project P2 by ID", - "PUT", - "/admin/v1/projects/" + pid2, - headers_json, - {"name": "P4"}, - 204, - None, - None, - ) - if res: - engine.test( - "Get Project P2 by new Name", - "GET", - "/admin/v1/projects/P4", - headers_json, - None, - 200, - None, - "json", - ) - - if uid1: - res = engine.test( - "Rename User U1 by Name", - "PUT", - "/admin/v1/users/U1", - headers_json, - {"username": "U3"}, - 204, - None, - None, - ) - if res: - engine.test( - "Get User U1 by new Name", - "GET", - "/admin/v1/users/U3", - headers_json, - None, - 200, - None, - "json", - ) - - if uid2: - res = engine.test( - "Rename User U2 by ID", - "PUT", - "/admin/v1/users/" + uid2, - headers_json, - {"username": "U4"}, - 204, - None, - None, - ) - if res: - engine.test( - "Get User U2 by new Name", - "GET", - "/admin/v1/users/U4", - headers_json, - None, - 200, - None, - "json", - ) - if uid1: - res = engine.test( - "Delete User U1 by Name", - "DELETE", - "/admin/v1/users/U3", - headers_json, - None, - 204, - None, - None, - ) - if res: - uid1 = None - - if uid2: - res = engine.test( - "Delete User U2 by ID", - "DELETE", - "/admin/v1/users/" + uid2, - headers_json, - None, - 204, - None, - None, - ) - if res: - uid2 = None - - if pid1: - res = engine.test( - "Delete Project P1 by Name", - "DELETE", - "/admin/v1/projects/P3", - headers_json, - None, - 204, - None, - None, - ) - if res: - pid1 = None - - if pid2: - res = engine.test( - "Delete Project P2 by ID", - "DELETE", - "/admin/v1/projects/" + pid2, - headers_json, - None, - 204, - None, - None, - ) - if res: - pid2 = None - - # END New Tests - Addressing Projects/Users by Name - - # CLEANUP - if pid1: - engine.test( - "Delete Project P1", - "DELETE", - "/admin/v1/projects/" + pid1, - headers_json, - None, - 204, - None, - None, - ) - if pid2: - engine.test( - "Delete Project P2", - "DELETE", - "/admin/v1/projects/" + pid2, - headers_json, - None, - 204, - None, - None, - ) - if uid1: - engine.test( - "Delete User U1", - "DELETE", - "/admin/v1/users/" + uid1, - headers_json, - None, - 204, - None, - None, - ) - if uid2: - engine.test( - "Delete User U2", - "DELETE", - "/admin/v1/users/" + uid2, - headers_json, - None, - 204, - None, - None, - ) - - engine.remove_authorization() # To finish - - -class TestProjectsDescriptors: - description = "test descriptors visibility among projects" - - @staticmethod - def run(engine, test_osm, manual_check, test_params=None): - vnfd_ids = [] - engine.set_test_name("ProjectDescriptors") - engine.get_autorization() - - project_admin_id = None - res = engine.test( - "Get my project Padmin", - "GET", - "/admin/v1/projects/{}".format(engine.project), - headers_json, - None, - 200, - r_header_json, - "json", - ) - if res: - response = res.json() - project_admin_id = response["_id"] - engine.test( - "Create project Padmin", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "Padmin", "admin": True}, - (201, 204), - {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, - "json", - ) - engine.test( - "Create project P2", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "P2"}, - (201, 204), - {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, - "json", - ) - engine.test( - "Create project P3", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "P3"}, - (201, 204), - {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, - "json", - ) - - engine.test( - "Create user U1", - "POST", - "/admin/v1/users", - headers_json, - { - "username": "U1", - "password": "pw1", - "project_role_mappings": [ - {"project": "Padmin", "role": "system_admin"}, - {"project": "P2", "role": "project_admin"}, - {"project": "P3", "role": "project_admin"}, - ], - }, - 201, - {"Location": "/admin/v1/users/", "Content-Type": "application/json"}, - "json", - ) - - engine.test( - "Onboard VNFD id1", - "POST", - "/vnfpkgm/v1/vnf_packages_content?id=id1", - headers_yaml, - TestDescriptors.vnfd_empty, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - vnfd_ids.append(engine.last_id) - engine.test( - "Onboard VNFD id2 PUBLIC", - "POST", - "/vnfpkgm/v1/vnf_packages_content?id=id2&PUBLIC=TRUE", - headers_yaml, - TestDescriptors.vnfd_empty, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - vnfd_ids.append(engine.last_id) - engine.test( - "Onboard VNFD id3", - "POST", - "/vnfpkgm/v1/vnf_packages_content?id=id3&PUBLIC=FALSE", - headers_yaml, - TestDescriptors.vnfd_empty, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - vnfd_ids.append(engine.last_id) - - res = engine.test( - "Get VNFD descriptors", - "GET", - "/vnfpkgm/v1/vnf_packages?id=id1,id2,id3", - headers_json, - None, - 200, - r_header_json, - "json", - ) - response = res.json() - if len(response) != 3: - logger.error( - "Only 3 vnfds should be present for project admin. {} listed".format( - len(response) - ) - ) - engine.failed_tests += 1 - - # Change to other project Padmin - res = engine.test( - "Change to user U1 project Padmin", - "POST", - "/admin/v1/tokens", - headers_json, - {"username": "U1", "password": "pw1", "project_id": "Padmin"}, - (200, 201), - r_header_json, - "json", - ) - if res: - response = res.json() - engine.set_header({"Authorization": "Bearer {}".format(response["id"])}) - - # list vnfds - res = engine.test( - "List VNFD descriptors for Padmin", - "GET", - "/vnfpkgm/v1/vnf_packages", - headers_json, - None, - 200, - r_header_json, - "json", - ) - response = res.json() - if len(response) != 0: - logger.error( - "Only 0 vnfds should be present for project Padmin. {} listed".format( - len(response) - ) - ) - engine.failed_tests += 1 - - # list Public vnfds - res = engine.test( - "List VNFD public descriptors", - "GET", - "/vnfpkgm/v1/vnf_packages?PUBLIC=True", - headers_json, - None, - 200, - r_header_json, - "json", - ) - response = res.json() - if len(response) != 1: - logger.error( - "Only 1 vnfds should be present for project Padmin. {} listed".format( - len(response) - ) - ) - engine.failed_tests += 1 - - # list vnfds belonging to project "admin" - res = engine.test( - "List VNFD of admin project", - "GET", - "/vnfpkgm/v1/vnf_packages?ADMIN={}".format(project_admin_id), - headers_json, - None, - 200, - r_header_json, - "json", - ) - if res: - response = res.json() - if len(response) != 3: - logger.error( - "Only 3 vnfds should be present for project Padmin. {} listed".format( - len(response) - ) - ) - engine.failed_tests += 1 - - # Get Public vnfds - engine.test( - "Get VNFD public descriptors", - "GET", - "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[1]), - headers_json, - None, - 200, - r_header_json, - "json", - ) - # Edit not owned vnfd - engine.test( - "Edit VNFD ", - "PATCH", - "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[0]), - headers_yaml, - "{name: pepe}", - 404, - r_header_yaml, - "yaml", - ) - - # Add to my catalog - engine.test( - "Add VNFD id2 to my catalog", - "PATCH", - "/vnfpkgm/v1/vnf_packages/{}?SET_PROJECT".format(vnfd_ids[1]), - headers_json, - None, - 204, - None, - 0, - ) - - # Add a new vnfd - engine.test( - "Onboard VNFD id4", - "POST", - "/vnfpkgm/v1/vnf_packages_content?id=id4", - headers_yaml, - TestDescriptors.vnfd_empty, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - vnfd_ids.append(engine.last_id) - - # list vnfds - res = engine.test( - "List VNFD public descriptors", - "GET", - "/vnfpkgm/v1/vnf_packages", - headers_json, - None, - 200, - r_header_json, - "json", - ) - response = res.json() - if len(response) != 2: - logger.error( - "Only 2 vnfds should be present for project Padmin. {} listed".format( - len(response) - ) - ) - engine.failed_tests += 1 - - if manual_check: - input( - "VNFDs have been omboarded. Perform manual check and press enter to resume" - ) - - test_rest.test( - "Delete VNFD id2", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[1]), - headers_yaml, - None, - 204, - None, - 0, - ) - - # change to admin project - engine.remove_authorization() # To force get authorization - engine.get_autorization() - test_rest.test( - "Delete VNFD id1", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[0]), - headers_yaml, - None, - 204, - None, - 0, - ) - test_rest.test( - "Delete VNFD id2", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[1]), - headers_yaml, - None, - 204, - None, - 0, - ) - test_rest.test( - "Delete VNFD id3", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[2]), - headers_yaml, - None, - 204, - None, - 0, - ) - test_rest.test( - "Delete VNFD id4", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_ids[3]), - headers_yaml, - None, - 404, - r_header_yaml, - "yaml", - ) - test_rest.test( - "Delete VNFD id4", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[3]), - headers_yaml, - None, - 204, - None, - 0, - ) - # Get Public vnfds - engine.test( - "Get VNFD deleted id1", - "GET", - "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[0]), - headers_json, - None, - 404, - r_header_json, - "json", - ) - engine.test( - "Get VNFD deleted id2", - "GET", - "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[1]), - headers_json, - None, - 404, - r_header_json, - "json", - ) - engine.test( - "Get VNFD deleted id3", - "GET", - "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[2]), - headers_json, - None, - 404, - r_header_json, - "json", - ) - engine.test( - "Get VNFD deleted id4", - "GET", - "/vnfpkgm/v1/vnf_packages/{}?ADMIN".format(vnfd_ids[3]), - headers_json, - None, - 404, - r_header_json, - "json", - ) - - engine.test( - "Delete user U1", - "DELETE", - "/admin/v1/users/U1", - headers_json, - None, - 204, - None, - None, - ) - engine.test( - "Delete project Padmin", - "DELETE", - "/admin/v1/projects/Padmin", - headers_json, - None, - 204, - None, - None, - ) - engine.test( - "Delete project P2", - "DELETE", - "/admin/v1/projects/P2", - headers_json, - None, - 204, - None, - None, - ) - engine.test( - "Delete project P3", - "DELETE", - "/admin/v1/projects/P3", - headers_json, - None, - 204, - None, - None, - ) - - -class TestFakeVim: - description = "Creates/edit/delete fake VIMs and SDN controllers" - - def __init__(self): - self.vim = { - "schema_version": "1.0", - "schema_type": "No idea", - "name": "myVim", - "description": "Descriptor name", - "vim_type": "openstack", - "vim_url": "http://localhost:/vim", - "vim_tenant_name": "vimTenant", - "vim_user": "user", - "vim_password": "password", - "config": {"config_param": 1}, - } - self.sdn = { - "name": "sdn-name", - "description": "sdn-description", - "dpid": "50:50:52:54:00:94:21:21", - "ip": "192.168.15.17", - "port": 8080, - "type": "opendaylight", - "version": "3.5.6", - "user": "user", - "password": "passwd", - } - self.port_mapping = [ - { - "compute_node": "compute node 1", - "ports": [ - { - "pci": "0000:81:00.0", - "switch_port": "port-2/1", - "switch_mac": "52:54:00:94:21:21", - }, - { - "pci": "0000:81:00.1", - "switch_port": "port-2/2", - "switch_mac": "52:54:00:94:21:22", - }, - ], - }, - { - "compute_node": "compute node 2", - "ports": [ - { - "pci": "0000:81:00.0", - "switch_port": "port-2/3", - "switch_mac": "52:54:00:94:21:23", - }, - { - "pci": "0000:81:00.1", - "switch_port": "port-2/4", - "switch_mac": "52:54:00:94:21:24", - }, - ], - }, - ] - - def run(self, engine, test_osm, manual_check, test_params=None): - vim_bad = self.vim.copy() - vim_bad.pop("name") - - engine.set_test_name("FakeVim") - engine.get_autorization() - engine.test( - "Create VIM", - "POST", - "/admin/v1/vim_accounts", - headers_json, - self.vim, - (201, 202), - {"Location": "/admin/v1/vim_accounts/", "Content-Type": "application/json"}, - "json", - ) - vim_id = engine.last_id - engine.test( - "Create VIM without name, bad schema", - "POST", - "/admin/v1/vim_accounts", - headers_json, - vim_bad, - 422, - None, - headers_json, - ) - engine.test( - "Create VIM name repeated", - "POST", - "/admin/v1/vim_accounts", - headers_json, - self.vim, - 409, - None, - headers_json, - ) - engine.test( - "Show VIMs", - "GET", - "/admin/v1/vim_accounts", - headers_yaml, - None, - 200, - r_header_yaml, - "yaml", - ) - engine.test( - "Show VIM", - "GET", - "/admin/v1/vim_accounts/{}".format(vim_id), - headers_yaml, - None, - 200, - r_header_yaml, - "yaml", - ) - if not test_osm: - # delete with FORCE - engine.test( - "Delete VIM", - "DELETE", - "/admin/v1/vim_accounts/{}?FORCE=True".format(vim_id), - headers_yaml, - None, - 202, - None, - 0, - ) - engine.test( - "Check VIM is deleted", - "GET", - "/admin/v1/vim_accounts/{}".format(vim_id), - headers_yaml, - None, - 404, - r_header_yaml, - "yaml", - ) - else: - # delete and wait until is really deleted - engine.test( - "Delete VIM", - "DELETE", - "/admin/v1/vim_accounts/{}".format(vim_id), - headers_yaml, - None, - 202, - None, - 0, - ) - engine.wait_until_delete( - "/admin/v1/vim_accounts/{}".format(vim_id), timeout - ) - - -class TestVIMSDN(TestFakeVim): - description = "Creates VIM with SDN editing SDN controllers and port_mapping" - - def __init__(self): - TestFakeVim.__init__(self) - self.wim = { - "schema_version": "1.0", - "schema_type": "No idea", - "name": "myWim", - "description": "Descriptor name", - "wim_type": "odl", - "wim_url": "http://localhost:/wim", - "user": "user", - "password": "password", - "config": {"config_param": 1}, - } - - def run(self, engine, test_osm, manual_check, test_params=None): - engine.set_test_name("VimSdn") - engine.get_autorization() - # Added SDN - engine.test( - "Create SDN", - "POST", - "/admin/v1/sdns", - headers_json, - self.sdn, - (201, 202), - {"Location": "/admin/v1/sdns/", "Content-Type": "application/json"}, - "json", - ) - sdnc_id = engine.last_id - # sleep(5) - # Edit SDN - engine.test( - "Edit SDN", - "PATCH", - "/admin/v1/sdns/{}".format(sdnc_id), - headers_json, - {"name": "new_sdn_name"}, - (202, 204), - None, - None, - ) - # sleep(5) - # VIM with SDN - self.vim["config"]["sdn-controller"] = sdnc_id - self.vim["config"]["sdn-port-mapping"] = self.port_mapping - engine.test( - "Create VIM", - "POST", - "/admin/v1/vim_accounts", - headers_json, - self.vim, - (200, 202, 201), - {"Location": "/admin/v1/vim_accounts/", "Content-Type": "application/json"}, - "json", - ), - - vim_id = engine.last_id - self.port_mapping[0]["compute_node"] = "compute node XX" - engine.test( - "Edit VIM change port-mapping", - "PUT", - "/admin/v1/vim_accounts/{}".format(vim_id), - headers_json, - {"config": {"sdn-port-mapping": self.port_mapping}}, - (202, 204), - None, - None, - ) - engine.test( - "Edit VIM remove port-mapping", - "PUT", - "/admin/v1/vim_accounts/{}".format(vim_id), - headers_json, - {"config": {"sdn-port-mapping": None}}, - (202, 204), - None, - None, - ) - - engine.test( - "Create WIM", - "POST", - "/admin/v1/wim_accounts", - headers_json, - self.wim, - (200, 202, 201), - {"Location": "/admin/v1/wim_accounts/", "Content-Type": "application/json"}, - "json", - ), - wim_id = engine.last_id - - if not test_osm: - # delete with FORCE - engine.test( - "Delete VIM remove port-mapping", - "DELETE", - "/admin/v1/vim_accounts/{}?FORCE=True".format(vim_id), - headers_json, - None, - 202, - None, - 0, - ) - engine.test( - "Delete SDNC", - "DELETE", - "/admin/v1/sdns/{}?FORCE=True".format(sdnc_id), - headers_json, - None, - 202, - None, - 0, - ) - - engine.test( - "Delete WIM", - "DELETE", - "/admin/v1/wim_accounts/{}?FORCE=True".format(wim_id), - headers_json, - None, - 202, - None, - 0, - ) - engine.test( - "Check VIM is deleted", - "GET", - "/admin/v1/vim_accounts/{}".format(vim_id), - headers_yaml, - None, - 404, - r_header_yaml, - "yaml", - ) - engine.test( - "Check SDN is deleted", - "GET", - "/admin/v1/sdns/{}".format(sdnc_id), - headers_yaml, - None, - 404, - r_header_yaml, - "yaml", - ) - engine.test( - "Check WIM is deleted", - "GET", - "/admin/v1/wim_accounts/{}".format(wim_id), - headers_yaml, - None, - 404, - r_header_yaml, - "yaml", - ) - else: - if manual_check: - input( - "VIM, SDN, WIM has been deployed. Perform manual check and press enter to resume" - ) - # delete and wait until is really deleted - engine.test( - "Delete VIM remove port-mapping", - "DELETE", - "/admin/v1/vim_accounts/{}".format(vim_id), - headers_json, - None, - (202, 201, 204), - None, - 0, - ) - engine.test( - "Delete SDN", - "DELETE", - "/admin/v1/sdns/{}".format(sdnc_id), - headers_json, - None, - (202, 201, 204), - None, - 0, - ) - engine.test( - "Delete VIM", - "DELETE", - "/admin/v1/wim_accounts/{}".format(wim_id), - headers_json, - None, - (202, 201, 204), - None, - 0, - ) - engine.wait_until_delete( - "/admin/v1/vim_accounts/{}".format(vim_id), timeout - ) - engine.wait_until_delete("/admin/v1/sdns/{}".format(sdnc_id), timeout) - engine.wait_until_delete( - "/admin/v1/wim_accounts/{}".format(wim_id), timeout - ) - - -class TestDeploy: - description = "Base class for downloading descriptors from ETSI, onboard and deploy in real VIM" - - def __init__(self): - self.test_name = "DEPLOY" - self.nsd_id = None - self.vim_id = None - self.ns_id = None - self.vnfds_id = [] - self.descriptor_url = ( - "https://osm-download.etsi.org/ftp/osm-3.0-three/2nd-hackfest/packages/" - ) - self.vnfd_filenames = ("cirros_vnf.tar.gz",) - self.nsd_filename = "cirros_2vnf_ns.tar.gz" - self.descriptor_edit = None - self.uses_configuration = False - self.users = {} - self.passwords = {} - self.commands = {} - self.keys = {} - self.timeout = 120 - self.qforce = "" - self.ns_params = None - self.vnfr_ip_list = {} - - def create_descriptors(self, engine): - temp_dir = os.path.dirname(os.path.abspath(__file__)) + "/temp/" - if not os.path.exists(temp_dir): - os.makedirs(temp_dir) - for vnfd_index, vnfd_filename in enumerate(self.vnfd_filenames): - if "/" in vnfd_filename: - vnfd_filename_path = vnfd_filename - if not os.path.exists(vnfd_filename_path): - raise TestException( - "File '{}' does not exist".format(vnfd_filename_path) - ) - else: - vnfd_filename_path = temp_dir + vnfd_filename - if not os.path.exists(vnfd_filename_path): - with open(vnfd_filename_path, "wb") as file: - response = requests.get(self.descriptor_url + vnfd_filename) - if response.status_code >= 300: - raise TestException( - "Error downloading descriptor from '{}': {}".format( - self.descriptor_url + vnfd_filename, - response.status_code, - ) - ) - file.write(response.content) - if vnfd_filename_path.endswith(".yaml"): - headers = headers_yaml - else: - headers = headers_zip_yaml - if randint(0, 1) == 0: - # vnfd CREATE AND UPLOAD in one step: - engine.test( - "Onboard VNFD in one step", - "POST", - "/vnfpkgm/v1/vnf_packages_content" + self.qforce, - headers, - "@b" + vnfd_filename_path, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - self.vnfds_id.append(engine.last_id) - else: - # vnfd CREATE AND UPLOAD ZIP - engine.test( - "Onboard VNFD step 1", - "POST", - "/vnfpkgm/v1/vnf_packages", - headers_json, - None, - 201, - { - "Location": "/vnfpkgm/v1/vnf_packages/", - "Content-Type": "application/json", - }, - "json", - ) - self.vnfds_id.append(engine.last_id) - engine.test( - "Onboard VNFD step 2 as ZIP", - "PUT", - "/vnfpkgm/v1/vnf_packages/<>/package_content" + self.qforce, - headers, - "@b" + vnfd_filename_path, - 204, - None, - 0, - ) - - if self.descriptor_edit: - if "vnfd{}".format(vnfd_index) in self.descriptor_edit: - # Modify VNFD - engine.test( - "Edit VNFD ", - "PATCH", - "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfds_id[-1]), - headers_yaml, - self.descriptor_edit["vnfd{}".format(vnfd_index)], - 204, - None, - None, - ) - - if "/" in self.nsd_filename: - nsd_filename_path = self.nsd_filename - if not os.path.exists(nsd_filename_path): - raise TestException( - "File '{}' does not exist".format(nsd_filename_path) - ) - else: - nsd_filename_path = temp_dir + self.nsd_filename - if not os.path.exists(nsd_filename_path): - with open(nsd_filename_path, "wb") as file: - response = requests.get(self.descriptor_url + self.nsd_filename) - if response.status_code >= 300: - raise TestException( - "Error downloading descriptor from '{}': {}".format( - self.descriptor_url + self.nsd_filename, - response.status_code, - ) - ) - file.write(response.content) - if nsd_filename_path.endswith(".yaml"): - headers = headers_yaml - else: - headers = headers_zip_yaml - - if randint(0, 1) == 0: - # nsd CREATE AND UPLOAD in one step: - engine.test( - "Onboard NSD in one step", - "POST", - "/nsd/v1/ns_descriptors_content" + self.qforce, - headers, - "@b" + nsd_filename_path, - 201, - r_headers_yaml_location_nsd, - yaml, - ) - self.nsd_id = engine.last_id - else: - # nsd CREATE AND UPLOAD ZIP - engine.test( - "Onboard NSD step 1", - "POST", - "/nsd/v1/ns_descriptors", - headers_json, - None, - 201, - { - "Location": "/nsd/v1/ns_descriptors/", - "Content-Type": "application/json", - }, - "json", - ) - self.nsd_id = engine.last_id - engine.test( - "Onboard NSD step 2 as ZIP", - "PUT", - "/nsd/v1/ns_descriptors/<>/nsd_content" + self.qforce, - headers, - "@b" + nsd_filename_path, - 204, - None, - 0, - ) - - if self.descriptor_edit and "nsd" in self.descriptor_edit: - # Modify NSD - engine.test( - "Edit NSD ", - "PATCH", - "/nsd/v1/ns_descriptors/{}".format(self.nsd_id), - headers_yaml, - self.descriptor_edit["nsd"], - 204, - None, - None, - ) - - def delete_descriptors(self, engine): - # delete descriptors - engine.test( - "Delete NSSD SOL005", - "DELETE", - "/nsd/v1/ns_descriptors/{}".format(self.nsd_id), - headers_yaml, - None, - 204, - None, - 0, - ) - for vnfd_id in self.vnfds_id: - engine.test( - "Delete VNFD SOL005", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(vnfd_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - def instantiate(self, engine, ns_data): - ns_data_text = yaml.safe_dump(ns_data, default_flow_style=True, width=256) - # create NS Two steps - r = engine.test( - "Create NS step 1", - "POST", - "/nslcm/v1/ns_instances", - headers_yaml, - ns_data_text, - (201, 202), - {"Location": "nslcm/v1/ns_instances/", "Content-Type": "application/yaml"}, - "yaml", - ) - if not r: - return - self.ns_id = engine.last_id - engine.test( - "Instantiate NS step 2", - "POST", - "/nslcm/v1/ns_instances/{}/instantiate".format(self.ns_id), - headers_yaml, - ns_data_text, - (201, 202), - r_headers_yaml_location_nslcmop, - "yaml", - ) - nslcmop_id = engine.last_id - - if test_osm: - # Wait until status is Ok - timeout = timeout_configure if self.uses_configuration else timeout_deploy - engine.wait_operation_ready("ns", nslcmop_id, timeout) - - def terminate(self, engine): - # remove deployment - if test_osm: - engine.test( - "Terminate NS", - "POST", - "/nslcm/v1/ns_instances/{}/terminate".format(self.ns_id), - headers_yaml, - None, - (201, 202), - r_headers_yaml_location_nslcmop, - "yaml", - ) - nslcmop2_id = engine.last_id - # Wait until status is Ok - engine.wait_operation_ready("ns", nslcmop2_id, timeout_deploy) - - engine.test( - "Delete NS", - "DELETE", - "/nslcm/v1/ns_instances/{}".format(self.ns_id), - headers_yaml, - None, - 204, - None, - 0, - ) - else: - engine.test( - "Delete NS with FORCE", - "DELETE", - "/nslcm/v1/ns_instances/{}?FORCE=True".format(self.ns_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - # check all it is deleted - engine.test( - "Check NS is deleted", - "GET", - "/nslcm/v1/ns_instances/{}".format(self.ns_id), - headers_yaml, - None, - 404, - None, - "yaml", - ) - r = engine.test( - "Check NSLCMOPs are deleted", - "GET", - "/nslcm/v1/ns_lcm_op_occs?nsInstanceId={}".format(self.ns_id), - headers_json, - None, - 200, - None, - "json", - ) - if not r: - return - nslcmops = r.json() - if not isinstance(nslcmops, list) or nslcmops: - raise TestException( - "NS {} deleted but with ns_lcm_op_occ active: {}".format( - self.ns_id, nslcmops - ) - ) - - def test_ns( - self, - engine, - test_osm, - commands=None, - users=None, - passwds=None, - keys=None, - timeout=0, - ): - r = engine.test( - "GET VNFR IDs", - "GET", - "/nslcm/v1/ns_instances/{}".format(self.ns_id), - headers_json, - None, - 200, - r_header_json, - "json", - ) - if not r: - return - ns_data = r.json() - - vnfr_list = ns_data["constituent-vnfr-ref"] - time = 0 - _commands = commands if commands is not None else self.commands - _users = users if users is not None else self.users - _passwds = passwds if passwds is not None else self.passwords - _keys = keys if keys is not None else self.keys - _timeout = timeout if timeout != 0 else self.timeout - - # vnfr_list=[d8272263-6bd3-4680-84ca-6a4be23b3f2d, 88b22e2f-994a-4b61-94fd-4a3c90de3dc4] - for vnfr_id in vnfr_list: - r = engine.test( - "Get VNFR to get IP_ADDRESS", - "GET", - "/nslcm/v1/vnfrs/{}".format(vnfr_id), - headers_json, - None, - 200, - r_header_json, - "json", - ) - if not r: - continue - vnfr_data = r.json() - - vnf_index = str(vnfr_data["member-vnf-index-ref"]) - - ip_address = self.get_vnfr_ip(engine, vnf_index) - description = "Exec command='{}' at VNFR={} IP={}".format( - _commands.get(vnf_index)[0], vnf_index, ip_address - ) - engine.step += 1 - test_description = "{}{} {}".format( - engine.test_name, engine.step, description - ) - logger.warning(test_description) - while _timeout >= time: - result, message = self.do_checks( - [ip_address], - vnf_index=vnfr_data["member-vnf-index-ref"], - commands=_commands.get(vnf_index), - user=_users.get(vnf_index), - passwd=_passwds.get(vnf_index), - key=_keys.get(vnf_index), - ) - if result == 1: - engine.passed_tests += 1 - logger.debug(message) - break - elif result == 0: - time += 20 - sleep(20) - elif result == -1: - engine.failed_tests += 1 - logger.error(message) - break - else: - time -= 20 - engine.failed_tests += 1 - logger.error(message) - else: - engine.failed_tests += 1 - logger.error( - "VNFR {} has not mgmt address. Check failed".format(vnf_index) - ) - - def do_checks(self, ip, vnf_index, commands=[], user=None, passwd=None, key=None): - try: - import urllib3 - from pssh.clients import ParallelSSHClient - from pssh.utils import load_private_key - from ssh2 import exceptions as ssh2Exception - except ImportError as e: - logger.critical( - "Package or/and is not installed. Please add them with 'pip3 install " - "parallel-ssh urllib3': {}".format(e) - ) - return -1, "install needed packages 'pip3 install parallel-ssh urllib3'" - urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning) - try: - p_host = os.environ.get("PROXY_HOST") - p_user = os.environ.get("PROXY_USER") - p_password = os.environ.get("PROXY_PASSWD") - - if key: - pkey = load_private_key(key) - else: - pkey = None - - client = ParallelSSHClient( - ip, - user=user, - password=passwd, - pkey=pkey, - proxy_host=p_host, - proxy_user=p_user, - proxy_password=p_password, - timeout=10, - num_retries=0, - ) - for cmd in commands: - output = client.run_command(cmd) - client.join(output) - if output[ip[0]].exit_code: - return -1, "VNFR {} command '{}' returns error: '{}'".format( - ip[0], cmd, "\n".join(output[ip[0]].stderr) - ) - else: - return 1, "VNFR {} command '{}' successful".format(ip[0], cmd) - except ( - ssh2Exception.ChannelFailure, - ssh2Exception.SocketDisconnectError, - ssh2Exception.SocketTimeout, - ssh2Exception.SocketRecvError, - ) as e: - return 0, "Timeout accessing the VNFR {}: {}".format(ip[0], str(e)) - except Exception as e: - return -1, "ERROR checking the VNFR {}: {}".format(ip[0], str(e)) - - def additional_operations(self, engine, test_osm, manual_check): - pass - - def run(self, engine, test_osm, manual_check, test_params=None): - engine.set_test_name(self.test_name) - engine.get_autorization() - nsname = os.environ.get("OSMNBITEST_NS_NAME", "OSMNBITEST") - if test_params: - if "vnfd-files" in test_params: - self.vnfd_filenames = test_params["vnfd-files"].split(",") - if "nsd-file" in test_params: - self.nsd_filename = test_params["nsd-file"] - if test_params.get("ns-name"): - nsname = test_params["ns-name"] - self.create_descriptors(engine) - - # create real VIM if not exist - self.vim_id = engine.get_create_vim(test_osm) - ns_data = { - "nsDescription": "default description", - "nsName": nsname, - "nsdId": self.nsd_id, - "vimAccountId": self.vim_id, - } - if self.ns_params: - ns_data.update(self.ns_params) - if test_params and test_params.get("ns-config"): - if isinstance(test_params["ns-config"], str): - ns_data.update(yaml.safe_load(test_params["ns-config"])) - else: - ns_data.update(test_params["ns-config"]) - self.instantiate(engine, ns_data) - - if manual_check: - input( - "NS has been deployed. Perform manual check and press enter to resume" - ) - if test_osm and self.commands: - self.test_ns(engine, test_osm) - self.additional_operations(engine, test_osm, manual_check) - self.terminate(engine) - self.delete_descriptors(engine) - - def get_first_ip(self, ip_string): - # When using a floating IP, the vnfr_data['ip-address'] contains a semicolon-separated list of IP:s. - first_ip = ip_string.split(";")[0] if ip_string else "" - return first_ip - - def get_vnfr_ip(self, engine, vnfr_index_wanted): - # If the IP address list has been obtained before, it has been stored in 'vnfr_ip_list' - ip = self.vnfr_ip_list.get(vnfr_index_wanted, "") - if ip: - return self.get_first_ip(ip) - r = engine.test( - "Get VNFR to get IP_ADDRESS", - "GET", - "/nslcm/v1/vnfrs?member-vnf-index-ref={}&nsr-id-ref={}".format( - vnfr_index_wanted, self.ns_id - ), - headers_json, - None, - 200, - r_header_json, - "json", - ) - if not r: - return "" - vnfr_data = r.json() - if not (vnfr_data and vnfr_data[0]): - return "" - # Store the IP (or list of IPs) in 'vnfr_ip_list' - ip_list = vnfr_data[0].get("ip-address", "") - if ip_list: - self.vnfr_ip_list[vnfr_index_wanted] = ip_list - ip = self.get_first_ip(ip_list) - return ip - - -class TestDeployHackfestCirros(TestDeploy): - description = "Load and deploy Hackfest cirros_2vnf_ns example" - - def __init__(self): - super().__init__() - self.test_name = "CIRROS" - self.vnfd_filenames = ("cirros_vnf.tar.gz",) - self.nsd_filename = "cirros_2vnf_ns.tar.gz" - self.commands = { - "1": [ - "ls -lrt", - ], - "2": [ - "ls -lrt", - ], - } - self.users = {"1": "cirros", "2": "cirros"} - self.passwords = {"1": "cubswin:)", "2": "cubswin:)"} - - def terminate(self, engine): - # Make a delete in one step, overriding the normal two step of TestDeploy that launched terminate and delete - if test_osm: - engine.test( - "Terminate and delete NS in one step", - "DELETE", - "/nslcm/v1/ns_instances_content/{}".format(self.ns_id), - headers_yaml, - None, - 202, - None, - "yaml", - ) - - engine.wait_until_delete( - "/nslcm/v1/ns_instances/{}".format(self.ns_id), timeout_deploy - ) - else: - engine.test( - "Delete NS with FORCE", - "DELETE", - "/nslcm/v1/ns_instances/{}?FORCE=True".format(self.ns_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - # check all it is deleted - engine.test( - "Check NS is deleted", - "GET", - "/nslcm/v1/ns_instances/{}".format(self.ns_id), - headers_yaml, - None, - 404, - None, - "yaml", - ) - r = engine.test( - "Check NSLCMOPs are deleted", - "GET", - "/nslcm/v1/ns_lcm_op_occs?nsInstanceId={}".format(self.ns_id), - headers_json, - None, - 200, - None, - "json", - ) - if not r: - return - nslcmops = r.json() - if not isinstance(nslcmops, list) or nslcmops: - raise TestException( - "NS {} deleted but with ns_lcm_op_occ active: {}".format( - self.ns_id, nslcmops - ) - ) - - -class TestDeployHackfest1(TestDeploy): - description = "Load and deploy Hackfest_1_vnfd example" - - def __init__(self): - super().__init__() - self.test_name = "HACKFEST1-" - self.vnfd_filenames = ("hackfest_1_vnfd.tar.gz",) - self.nsd_filename = "hackfest_1_nsd.tar.gz" - # self.commands = {'1': ['ls -lrt', ], '2': ['ls -lrt', ]} - # self.users = {'1': "cirros", '2': "cirros"} - # self.passwords = {'1': "cubswin:)", '2': "cubswin:)"} - - -class TestDeployHackfestCirrosScaling(TestDeploy): - description = ( - "Load and deploy Hackfest cirros_2vnf_ns example with scaling modifications" - ) - - def __init__(self): - super().__init__() - self.test_name = "CIRROS-SCALE" - self.vnfd_filenames = ("cirros_vnf.tar.gz",) - self.nsd_filename = "cirros_2vnf_ns.tar.gz" - # Modify VNFD to add scaling and count=2 - self.descriptor_edit = { - "vnfd0": { - "vdu": {"$id: 'cirros_vnfd-VM'": {"count": 2}}, - "scaling-group-descriptor": [ - { - "name": "scale_cirros", - "max-instance-count": 2, - "vdu": [{"vdu-id-ref": "cirros_vnfd-VM", "count": 2}], - } - ], - } - } - - def additional_operations(self, engine, test_osm, manual_check): - if not test_osm: - return - # 2 perform scale out twice - payload = ( - "{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_OUT, scaleByStepData: " - '{scaling-group-descriptor: scale_cirros, member-vnf-index: "1"}}}' - ) - for i in range(0, 2): - engine.test( - "Execute scale action over NS", - "POST", - "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), - headers_yaml, - payload, - (201, 202), - r_headers_yaml_location_nslcmop, - "yaml", - ) - nslcmop2_scale_out = engine.last_id - engine.wait_operation_ready("ns", nslcmop2_scale_out, timeout_deploy) - if manual_check: - input("NS scale out done. Check that two more vdus are there") - # TODO check automatic - - # 2 perform scale in - payload = ( - "{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_IN, scaleByStepData: " - '{scaling-group-descriptor: scale_cirros, member-vnf-index: "1"}}}' - ) - for i in range(0, 2): - engine.test( - "Execute scale IN action over NS", - "POST", - "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), - headers_yaml, - payload, - (201, 202), - r_headers_yaml_location_nslcmop, - "yaml", - ) - nslcmop2_scale_in = engine.last_id - engine.wait_operation_ready("ns", nslcmop2_scale_in, timeout_deploy) - if manual_check: - input("NS scale in done. Check that two less vdus are there") - # TODO check automatic - - # perform scale in that must fail as reached limit - engine.test( - "Execute scale IN out of limit action over NS", - "POST", - "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), - headers_yaml, - payload, - (201, 202), - r_headers_yaml_location_nslcmop, - "yaml", - ) - nslcmop2_scale_in = engine.last_id - engine.wait_operation_ready( - "ns", nslcmop2_scale_in, timeout_deploy, expected_fail=True - ) - - -class TestDeployIpMac(TestDeploy): - description = "Load and deploy descriptor examples setting mac, ip address at descriptor and instantiate params" - - def __init__(self): - super().__init__() - self.test_name = "SetIpMac" - self.vnfd_filenames = ( - "vnfd_2vdu_set_ip_mac2.yaml", - "vnfd_2vdu_set_ip_mac.yaml", - ) - self.nsd_filename = "scenario_2vdu_set_ip_mac.yaml" - self.descriptor_url = "https://osm.etsi.org/gitweb/?p=osm/RO.git;a=blob_plain;f=test/RO_tests/v3_2vdu_set_ip_mac/" - self.commands = { - "1": [ - "ls -lrt", - ], - "2": [ - "ls -lrt", - ], - } - self.users = {"1": "osm", "2": "osm"} - self.passwords = {"1": "osm4u", "2": "osm4u"} - self.timeout = 360 - - def run(self, engine, test_osm, manual_check, test_params=None): - # super().run(engine, test_osm, manual_check, test_params) - # run again setting IPs with instantiate parameters - instantiation_params = { - "vnf": [ - { - "member-vnf-index": "1", - "internal-vld": [ - { - "name": "internal_vld1", # net_internal - "ip-profile": { - "ip-version": "ipv4", - "subnet-address": "10.9.8.0/24", - "dhcp-params": { - "count": 100, - "start-address": "10.9.8.100", - }, - }, - "internal-connection-point": [ - { - "id-ref": "eth2", - "ip-address": "10.9.8.2", - }, - { - "id-ref": "eth3", - "ip-address": "10.9.8.3", - }, - ], - }, - ], - "vdu": [ - { - "id": "VM1", - "interface": [ - # { - # "name": "iface11", - # "floating-ip-required": True, - # }, - {"name": "iface13", "mac-address": "52:33:44:55:66:13"}, - ], - }, - { - "id": "VM2", - "interface": [ - { - "name": "iface21", - "ip-address": "10.31.31.22", - "mac-address": "52:33:44:55:66:21", - }, - ], - }, - ], - }, - ] - } - - super().run( - engine, - test_osm, - manual_check, - test_params={"ns-config": instantiation_params}, - ) - - -class TestDeployHackfest4(TestDeploy): - description = "Load and deploy Hackfest 4 example." - - def __init__(self): - super().__init__() - self.test_name = "HACKFEST4-" - self.vnfd_filenames = ("hackfest_4_vnfd.tar.gz",) - self.nsd_filename = "hackfest_4_nsd.tar.gz" - self.uses_configuration = True - self.commands = { - "1": [ - "ls -lrt", - ], - "2": [ - "ls -lrt", - ], - } - self.users = {"1": "ubuntu", "2": "ubuntu"} - self.passwords = {"1": "osm4u", "2": "osm4u"} - # Modify VNFD to add scaling - # self.descriptor_edit = { - # "vnfd0": { - # 'vnf-configuration': { - # 'config-primitive': [{ - # 'name': 'touch', - # 'parameter': [{ - # 'name': 'filename', - # 'data-type': 'STRING', - # 'default-value': '/home/ubuntu/touched' - # }] - # }] - # }, - # 'scaling-group-descriptor': [{ - # 'name': 'scale_dataVM', - # 'scaling-policy': [{ - # 'threshold-time': 0, - # 'name': 'auto_cpu_util_above_threshold', - # 'scaling-type': 'automatic', - # 'scaling-criteria': [{ - # 'name': 'cpu_util_above_threshold', - # 'vnf-monitoring-param-ref': 'all_aaa_cpu_util', - # 'scale-out-relational-operation': 'GE', - # 'scale-in-threshold': 15, - # 'scale-out-threshold': 60, - # 'scale-in-relational-operation': 'LE' - # }], - # 'cooldown-time': 60 - # }], - # 'max-instance-count': 10, - # 'scaling-config-action': [ - # {'vnf-config-primitive-name-ref': 'touch', - # 'trigger': 'post-scale-out'}, - # {'vnf-config-primitive-name-ref': 'touch', - # 'trigger': 'pre-scale-in'} - # ], - # 'vdu': [{ - # 'vdu-id-ref': 'dataVM', - # 'count': 1 - # }] - # }] - # } - # } - - -class TestDeployHackfest3Charmed(TestDeploy): - description = "Load and deploy Hackfest 3charmed_ns example" - - def __init__(self): - super().__init__() - self.test_name = "HACKFEST3-" - self.vnfd_filenames = ("hackfest_3charmed_vnfd.tar.gz",) - self.nsd_filename = "hackfest_3charmed_nsd.tar.gz" - self.uses_configuration = True - self.commands = { - "1": ["ls -lrt /home/ubuntu/first-touch"], - "2": ["ls -lrt /home/ubuntu/first-touch"], - } - self.users = {"1": "ubuntu", "2": "ubuntu"} - self.passwords = {"1": "osm4u", "2": "osm4u"} - self.descriptor_edit = { - "vnfd0": yaml.safe_load( - """ - vnf-configuration: - terminate-config-primitive: - - seq: '1' - name: touch - parameter: - - name: filename - value: '/home/ubuntu/last-touch1' - - seq: '3' - name: touch - parameter: - - name: filename - value: '/home/ubuntu/last-touch3' - - seq: '2' - name: touch - parameter: - - name: filename - value: '/home/ubuntu/last-touch2' - """ - ) - } - - def additional_operations(self, engine, test_osm, manual_check): - if not test_osm: - return - # 1 perform action - vnfr_index_selected = "2" - payload = '{member_vnf_index: "2", primitive: touch, primitive_params: { filename: /home/ubuntu/OSMTESTNBI }}' - engine.test( - "Exec service primitive over NS", - "POST", - "/nslcm/v1/ns_instances/{}/action".format(self.ns_id), - headers_yaml, - payload, - (201, 202), - r_headers_yaml_location_nslcmop, - "yaml", - ) - nslcmop2_action = engine.last_id - # Wait until status is Ok - engine.wait_operation_ready("ns", nslcmop2_action, timeout_deploy) - vnfr_ip = self.get_vnfr_ip(engine, vnfr_index_selected) - if manual_check: - input( - "NS service primitive has been executed." - "Check that file /home/ubuntu/OSMTESTNBI is present at {}".format( - vnfr_ip - ) - ) - if test_osm: - commands = { - "1": [""], - "2": [ - "ls -lrt /home/ubuntu/OSMTESTNBI", - ], - } - self.test_ns(engine, test_osm, commands=commands) - - # # 2 perform scale out - # payload = '{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_OUT, scaleByStepData: ' \ - # '{scaling-group-descriptor: scale_dataVM, member-vnf-index: "1"}}}' - # engine.test("Execute scale action over NS", "POST", - # "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), headers_yaml, payload, - # (201, 202), r_headers_yaml_location_nslcmop, "yaml") - # nslcmop2_scale_out = engine.last_id - # engine.wait_operation_ready("ns", nslcmop2_scale_out, timeout_deploy) - # if manual_check: - # input('NS scale out done. Check that file /home/ubuntu/touched is present and new VM is created') - # # TODO check automatic - # - # # 2 perform scale in - # payload = '{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_IN, scaleByStepData: ' \ - # '{scaling-group-descriptor: scale_dataVM, member-vnf-index: "1"}}}' - # engine.test("Execute scale action over NS", "POST", - # "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), headers_yaml, payload, - # (201, 202), r_headers_yaml_location_nslcmop, "yaml") - # nslcmop2_scale_in = engine.last_id - # engine.wait_operation_ready("ns", nslcmop2_scale_in, timeout_deploy) - # if manual_check: - # input('NS scale in done. Check that file /home/ubuntu/touched is updated and new VM is deleted') - # # TODO check automatic - - -class TestDeployHackfest3Charmed2(TestDeployHackfest3Charmed): - description = ( - "Load and deploy Hackfest 3charmed_ns example modified version of descriptors to have dots in " - "ids and member-vnf-index." - ) - - def __init__(self): - super().__init__() - self.test_name = "HACKFEST3v2-" - self.qforce = "?FORCE=True" - self.descriptor_edit = { - "vnfd0": { - "vdu": { - "$[0]": { - "interface": { - "$[0]": {"external-connection-point-ref": "pdu-mgmt"} - } - }, - "$[1]": None, - }, - "vnf-configuration": None, - "connection-point": { - "$[0]": { - "id": "pdu-mgmt", - "name": "pdu-mgmt", - "short-name": "pdu-mgmt", - }, - "$[1]": None, - }, - "mgmt-interface": {"cp": "pdu-mgmt"}, - "description": "A vnf single vdu to be used as PDU", - "id": "vdu-as-pdu", - "internal-vld": { - "$[0]": { - "id": "pdu_internal", - "name": "pdu_internal", - "internal-connection-point": {"$[1]": None}, - "short-name": "pdu_internal", - "type": "ELAN", - } - }, - }, - # Modify NSD accordingly - "nsd": { - "constituent-vnfd": { - "$[0]": {"vnfd-id-ref": "vdu-as-pdu"}, - "$[1]": None, - }, - "description": "A nsd to deploy the vnf to act as as PDU", - "id": "nsd-as-pdu", - "name": "nsd-as-pdu", - "short-name": "nsd-as-pdu", - "vld": { - "$[0]": { - "id": "mgmt_pdu", - "name": "mgmt_pdu", - "short-name": "mgmt_pdu", - "vnfd-connection-point-ref": { - "$[0]": { - "vnfd-connection-point-ref": "pdu-mgmt", - "vnfd-id-ref": "vdu-as-pdu", - }, - "$[1]": None, - }, - "type": "ELAN", - }, - "$[1]": None, - }, - }, - } - - -class TestDeployHackfest3Charmed3(TestDeployHackfest3Charmed): - description = "Load and deploy Hackfest 3charmed_ns example modified version to test scaling and NS parameters" - - def __init__(self): - super().__init__() - self.test_name = "HACKFEST3v3-" - self.commands = { - "1": ["ls -lrt /home/ubuntu/first-touch-1"], - "2": ["ls -lrt /home/ubuntu/first-touch-2"], - } - self.descriptor_edit = { - "vnfd0": yaml.safe_load( - """ - scaling-group-descriptor: - - name: "scale_dataVM" - max-instance-count: 10 - scaling-policy: - - name: "auto_cpu_util_above_threshold" - scaling-type: "automatic" - threshold-time: 0 - cooldown-time: 60 - scaling-criteria: - - name: "cpu_util_above_threshold" - scale-in-threshold: 15 - scale-in-relational-operation: "LE" - scale-out-threshold: 60 - scale-out-relational-operation: "GE" - vnf-monitoring-param-ref: "monitor1" - vdu: - - vdu-id-ref: dataVM - count: 1 - scaling-config-action: - - trigger: post-scale-out - vnf-config-primitive-name-ref: touch - - trigger: pre-scale-in - vnf-config-primitive-name-ref: touch - vdu: - "$id: dataVM": - monitoring-param: - - id: "dataVM_cpu_util" - nfvi-metric: "cpu_utilization" - - monitoring-param: - - id: "monitor1" - name: "monitor1" - aggregation-type: AVERAGE - vdu-monitoring-param: - vdu-ref: "dataVM" - vdu-monitoring-param-ref: "dataVM_cpu_util" - vnf-configuration: - initial-config-primitive: - "$[1]": - parameter: - "$[0]": - value: "" # default-value: /home/ubuntu/first-touch - config-primitive: - "$[0]": - parameter: - "$[0]": - default-value: "" - """, - ) - } - self.ns_params = { - "additionalParamsForVnf": [ - { - "member-vnf-index": "1", - "additionalParams": { - "touch_filename": "/home/ubuntu/first-touch-1", - "touch_filename2": "/home/ubuntu/second-touch-1", - }, - }, - { - "member-vnf-index": "2", - "additionalParams": { - "touch_filename": "/home/ubuntu/first-touch-2", - "touch_filename2": "/home/ubuntu/second-touch-2", - }, - }, - ] - } - - def additional_operations(self, engine, test_osm, manual_check): - super().additional_operations(engine, test_osm, manual_check) - if not test_osm: - return - - # 2 perform scale out - payload = ( - "{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_OUT, scaleByStepData: " - '{scaling-group-descriptor: scale_dataVM, member-vnf-index: "1"}}}' - ) - engine.test( - "Execute scale action over NS", - "POST", - "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), - headers_yaml, - payload, - (201, 202), - r_headers_yaml_location_nslcmop, - "yaml", - ) - nslcmop2_scale_out = engine.last_id - engine.wait_operation_ready("ns", nslcmop2_scale_out, timeout_deploy) - if manual_check: - input( - "NS scale out done. Check that file /home/ubuntu/second-touch-1 is present and new VM is created" - ) - if test_osm: - commands = { - "1": [ - "ls -lrt /home/ubuntu/second-touch-1", - ] - } - self.test_ns(engine, test_osm, commands=commands) - # TODO check automatic connection to scaled VM - - # 2 perform scale in - payload = ( - "{scaleType: SCALE_VNF, scaleVnfData: {scaleVnfType: SCALE_IN, scaleByStepData: " - '{scaling-group-descriptor: scale_dataVM, member-vnf-index: "1"}}}' - ) - engine.test( - "Execute scale action over NS", - "POST", - "/nslcm/v1/ns_instances/{}/scale".format(self.ns_id), - headers_yaml, - payload, - (201, 202), - r_headers_yaml_location_nslcmop, - "yaml", - ) - nslcmop2_scale_in = engine.last_id - engine.wait_operation_ready("ns", nslcmop2_scale_in, timeout_deploy) - if manual_check: - input( - "NS scale in done. Check that file /home/ubuntu/second-touch-1 is updated and new VM is deleted" - ) - # TODO check automatic - - -class TestDeploySimpleCharm(TestDeploy): - description = "Deploy hackfest-4 hackfest_simplecharm example" - - def __init__(self): - super().__init__() - self.test_name = "HACKFEST-SIMPLE" - self.descriptor_url = ( - "https://osm-download.etsi.org/ftp/osm-4.0-four/4th-hackfest/packages/" - ) - self.vnfd_filenames = ("hackfest_simplecharm_vnf.tar.gz",) - self.nsd_filename = "hackfest_simplecharm_ns.tar.gz" - self.uses_configuration = True - self.commands = { - "1": [""], - "2": [ - "ls -lrt /home/ubuntu/first-touch", - ], - } - self.users = {"1": "ubuntu", "2": "ubuntu"} - self.passwords = {"1": "osm4u", "2": "osm4u"} - - -class TestDeploySimpleCharm2(TestDeploySimpleCharm): - description = ( - "Deploy hackfest-4 hackfest_simplecharm example changing naming to contain dots on ids and " - "vnf-member-index" - ) - - def __init__(self): - super().__init__() - self.test_name = "HACKFEST-SIMPLE2-" - self.qforce = "?FORCE=True" - self.descriptor_edit = { - "vnfd0": {"id": "hackfest.simplecharm.vnf"}, - "nsd": { - "id": "hackfest.simplecharm.ns", - "constituent-vnfd": { - "$[0]": { - "vnfd-id-ref": "hackfest.simplecharm.vnf", - "member-vnf-index": "$1", - }, - "$[1]": { - "vnfd-id-ref": "hackfest.simplecharm.vnf", - "member-vnf-index": "$2", - }, - }, - "vld": { - "$[0]": { - "vnfd-connection-point-ref": { - "$[0]": { - "member-vnf-index-ref": "$1", - "vnfd-id-ref": "hackfest.simplecharm.vnf", - }, - "$[1]": { - "member-vnf-index-ref": "$2", - "vnfd-id-ref": "hackfest.simplecharm.vnf", - }, - }, - }, - "$[1]": { - "vnfd-connection-point-ref": { - "$[0]": { - "member-vnf-index-ref": "$1", - "vnfd-id-ref": "hackfest.simplecharm.vnf", - }, - "$[1]": { - "member-vnf-index-ref": "$2", - "vnfd-id-ref": "hackfest.simplecharm.vnf", - }, - }, - }, - }, - }, - } - - -class TestDeploySingleVdu(TestDeployHackfest3Charmed): - description = ( - "Generate a single VDU base on editing Hackfest3Charmed descriptors and deploy" - ) - - def __init__(self): - super().__init__() - self.test_name = "SingleVDU" - self.qforce = "?FORCE=True" - self.descriptor_edit = { - # Modify VNFD to remove one VDU - "vnfd0": { - "vdu": { - "$[0]": { - "interface": { - "$[0]": {"external-connection-point-ref": "pdu-mgmt"} - } - }, - "$[1]": None, - }, - "vnf-configuration": None, - "connection-point": { - "$[0]": { - "id": "pdu-mgmt", - "name": "pdu-mgmt", - "short-name": "pdu-mgmt", - }, - "$[1]": None, - }, - "mgmt-interface": {"cp": "pdu-mgmt"}, - "description": "A vnf single vdu to be used as PDU", - "id": "vdu-as-pdu", - "internal-vld": { - "$[0]": { - "id": "pdu_internal", - "name": "pdu_internal", - "internal-connection-point": {"$[1]": None}, - "short-name": "pdu_internal", - "type": "ELAN", - } - }, - }, - # Modify NSD accordingly - "nsd": { - "constituent-vnfd": { - "$[0]": {"vnfd-id-ref": "vdu-as-pdu"}, - "$[1]": None, - }, - "description": "A nsd to deploy the vnf to act as as PDU", - "id": "nsd-as-pdu", - "name": "nsd-as-pdu", - "short-name": "nsd-as-pdu", - "vld": { - "$[0]": { - "id": "mgmt_pdu", - "name": "mgmt_pdu", - "short-name": "mgmt_pdu", - "vnfd-connection-point-ref": { - "$[0]": { - "vnfd-connection-point-ref": "pdu-mgmt", - "vnfd-id-ref": "vdu-as-pdu", - }, - "$[1]": None, - }, - "type": "ELAN", - }, - "$[1]": None, - }, - }, - } - - -class TestDeployHnfd(TestDeployHackfest3Charmed): - description = ( - "Generate a HNFD base on editing Hackfest3Charmed descriptors and deploy" - ) - - def __init__(self): - super().__init__() - self.test_name = "HNFD" - self.pduDeploy = TestDeploySingleVdu() - self.pdu_interface_0 = {} - self.pdu_interface_1 = {} - - self.pdu_id = None - # self.vnf_to_pdu = """ - # vdu: - # "$[0]": - # pdu-type: PDU-TYPE-1 - # interface: - # "$[0]": - # name: mgmt-iface - # "$[1]": - # name: pdu-iface-internal - # id: hfn1 - # description: HFND, one PDU + One VDU - # name: hfn1 - # short-name: hfn1 - # - # """ - - self.pdu_descriptor = { - "name": "my-PDU", - "type": "PDU-TYPE-1", - "vim_accounts": "to-override", - "interfaces": [ - { - "name": "mgmt-iface", - "mgmt": True, - "type": "overlay", - "ip-address": "to override", - "mac-address": "mac_address", - "vim-network-name": "mgmt", - }, - { - "name": "pdu-iface-internal", - "mgmt": False, - "type": "overlay", - "ip-address": "to override", - "mac-address": "mac_address", - "vim-network-name": "pdu_internal", # OSMNBITEST-PDU-pdu_internal - }, - ], - } - self.vnfd_filenames = ( - "hackfest_3charmed_vnfd.tar.gz", - "hackfest_3charmed_vnfd.tar.gz", - ) - - self.descriptor_edit = { - "vnfd0": { - "id": "hfnd1", - "name": "hfn1", - "short-name": "hfn1", - "vdu": { - "$[0]": { - "pdu-type": "PDU-TYPE-1", - "interface": { - "$[0]": {"name": "mgmt-iface"}, - "$[1]": {"name": "pdu-iface-internal"}, - }, - } - }, - }, - "nsd": { - "constituent-vnfd": {"$[1]": {"vnfd-id-ref": "hfnd1"}}, - "vld": { - "$[0]": { - "vnfd-connection-point-ref": {"$[1]": {"vnfd-id-ref": "hfnd1"}} - }, - "$[1]": { - "vnfd-connection-point-ref": {"$[1]": {"vnfd-id-ref": "hfnd1"}} - }, - }, - }, - } - - def create_descriptors(self, engine): - super().create_descriptors(engine) - - # Create PDU - self.pdu_descriptor["interfaces"][0].update(self.pdu_interface_0) - self.pdu_descriptor["interfaces"][1].update(self.pdu_interface_1) - self.pdu_descriptor["vim_accounts"] = [self.vim_id] - # TODO get vim-network-name from vnfr.vld.name - self.pdu_descriptor["interfaces"][1]["vim-network-name"] = "{}-{}-{}".format( - os.environ.get("OSMNBITEST_NS_NAME", "OSMNBITEST"), - "PDU", - self.pdu_descriptor["interfaces"][1]["vim-network-name"], - ) - engine.test( - "Onboard PDU descriptor", - "POST", - "/pdu/v1/pdu_descriptors", - { - "Location": "/pdu/v1/pdu_descriptors/", - "Content-Type": "application/yaml", - }, - self.pdu_descriptor, - 201, - r_header_yaml, - "yaml", - ) - self.pdu_id = engine.last_id - - def run(self, engine, test_osm, manual_check, test_params=None): - engine.get_autorization() - engine.set_test_name(self.test_name) - nsname = os.environ.get("OSMNBITEST_NS_NAME", "OSMNBITEST") - - # create real VIM if not exist - self.vim_id = engine.get_create_vim(test_osm) - # instantiate PDU - self.pduDeploy.create_descriptors(engine) - self.pduDeploy.instantiate( - engine, - { - "nsDescription": "to be used as PDU", - "nsName": nsname + "-PDU", - "nsdId": self.pduDeploy.nsd_id, - "vimAccountId": self.vim_id, - }, - ) - if manual_check: - input( - "VNF to be used as PDU has been deployed. Perform manual check and press enter to resume" - ) - if test_osm: - self.pduDeploy.test_ns(engine, test_osm) - - if test_osm: - r = engine.test( - "Get VNFR to obtain IP_ADDRESS", - "GET", - "/nslcm/v1/vnfrs?nsr-id-ref={}".format(self.pduDeploy.ns_id), - headers_json, - None, - 200, - r_header_json, - "json", - ) - if not r: - return - vnfr_data = r.json() - # print(vnfr_data) - - self.pdu_interface_0["ip-address"] = vnfr_data[0]["vdur"][0]["interfaces"][ - 0 - ].get("ip-address") - self.pdu_interface_1["ip-address"] = vnfr_data[0]["vdur"][0]["interfaces"][ - 1 - ].get("ip-address") - self.pdu_interface_0["mac-address"] = vnfr_data[0]["vdur"][0]["interfaces"][ - 0 - ].get("mac-address") - self.pdu_interface_1["mac-address"] = vnfr_data[0]["vdur"][0]["interfaces"][ - 1 - ].get("mac-address") - if not self.pdu_interface_0["ip-address"]: - raise TestException("Vnfr has not managment ip address") - else: - self.pdu_interface_0["ip-address"] = "192.168.10.10" - self.pdu_interface_1["ip-address"] = "192.168.11.10" - self.pdu_interface_0["mac-address"] = "52:33:44:55:66:13" - self.pdu_interface_1["mac-address"] = "52:33:44:55:66:14" - - self.create_descriptors(engine) - - ns_data = { - "nsDescription": "default description", - "nsName": nsname, - "nsdId": self.nsd_id, - "vimAccountId": self.vim_id, - } - if test_params and test_params.get("ns-config"): - if isinstance(test_params["ns-config"], str): - ns_data.update(yaml.safe_load(test_params["ns-config"])) - else: - ns_data.update(test_params["ns-config"]) - - self.instantiate(engine, ns_data) - if manual_check: - input( - "NS has been deployed. Perform manual check and press enter to resume" - ) - if test_osm: - self.test_ns(engine, test_osm) - self.additional_operations(engine, test_osm, manual_check) - self.terminate(engine) - self.pduDeploy.terminate(engine) - self.delete_descriptors(engine) - self.pduDeploy.delete_descriptors(engine) - - def delete_descriptors(self, engine): - super().delete_descriptors(engine) - # delete pdu - engine.test( - "Delete PDU SOL005", - "DELETE", - "/pdu/v1/pdu_descriptors/{}".format(self.pdu_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - -class TestDescriptors: - description = "Test VNFD, NSD, PDU descriptors CRUD and dependencies" - vnfd_empty = """vnfd:vnfd-catalog: - vnfd: - - name: prova - short-name: prova - id: prova - """ - vnfd_prova = """vnfd:vnfd-catalog: - vnfd: - - connection-point: - - name: cp_0h8m - type: VPORT - id: prova - name: prova - short-name: prova - vdu: - - id: vdu_z4bm - image: ubuntu - interface: - - external-connection-point-ref: cp_0h8m - name: eth0 - virtual-interface: - type: VIRTIO - name: vdu_z4bm - version: '1.0' - """ - - def __init__(self): - self.vnfd_filename = "hackfest_3charmed_vnfd.tar.gz" - self.nsd_filename = "hackfest_3charmed_nsd.tar.gz" - self.descriptor_url = ( - "https://osm-download.etsi.org/ftp/osm-3.0-three/2nd-hackfest/packages/" - ) - self.vnfd_id = None - self.nsd_id = None - - def run(self, engine, test_osm, manual_check, test_params=None): - engine.set_test_name("Descriptors") - engine.get_autorization() - temp_dir = os.path.dirname(os.path.abspath(__file__)) + "/temp/" - if not os.path.exists(temp_dir): - os.makedirs(temp_dir) - - # download files - for filename in (self.vnfd_filename, self.nsd_filename): - filename_path = temp_dir + filename - if not os.path.exists(filename_path): - with open(filename_path, "wb") as file: - response = requests.get(self.descriptor_url + filename) - if response.status_code >= 300: - raise TestException( - "Error downloading descriptor from '{}': {}".format( - self.descriptor_url + filename, response.status_code - ) - ) - file.write(response.content) - - vnfd_filename_path = temp_dir + self.vnfd_filename - nsd_filename_path = temp_dir + self.nsd_filename - - engine.test( - "Onboard empty VNFD in one step", - "POST", - "/vnfpkgm/v1/vnf_packages_content", - headers_yaml, - self.vnfd_empty, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - self.vnfd_id = engine.last_id - - # test bug 605 - engine.test( - "Upload invalid VNFD ", - "PUT", - "/vnfpkgm/v1/vnf_packages/{}/package_content".format(self.vnfd_id), - headers_yaml, - self.vnfd_prova, - 422, - r_header_yaml, - "yaml", - ) - - engine.test( - "Upload VNFD {}".format(self.vnfd_filename), - "PUT", - "/vnfpkgm/v1/vnf_packages/{}/package_content".format(self.vnfd_id), - headers_zip_yaml, - "@b" + vnfd_filename_path, - 204, - None, - 0, - ) - - queries = [ - "mgmt-interface.cp=mgmt", - "vdu.0.interface.0.external-connection-point-ref=mgmt", - "vdu.0.interface.1.internal-connection-point-ref=internal", - "internal-vld.0.internal-connection-point.0.id-ref=internal", - # Detection of duplicated VLD names in VNF Descriptors - # URL: internal-vld=[ - # {id: internal1, name: internal, type:ELAN, - # internal-connection-point: [{id-ref: mgmtVM-internal}, {id-ref: dataVM-internal}]}, - # {id: internal2, name: internal, type:ELAN, - # internal-connection-point: [{id-ref: mgmtVM-internal}, {id-ref: dataVM-internal}]} - # ] - "internal-vld=%5B%7Bid%3A%20internal1%2C%20name%3A%20internal%2C%20type%3A%20ELAN%2C%20" - "internal-connection-point%3A%20%5B%7Bid-ref%3A%20mgmtVM-internal%7D%2C%20%7Bid-ref%3A%20" - "dataVM-internal%7D%5D%7D%2C%20%7Bid%3A%20internal2%2C%20name%3A%20internal%2C%20type%3A%20" - "ELAN%2C%20internal-connection-point%3A%20%5B%7Bid-ref%3A%20mgmtVM-internal%7D%2C%20%7B" - "id-ref%3A%20dataVM-internal%7D%5D%7D%5D", - ] - for query in queries: - engine.test( - "Upload invalid VNFD ", - "PUT", - "/vnfpkgm/v1/vnf_packages/{}/package_content?{}".format( - self.vnfd_id, query - ), - headers_zip_yaml, - "@b" + vnfd_filename_path, - 422, - r_header_yaml, - "yaml", - ) - - # test bug 605 - engine.test( - "Upload invalid VNFD ", - "PUT", - "/vnfpkgm/v1/vnf_packages/{}/package_content".format(self.vnfd_id), - headers_yaml, - self.vnfd_prova, - 422, - r_header_yaml, - "yaml", - ) - - # get vnfd descriptor - engine.test( - "Get VNFD descriptor", - "GET", - "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_id), - headers_yaml, - None, - 200, - r_header_yaml, - "yaml", - ) - - # get vnfd file descriptor - engine.test( - "Get VNFD file descriptor", - "GET", - "/vnfpkgm/v1/vnf_packages/{}/vnfd".format(self.vnfd_id), - headers_text, - None, - 200, - r_header_text, - "text", - temp_dir + "vnfd-yaml", - ) - # TODO compare files: diff vnfd-yaml hackfest_3charmed_vnfd/hackfest_3charmed_vnfd.yaml - - # get vnfd zip file package - engine.test( - "Get VNFD zip package", - "GET", - "/vnfpkgm/v1/vnf_packages/{}/package_content".format(self.vnfd_id), - headers_zip, - None, - 200, - r_header_zip, - "zip", - temp_dir + "vnfd-zip", - ) - # TODO compare files: diff vnfd-zip hackfest_3charmed_vnfd.tar.gz - - # get vnfd artifact - engine.test( - "Get VNFD artifact package", - "GET", - "/vnfpkgm/v1/vnf_packages/{}/artifacts/icons/osm.png".format(self.vnfd_id), - headers_zip, - None, - 200, - r_header_octect, - "octet-string", - temp_dir + "vnfd-icon", - ) - # TODO compare files: diff vnfd-icon hackfest_3charmed_vnfd/icons/osm.png - - # nsd CREATE AND UPLOAD in one step: - engine.test( - "Onboard NSD in one step", - "POST", - "/nsd/v1/ns_descriptors_content", - headers_zip_yaml, - "@b" + nsd_filename_path, - 201, - r_headers_yaml_location_nsd, - "yaml", - ) - self.nsd_id = engine.last_id - - queries = ["vld.0.vnfd-connection-point-ref.0.vnfd-id-ref=hf"] - for query in queries: - engine.test( - "Upload invalid NSD ", - "PUT", - "/nsd/v1/ns_descriptors/{}/nsd_content?{}".format(self.nsd_id, query), - headers_zip_yaml, - "@b" + nsd_filename_path, - 422, - r_header_yaml, - "yaml", - ) - - # get nsd descriptor - engine.test( - "Get NSD descriptor", - "GET", - "/nsd/v1/ns_descriptors/{}".format(self.nsd_id), - headers_yaml, - None, - 200, - r_header_yaml, - "yaml", - ) - - # get nsd file descriptor - engine.test( - "Get NSD file descriptor", - "GET", - "/nsd/v1/ns_descriptors/{}/nsd".format(self.nsd_id), - headers_text, - None, - 200, - r_header_text, - "text", - temp_dir + "nsd-yaml", - ) - # TODO compare files: diff nsd-yaml hackfest_3charmed_nsd/hackfest_3charmed_nsd.yaml - - # get nsd zip file package - engine.test( - "Get NSD zip package", - "GET", - "/nsd/v1/ns_descriptors/{}/nsd_content".format(self.nsd_id), - headers_zip, - None, - 200, - r_header_zip, - "zip", - temp_dir + "nsd-zip", - ) - # TODO compare files: diff nsd-zip hackfest_3charmed_nsd.tar.gz - - # get nsd artifact - engine.test( - "Get NSD artifact package", - "GET", - "/nsd/v1/ns_descriptors/{}/artifacts/icons/osm.png".format(self.nsd_id), - headers_zip, - None, - 200, - r_header_octect, - "octet-string", - temp_dir + "nsd-icon", - ) - # TODO compare files: diff nsd-icon hackfest_3charmed_nsd/icons/osm.png - - # vnfd DELETE - test_rest.test( - "Delete VNFD conflict", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_id), - headers_yaml, - None, - 409, - None, - None, - ) - - test_rest.test( - "Delete VNFD force", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}?FORCE=TRUE".format(self.vnfd_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - # nsd DELETE - test_rest.test( - "Delete NSD", - "DELETE", - "/nsd/v1/ns_descriptors/{}".format(self.nsd_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - -class TestNetSliceTemplates: - description = "Upload a NST to OSM" - - def __init__(self): - self.vnfd_filename = "@./slice_shared/vnfd/slice_shared_vnfd.yaml" - self.vnfd_filename_middle = "@./slice_shared/vnfd/slice_shared_middle_vnfd.yaml" - self.nsd_filename = "@./slice_shared/nsd/slice_shared_nsd.yaml" - self.nsd_filename_middle = "@./slice_shared/nsd/slice_shared_middle_nsd.yaml" - self.nst_filenames = "@./slice_shared/slice_shared_nstd.yaml" - - def run(self, engine, test_osm, manual_check, test_params=None): - # nst CREATE - engine.set_test_name("NST step ") - engine.get_autorization() - temp_dir = os.path.dirname(os.path.abspath(__file__)) + "/temp/" - if not os.path.exists(temp_dir): - os.makedirs(temp_dir) - - # Onboard VNFDs - engine.test( - "Onboard edge VNFD", - "POST", - "/vnfpkgm/v1/vnf_packages_content", - headers_yaml, - self.vnfd_filename, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - self.vnfd_edge_id = engine.last_id - - engine.test( - "Onboard middle VNFD", - "POST", - "/vnfpkgm/v1/vnf_packages_content", - headers_yaml, - self.vnfd_filename_middle, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - self.vnfd_middle_id = engine.last_id - - # Onboard NSDs - engine.test( - "Onboard NSD edge", - "POST", - "/nsd/v1/ns_descriptors_content", - headers_yaml, - self.nsd_filename, - 201, - r_headers_yaml_location_nsd, - "yaml", - ) - self.nsd_edge_id = engine.last_id - - engine.test( - "Onboard NSD middle", - "POST", - "/nsd/v1/ns_descriptors_content", - headers_yaml, - self.nsd_filename_middle, - 201, - r_headers_yaml_location_nsd, - "yaml", - ) - self.nsd_middle_id = engine.last_id - - # Onboard NST - engine.test( - "Onboard NST", - "POST", - "/nst/v1/netslice_templates_content", - headers_yaml, - self.nst_filenames, - 201, - r_headers_yaml_location_nst, - "yaml", - ) - nst_id = engine.last_id - - # nstd SHOW OSM format - engine.test( - "Show NSTD OSM format", - "GET", - "/nst/v1/netslice_templates/{}".format(nst_id), - headers_json, - None, - 200, - r_header_json, - "json", - ) - - # nstd DELETE - engine.test( - "Delete NSTD", - "DELETE", - "/nst/v1/netslice_templates/{}".format(nst_id), - headers_json, - None, - 204, - None, - 0, - ) - - # NSDs DELETE - test_rest.test( - "Delete NSD middle", - "DELETE", - "/nsd/v1/ns_descriptors/{}".format(self.nsd_middle_id), - headers_json, - None, - 204, - None, - 0, - ) - - test_rest.test( - "Delete NSD edge", - "DELETE", - "/nsd/v1/ns_descriptors/{}".format(self.nsd_edge_id), - headers_json, - None, - 204, - None, - 0, - ) - - # VNFDs DELETE - test_rest.test( - "Delete VNFD edge", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_edge_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - test_rest.test( - "Delete VNFD middle", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_middle_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - -class TestNetSliceInstances: - """ - Test procedure: - 1. Populate databases with VNFD, NSD, NST with the following scenario - +-----------------management-----------------+ - | | | - +--+---+ +----+----+ +---+--+ - | | | | | | - | edge +---data1----+ middle +---data2-----+ edge | - | | | | | | - +------+ +---------+ +------+ - shared-nss - 2. Create NSI-1 - 3. Instantiate NSI-1 - 4. Create NSI-2 - 5. Instantiate NSI-2 - Manual check - Are 2 slices instantiated correctly? - NSI-1 3 nss (2 nss-edges + 1 nss-middle) - NSI-2 2 nss (2 nss-edge sharing nss-middle) - 6. Terminate NSI-1 - 7. Delete NSI-1 - Manual check - Is slice NSI-1 deleted correctly? - NSI-2 with 2 nss-edge + 1 nss-middle (The one from NSI-1) - 8. Create NSI-3 - 9. Instantiate NSI-3 - Manual check - Is slice NSI-3 instantiated correctly? - NSI-3 reuse nss-middle. NSI-3 only create 2 nss-edge - 10. Delete NSI-2 - 11. Terminate NSI-2 - 12. Delete NSI-3 - 13. Terminate NSI-3 - Manual check - All cleaned correctly? - NSI-2 and NSI-3 were terminated and deleted - 14. Cleanup database - """ - - description = "Upload a NST to OSM" - - def __init__(self): - self.vim_id = None - self.vnfd_filename = "@./slice_shared/vnfd/slice_shared_vnfd.yaml" - self.vnfd_filename_middle = "@./slice_shared/vnfd/slice_shared_middle_vnfd.yaml" - self.nsd_filename = "@./slice_shared/nsd/slice_shared_nsd.yaml" - self.nsd_filename_middle = "@./slice_shared/nsd/slice_shared_middle_nsd.yaml" - self.nst_filenames = "@./slice_shared/slice_shared_nstd.yaml" - - def create_slice(self, engine, nsi_data, name): - ns_data_text = yaml.safe_dump(nsi_data, default_flow_style=True, width=256) - r = engine.test( - name, - "POST", - "/nsilcm/v1/netslice_instances", - headers_yaml, - ns_data_text, - (201, 202), - { - "Location": "nsilcm/v1/netslice_instances/", - "Content-Type": "application/yaml", - }, - "yaml", - ) - return r - - def instantiate_slice(self, engine, nsi_data, nsi_id, name): - ns_data_text = yaml.safe_dump(nsi_data, default_flow_style=True, width=256) - engine.test( - name, - "POST", - "/nsilcm/v1/netslice_instances/{}/instantiate".format(nsi_id), - headers_yaml, - ns_data_text, - (201, 202), - r_headers_yaml_location_nsilcmop, - "yaml", - ) - - def terminate_slice(self, engine, nsi_id, name): - engine.test( - name, - "POST", - "/nsilcm/v1/netslice_instances/{}/terminate".format(nsi_id), - headers_yaml, - None, - (201, 202), - r_headers_yaml_location_nsilcmop, - "yaml", - ) - - def delete_slice(self, engine, nsi_id, name): - engine.test( - name, - "DELETE", - "/nsilcm/v1/netslice_instances/{}".format(nsi_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - def run(self, engine, test_osm, manual_check, test_params=None): - # nst CREATE - engine.set_test_name("NSI") - engine.get_autorization() - - # Onboard VNFDs - engine.test( - "Onboard edge VNFD", - "POST", - "/vnfpkgm/v1/vnf_packages_content", - headers_yaml, - self.vnfd_filename, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - self.vnfd_edge_id = engine.last_id - - engine.test( - "Onboard middle VNFD", - "POST", - "/vnfpkgm/v1/vnf_packages_content", - headers_yaml, - self.vnfd_filename_middle, - 201, - r_headers_yaml_location_vnfd, - "yaml", - ) - self.vnfd_middle_id = engine.last_id - - # Onboard NSDs - engine.test( - "Onboard NSD edge", - "POST", - "/nsd/v1/ns_descriptors_content", - headers_yaml, - self.nsd_filename, - 201, - r_headers_yaml_location_nsd, - "yaml", - ) - self.nsd_edge_id = engine.last_id - - engine.test( - "Onboard NSD middle", - "POST", - "/nsd/v1/ns_descriptors_content", - headers_yaml, - self.nsd_filename_middle, - 201, - r_headers_yaml_location_nsd, - "yaml", - ) - self.nsd_middle_id = engine.last_id - - # Onboard NST - engine.test( - "Onboard NST", - "POST", - "/nst/v1/netslice_templates_content", - headers_yaml, - self.nst_filenames, - 201, - r_headers_yaml_location_nst, - "yaml", - ) - nst_id = engine.last_id - - self.vim_id = engine.get_create_vim(test_osm) - - # CREATE NSI-1 - ns_data = { - "nsiName": "Deploy-NSI-1", - "vimAccountId": self.vim_id, - "nstId": nst_id, - "nsiDescription": "default", - } - r = self.create_slice(engine, ns_data, "Create NSI-1 step 1") - if not r: - return - self.nsi_id1 = engine.last_id - - # INSTANTIATE NSI-1 - self.instantiate_slice( - engine, ns_data, self.nsi_id1, "Instantiate NSI-1 step 2" - ) - nsilcmop_id1 = engine.last_id - - # Waiting for NSI-1 - if test_osm: - engine.wait_operation_ready("nsi", nsilcmop_id1, timeout_deploy) - - # CREATE NSI-2 - ns_data = { - "nsiName": "Deploy-NSI-2", - "vimAccountId": self.vim_id, - "nstId": nst_id, - "nsiDescription": "default", - } - r = self.create_slice(engine, ns_data, "Create NSI-2 step 1") - if not r: - return - self.nsi_id2 = engine.last_id - - # INSTANTIATE NSI-2 - self.instantiate_slice( - engine, ns_data, self.nsi_id2, "Instantiate NSI-2 step 2" - ) - nsilcmop_id2 = engine.last_id - - # Waiting for NSI-2 - if test_osm: - engine.wait_operation_ready("nsi", nsilcmop_id2, timeout_deploy) - - if manual_check: - input( - "NSI-1 AND NSI-2 has been deployed. Perform manual check and press enter to resume" - ) - - # TERMINATE NSI-1 - if test_osm: - self.terminate_slice(engine, self.nsi_id1, "Terminate NSI-1") - nsilcmop1_id = engine.last_id - - # Wait terminate NSI-1 - engine.wait_operation_ready("nsi", nsilcmop1_id, timeout_deploy) - - # DELETE NSI-1 - self.delete_slice(engine, self.nsi_id1, "Delete NS") - - if manual_check: - input( - "NSI-1 has been deleted. Perform manual check and press enter to resume" - ) - - # CREATE NSI-3 - ns_data = { - "nsiName": "Deploy-NSI-3", - "vimAccountId": self.vim_id, - "nstId": nst_id, - "nsiDescription": "default", - } - r = self.create_slice(engine, ns_data, "Create NSI-3 step 1") - - if not r: - return - self.nsi_id3 = engine.last_id - - # INSTANTIATE NSI-3 - self.instantiate_slice( - engine, ns_data, self.nsi_id3, "Instantiate NSI-3 step 2" - ) - nsilcmop_id3 = engine.last_id - - # Wait Instantiate NSI-3 - if test_osm: - engine.wait_operation_ready("nsi", nsilcmop_id3, timeout_deploy) - - if manual_check: - input( - "NSI-3 has been deployed. Perform manual check and press enter to resume" - ) - - # TERMINATE NSI-2 - if test_osm: - self.terminate_slice(engine, self.nsi_id2, "Terminate NSI-2") - nsilcmop2_id = engine.last_id - - # Wait terminate NSI-2 - engine.wait_operation_ready("nsi", nsilcmop2_id, timeout_deploy) - - # DELETE NSI-2 - self.delete_slice(engine, self.nsi_id2, "DELETE NSI-2") - - # TERMINATE NSI-3 - if test_osm: - self.terminate_slice(engine, self.nsi_id3, "Terminate NSI-3") - nsilcmop3_id = engine.last_id - - # Wait terminate NSI-3 - engine.wait_operation_ready("nsi", nsilcmop3_id, timeout_deploy) - - # DELETE NSI-3 - self.delete_slice(engine, self.nsi_id3, "DELETE NSI-3") - - if manual_check: - input( - "NSI-2 and NSI-3 has been deleted. Perform manual check and press enter to resume" - ) - - # nstd DELETE - engine.test( - "Delete NSTD", - "DELETE", - "/nst/v1/netslice_templates/{}".format(nst_id), - headers_json, - None, - 204, - None, - 0, - ) - - # NSDs DELETE - test_rest.test( - "Delete NSD middle", - "DELETE", - "/nsd/v1/ns_descriptors/{}".format(self.nsd_middle_id), - headers_json, - None, - 204, - None, - 0, - ) - - test_rest.test( - "Delete NSD edge", - "DELETE", - "/nsd/v1/ns_descriptors/{}".format(self.nsd_edge_id), - headers_json, - None, - 204, - None, - 0, - ) - - # VNFDs DELETE - test_rest.test( - "Delete VNFD edge", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_edge_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - test_rest.test( - "Delete VNFD middle", - "DELETE", - "/vnfpkgm/v1/vnf_packages/{}".format(self.vnfd_middle_id), - headers_yaml, - None, - 204, - None, - 0, - ) - - -class TestAuthentication: - description = "Test Authentication" - - @staticmethod - def run(engine, test_osm, manual_check, test_params=None): - engine.set_test_name("Authentication") - # backend = test_params.get("backend") if test_params else None # UNUSED - - admin_project_id = test_project_id = None - project_admin_role_id = project_user_role_id = None - test_user_id = empty_user_id = None - default_role_id = empty_role_id = token_role_id = None - - engine.get_autorization() - - # GET - engine.test( - "Get tokens", - "GET", - "/admin/v1/tokens", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - engine.test( - "Get projects", - "GET", - "/admin/v1/projects", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - engine.test( - "Get users", - "GET", - "/admin/v1/users", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - engine.test( - "Get roles", - "GET", - "/admin/v1/roles", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - res = engine.test( - "Get admin project", - "GET", - "/admin/v1/projects?name=admin", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - admin_project_id = res.json()[0]["_id"] if res else None - res = engine.test( - "Get project admin role", - "GET", - "/admin/v1/roles?name=project_admin", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - project_admin_role_id = res.json()[0]["_id"] if res else None - res = engine.test( - "Get project user role", - "GET", - "/admin/v1/roles?name=project_user", - headers_json, - {}, - (200), - {"Content-Type": "application/json"}, - "json", - ) - project_user_role_id = res.json()[0]["_id"] if res else None - - # POST - res = engine.test( - "Create test project", - "POST", - "/admin/v1/projects", - headers_json, - {"name": "test"}, - (201), - {"Location": "/admin/v1/projects/", "Content-Type": "application/json"}, - "json", - ) - test_project_id = engine.last_id if res else None - res = engine.test( - "Create role without permissions", - "POST", - "/admin/v1/roles", - headers_json, - {"name": "empty"}, - (201), - {"Content-Type": "application/json"}, - "json", - ) - empty_role_id = engine.last_id if res else None - res = engine.test( - "Create role with default permissions", - "POST", - "/admin/v1/roles", - headers_json, - {"name": "default", "permissions": {"default": True}}, - (201), - {"Location": "/admin/v1/roles/", "Content-Type": "application/json"}, - "json", - ) - default_role_id = engine.last_id if res else None - res = engine.test( - "Create role with token permissions", - "POST", - "/admin/v1/roles", - headers_json, - { - "name": "tokens", - "permissions": {"tokens": True}, - }, # is default required ? - (201), - {"Location": "/admin/v1/roles/", "Content-Type": "application/json"}, - "json", - ) - token_role_id = engine.last_id if res else None - pr = "project-role mappings" - res = engine.test( - "Create user without " + pr, - "POST", - "/admin/v1/users", - headers_json, - {"username": "empty", "password": "empty"}, - 201, - {"Content-Type": "application/json"}, - "json", - ) - empty_user_id = engine.last_id if res else None - if ( - admin_project_id - and test_project_id - and project_admin_role_id - and project_user_role_id - ): - data = {"username": "test", "password": "test"} - data["project_role_mappings"] = [ - {"project": test_project_id, "role": project_admin_role_id}, - {"project": admin_project_id, "role": project_user_role_id}, - ] - res = engine.test( - "Create user with " + pr, - "POST", - "/admin/v1/users", - headers_json, - data, - (201), - {"Content-Type": "application/json"}, - "json", - ) - test_user_id = engine.last_id if res else None - - # PUT - if test_user_id: - engine.test( - "Modify test user's password", - "PUT", - "/admin/v1/users/" + test_user_id, - headers_json, - {"password": "password"}, - (204), - {}, - 0, - ) - if ( - empty_user_id - and admin_project_id - and test_project_id - and project_admin_role_id - and project_user_role_id - ): - data = { - "project_role_mappings": [ - {"project": test_project_id, "role": project_admin_role_id}, - {"project": admin_project_id, "role": project_user_role_id}, - ] - } - engine.test( - "Modify empty user's " + pr, - "PUT", - "/admin/v1/users/" + empty_user_id, - headers_json, - data, - (204), - {}, - 0, - ) - - # DELETE - if empty_user_id: - engine.test( - "Delete empty user", - "DELETE", - "/admin/v1/users/" + empty_user_id, - headers_json, - {}, - (204), - {}, - 0, - ) - if test_user_id: - engine.test( - "Delete test user", - "DELETE", - "/admin/v1/users/" + test_user_id, - headers_json, - {}, - (204), - {}, - 0, - ) - if empty_role_id: - engine.test( - "Delete empty role", - "DELETE", - "/admin/v1/roles/" + empty_role_id, - headers_json, - {}, - (204), - {}, - 0, - ) - if default_role_id: - engine.test( - "Delete default role", - "DELETE", - "/admin/v1/roles/" + default_role_id, - headers_json, - {}, - (204), - {}, - 0, - ) - if token_role_id: - engine.test( - "Delete token role", - "DELETE", - "/admin/v1/roles/" + token_role_id, - headers_json, - {}, - (204), - {}, - 0, - ) - if test_project_id: - engine.test( - "Delete test project", - "DELETE", - "/admin/v1/projects/" + test_project_id, - headers_json, - {}, - (204), - {}, - 0, - ) - - # END Tests - - engine.remove_authorization() # To finish - - -class TestNbiQuotas: - description = "Test NBI Quotas" - - @staticmethod - def run(engine, test_osm, manual_check, test_params=None): - engine.set_test_name("NBI-Quotas_") - # backend = test_params.get("backend") if test_params else None # UNUSED - - test_username = "test-nbi-quotas" - test_password = "test-nbi-quotas" - test_project = "test-nbi-quotas" - - test_vim = "test-nbi-quotas" - test_wim = "test-nbi-quotas" - test_sdn = "test-nbi-quotas" - - test_user_id = None - test_project_id = None - - test_vim_ids = [] - test_wim_ids = [] - test_sdn_ids = [] - test_vnfd_ids = [] - test_nsd_ids = [] - test_nst_ids = [] - test_pdu_ids = [] - test_nsr_ids = [] - test_nsi_ids = [] - - # Save admin access data - admin_username = engine.user - admin_password = engine.password - admin_project = engine.project - - # Get admin access - engine.get_autorization() - admin_token = engine.last_id - - # Check that test project,user do not exist - res1 = engine.test( - "Check that test project doesn't exist", - "GET", - "/admin/v1/projects/" + test_project, - headers_json, - {}, - (404), - {}, - True, - ) - res2 = engine.test( - "Check that test user doesn't exist", - "GET", - "/admin/v1/users/" + test_username, - headers_json, - {}, - (404), - {}, - True, - ) - if None in [res1, res2]: - engine.remove_authorization() - logger.error("Test project and/or user already exist") - return - - # Create test project&user - res = engine.test( - "Create test project", - "POST", - "/admin/v1/projects", - headers_json, - { - "name": test_username, - "quotas": { - "vnfds": 2, - "nsds": 2, - "nsts": 1, - "pdus": 1, - "nsrs": 2, - "nsis": 1, - "vim_accounts": 1, - "wim_accounts": 1, - "sdns": 1, - }, - }, - (201), - r_header_json, - "json", - ) - test_project_id = engine.last_id if res else None - res = engine.test( - "Create test user", - "POST", - "/admin/v1/users", - headers_json, - { - "username": test_username, - "password": test_password, - "project_role_mappings": [ - {"project": test_project, "role": "project_admin"} - ], - }, - (201), - r_header_json, - "json", - ) - test_user_id = engine.last_id if res else None - - if test_project_id and test_user_id: - # Get user access - engine.token = None - engine.user = test_username - engine.password = test_password - engine.project = test_project - engine.get_autorization() - user_token = engine.last_id - - # Create test VIM - res = engine.test( - "Create test VIM", - "POST", - "/admin/v1/vim_accounts", - headers_json, - { - "name": test_vim, - "vim_type": "openvim", - "vim_user": test_username, - "vim_password": test_password, - "vim_tenant_name": test_project, - "vim_url": "https://0.0.0.0:0/v0.0", - }, - (202), - r_header_json, - "json", - ) - test_vim_ids += [engine.last_id if res else None] - - res = engine.test( - "Try to create second test VIM", - "POST", - "/admin/v1/vim_accounts", - headers_json, - { - "name": test_vim + "_2", - "vim_type": "openvim", - "vim_user": test_username, - "vim_password": test_password, - "vim_tenant_name": test_project, - "vim_url": "https://0.0.0.0:0/v0.0", - }, - (422), - r_header_json, - "json", - ) - test_vim_ids += [engine.last_id if res is None else None] - - res = engine.test( - "Try to create second test VIM with FORCE", - "POST", - "/admin/v1/vim_accounts?FORCE", - headers_json, - { - "name": test_vim + "_3", - "vim_type": "openvim", - "vim_user": test_username, - "vim_password": test_password, - "vim_tenant_name": test_project, - "vim_url": "https://0.0.0.0:0/v0.0", - }, - (202), - r_header_json, - "json", - ) - test_vim_ids += [engine.last_id if res else None] - - if test_vim_ids[0]: - # Download descriptor files (if required) - test_dir = "/tmp/" + test_username + "/" - test_url = "https://osm-download.etsi.org/ftp/osm-6.0-six/7th-hackfest/packages/" - vnfd_filenames = [ - "slice_hackfest_vnfd.tar.gz", - "slice_hackfest_middle_vnfd.tar.gz", - ] - nsd_filenames = [ - "slice_hackfest_nsd.tar.gz", - "slice_hackfest_middle_nsd.tar.gz", - ] - nst_filenames = ["slice_hackfest_nstd.yaml"] - pdu_filenames = ["PDU_router.yaml"] - desc_filenames = ( - vnfd_filenames + nsd_filenames + nst_filenames + pdu_filenames - ) - if not os.path.exists(test_dir): - os.makedirs(test_dir) - for filename in desc_filenames: - if not os.path.exists(test_dir + filename): - res = requests.get(test_url + filename) - if res.status_code < 300: - with open(test_dir + filename, "wb") as file: - file.write(res.content) - - if all([os.path.exists(test_dir + p) for p in desc_filenames]): - # Test VNFD Quotas - res = engine.test( - "Create test VNFD #1", - "POST", - "/vnfpkgm/v1/vnf_packages_content", - headers_zip_json, - "@b" + test_dir + vnfd_filenames[0], - (201), - r_header_json, - "json", - ) - test_vnfd_ids += [engine.last_id if res else None] - res = engine.test( - "Create test VNFD #2", - "POST", - "/vnfpkgm/v1/vnf_packages_content", - headers_zip_json, - "@b" + test_dir + vnfd_filenames[1], - (201), - r_header_json, - "json", - ) - test_vnfd_ids += [engine.last_id if res else None] - res = engine.test( - "Try to create extra test VNFD", - "POST", - "/vnfpkgm/v1/vnf_packages_content", - headers_zip_json, - "@b" + test_dir + vnfd_filenames[0], - (422), - r_header_json, - "json", - ) - test_vnfd_ids += [engine.last_id if res is None else None] - res = engine.test( - "Try to create extra test VNFD with FORCE", - "POST", - "/vnfpkgm/v1/vnf_packages_content?FORCE", - headers_zip_json, - "@b" + test_dir + vnfd_filenames[0], - (201), - r_header_json, - "json", - ) - test_vnfd_ids += [engine.last_id if res else None] - - # Remove extra VNFDs to prevent further errors - for i in [2, 3]: - if test_vnfd_ids[i]: - res = engine.test( - "Delete test VNFD #" + str(i), - "DELETE", - "/vnfpkgm/v1/vnf_packages_content/" - + test_vnfd_ids[i] - + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - if res: - test_vnfd_ids[i] = None - - if test_vnfd_ids[0] and test_vnfd_ids[1]: - # Test NSD Quotas - res = engine.test( - "Create test NSD #1", - "POST", - "/nsd/v1/ns_descriptors_content", - headers_zip_json, - "@b" + test_dir + nsd_filenames[0], - (201), - r_header_json, - "json", - ) - test_nsd_ids += [engine.last_id if res else None] - res = engine.test( - "Create test NSD #2", - "POST", - "/nsd/v1/ns_descriptors_content", - headers_zip_json, - "@b" + test_dir + nsd_filenames[1], - (201), - r_header_json, - "json", - ) - test_nsd_ids += [engine.last_id if res else None] - res = engine.test( - "Try to create extra test NSD", - "POST", - "/nsd/v1/ns_descriptors_content", - headers_zip_json, - "@b" + test_dir + nsd_filenames[0], - (422), - r_header_json, - "json", - ) - test_nsd_ids += [engine.last_id if res is None else None] - res = engine.test( - "Try to create extra test NSD with FORCE", - "POST", - "/nsd/v1/ns_descriptors_content?FORCE", - headers_zip_json, - "@b" + test_dir + nsd_filenames[0], - (201), - r_header_json, - "json", - ) - test_nsd_ids += [engine.last_id if res else None] - - # Remove extra NSDs to prevent further errors - for i in [2, 3]: - if test_nsd_ids[i]: - res = engine.test( - "Delete test NSD #" + str(i), - "DELETE", - "/nsd/v1/ns_descriptors_content/" - + test_nsd_ids[i] - + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - if res: - test_nsd_ids[i] = None - - if test_nsd_ids[0] and test_nsd_ids[1]: - # Test NSR Quotas - res = engine.test( - "Create test NSR #1", - "POST", - "/nslcm/v1/ns_instances_content", - headers_json, - { - "nsName": test_username + "_1", - "nsdId": test_nsd_ids[0], - "vimAccountId": test_vim_ids[0], - }, - (201), - r_header_json, - "json", - ) - test_nsr_ids += [engine.last_id if res else None] - res = engine.test( - "Create test NSR #2", - "POST", - "/nslcm/v1/ns_instances_content", - headers_json, - { - "nsName": test_username + "_2", - "nsdId": test_nsd_ids[1], - "vimAccountId": test_vim_ids[0], - }, - (201), - r_header_json, - "json", - ) - test_nsr_ids += [engine.last_id if res else None] - res = engine.test( - "Try to create extra test NSR", - "POST", - "/nslcm/v1/ns_instances_content", - headers_json, - { - "nsName": test_username + "_3", - "nsdId": test_nsd_ids[0], - "vimAccountId": test_vim_ids[0], - }, - (422), - r_header_json, - "json", - ) - test_nsr_ids += [engine.last_id if res is None else None] - res = engine.test( - "Try to create test NSR with FORCE", - "POST", - "/nslcm/v1/ns_instances_content?FORCE", - headers_json, - { - "nsName": test_username + "_4", - "nsdId": test_nsd_ids[0], - "vimAccountId": test_vim_ids[0], - }, - (201), - r_header_json, - "json", - ) - test_nsr_ids += [engine.last_id if res else None] - - # Test NST Quotas - res = engine.test( - "Create test NST", - "POST", - "/nst/v1/netslice_templates_content", - headers_txt_json, - "@b" + test_dir + nst_filenames[0], - (201), - r_header_json, - "json", - ) - test_nst_ids += [engine.last_id if res else None] - res = engine.test( - "Try to create extra test NST", - "POST", - "/nst/v1/netslice_templates_content", - headers_txt_json, - "@b" + test_dir + nst_filenames[0], - (422), - r_header_json, - "json", - ) - test_nst_ids += [engine.last_id if res is None else None] - res = engine.test( - "Try to create extra test NST with FORCE", - "POST", - "/nst/v1/netslice_templates_content?FORCE", - headers_txt_json, - "@b" + test_dir + nst_filenames[0], - (201), - r_header_json, - "json", - ) - test_nst_ids += [engine.last_id if res else None] - - if test_nst_ids[0]: - # Remove NSR Quota - engine.set_header( - {"Authorization": "Bearer {}".format(admin_token)} - ) - res = engine.test( - "Remove NSR Quota", - "PUT", - "/admin/v1/projects/" + test_project_id, - headers_json, - {"quotas": {"nsrs": None}}, - (204), - {}, - 0, - ) - engine.set_header( - {"Authorization": "Bearer {}".format(user_token)} - ) - if res: - # Test NSI Quotas - res = engine.test( - "Create test NSI", - "POST", - "/nsilcm/v1/netslice_instances_content", - headers_json, - { - "nsiName": test_username, - "nstId": test_nst_ids[0], - "vimAccountId": test_vim_ids[0], - }, - (201), - r_header_json, - "json", - ) - test_nsi_ids += [engine.last_id if res else None] - res = engine.test( - "Try to create extra test NSI", - "POST", - "/nsilcm/v1/netslice_instances_content", - headers_json, - { - "nsiName": test_username, - "nstId": test_nst_ids[0], - "vimAccountId": test_vim_ids[0], - }, - (400), - r_header_json, - "json", - ) - test_nsi_ids += [ - engine.last_id if res is None else None - ] - res = engine.test( - "Try to create extra test NSI with FORCE", - "POST", - "/nsilcm/v1/netslice_instances_content?FORCE", - headers_json, - { - "nsiName": test_username, - "nstId": test_nst_ids[0], - "vimAccountId": test_vim_ids[0], - }, - (201), - r_header_json, - "json", - ) - test_nsi_ids += [engine.last_id if res else None] - - # Test PDU Quotas - with open(test_dir + pdu_filenames[0], "rb") as file: - pdu_text = re.sub( - r"ip-address: *\[[^\]]*\]", - "ip-address: '0.0.0.0'", - file.read().decode("utf-8"), - ) - with open(test_dir + pdu_filenames[0], "wb") as file: - file.write(pdu_text.encode("utf-8")) - res = engine.test( - "Create test PDU", - "POST", - "/pdu/v1/pdu_descriptors", - headers_yaml, - "@b" + test_dir + pdu_filenames[0], - (201), - r_header_yaml, - "yaml", - ) - test_pdu_ids += [engine.last_id if res else None] - res = engine.test( - "Try to create extra test PDU", - "POST", - "/pdu/v1/pdu_descriptors", - headers_yaml, - "@b" + test_dir + pdu_filenames[0], - (422), - r_header_yaml, - "yaml", - ) - test_pdu_ids += [engine.last_id if res is None else None] - res = engine.test( - "Try to create extra test PDU with FORCE", - "POST", - "/pdu/v1/pdu_descriptors?FORCE", - headers_yaml, - "@b" + test_dir + pdu_filenames[0], - (201), - r_header_yaml, - "yaml", - ) - test_pdu_ids += [engine.last_id if res else None] - - # Cleanup - for i, id in enumerate(test_nsi_ids): - if id: - engine.test( - "Delete test NSI #" + str(i), - "DELETE", - "/nsilcm/v1/netslice_instances_content/" - + id - + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - for i, id in enumerate(test_nsr_ids): - if id: - engine.test( - "Delete test NSR #" + str(i), - "DELETE", - "/nslcm/v1/ns_instances_content/" + id + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - for i, id in enumerate(test_nst_ids): - if id: - engine.test( - "Delete test NST #" + str(i), - "DELETE", - "/nst/v1/netslice_templates_content/" + id + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - for i, id in enumerate(test_nsd_ids): - if id: - engine.test( - "Delete test NSD #" + str(i), - "DELETE", - "/nsd/v1/ns_descriptors_content/" + id + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - for i, id in enumerate(test_vnfd_ids): - if id: - engine.test( - "Delete test VNFD #" + str(i), - "DELETE", - "/vnfpkgm/v1/vnf_packages_content/" + id + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - for i, id in enumerate(test_pdu_ids): - if id: - engine.test( - "Delete test PDU #" + str(i), - "DELETE", - "/pdu/v1/pdu_descriptors/" + id + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - - # END Test NBI Quotas - - # Test WIM Quotas - res = engine.test( - "Create test WIM", - "POST", - "/admin/v1/wim_accounts", - headers_json, - { - "name": test_wim, - "wim_type": "onos", - "wim_url": "https://0.0.0.0:0/v0.0", - }, - (202), - r_header_json, - "json", - ) - test_wim_ids += [engine.last_id if res else None] - res = engine.test( - "Try to create second test WIM", - "POST", - "/admin/v1/wim_accounts", - headers_json, - { - "name": test_wim + "_2", - "wim_type": "onos", - "wim_url": "https://0.0.0.0:0/v0.0", - }, - (422), - r_header_json, - "json", - ) - test_wim_ids += [engine.last_id if res is None else None] - res = engine.test( - "Try to create second test WIM with FORCE", - "POST", - "/admin/v1/wim_accounts?FORCE", - headers_json, - { - "name": test_wim + "_3", - "wim_type": "onos", - "wim_url": "https://0.0.0.0:0/v0.0", - }, - (202), - r_header_json, - "json", - ) - test_wim_ids += [engine.last_id if res else None] - - # Test SDN Quotas - res = engine.test( - "Create test SDN", - "POST", - "/admin/v1/sdns", - headers_json, - { - "name": test_sdn, - "type": "onos", - "ip": "0.0.0.0", - "port": 9999, - "dpid": "00:00:00:00:00:00:00:00", - }, - (202), - r_header_json, - "json", - ) - test_sdn_ids += [engine.last_id if res else None] - res = engine.test( - "Try to create second test SDN", - "POST", - "/admin/v1/sdns", - headers_json, - { - "name": test_sdn + "_2", - "type": "onos", - "ip": "0.0.0.0", - "port": 9999, - "dpid": "00:00:00:00:00:00:00:00", - }, - (422), - r_header_json, - "json", - ) - test_sdn_ids += [engine.last_id if res is None else None] - res = engine.test( - "Try to create second test SDN with FORCE", - "POST", - "/admin/v1/sdns?FORCE", - headers_json, - { - "name": test_sdn + "_3", - "type": "onos", - "ip": "0.0.0.0", - "port": 9999, - "dpid": "00:00:00:00:00:00:00:00", - }, - (202), - r_header_json, - "json", - ) - test_sdn_ids += [engine.last_id if res else None] - - # Cleanup - for i, id in enumerate(test_vim_ids): - if id: - engine.test( - "Delete test VIM #" + str(i), - "DELETE", - "/admin/v1/vim_accounts/" + id + "?FORCE", - headers_json, - {}, - (202), - {}, - 0, - ) - for i, id in enumerate(test_wim_ids): - if id: - engine.test( - "Delete test WIM #" + str(i), - "DELETE", - "/admin/v1/wim_accounts/" + id + "?FORCE", - headers_json, - {}, - (202), - {}, - 0, - ) - for i, id in enumerate(test_sdn_ids): - if id: - engine.test( - "Delete test SDN #" + str(i), - "DELETE", - "/admin/v1/sdns/" + id + "?FORCE", - headers_json, - {}, - (202), - {}, - 0, - ) - - # Release user access - engine.remove_authorization() - - # Cleanup - engine.user = admin_username - engine.password = admin_password - engine.project = admin_project - engine.get_autorization() - if test_user_id: - engine.test( - "Delete test user", - "DELETE", - "/admin/v1/users/" + test_user_id + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - if test_project_id: - engine.test( - "Delete test project", - "DELETE", - "/admin/v1/projects/" + test_project_id + "?FORCE", - headers_json, - {}, - (204), - {}, - 0, - ) - engine.remove_authorization() - - # END class TestNbiQuotas - - -if __name__ == "__main__": - global logger - test = "" - - # Disable warnings from self-signed certificates. - requests.packages.urllib3.disable_warnings() - try: - logging.basicConfig(format="%(levelname)s %(message)s", level=logging.ERROR) - logger = logging.getLogger("NBI") - # load parameters and configuration - opts, args = getopt.getopt( - sys.argv[1:], - "hvu:p:", - [ - "url=", - "user=", - "password=", - "help", - "version", - "verbose", - "no-verbose", - "project=", - "insecure", - "timeout", - "timeout-deploy", - "timeout-configure", - "test=", - "list", - "test-osm", - "manual-check", - "params=", - "fail-fast", - ], - ) - url = "https://localhost:9999/osm" - user = password = project = "admin" - test_osm = False - manual_check = False - verbose = 0 - verify = True - fail_fast = False - test_classes = { - "NonAuthorized": TestNonAuthorized, - "FakeVIM": TestFakeVim, - "Users-Projects": TestUsersProjects, - "Projects-Descriptors": TestProjectsDescriptors, - "VIM-SDN": TestVIMSDN, - "Deploy-Custom": TestDeploy, - "Deploy-Hackfest-Cirros": TestDeployHackfestCirros, - "Deploy-Hackfest-Cirros-Scaling": TestDeployHackfestCirrosScaling, - "Deploy-Hackfest-3Charmed": TestDeployHackfest3Charmed, - "Deploy-Hackfest-3Charmed2": TestDeployHackfest3Charmed2, - "Deploy-Hackfest-3Charmed3": TestDeployHackfest3Charmed3, - "Deploy-Hackfest-4": TestDeployHackfest4, - "Deploy-CirrosMacIp": TestDeployIpMac, - "Descriptors": TestDescriptors, - "Deploy-Hackfest1": TestDeployHackfest1, - # "Deploy-MultiVIM": TestDeployMultiVIM, - "Deploy-SingleVdu": TestDeploySingleVdu, - "Deploy-Hnfd": TestDeployHnfd, - "Upload-Slice-Template": TestNetSliceTemplates, - "Deploy-Slice-Instance": TestNetSliceInstances, - "Deploy-SimpleCharm": TestDeploySimpleCharm, - "Deploy-SimpleCharm2": TestDeploySimpleCharm2, - "Authentication": TestAuthentication, - "NBI-Quotas": TestNbiQuotas, - } - test_to_do = [] - test_params = {} - - for o, a in opts: - # print("parameter:", o, a) - if o == "--version": - print("test version " + __version__ + " " + version_date) - exit() - elif o == "--list": - for test, test_class in sorted(test_classes.items()): - print("{:32} {}".format(test + ":", test_class.description)) - exit() - elif o in ("-v", "--verbose"): - verbose += 1 - elif o == "no-verbose": - verbose = -1 - elif o in ("-h", "--help"): - usage() - sys.exit() - elif o == "--test-osm": - test_osm = True - elif o == "--manual-check": - manual_check = True - elif o == "--url": - url = a - elif o in ("-u", "--user"): - user = a - elif o in ("-p", "--password"): - password = a - elif o == "--project": - project = a - elif o == "--fail-fast": - fail_fast = True - elif o == "--test": - for _test in a.split(","): - if _test not in test_classes: - print( - "Invalid test name '{}'. Use option '--list' to show available tests".format( - _test - ), - file=sys.stderr, - ) - exit(1) - test_to_do.append(_test) - elif o == "--params": - param_key, _, param_value = a.partition("=") - text_index = len(test_to_do) - if text_index not in test_params: - test_params[text_index] = {} - test_params[text_index][param_key] = param_value - elif o == "--insecure": - verify = False - elif o == "--timeout": - timeout = int(a) - elif o == "--timeout-deploy": - timeout_deploy = int(a) - elif o == "--timeout-configure": - timeout_configure = int(a) - else: - assert False, "Unhandled option" - if verbose == 0: - logger.setLevel(logging.WARNING) - elif verbose > 1: - logger.setLevel(logging.DEBUG) - else: - logger.setLevel(logging.ERROR) - - test_rest = TestRest(url, user=user, password=password, project=project) - # print("tests to do:", test_to_do) - if test_to_do: - text_index = 0 - for test in test_to_do: - if fail_fast and test_rest.failed_tests: - break - text_index += 1 - test_class = test_classes[test] - test_class().run( - test_rest, test_osm, manual_check, test_params.get(text_index) - ) - else: - for test, test_class in sorted(test_classes.items()): - if fail_fast and test_rest.failed_tests: - break - test_class().run(test_rest, test_osm, manual_check, test_params.get(0)) - test_rest.print_results() - exit(1 if test_rest.failed_tests else 0) - - except TestException as e: - logger.error(test + "Test {} Exception: {}".format(test, str(e))) - exit(1) - except getopt.GetoptError as e: - logger.error(e) - print(e, file=sys.stderr) - exit(1) - except Exception as e: - logger.critical(test + " Exception: " + str(e), exc_info=True) diff --git a/pyangbind.patch b/pyangbind.patch new file mode 100644 index 0000000..3077299 --- /dev/null +++ b/pyangbind.patch @@ -0,0 +1,46 @@ +####################################################################################### +# Copyright ETSI Contributors and Others. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or +# implied. +# See the License for the specific language governing permissions and +# limitations under the License. +####################################################################################### + +*** .tox/cover/lib/python3.10/site-packages/pyangbind/lib/yangtypes.py 2023-05-10 06:50:57.876027148 -0400 +--- .tox/cover/lib/python3.10/site-packages/pyangbind/lib/yangtypes.py 2023-05-10 06:51:11.772022417 -0400 +*************** limitations under the License. +*** 22,27 **** +--- 22,28 ---- + from __future__ import unicode_literals + + import collections ++ from six.moves import collections_abc + import copy + import uuid + from decimal import Decimal +*************** def TypedListType(*args, **kwargs): +*** 372,378 **** + if not isinstance(allowed_type, list): + allowed_type = [allowed_type] + +! class TypedList(collections.MutableSequence): + _pybind_generated_by = "TypedListType" + _list = list() + +--- 373,379 ---- + if not isinstance(allowed_type, list): + allowed_type = [allowed_type] + +! class TypedList(collections_abc.MutableSequence): + _pybind_generated_by = "TypedListType" + _list = list() + diff --git a/requirements-dev.txt b/requirements-dev.txt index 9ef75e0..8bbb75d 100644 --- a/requirements-dev.txt +++ b/requirements-dev.txt @@ -20,12 +20,16 @@ async-timeout==4.0.2 # via # -r https://osm.etsi.org/gitweb/?p=osm/common.git;a=blob_plain;f=requirements.txt;hb=master # aiokafka -bitarray==2.6.2 +bitarray==2.7.3 # via # -r https://osm.etsi.org/gitweb/?p=osm/IM.git;a=blob_plain;f=requirements.txt;hb=master # pyangbind dataclasses==0.6 # via -r https://osm.etsi.org/gitweb/?p=osm/common.git;a=blob_plain;f=requirements.txt;hb=master +dnspython==2.3.0 + # via + # -r https://osm.etsi.org/gitweb/?p=osm/common.git;a=blob_plain;f=requirements.txt;hb=master + # pymongo enum34==1.1.10 # via # -r https://osm.etsi.org/gitweb/?p=osm/IM.git;a=blob_plain;f=requirements.txt;hb=master @@ -39,13 +43,13 @@ lxml==4.9.2 # -r https://osm.etsi.org/gitweb/?p=osm/IM.git;a=blob_plain;f=requirements.txt;hb=master # pyang # pyangbind -motor==1.3.1 +motor==3.1.2 # via -r https://osm.etsi.org/gitweb/?p=osm/common.git;a=blob_plain;f=requirements.txt;hb=master osm-common @ git+https://osm.etsi.org/gerrit/osm/common.git@master # via -r requirements-dev.in osm-im @ git+https://osm.etsi.org/gerrit/osm/IM.git@master # via -r requirements-dev.in -packaging==23.0 +packaging==23.1 # via # -r https://osm.etsi.org/gitweb/?p=osm/common.git;a=blob_plain;f=requirements.txt;hb=master # aiokafka @@ -57,7 +61,7 @@ pyangbind==0.8.1 # via -r https://osm.etsi.org/gitweb/?p=osm/IM.git;a=blob_plain;f=requirements.txt;hb=master pycryptodome==3.17 # via -r https://osm.etsi.org/gitweb/?p=osm/common.git;a=blob_plain;f=requirements.txt;hb=master -pymongo==3.13.0 +pymongo==4.3.3 # via # -r https://osm.etsi.org/gitweb/?p=osm/common.git;a=blob_plain;f=requirements.txt;hb=master # motor @@ -65,7 +69,7 @@ pyyaml==5.4.1 # via # -r https://osm.etsi.org/gitweb/?p=osm/IM.git;a=blob_plain;f=requirements.txt;hb=master # -r https://osm.etsi.org/gitweb/?p=osm/common.git;a=blob_plain;f=requirements.txt;hb=master -regex==2022.10.31 +regex==2023.5.5 # via # -r https://osm.etsi.org/gitweb/?p=osm/IM.git;a=blob_plain;f=requirements.txt;hb=master # pyangbind diff --git a/requirements-test.txt b/requirements-test.txt index 0c59b6e..c81781d 100644 --- a/requirements-test.txt +++ b/requirements-test.txt @@ -14,7 +14,7 @@ # limitations under the License. -aiohttp==3.8.3 +aiohttp==3.8.4 # via aioresponses aioresponses==0.7.4 # via -r requirements-test.in @@ -24,13 +24,13 @@ async-timeout==4.0.2 # via aiohttp asynctest==0.13.0 # via -r requirements-test.in -attrs==22.2.0 +attrs==23.1.0 # via aiohttp -charset-normalizer==2.1.1 +charset-normalizer==3.1.0 # via aiohttp -coverage==7.1.0 +coverage==7.2.5 # via -r requirements-test.in -deepdiff==6.2.3 +deepdiff==6.3.0 # via -r requirements-test.in frozenlist==1.3.3 # via @@ -42,11 +42,9 @@ multidict==6.0.4 # via # aiohttp # yarl -nose2==0.12.0 +nose2==0.13.0 # via -r requirements-test.in ordered-set==4.1.0 # via deepdiff -orjson==3.8.5 - # via deepdiff -yarl==1.8.2 +yarl==1.9.2 # via aiohttp diff --git a/requirements.txt b/requirements.txt index 941e6aa..46d9b27 100644 --- a/requirements.txt +++ b/requirements.txt @@ -14,21 +14,21 @@ # limitations under the License. -aiohttp==3.8.3 +aiohttp==3.8.4 # via -r requirements.in aiosignal==1.3.1 # via aiohttp async-timeout==4.0.2 # via aiohttp -attrs==22.2.0 +attrs==23.1.0 # via # aiohttp # jsonschema autocommand==2.2.2 # via jaraco-text -certifi==2022.12.7 +certifi==2023.5.7 # via requests -charset-normalizer==2.1.1 +charset-normalizer==3.1.0 # via # aiohttp # requests @@ -41,7 +41,7 @@ debtcollector==2.5.0 # oslo-config # oslo-utils # python-keystoneclient -deepdiff==6.2.3 +deepdiff==6.3.0 # via -r requirements.in frozenlist==1.3.3 # via @@ -51,41 +51,34 @@ idna==3.4 # via # requests # yarl -importlib-resources==5.10.2 - # via - # jaraco-text - # jsonschema -inflect==6.0.2 +inflect==6.0.4 # via jaraco-text iso8601==1.1.0 # via # keystoneauth1 # oslo-utils -jaraco-classes==3.2.3 - # via jaraco-collections -jaraco-collections==3.8.0 +jaraco-collections==4.1.0 # via cherrypy jaraco-context==4.3.0 # via jaraco-text -jaraco-functools==3.5.2 +jaraco-functools==3.6.0 # via # cheroot # jaraco-text # tempora -jaraco-text==3.11.0 +jaraco-text==3.11.1 # via jaraco-collections jsonschema==4.17.3 # via -r requirements.in -keystoneauth1==5.1.1 +keystoneauth1==5.1.2 # via python-keystoneclient -more-itertools==9.0.0 +more-itertools==9.1.0 # via # cheroot # cherrypy - # jaraco-classes # jaraco-functools # jaraco-text -msgpack==1.0.4 +msgpack==1.0.5 # via oslo-serialization multidict==6.0.4 # via @@ -99,24 +92,22 @@ netifaces==0.11.0 # via oslo-utils ordered-set==4.1.0 # via deepdiff -orjson==3.8.5 - # via deepdiff os-service-types==1.7.0 # via keystoneauth1 -oslo-config==9.1.0 +oslo-config==9.1.1 # via python-keystoneclient -oslo-i18n==5.1.0 +oslo-i18n==6.0.0 # via # oslo-config # oslo-utils # python-keystoneclient -oslo-serialization==5.0.0 +oslo-serialization==5.1.1 # via python-keystoneclient oslo-utils==6.1.0 # via # oslo-serialization # python-keystoneclient -packaging==23.0 +packaging==23.1 # via # oslo-utils # python-keystoneclient @@ -128,19 +119,17 @@ pbr==5.11.1 # oslo-serialization # python-keystoneclient # stevedore -pkgutil-resolve-name==1.3.10 - # via jsonschema portend==3.1.0 # via cherrypy -pydantic==1.10.4 +pydantic==1.10.7 # via inflect pyparsing==3.0.9 # via oslo-utils pyrsistent==0.19.3 # via jsonschema -python-keystoneclient==5.0.1 +python-keystoneclient==5.1.0 # via -r requirements.in -pytz==2022.7.1 +pytz==2023.3 # via # oslo-serialization # oslo-utils @@ -149,7 +138,7 @@ pyyaml==5.4.1 # via # -r requirements.in # oslo-config -requests==2.28.2 +requests==2.30.0 # via # -r requirements.in # keystoneauth1 @@ -163,27 +152,25 @@ six==1.16.0 # keystoneauth1 # python-keystoneclient # tacacs-plus -stevedore==4.1.1 +stevedore==5.0.0 # via # keystoneauth1 # oslo-config # python-keystoneclient tacacs-plus==2.6 # via -r requirements.in -tempora==5.2.1 +tempora==5.2.2 # via portend -typing-extensions==4.4.0 +typing-extensions==4.5.0 # via pydantic -urllib3==1.26.14 +urllib3==2.0.2 # via requests -wrapt==1.14.1 +wrapt==1.15.0 # via debtcollector -yarl==1.8.2 +yarl==1.9.2 # via aiohttp -zc-lockfile==2.0 +zc-lockfile==3.0.post1 # via cherrypy -zipp==3.11.0 - # via importlib-resources # The following packages are considered to be unsafe in a requirements file: # setuptools diff --git a/tox.ini b/tox.ini index c77d921..4c8fd06 100644 --- a/tox.ini +++ b/tox.ini @@ -22,7 +22,7 @@ toxworkdir = /tmp/.tox [testenv] usedevelop = True -basepython = python3.8 +basepython = python3.10 setenv = VIRTUAL_ENV={envdir} PYTHONDONTWRITEBYTECODE = 1 deps = -r{toxinidir}/requirements.txt @@ -42,9 +42,11 @@ deps = {[testenv]deps} -r{toxinidir}/requirements-dev.txt -r{toxinidir}/requirements-test.txt commands = + sh -c "patch {toxworkdir}/cover/lib/python3.10/site-packages/pyangbind/lib/yangtypes.py < pyangbind.patch" sh -c 'rm -f nosetests.xml' coverage erase nose2 -C --coverage osm_nbi -s osm_nbi/tests + sh -c "patch -R {toxworkdir}/cover/lib/python3.10/site-packages/pyangbind/lib/yangtypes.py < pyangbind.patch" coverage report --omit='*tests*' coverage html -d ./cover --omit='*tests*' coverage xml -o coverage.xml --omit=*tests*