| # |
| # Licensed under the Apache License, Version 2.0 (the "License"); you may |
| # not use this file except in compliance with the License. You may obtain |
| # a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT |
| # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the |
| # License for the specific language governing permissions and limitations |
| # under the License. |
| # |
| # For those usages not covered by the Apache License, Version 2.0 please |
| # contact: alfonso.tiernosepulveda@telefonica.com |
| ## |
| |
| |
| import asynctest # pip3 install asynctest --user |
| import asyncio |
| import yaml |
| import copy |
| from os import getenv |
| from osm_lcm import ns |
| from osm_common.msgkafka import MsgKafka |
| from osm_lcm.lcm_utils import TaskRegistry |
| from osm_lcm.ng_ro import NgRoClient |
| from osm_lcm.data_utils.database.database import Database |
| from osm_lcm.data_utils.filesystem.filesystem import Filesystem |
| from uuid import uuid4 |
| |
| from osm_lcm.tests import test_db_descriptors as descriptors |
| |
| __author__ = "Alfonso Tierno <alfonso.tiernosepulveda@telefonica.com>" |
| |
| """ Perform unittests using asynctest of osm_lcm.ns module |
| It allows, if some testing ENV are supplied, testing without mocking some external libraries for debugging: |
| OSMLCMTEST_NS_PUBKEY: public ssh-key returned by N2VC to inject to VMs |
| OSMLCMTEST_NS_NAME: change name of NS |
| OSMLCMTEST_PACKAGES_PATH: path where the vnf-packages are stored (de-compressed), each one on a 'vnfd_id' folder |
| OSMLCMTEST_NS_IPADDRESS: IP address where emulated VMs are reached. Comma separate list |
| OSMLCMTEST_RO_VIMID: VIM id of RO target vim IP. Obtain it with openmano datcenter-list on RO container |
| OSMLCMTEST_VCA_NOMOCK: Do no mock the VCA, N2VC library, for debugging it |
| OSMLCMTEST_RO_NOMOCK: Do no mock the ROClient library, for debugging it |
| OSMLCMTEST_DB_NOMOCK: Do no mock the database library, for debugging it |
| OSMLCMTEST_FS_NOMOCK: Do no mock the File Storage library, for debugging it |
| OSMLCMTEST_LOGGING_NOMOCK: Do no mock the logging |
| OSMLCM_VCA_XXX: configuration of N2VC |
| OSMLCM_RO_XXX: configuration of RO |
| """ |
| |
| lcm_config = { |
| "global": {"loglevel": "DEBUG"}, |
| "timeout": {}, |
| "VCA": { # TODO replace with os.get_env to get other configurations |
| "host": getenv("OSMLCM_VCA_HOST", "vca"), |
| "port": getenv("OSMLCM_VCA_PORT", 17070), |
| "user": getenv("OSMLCM_VCA_USER", "admin"), |
| "secret": getenv("OSMLCM_VCA_SECRET", "vca"), |
| "public_key": getenv("OSMLCM_VCA_PUBKEY", None), |
| "ca_cert": getenv("OSMLCM_VCA_CACERT", None), |
| "apiproxy": getenv("OSMLCM_VCA_APIPROXY", "192.168.1.1"), |
| }, |
| "ro_config": { |
| "uri": "http://{}:{}/openmano".format( |
| getenv("OSMLCM_RO_HOST", "ro"), getenv("OSMLCM_RO_PORT", "9090") |
| ), |
| "tenant": getenv("OSMLCM_RO_TENANT", "osm"), |
| "logger_name": "lcm.ROclient", |
| "loglevel": "DEBUG", |
| "ng": True, |
| }, |
| } |
| |
| |
| class TestMyNS(asynctest.TestCase): |
| async def _n2vc_DeployCharms( |
| self, |
| model_name, |
| application_name, |
| vnfd, |
| charm_path, |
| params={}, |
| machine_spec={}, |
| callback=None, |
| *callback_args |
| ): |
| if callback: |
| for status, message in ( |
| ("maintenance", "installing sofwware"), |
| ("active", "Ready!"), |
| ): |
| # call callback after some time |
| asyncio.sleep(5, loop=self.loop) |
| callback(model_name, application_name, status, message, *callback_args) |
| |
| @staticmethod |
| def _n2vc_FormatApplicationName(*args): |
| num_calls = 0 |
| while True: |
| yield "app_name-{}".format(num_calls) |
| num_calls += 1 |
| |
| def _n2vc_CreateExecutionEnvironment( |
| self, namespace, reuse_ee_id, db_dict, *args, **kwargs |
| ): |
| k_list = namespace.split(".") |
| ee_id = k_list[1] + "." |
| if len(k_list) >= 2: |
| for k in k_list[2:4]: |
| ee_id += k[:8] |
| else: |
| ee_id += "_NS_" |
| return ee_id, {} |
| |
| def _ro_status(self, *args, **kwargs): |
| print("Args > {}".format(args)) |
| print("kwargs > {}".format(kwargs)) |
| if kwargs.get("delete"): |
| ro_ns_desc = yaml.load( |
| descriptors.ro_delete_action_text, Loader=yaml.Loader |
| ) |
| while True: |
| yield ro_ns_desc |
| |
| ro_ns_desc = yaml.load(descriptors.ro_ns_text, Loader=yaml.Loader) |
| |
| # if ip address provided, replace descriptor |
| ip_addresses = getenv("OSMLCMTEST_NS_IPADDRESS", "") |
| if ip_addresses: |
| ip_addresses_list = ip_addresses.split(",") |
| for vnf in ro_ns_desc["vnfs"]: |
| if not ip_addresses_list: |
| break |
| vnf["ip_address"] = ip_addresses_list[0] |
| for vm in vnf["vms"]: |
| if not ip_addresses_list: |
| break |
| vm["ip_address"] = ip_addresses_list.pop(0) |
| |
| while True: |
| yield ro_ns_desc |
| for net in ro_ns_desc["nets"]: |
| if net["status"] != "ACTIVE": |
| net["status"] = "ACTIVE" |
| break |
| else: |
| for vnf in ro_ns_desc["vnfs"]: |
| for vm in vnf["vms"]: |
| if vm["status"] != "ACTIVE": |
| vm["status"] = "ACTIVE" |
| break |
| |
| def _ro_deploy(self, *args, **kwargs): |
| return {"action_id": args[1]["action_id"], "nsr_id": args[0], "status": "ok"} |
| |
| def _return_uuid(self, *args, **kwargs): |
| return str(uuid4()) |
| |
| async def setUp(self): |
| |
| # Mock DB |
| if not getenv("OSMLCMTEST_DB_NOMOCK"): |
| # Cleanup singleton Database instance |
| Database.instance = None |
| |
| self.db = Database({"database": {"driver": "memory"}}).instance.db |
| self.db.create_list( |
| "vnfds", yaml.load(descriptors.db_vnfds_text, Loader=yaml.Loader) |
| ) |
| self.db.create_list( |
| "nsds", yaml.load(descriptors.db_nsds_text, Loader=yaml.Loader) |
| ) |
| self.db.create_list( |
| "nsrs", yaml.load(descriptors.db_nsrs_text, Loader=yaml.Loader) |
| ) |
| self.db.create_list( |
| "vim_accounts", |
| yaml.load(descriptors.db_vim_accounts_text, Loader=yaml.Loader), |
| ) |
| self.db.create_list( |
| "k8sclusters", |
| yaml.load(descriptors.db_k8sclusters_text, Loader=yaml.Loader), |
| ) |
| self.db.create_list( |
| "nslcmops", yaml.load(descriptors.db_nslcmops_text, Loader=yaml.Loader) |
| ) |
| self.db.create_list( |
| "vnfrs", yaml.load(descriptors.db_vnfrs_text, Loader=yaml.Loader) |
| ) |
| self.db_vim_accounts = yaml.load( |
| descriptors.db_vim_accounts_text, Loader=yaml.Loader |
| ) |
| |
| # Mock kafka |
| self.msg = asynctest.Mock(MsgKafka()) |
| |
| # Mock filesystem |
| if not getenv("OSMLCMTEST_FS_NOMOCK"): |
| self.fs = asynctest.Mock( |
| Filesystem({"storage": {"driver": "local", "path": "/"}}).instance.fs |
| ) |
| self.fs.get_params.return_value = { |
| "path": getenv("OSMLCMTEST_PACKAGES_PATH", "./test/temp/packages") |
| } |
| self.fs.file_open = asynctest.mock_open() |
| # self.fs.file_open.return_value.__enter__.return_value = asynctest.MagicMock() # called on a python "with" |
| # self.fs.file_open.return_value.__enter__.return_value.read.return_value = "" # empty file |
| |
| # Mock TaskRegistry |
| self.lcm_tasks = asynctest.Mock(TaskRegistry()) |
| self.lcm_tasks.lock_HA.return_value = True |
| self.lcm_tasks.waitfor_related_HA.return_value = None |
| self.lcm_tasks.lookfor_related.return_value = ("", []) |
| |
| # Mock VCA - K8s |
| if not getenv("OSMLCMTEST_VCA_K8s_NOMOCK"): |
| ns.K8sJujuConnector = asynctest.MagicMock(ns.K8sJujuConnector) |
| ns.K8sHelmConnector = asynctest.MagicMock(ns.K8sHelmConnector) |
| ns.K8sHelm3Connector = asynctest.MagicMock(ns.K8sHelm3Connector) |
| |
| if not getenv("OSMLCMTEST_VCA_NOMOCK"): |
| ns.N2VCJujuConnector = asynctest.MagicMock(ns.N2VCJujuConnector) |
| ns.LCMHelmConn = asynctest.MagicMock(ns.LCMHelmConn) |
| |
| # Create NsLCM class |
| self.my_ns = ns.NsLcm(self.msg, self.lcm_tasks, lcm_config, self.loop) |
| self.my_ns.fs = self.fs |
| self.my_ns.db = self.db |
| self.my_ns._wait_dependent_n2vc = asynctest.CoroutineMock() |
| |
| # Mock logging |
| if not getenv("OSMLCMTEST_LOGGING_NOMOCK"): |
| self.my_ns.logger = asynctest.Mock(self.my_ns.logger) |
| |
| # Mock VCA - N2VC |
| if not getenv("OSMLCMTEST_VCA_NOMOCK"): |
| pub_key = getenv("OSMLCMTEST_NS_PUBKEY", "ssh-rsa test-pub-key t@osm.com") |
| # self.my_ns.n2vc = asynctest.Mock(N2VC()) |
| self.my_ns.n2vc.GetPublicKey.return_value = getenv( |
| "OSMLCM_VCA_PUBKEY", "public_key" |
| ) |
| # allow several versions of n2vc |
| self.my_ns.n2vc.FormatApplicationName = asynctest.Mock( |
| side_effect=self._n2vc_FormatApplicationName() |
| ) |
| self.my_ns.n2vc.DeployCharms = asynctest.CoroutineMock( |
| side_effect=self._n2vc_DeployCharms |
| ) |
| self.my_ns.n2vc.create_execution_environment = asynctest.CoroutineMock( |
| side_effect=self._n2vc_CreateExecutionEnvironment |
| ) |
| self.my_ns.n2vc.install_configuration_sw = asynctest.CoroutineMock( |
| return_value=pub_key |
| ) |
| self.my_ns.n2vc.get_ee_ssh_public__key = asynctest.CoroutineMock( |
| return_value=pub_key |
| ) |
| self.my_ns.n2vc.exec_primitive = asynctest.CoroutineMock( |
| side_effect=self._return_uuid |
| ) |
| self.my_ns.n2vc.exec_primitive = asynctest.CoroutineMock( |
| side_effect=self._return_uuid |
| ) |
| self.my_ns.n2vc.GetPrimitiveStatus = asynctest.CoroutineMock( |
| return_value="completed" |
| ) |
| self.my_ns.n2vc.GetPrimitiveOutput = asynctest.CoroutineMock( |
| return_value={"result": "ok", "pubkey": pub_key} |
| ) |
| self.my_ns.n2vc.delete_execution_environment = asynctest.CoroutineMock( |
| return_value=None |
| ) |
| self.my_ns.n2vc.get_public_key = asynctest.CoroutineMock( |
| return_value=getenv("OSMLCM_VCA_PUBKEY", "public_key") |
| ) |
| self.my_ns.n2vc.delete_namespace = asynctest.CoroutineMock( |
| return_value=None |
| ) |
| |
| # Mock RO |
| if not getenv("OSMLCMTEST_RO_NOMOCK"): |
| self.my_ns.RO = asynctest.Mock( |
| NgRoClient(self.loop, **lcm_config["ro_config"]) |
| ) |
| # TODO first time should be empty list, following should return a dict |
| # self.my_ns.RO.get_list = asynctest.CoroutineMock(self.my_ns.RO.get_list, return_value=[]) |
| self.my_ns.RO.deploy = asynctest.CoroutineMock( |
| self.my_ns.RO.deploy, side_effect=self._ro_deploy |
| ) |
| # self.my_ns.RO.status = asynctest.CoroutineMock(self.my_ns.RO.status, side_effect=self._ro_status) |
| # self.my_ns.RO.create_action = asynctest.CoroutineMock(self.my_ns.RO.create_action, |
| # return_value={"vm-id": {"vim_result": 200, |
| # "description": "done"}}) |
| self.my_ns.RO.delete = asynctest.CoroutineMock(self.my_ns.RO.delete) |
| |
| # @asynctest.fail_on(active_handles=True) # all async tasks must be completed |
| # async def test_instantiate(self): |
| # nsr_id = descriptors.test_ids["TEST-A"]["ns"] |
| # nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| # # print("Test instantiate started") |
| |
| # # delete deployed information of database |
| # if not getenv("OSMLCMTEST_DB_NOMOCK"): |
| # if self.db.get_list("nsrs")[0]["_admin"].get("deployed"): |
| # del self.db.get_list("nsrs")[0]["_admin"]["deployed"] |
| # for db_vnfr in self.db.get_list("vnfrs"): |
| # db_vnfr.pop("ip_address", None) |
| # for db_vdur in db_vnfr["vdur"]: |
| # db_vdur.pop("ip_address", None) |
| # db_vdur.pop("mac_address", None) |
| # if getenv("OSMLCMTEST_RO_VIMID"): |
| # self.db.get_list("vim_accounts")[0]["_admin"]["deployed"]["RO"] = getenv("OSMLCMTEST_RO_VIMID") |
| # if getenv("OSMLCMTEST_RO_VIMID"): |
| # self.db.get_list("nsrs")[0]["_admin"]["deployed"]["RO"] = getenv("OSMLCMTEST_RO_VIMID") |
| |
| # await self.my_ns.instantiate(nsr_id, nslcmop_id) |
| |
| # self.msg.aiowrite.assert_called_once_with("ns", "instantiated", |
| # {"nsr_id": nsr_id, "nslcmop_id": nslcmop_id, |
| # "operationState": "COMPLETED"}, |
| # loop=self.loop) |
| # self.lcm_tasks.lock_HA.assert_called_once_with('ns', 'nslcmops', nslcmop_id) |
| # if not getenv("OSMLCMTEST_LOGGING_NOMOCK"): |
| # self.assertTrue(self.my_ns.logger.debug.called, "Debug method not called") |
| # self.my_ns.logger.error.assert_not_called() |
| # self.my_ns.logger.exception().assert_not_called() |
| |
| # if not getenv("OSMLCMTEST_DB_NOMOCK"): |
| # self.assertTrue(self.db.set_one.called, "db.set_one not called") |
| # db_nsr = self.db.get_one("nsrs", {"_id": nsr_id}) |
| # db_vnfrs_list = self.db.get_list("vnfrs", {"nsr-id-ref": nsr_id}) |
| # self.assertEqual(db_nsr["_admin"].get("nsState"), "INSTANTIATED", "Not instantiated") |
| # for vnfr in db_vnfrs_list: |
| # self.assertEqual(vnfr["_admin"].get("nsState"), "INSTANTIATED", "Not instantiated") |
| |
| # if not getenv("OSMLCMTEST_VCA_NOMOCK"): |
| # # check intial-primitives called |
| # self.assertTrue(self.my_ns.n2vc.exec_primitive.called, |
| # "Exec primitive not called for initial config primitive") |
| # for _call in self.my_ns.n2vc.exec_primitive.call_args_list: |
| # self.assertIn(_call[1]["primitive_name"], ("config", "touch"), |
| # "called exec primitive with a primitive different than config or touch") |
| |
| # # TODO add more checks of called methods |
| # # TODO add a terminate |
| |
| # async def test_instantiate_ee_list(self): |
| # # Using modern IM where configuration is in the new format of execution_environment_list |
| # ee_descriptor_id = "charm_simple" |
| # non_used_initial_primitive = { |
| # "name": "not_to_be_called", |
| # "seq": 3, |
| # "execution-environment-ref": "not_used_ee" |
| # } |
| # ee_list = [ |
| # { |
| # "id": ee_descriptor_id, |
| # "juju": {"charm": "simple"}, |
| |
| # }, |
| # ] |
| |
| # self.db.set_one( |
| # "vnfds", |
| # q_filter={"_id": "7637bcf8-cf14-42dc-ad70-c66fcf1e6e77"}, |
| # update_dict={"vnf-configuration.0.execution-environment-list": ee_list, |
| # "vnf-configuration.0.initial-config-primitive.0.execution-environment-ref": ee_descriptor_id, |
| # "vnf-configuration.0.initial-config-primitive.1.execution-environment-ref": ee_descriptor_id, |
| # "vnf-configuration.0.initial-config-primitive.2": non_used_initial_primitive, |
| # "vnf-configuration.0.config-primitive.0.execution-environment-ref": ee_descriptor_id, |
| # "vnf-configuration.0.config-primitive.0.execution-environment-primitive": "touch_charm", |
| # }, |
| # unset={"vnf-configuration.juju": None}) |
| # await self.test_instantiate() |
| # # this will check that the initial-congig-primitive 'not_to_be_called' is not called |
| |
| # Test scale() and related methods |
| @asynctest.fail_on(active_handles=True) # all async tasks must be completed |
| async def test_scale(self): |
| # print("Test scale started") |
| |
| # TODO: Add more higher-lever tests here, for example: |
| # scale-out/scale-in operations with success/error result |
| |
| # Test scale() with missing 'scaleVnfData', should return operationState = 'FAILED' |
| nsr_id = descriptors.test_ids["TEST-A"]["ns"] |
| nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| await self.my_ns.scale(nsr_id, nslcmop_id) |
| expected_value = "FAILED" |
| return_value = self.db.get_one("nslcmops", {"_id": nslcmop_id}).get( |
| "operationState" |
| ) |
| self.assertEqual(return_value, expected_value) |
| # print("scale_result: {}".format(self.db.get_one("nslcmops", {"_id": nslcmop_id}).get("detailed-status"))) |
| |
| # Test scale() for native kdu |
| # this also includes testing _scale_kdu() |
| nsr_id = descriptors.test_ids["TEST-NATIVE-KDU"]["ns"] |
| nslcmop_id = descriptors.test_ids["TEST-NATIVE-KDU"]["instantiate"] |
| |
| self.my_ns.k8sclusterjuju.scale = asynctest.mock.CoroutineMock() |
| self.my_ns.k8sclusterjuju.exec_primitive = asynctest.mock.CoroutineMock() |
| self.my_ns.k8sclusterjuju.get_scale_count = asynctest.mock.CoroutineMock( |
| return_value=1 |
| ) |
| await self.my_ns.scale(nsr_id, nslcmop_id) |
| expected_value = "COMPLETED" |
| return_value = self.db.get_one("nslcmops", {"_id": nslcmop_id}).get( |
| "operationState" |
| ) |
| self.assertEqual(return_value, expected_value) |
| self.my_ns.k8sclusterjuju.scale.assert_called_once() |
| |
| # Test scale() for native kdu with 2 resource |
| nsr_id = descriptors.test_ids["TEST-NATIVE-KDU-2"]["ns"] |
| nslcmop_id = descriptors.test_ids["TEST-NATIVE-KDU-2"]["instantiate"] |
| |
| self.my_ns.k8sclusterjuju.get_scale_count.return_value = 2 |
| await self.my_ns.scale(nsr_id, nslcmop_id) |
| expected_value = "COMPLETED" |
| return_value = self.db.get_one("nslcmops", {"_id": nslcmop_id}).get( |
| "operationState" |
| ) |
| self.assertEqual(return_value, expected_value) |
| self.my_ns.k8sclusterjuju.scale.assert_called() |
| |
| async def test_vca_status_refresh(self): |
| nsr_id = descriptors.test_ids["TEST-A"]["ns"] |
| nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| await self.my_ns.vca_status_refresh(nsr_id, nslcmop_id) |
| expected_value = dict() |
| return_value = dict() |
| vnf_descriptors = self.db.get_list("vnfds") |
| for i, _ in enumerate(vnf_descriptors): |
| for j, value in enumerate(vnf_descriptors[i]["df"]): |
| if "lcm-operations-configuration" in vnf_descriptors[i]["df"][j]: |
| if ( |
| "day1-2" |
| in value["lcm-operations-configuration"][ |
| "operate-vnf-op-config" |
| ] |
| ): |
| for k, v in enumerate( |
| value["lcm-operations-configuration"][ |
| "operate-vnf-op-config" |
| ]["day1-2"] |
| ): |
| if ( |
| v.get("execution-environment-list") |
| and "juju" in v["execution-environment-list"][k] |
| ): |
| expected_value = self.db.get_list("nsrs")[i][ |
| "vcaStatus" |
| ] |
| await self.my_ns._on_update_n2vc_db( |
| "nsrs", {"_id": nsr_id}, "_admin.deployed.VCA.0", {} |
| ) |
| return_value = self.db.get_list("nsrs")[i]["vcaStatus"] |
| self.assertEqual(return_value, expected_value) |
| |
| # Test _retry_or_skip_suboperation() |
| # Expected result: |
| # - if a suboperation's 'operationState' is marked as 'COMPLETED', SUBOPERATION_STATUS_SKIP is expected |
| # - if marked as anything but 'COMPLETED', the suboperation index is expected |
| def test_scale_retry_or_skip_suboperation(self): |
| # Load an alternative 'nslcmops' YAML for this test |
| nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) |
| op_index = 2 |
| # Test when 'operationState' is 'COMPLETED' |
| db_nslcmop["_admin"]["operations"][op_index]["operationState"] = "COMPLETED" |
| return_value = self.my_ns._retry_or_skip_suboperation(db_nslcmop, op_index) |
| expected_value = self.my_ns.SUBOPERATION_STATUS_SKIP |
| self.assertEqual(return_value, expected_value) |
| # Test when 'operationState' is not 'COMPLETED' |
| db_nslcmop["_admin"]["operations"][op_index]["operationState"] = None |
| return_value = self.my_ns._retry_or_skip_suboperation(db_nslcmop, op_index) |
| expected_value = op_index |
| self.assertEqual(return_value, expected_value) |
| |
| # Test _find_suboperation() |
| # Expected result: index of the found sub-operation, or SUBOPERATION_STATUS_NOT_FOUND if not found |
| def test_scale_find_suboperation(self): |
| # Load an alternative 'nslcmops' YAML for this test |
| nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) |
| # Find this sub-operation |
| op_index = 2 |
| vnf_index = db_nslcmop["_admin"]["operations"][op_index]["member_vnf_index"] |
| primitive = db_nslcmop["_admin"]["operations"][op_index]["primitive"] |
| primitive_params = db_nslcmop["_admin"]["operations"][op_index][ |
| "primitive_params" |
| ] |
| match = { |
| "member_vnf_index": vnf_index, |
| "primitive": primitive, |
| "primitive_params": primitive_params, |
| } |
| found_op_index = self.my_ns._find_suboperation(db_nslcmop, match) |
| self.assertEqual(found_op_index, op_index) |
| # Test with not-matching params |
| match = { |
| "member_vnf_index": vnf_index, |
| "primitive": "", |
| "primitive_params": primitive_params, |
| } |
| found_op_index = self.my_ns._find_suboperation(db_nslcmop, match) |
| self.assertEqual(found_op_index, self.my_ns.SUBOPERATION_STATUS_NOT_FOUND) |
| # Test with None |
| match = None |
| found_op_index = self.my_ns._find_suboperation(db_nslcmop, match) |
| self.assertEqual(found_op_index, self.my_ns.SUBOPERATION_STATUS_NOT_FOUND) |
| |
| # Test _update_suboperation_status() |
| def test_scale_update_suboperation_status(self): |
| self.db.set_one = asynctest.Mock() |
| nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) |
| op_index = 0 |
| # Force the initial values to be distinct from the updated ones |
| q_filter = {"_id": db_nslcmop["_id"]} |
| # Test to change 'operationState' and 'detailed-status' |
| operationState = "COMPLETED" |
| detailed_status = "Done" |
| expected_update_dict = { |
| "_admin.operations.0.operationState": operationState, |
| "_admin.operations.0.detailed-status": detailed_status, |
| } |
| self.my_ns._update_suboperation_status( |
| db_nslcmop, op_index, operationState, detailed_status |
| ) |
| self.db.set_one.assert_called_once_with( |
| "nslcmops", |
| q_filter=q_filter, |
| update_dict=expected_update_dict, |
| fail_on_empty=False, |
| ) |
| |
| def test_scale_add_suboperation(self): |
| nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) |
| vnf_index = "1" |
| num_ops_before = len(db_nslcmop.get("_admin", {}).get("operations", [])) - 1 |
| vdu_id = None |
| vdu_count_index = None |
| vdu_name = None |
| primitive = "touch" |
| mapped_primitive_params = { |
| "parameter": [ |
| { |
| "data-type": "STRING", |
| "name": "filename", |
| "default-value": "<touch_filename2>", |
| } |
| ], |
| "name": "touch", |
| } |
| operationState = "PROCESSING" |
| detailed_status = "In progress" |
| operationType = "PRE-SCALE" |
| # Add a 'pre-scale' suboperation |
| op_index_after = self.my_ns._add_suboperation( |
| db_nslcmop, |
| vnf_index, |
| vdu_id, |
| vdu_count_index, |
| vdu_name, |
| primitive, |
| mapped_primitive_params, |
| operationState, |
| detailed_status, |
| operationType, |
| ) |
| self.assertEqual(op_index_after, num_ops_before + 1) |
| |
| # Delete all suboperations and add the same operation again |
| del db_nslcmop["_admin"]["operations"] |
| op_index_zero = self.my_ns._add_suboperation( |
| db_nslcmop, |
| vnf_index, |
| vdu_id, |
| vdu_count_index, |
| vdu_name, |
| primitive, |
| mapped_primitive_params, |
| operationState, |
| detailed_status, |
| operationType, |
| ) |
| self.assertEqual(op_index_zero, 0) |
| |
| # Add a 'RO' suboperation |
| RO_nsr_id = "1234567890" |
| RO_scaling_info = [ |
| { |
| "type": "create", |
| "count": 1, |
| "member-vnf-index": "1", |
| "osm_vdu_id": "dataVM", |
| } |
| ] |
| op_index = self.my_ns._add_suboperation( |
| db_nslcmop, |
| vnf_index, |
| vdu_id, |
| vdu_count_index, |
| vdu_name, |
| primitive, |
| mapped_primitive_params, |
| operationState, |
| detailed_status, |
| operationType, |
| RO_nsr_id, |
| RO_scaling_info, |
| ) |
| db_RO_nsr_id = db_nslcmop["_admin"]["operations"][op_index]["RO_nsr_id"] |
| self.assertEqual(op_index, 1) |
| self.assertEqual(RO_nsr_id, db_RO_nsr_id) |
| |
| # Try to add an invalid suboperation, should return SUBOPERATION_STATUS_NOT_FOUND |
| op_index_invalid = self.my_ns._add_suboperation( |
| None, None, None, None, None, None, None, None, None, None, None |
| ) |
| self.assertEqual(op_index_invalid, self.my_ns.SUBOPERATION_STATUS_NOT_FOUND) |
| |
| # Test _check_or_add_scale_suboperation() and _check_or_add_scale_suboperation_RO() |
| # check the possible return values: |
| # - SUBOPERATION_STATUS_NEW: This is a new sub-operation |
| # - op_index (non-negative number): This is an existing sub-operation, operationState != 'COMPLETED' |
| # - SUBOPERATION_STATUS_SKIP: This is an existing sub-operation, operationState == 'COMPLETED' |
| def test_scale_check_or_add_scale_suboperation(self): |
| nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) |
| operationType = "PRE-SCALE" |
| vnf_index = "1" |
| primitive = "touch" |
| primitive_params = { |
| "parameter": [ |
| { |
| "data-type": "STRING", |
| "name": "filename", |
| "default-value": "<touch_filename2>", |
| } |
| ], |
| "name": "touch", |
| } |
| |
| # Delete all sub-operations to be sure this is a new sub-operation |
| del db_nslcmop["_admin"]["operations"] |
| |
| # Add a new sub-operation |
| # For new sub-operations, operationState is set to 'PROCESSING' by default |
| op_index_new = self.my_ns._check_or_add_scale_suboperation( |
| db_nslcmop, vnf_index, primitive, primitive_params, operationType |
| ) |
| self.assertEqual(op_index_new, self.my_ns.SUBOPERATION_STATUS_NEW) |
| |
| # Use the same parameters again to match the already added sub-operation |
| # which has status 'PROCESSING' (!= 'COMPLETED') by default |
| # The expected return value is a non-negative number |
| op_index_existing = self.my_ns._check_or_add_scale_suboperation( |
| db_nslcmop, vnf_index, primitive, primitive_params, operationType |
| ) |
| self.assertTrue(op_index_existing >= 0) |
| |
| # Change operationState 'manually' for this sub-operation |
| db_nslcmop["_admin"]["operations"][op_index_existing][ |
| "operationState" |
| ] = "COMPLETED" |
| # Then use the same parameters again to match the already added sub-operation, |
| # which now has status 'COMPLETED' |
| # The expected return value is SUBOPERATION_STATUS_SKIP |
| op_index_skip = self.my_ns._check_or_add_scale_suboperation( |
| db_nslcmop, vnf_index, primitive, primitive_params, operationType |
| ) |
| self.assertEqual(op_index_skip, self.my_ns.SUBOPERATION_STATUS_SKIP) |
| |
| # RO sub-operation test: |
| # Repeat tests for the very similar _check_or_add_scale_suboperation_RO(), |
| RO_nsr_id = "1234567890" |
| RO_scaling_info = [ |
| { |
| "type": "create", |
| "count": 1, |
| "member-vnf-index": "1", |
| "osm_vdu_id": "dataVM", |
| } |
| ] |
| op_index_new_RO = self.my_ns._check_or_add_scale_suboperation( |
| db_nslcmop, vnf_index, None, None, "SCALE-RO", RO_nsr_id, RO_scaling_info |
| ) |
| self.assertEqual(op_index_new_RO, self.my_ns.SUBOPERATION_STATUS_NEW) |
| |
| # Use the same parameters again to match the already added RO sub-operation |
| op_index_existing_RO = self.my_ns._check_or_add_scale_suboperation( |
| db_nslcmop, vnf_index, None, None, "SCALE-RO", RO_nsr_id, RO_scaling_info |
| ) |
| self.assertTrue(op_index_existing_RO >= 0) |
| |
| # Change operationState 'manually' for this RO sub-operation |
| db_nslcmop["_admin"]["operations"][op_index_existing_RO][ |
| "operationState" |
| ] = "COMPLETED" |
| # Then use the same parameters again to match the already added sub-operation, |
| # which now has status 'COMPLETED' |
| # The expected return value is SUBOPERATION_STATUS_SKIP |
| op_index_skip_RO = self.my_ns._check_or_add_scale_suboperation( |
| db_nslcmop, vnf_index, None, None, "SCALE-RO", RO_nsr_id, RO_scaling_info |
| ) |
| self.assertEqual(op_index_skip_RO, self.my_ns.SUBOPERATION_STATUS_SKIP) |
| |
| async def test_deploy_kdus(self): |
| nsr_id = descriptors.test_ids["TEST-KDU"]["ns"] |
| nslcmop_id = descriptors.test_ids["TEST-KDU"]["instantiate"] |
| db_nsr = self.db.get_one("nsrs", {"_id": nsr_id}) |
| db_vnfr = self.db.get_one( |
| "vnfrs", {"nsr-id-ref": nsr_id, "member-vnf-index-ref": "multikdu"} |
| ) |
| db_vnfrs = {"multikdu": db_vnfr} |
| db_vnfd = self.db.get_one("vnfds", {"_id": db_vnfr["vnfd-id"]}) |
| db_vnfds = [db_vnfd] |
| task_register = {} |
| logging_text = "KDU" |
| self.my_ns.k8sclusterhelm3.generate_kdu_instance_name = asynctest.mock.Mock() |
| self.my_ns.k8sclusterhelm3.generate_kdu_instance_name.return_value = "k8s_id" |
| self.my_ns.k8sclusterhelm3.install = asynctest.CoroutineMock() |
| self.my_ns.k8sclusterhelm3.synchronize_repos = asynctest.CoroutineMock( |
| return_value=("", "") |
| ) |
| self.my_ns.k8sclusterhelm3.get_services = asynctest.CoroutineMock( |
| return_value=([]) |
| ) |
| await self.my_ns.deploy_kdus( |
| logging_text, nsr_id, nslcmop_id, db_vnfrs, db_vnfds, task_register |
| ) |
| await asyncio.wait(list(task_register.keys()), timeout=100) |
| db_nsr = self.db.get_list("nsrs")[1] |
| self.assertIn( |
| "K8s", |
| db_nsr["_admin"]["deployed"], |
| "K8s entry not created at '_admin.deployed'", |
| ) |
| self.assertIsInstance( |
| db_nsr["_admin"]["deployed"]["K8s"], list, "K8s entry is not of type list" |
| ) |
| self.assertEqual( |
| len(db_nsr["_admin"]["deployed"]["K8s"]), 2, "K8s entry is not of type list" |
| ) |
| k8s_instace_info = { |
| "kdu-instance": "k8s_id", |
| "k8scluster-uuid": "73d96432-d692-40d2-8440-e0c73aee209c", |
| "k8scluster-type": "helm-chart-v3", |
| "kdu-name": "ldap", |
| "member-vnf-index": "multikdu", |
| "namespace": None, |
| "kdu-deployment-name": None, |
| } |
| |
| nsr_result = copy.deepcopy(db_nsr["_admin"]["deployed"]["K8s"][0]) |
| nsr_kdu_model_result = nsr_result.pop("kdu-model") |
| expected_kdu_model = "stable/openldap:1.2.1" |
| self.assertEqual(nsr_result, k8s_instace_info) |
| self.assertTrue( |
| nsr_kdu_model_result in expected_kdu_model |
| or expected_kdu_model in nsr_kdu_model_result |
| ) |
| nsr_result = copy.deepcopy(db_nsr["_admin"]["deployed"]["K8s"][1]) |
| nsr_kdu_model_result = nsr_result.pop("kdu-model") |
| k8s_instace_info["kdu-name"] = "mongo" |
| expected_kdu_model = "stable/mongodb" |
| self.assertEqual(nsr_result, k8s_instace_info) |
| self.assertTrue( |
| nsr_kdu_model_result in expected_kdu_model |
| or expected_kdu_model in nsr_kdu_model_result |
| ) |
| |
| # async def test_instantiate_pdu(self): |
| # nsr_id = descriptors.test_ids["TEST-A"]["ns"] |
| # nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"] |
| # # Modify vnfd/vnfr to change KDU for PDU. Adding keys that NBI will already set |
| # self.db.set_one("vnfrs", {"nsr-id-ref": nsr_id, "member-vnf-index-ref": "1"}, |
| # update_dict={"ip-address": "10.205.1.46", |
| # "vdur.0.pdu-id": "53e1ec21-2464-451e-a8dc-6e311d45b2c8", |
| # "vdur.0.pdu-type": "PDU-TYPE-1", |
| # "vdur.0.ip-address": "10.205.1.46", |
| # }, |
| # unset={"vdur.status": None}) |
| # self.db.set_one("vnfrs", {"nsr-id-ref": nsr_id, "member-vnf-index-ref": "2"}, |
| # update_dict={"ip-address": "10.205.1.47", |
| # "vdur.0.pdu-id": "53e1ec21-2464-451e-a8dc-6e311d45b2c8", |
| # "vdur.0.pdu-type": "PDU-TYPE-1", |
| # "vdur.0.ip-address": "10.205.1.47", |
| # }, |
| # unset={"vdur.status": None}) |
| |
| # await self.my_ns.instantiate(nsr_id, nslcmop_id) |
| # db_nsr = self.db.get_one("nsrs", {"_id": nsr_id}) |
| # self.assertEqual(db_nsr.get("nsState"), "READY", str(db_nsr.get("errorDescription "))) |
| # self.assertEqual(db_nsr.get("currentOperation"), "IDLE", "currentOperation different than 'IDLE'") |
| # self.assertEqual(db_nsr.get("currentOperationID"), None, "currentOperationID different than None") |
| # self.assertEqual(db_nsr.get("errorDescription "), None, "errorDescription different than None") |
| # self.assertEqual(db_nsr.get("errorDetail"), None, "errorDetail different than None") |
| |
| # @asynctest.fail_on(active_handles=True) # all async tasks must be completed |
| # async def test_terminate_without_configuration(self): |
| # nsr_id = descriptors.test_ids["TEST-A"]["ns"] |
| # nslcmop_id = descriptors.test_ids["TEST-A"]["terminate"] |
| # # set instantiation task as completed |
| # self.db.set_list("nslcmops", {"nsInstanceId": nsr_id, "_id.ne": nslcmop_id}, |
| # update_dict={"operationState": "COMPLETED"}) |
| # self.db.set_one("nsrs", {"_id": nsr_id}, |
| # update_dict={"_admin.deployed.VCA.0": None, "_admin.deployed.VCA.1": None}) |
| |
| # await self.my_ns.terminate(nsr_id, nslcmop_id) |
| # db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) |
| # self.assertEqual(db_nslcmop.get("operationState"), 'COMPLETED', db_nslcmop.get("detailed-status")) |
| # db_nsr = self.db.get_one("nsrs", {"_id": nsr_id}) |
| # self.assertEqual(db_nsr.get("nsState"), "NOT_INSTANTIATED", str(db_nsr.get("errorDescription "))) |
| # self.assertEqual(db_nsr["_admin"].get("nsState"), "NOT_INSTANTIATED", str(db_nsr.get("errorDescription "))) |
| # self.assertEqual(db_nsr.get("currentOperation"), "IDLE", "currentOperation different than 'IDLE'") |
| # self.assertEqual(db_nsr.get("currentOperationID"), None, "currentOperationID different than None") |
| # self.assertEqual(db_nsr.get("errorDescription "), None, "errorDescription different than None") |
| # self.assertEqual(db_nsr.get("errorDetail"), None, "errorDetail different than None") |
| # db_vnfrs_list = self.db.get_list("vnfrs", {"nsr-id-ref": nsr_id}) |
| # for vnfr in db_vnfrs_list: |
| # self.assertEqual(vnfr["_admin"].get("nsState"), "NOT_INSTANTIATED", "Not instantiated") |
| |
| # @asynctest.fail_on(active_handles=True) # all async tasks must be completed |
| # async def test_terminate_primitive(self): |
| # nsr_id = descriptors.test_ids["TEST-A"]["ns"] |
| # nslcmop_id = descriptors.test_ids["TEST-A"]["terminate"] |
| # # set instantiation task as completed |
| # self.db.set_list("nslcmops", {"nsInstanceId": nsr_id, "_id.ne": nslcmop_id}, |
| # update_dict={"operationState": "COMPLETED"}) |
| |
| # # modify vnfd descriptor to include terminate_primitive |
| # terminate_primitive = [{ |
| # "name": "touch", |
| # "parameter": [{"name": "filename", "value": "terminate_filename"}], |
| # "seq": '1' |
| # }] |
| # db_vnfr = self.db.get_one("vnfrs", {"nsr-id-ref": nsr_id, "member-vnf-index-ref": "1"}) |
| # self.db.set_one("vnfds", {"_id": db_vnfr["vnfd-id"]}, |
| # {"vnf-configuration.0.terminate-config-primitive": terminate_primitive}) |
| |
| # await self.my_ns.terminate(nsr_id, nslcmop_id) |
| # db_nslcmop = self.db.get_one("nslcmops", {"_id": nslcmop_id}) |
| # self.assertEqual(db_nslcmop.get("operationState"), 'COMPLETED', db_nslcmop.get("detailed-status")) |
| # db_nsr = self.db.get_one("nsrs", {"_id": nsr_id}) |
| # self.assertEqual(db_nsr.get("nsState"), "NOT_INSTANTIATED", str(db_nsr.get("errorDescription "))) |
| # self.assertEqual(db_nsr["_admin"].get("nsState"), "NOT_INSTANTIATED", str(db_nsr.get("errorDescription "))) |
| # self.assertEqual(db_nsr.get("currentOperation"), "IDLE", "currentOperation different than 'IDLE'") |
| # self.assertEqual(db_nsr.get("currentOperationID"), None, "currentOperationID different than None") |
| # self.assertEqual(db_nsr.get("errorDescription "), None, "errorDescription different than None") |
| # self.assertEqual(db_nsr.get("errorDetail"), None, "errorDetail different than None") |
| |
| |
| if __name__ == "__main__": |
| asynctest.main() |