operationState: FAILED
startTime: 1566823354.414689
statusEnteredTime: 1566824534.5112448
+
+- _admin:
+ created: 1600000000.0000000
+ modified: 1600000000.0000000
+ projects_read:
+ - 25b5aebf-3da1-49ed-99de-1d2b4a86d6e4
+ projects_write:
+ - 25b5aebf-3da1-49ed-99de-1d2b4a86d6e4
+ worker: 86434c2948e2
+ _id: a639fac7-e0bb-4225-ffff-c1f8efcc125e
+ detailed-status: None
+ lcmOperationType: terminate
+ nsInstanceId: f48163a6-c807-47bc-9682-f72caef5af85
+ operationParams: {}
+ operationState: PROCESSING
+ startTime: 1600000000.0000000
+ statusEnteredTime: 1600000000.0000000
+
- _admin:
created: 1575034637.044651
modified: 1575034637.044651
- 7ab0d10d-8ce2-4c68-aef6-cc5a437a9c62
"""
-db_ro_ns_text = """
+ro_ns_text = """
datacenter_tenant_id: dc5c67fa-c7f2-11e9-b9c0-02420aff0004
description: null
name: ALF
vnfd_osm_id: f48163a6-c807-47bc-9682-f72caef5af85.1.2
"""
+ro_delete_action_text = """
+actions:
+- created_at: 1580140763.1099188
+ description: DELETE
+ instance_id: c9fe9908-3180-430d-b633-fca2f68db008
+ modified_at: 1580140763.253148
+ number_done: 1
+ number_failed: 0
+ number_tasks: 1
+ tenant_id: 0ea38bd0-2729-47a9-ae07-c6ce76115eb2
+ uuid: delete
+ vim_wim_actions:
+ - action: DELETE
+ created_at: 1580140763.1099188
+ datacenter_vim_id: dc5c67fa-c7f2-11e9-b9c0-02420aff0004
+ error_msg: null
+ extra: '{params: [9b6a2ac4-767e-4ec9-9497-8ba63084c77f, null]}'
+ instance_action_id: ACTION-1580140763.054037
+ item: instance_nets
+ item_id: 8cb06b72-c71d-4b58-b419-95025fa651d3
+ modified_at: 1580140763.1099188
+ related: 8cb06b72-c71d-4b58-b419-95025fa651d3
+ status: SUPERSEDED
+ task_index: 0
+ vim_id: null
+ wim_account_id: null
+ wim_internal_id: null
+ worker: null
+"""
+
db_vim_accounts_text = """
---
- _admin:
operationState: COMPLETED
detailed-status: done
"""
+
+test_ids = {
+ # contains the ids of ns and operations of every test
+ "TEST-A": {
+ "ns": "f48163a6-c807-47bc-9682-f72caef5af85",
+ "instantiate": "a639fac7-e0bb-4225-8ecb-c1f8efcc125e",
+ "terminate": "a639fac7-e0bb-4225-ffff-c1f8efcc125e",
+ },
+ "TEST-KDU": {
+ "ns": "0bcb701c-ee4d-41ab-8ee6-f4156f7f114d",
+ "instantiate": "cf3aa178-7640-4174-b921-2330e6f2aad6",
+ "terminate": None,
+ }
+}
import asynctest # pip3 install asynctest --user
import asyncio
import yaml
-# import logging
from os import getenv
-from osm_lcm.ns import NsLcm
+from osm_lcm import ns
from osm_common.dbmemory import DbMemory
from osm_common.msgkafka import MsgKafka
from osm_common.fslocal import FsLocal
from osm_lcm.lcm_utils import TaskRegistry
-from n2vc.vnf import N2VC
-# from n2vc.k8s_helm_conn import K8sHelmConnector
+# from osm_lcm.ROclient import ROClient
from uuid import uuid4
-from asynctest.mock import patch
+# from asynctest.mock import patch
from osm_lcm.tests import test_db_descriptors as descriptors
return ee_id, {}
def _ro_show(self, *args, **kwargs):
- ro_ns_desc = yaml.load(descriptors.db_ro_ns_text, Loader=yaml.Loader)
+ if kwargs.get("delete"):
+ ro_ns_desc = yaml.load(descriptors.ro_delete_action_text, Loader=yaml.Loader)
+ while True:
+ yield ro_ns_desc
+
+ ro_ns_desc = yaml.load(descriptors.ro_ns_text, Loader=yaml.Loader)
# if ip address provided, replace descriptor
ip_addresses = getenv("OSMLCMTEST_NS_IPADDRESS", "")
def _return_uuid(self, *args, **kwargs):
return str(uuid4())
- @patch("osm_lcm.ns.N2VCJujuConnector")
- @patch("osm_lcm.ns.K8sHelmConnector")
- async def setUp(self, k8s_mock, n2vc_mock):
+ async def setUp(self):
+
# Mock DB
if not getenv("OSMLCMTEST_DB_NOMOCK"):
self.db = DbMemory()
self.db.create_list("k8sclusters", yaml.load(descriptors.db_k8sclusters_text, Loader=yaml.Loader))
self.db.create_list("nslcmops", yaml.load(descriptors.db_nslcmops_text, Loader=yaml.Loader))
self.db.create_list("vnfrs", yaml.load(descriptors.db_vnfrs_text, Loader=yaml.Loader))
-
self.db_vim_accounts = yaml.load(descriptors.db_vim_accounts_text, Loader=yaml.Loader)
# Mock kafka
self.lcm_tasks.waitfor_related_HA.return_value = None
self.lcm_tasks.lookfor_related.return_value = ("", [])
+ # Mock VCA - K8s
+ if not getenv("OSMLCMTEST_VCA_K8s_NOMOCK"):
+ ns.K8sJujuConnector = asynctest.MagicMock(ns.K8sJujuConnector)
+ ns.K8sHelmConnector = asynctest.MagicMock(ns.K8sHelmConnector)
+
+ if not getenv("OSMLCMTEST_VCA_NOMOCK"):
+ ns.N2VCJujuConnector = asynctest.MagicMock(ns.N2VCJujuConnector)
+
# Create NsLCM class
- self.my_ns = NsLcm(self.db, self.msg, self.fs, self.lcm_tasks, lcm_config, self.loop)
+ self.my_ns = ns.NsLcm(self.db, self.msg, self.fs, self.lcm_tasks, lcm_config, self.loop)
self.my_ns._wait_dependent_n2vc = asynctest.CoroutineMock()
# Mock logging
# Mock VCA - N2VC
if not getenv("OSMLCMTEST_VCA_NOMOCK"):
pub_key = getenv("OSMLCMTEST_NS_PUBKEY", "ssh-rsa test-pub-key t@osm.com")
- self.my_ns.n2vc = asynctest.Mock(N2VC())
+ # self.my_ns.n2vc = asynctest.Mock(N2VC())
self.my_ns.n2vc.GetPublicKey.return_value = getenv("OSMLCM_VCA_PUBKEY", "public_key")
# allow several versions of n2vc
self.my_ns.n2vc.FormatApplicationName = asynctest.Mock(side_effect=self._n2vc_FormatApplicationName())
"pubkey": pub_key})
self.my_ns.n2vc.get_public_key = asynctest.CoroutineMock(
return_value=getenv("OSMLCM_VCA_PUBKEY", "public_key"))
-
- # # Mock VCA - K8s
- # if not getenv("OSMLCMTEST_VCA_K8s_NOMOCK"):
- # pub_key = getenv("OSMLCMTEST_NS_PUBKEY", "ssh-rsa test-pub-key t@osm.com")
- # self.my_ns.k8sclusterhelm = asynctest.Mock(K8sHelmConnector())
+ self.my_ns.n2vc.delete_namespace = asynctest.CoroutineMock(return_value=None)
# Mock RO
if not getenv("OSMLCMTEST_RO_NOMOCK"):
self.my_ns.RO.create_action = asynctest.CoroutineMock(self.my_ns.RO.create_action,
return_value={"vm-id": {"vim_result": 200,
"description": "done"}})
+ self.my_ns.RO.delete = asynctest.CoroutineMock(self.my_ns.RO.delete, return_value={"action_id": "del"})
# self.my_ns.wait_vm_up_insert_key_ro = asynctest.CoroutineMock(return_value="ip-address")
@asynctest.fail_on(active_handles=True) # all async tasks must be completed
async def test_instantiate(self):
self.db.set_one = asynctest.Mock()
- nsr_id = self.db.get_list("nsrs")[0]["_id"]
- nslcmop_id = self.db.get_list("nslcmops")[0]["_id"]
+ nsr_id = descriptors.test_ids["TEST-A"]["ns"]
+ nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"]
# print("Test instantiate started")
# delete deployed information of database
# scale-out/scale-in operations with success/error result
# Test scale() with missing 'scaleVnfData', should return operationState = 'FAILED'
- nsr_id = self.db.get_list("nsrs")[0]["_id"]
- nslcmop_id = self.db.get_list("nslcmops")[0]["_id"]
+ nsr_id = descriptors.test_ids["TEST-A"]["ns"]
+ nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"]
await self.my_ns.scale(nsr_id, nslcmop_id)
expected_value = 'FAILED'
return_value = self.db.get_one("nslcmops", {"_id": nslcmop_id}).get("operationState")
# - if marked as anything but 'COMPLETED', the suboperation index is expected
def test_scale_reintent_or_skip_suboperation(self):
# Load an alternative 'nslcmops' YAML for this test
- db_nslcmop = self.db.get_list('nslcmops')[0]
+ nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"]
+ db_nslcmop = self.db.get_one('nslcmops', {"_id": nslcmop_id})
op_index = 2
# Test when 'operationState' is 'COMPLETED'
db_nslcmop['_admin']['operations'][op_index]['operationState'] = 'COMPLETED'
# Expected result: index of the found sub-operation, or SUBOPERATION_STATUS_NOT_FOUND if not found
def test_scale_find_suboperation(self):
# Load an alternative 'nslcmops' YAML for this test
- db_nslcmop = self.db.get_list('nslcmops')[0]
+ nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"]
+ db_nslcmop = self.db.get_one('nslcmops', {"_id": nslcmop_id})
# Find this sub-operation
op_index = 2
vnf_index = db_nslcmop['_admin']['operations'][op_index]['member_vnf_index']
# Test _update_suboperation_status()
def test_scale_update_suboperation_status(self):
self.db.set_one = asynctest.Mock()
- db_nslcmop = self.db.get_list('nslcmops')[0]
+ nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"]
+ db_nslcmop = self.db.get_one('nslcmops', {"_id": nslcmop_id})
op_index = 0
# Force the initial values to be distinct from the updated ones
q_filter = {"_id": db_nslcmop["_id"]}
fail_on_empty=False)
def test_scale_add_suboperation(self):
- db_nslcmop = self.db.get_list('nslcmops')[0]
+ nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"]
+ db_nslcmop = self.db.get_one('nslcmops', {"_id": nslcmop_id})
vnf_index = '1'
num_ops_before = len(db_nslcmop.get('_admin', {}).get('operations', [])) - 1
vdu_id = None
# - op_index (non-negative number): This is an existing sub-operation, operationState != 'COMPLETED'
# - SUBOPERATION_STATUS_SKIP: This is an existing sub-operation, operationState == 'COMPLETED'
def test_scale_check_or_add_scale_suboperation(self):
- db_nslcmop = self.db.get_list('nslcmops')[0]
+ nslcmop_id = descriptors.test_ids["TEST-A"]["instantiate"]
+ db_nslcmop = self.db.get_one('nslcmops', {"_id": nslcmop_id})
operationType = 'PRE-SCALE'
vnf_index = '1'
primitive = 'touch'
self.assertEqual(op_index_skip_RO, self.my_ns.SUBOPERATION_STATUS_SKIP)
async def test_deploy_kdus(self):
- db_nsr = self.db.get_list("nsrs")[1]
- db_vnfr = self.db.get_list("vnfrs")[2]
+ nsr_id = descriptors.test_ids["TEST-KDU"]["ns"]
+ # nslcmop_id = descriptors.test_ids["TEST-KDU"]["instantiate"]
+ db_nsr = self.db.get_one("nsrs", {"_id": nsr_id})
+ db_vnfr = self.db.get_one("vnfrs", {"nsr-id-ref": nsr_id, "member-vnf-index-ref": "multikdu"})
db_vnfrs = {"multikdu": db_vnfr}
- db_vnfd = self.db.get_list("vnfds")[1]
+ db_vnfd = self.db.get_one("vnfds", {"_id": db_vnfr["vnfd-id"]})
db_vnfds = {db_vnfd["_id"]: db_vnfd}
- nsr_id = db_nsr["_id"]
- # nslcmop_id = self.db.get_list("nslcmops")[1]["_id"]
logging_text = "KDU"
self.my_ns.k8sclusterhelm.install = asynctest.CoroutineMock(return_value="k8s_id")
await self.my_ns.deploy_kdus(logging_text, nsr_id, db_nsr, db_vnfrs, db_vnfds)