Console Output
[LCM-stage_2-merge_v13.0] Running shell script
+ runuser jenkins -c devops-stages/stage-test.sh
Launching tox
✔ OK flake8 in 23.062 seconds
flake8 create: /tmp/.tox/flake8
flake8 installdeps: flake8==5.0.4
flake8 develop-inst: /home/jenkins/workspace/LCM-stage_2-merge_v13.0
flake8 installed: flake8==5.0.4,mccabe==0.7.0,-e git+https://osm.etsi.org/gerrit/osm/LCM.git@c96cb8bb9cef9b19c51baefe6a67c22fa8a71830#egg=osm_lcm,pycodestyle==2.9.1,pyflakes==2.5.0
flake8 run-test-pre: PYTHONHASHSEED='3576291598'
flake8 run-test: commands[0] | flake8 osm_lcm/ setup.py
ERROR: invocation failed (exit code 1), logfile: /tmp/.tox/black/log/black-0.log
================================== log start ===================================
black create: /tmp/.tox/black
black installdeps: black
black installed: black==24.1.1,click==8.1.7,mypy-extensions==1.0.0,packaging==23.2,pathspec==0.12.1,platformdirs==4.1.0,tomli==2.0.1,typing_extensions==4.9.0
black run-test-pre: PYTHONHASHSEED='1218391369'
black run-test: commands[0] | black --check --diff osm_lcm/
--- /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/frontend_pb2.py 2024-01-26 09:50:35.644553+00:00
+++ /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/frontend_pb2.py 2024-01-28 09:51:15.636820+00:00
@@ -237,44 +237,44 @@
PrimitiveRequest = _reflection.GeneratedProtocolMessageType(
"PrimitiveRequest",
(_message.Message,),
{
"DESCRIPTOR": _PRIMITIVEREQUEST,
- "__module__": "osm_lcm.frontend_pb2"
+ "__module__": "osm_lcm.frontend_pb2",
# @@protoc_insertion_point(class_scope:osm_ee.PrimitiveRequest)
},
)
_sym_db.RegisterMessage(PrimitiveRequest)
PrimitiveReply = _reflection.GeneratedProtocolMessageType(
"PrimitiveReply",
(_message.Message,),
{
"DESCRIPTOR": _PRIMITIVEREPLY,
- "__module__": "osm_lcm.frontend_pb2"
+ "__module__": "osm_lcm.frontend_pb2",
# @@protoc_insertion_point(class_scope:osm_ee.PrimitiveReply)
},
)
_sym_db.RegisterMessage(PrimitiveReply)
SshKeyRequest = _reflection.GeneratedProtocolMessageType(
"SshKeyRequest",
(_message.Message,),
{
"DESCRIPTOR": _SSHKEYREQUEST,
- "__module__": "osm_lcm.frontend_pb2"
+ "__module__": "osm_lcm.frontend_pb2",
# @@protoc_insertion_point(class_scope:osm_ee.SshKeyRequest)
},
)
_sym_db.RegisterMessage(SshKeyRequest)
SshKeyReply = _reflection.GeneratedProtocolMessageType(
"SshKeyReply",
(_message.Message,),
{
"DESCRIPTOR": _SSHKEYREPLY,
- "__module__": "osm_lcm.frontend_pb2"
+ "__module__": "osm_lcm.frontend_pb2",
# @@protoc_insertion_point(class_scope:osm_ee.SshKeyReply)
},
)
_sym_db.RegisterMessage(SshKeyReply)
would reformat /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/frontend_pb2.py
--- /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/lcm.py 2024-01-26 09:50:35.644553+00:00
+++ /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/lcm.py 2024-01-28 09:51:16.203866+00:00
@@ -95,15 +95,13 @@
self.main_config.load_from_env()
self.logger.critical("Loaded configuration:" + str(self.main_config.to_dict()))
# TODO: check if lcm_hc.py is necessary
self.health_check_file = get_health_check_file(self.main_config.to_dict())
self.loop = loop or asyncio.get_event_loop()
- self.ns = (
- self.netslice
- ) = (
- self.vim
- ) = self.wim = self.sdn = self.k8scluster = self.vca = self.k8srepo = None
+ self.ns = self.netslice = self.vim = self.wim = self.sdn = self.k8scluster = (
+ self.vca
+ ) = self.k8srepo = None
# logging
log_format_simple = (
"%(asctime)s %(levelname)s %(name)s %(filename)s:%(lineno)s %(message)s"
)
would reformat /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/lcm.py
--- /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/lcm_utils.py 2024-01-26 09:50:35.644553+00:00
+++ /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/lcm_utils.py 2024-01-28 09:51:16.203747+00:00
@@ -125,23 +125,27 @@
if base_folder.get("pkg-dir"):
artifact_path = "{}/{}/{}/{}".format(
base_folder["folder"].split(":")[0] + extension,
base_folder["pkg-dir"],
- "charms"
- if charm_type in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
- else "helm-charts",
+ (
+ "charms"
+ if charm_type in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
+ else "helm-charts"
+ ),
charm_name,
)
else:
# For SOL004 packages
artifact_path = "{}/Scripts/{}/{}".format(
base_folder["folder"].split(":")[0] + extension,
- "charms"
- if charm_type in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
- else "helm-charts",
+ (
+ "charms"
+ if charm_type in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
+ else "helm-charts"
+ ),
charm_name,
)
return artifact_path
would reformat /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/lcm_utils.py
--- /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/netslice.py 2024-01-26 09:50:35.648553+00:00
+++ /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/netslice.py 2024-01-28 09:51:16.602934+00:00
@@ -504,13 +504,13 @@
raise LcmException("Timeout waiting nsi to be ready.")
db_nsir_update["operational-status"] = "running"
db_nsir_update["detailed-status"] = "done"
db_nsir_update["config-status"] = "configured"
- db_nsilcmop_update[
- "operationState"
- ] = nsilcmop_operation_state = "COMPLETED"
+ db_nsilcmop_update["operationState"] = nsilcmop_operation_state = (
+ "COMPLETED"
+ )
db_nsilcmop_update["statusEnteredTime"] = time()
db_nsilcmop_update["detailed-status"] = "done"
return
except (LcmException, DbException) as e:
@@ -538,13 +538,13 @@
db_nsir_update["config-status"] = "configured"
if db_nsilcmop:
db_nsilcmop_update["detailed-status"] = "FAILED {}: {}".format(
step, exc
)
- db_nsilcmop_update[
- "operationState"
- ] = nsilcmop_operation_state = "FAILED"
+ db_nsilcmop_update["operationState"] = nsilcmop_operation_state = (
+ "FAILED"
+ )
db_nsilcmop_update["statusEnteredTime"] = time()
try:
if db_nsir:
db_nsir_update["_admin.nsilcmop"] = None
self.update_db_2("nsis", nsir_id, db_nsir_update)
@@ -736,16 +736,16 @@
RO_nsir_id = RO_delete_action = None
for nsir_deployed_RO in get_iterable(nsir_deployed, "RO"):
RO_nsir_id = nsir_deployed_RO.get("netslice_scenario_id")
try:
if not self.ro_config["ng"]:
- step = db_nsir_update[
- "detailed-status"
- ] = "Deleting netslice-vld at RO"
- db_nsilcmop_update[
- "detailed-status"
- ] = "Deleting netslice-vld at RO"
+ step = db_nsir_update["detailed-status"] = (
+ "Deleting netslice-vld at RO"
+ )
+ db_nsilcmop_update["detailed-status"] = (
+ "Deleting netslice-vld at RO"
+ )
self.logger.debug(logging_text + step)
desc = await RO.delete("ns", RO_nsir_id)
RO_delete_action = desc["action_id"]
nsir_deployed_RO["vld_delete_action_id"] = RO_delete_action
nsir_deployed_RO["vld_status"] = "DELETING"
@@ -781,21 +781,21 @@
db_nsir_update["operational-status"] = "failed"
db_nsir_update["detailed-status"] = "Deletion errors " + "; ".join(
failed_detail
)
db_nsilcmop_update["detailed-status"] = "; ".join(failed_detail)
- db_nsilcmop_update[
- "operationState"
- ] = nsilcmop_operation_state = "FAILED"
+ db_nsilcmop_update["operationState"] = nsilcmop_operation_state = (
+ "FAILED"
+ )
db_nsilcmop_update["statusEnteredTime"] = time()
else:
db_nsir_update["operational-status"] = "terminating"
db_nsir_update["config-status"] = "terminating"
db_nsir_update["_admin.nsiState"] = "NOT_INSTANTIATED"
- db_nsilcmop_update[
- "operationState"
- ] = nsilcmop_operation_state = "COMPLETED"
+ db_nsilcmop_update["operationState"] = nsilcmop_operation_state = (
+ "COMPLETED"
+ )
db_nsilcmop_update["statusEnteredTime"] = time()
if db_nsilcmop["operationParams"].get("autoremove"):
autoremove = True
db_nsir_update["detailed-status"] = "done"
@@ -830,13 +830,13 @@
db_nsir_update["operational-status"] = "failed"
if db_nsilcmop:
db_nsilcmop_update["detailed-status"] = "FAILED {}: {}".format(
step, exc
)
- db_nsilcmop_update[
- "operationState"
- ] = nsilcmop_operation_state = "FAILED"
+ db_nsilcmop_update["operationState"] = nsilcmop_operation_state = (
+ "FAILED"
+ )
db_nsilcmop_update["statusEnteredTime"] = time()
try:
if db_nsir:
db_nsir_update["_admin.deployed"] = nsir_deployed
db_nsir_update["_admin.nsilcmop"] = None
would reformat /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/netslice.py
--- /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/vim_sdn.py 2024-01-26 09:50:35.652554+00:00
+++ /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/vim_sdn.py 2024-01-28 09:51:17.325324+00:00
@@ -1221,13 +1221,13 @@
)
db_k8scluster_update[
"_admin.{}.error_msg".format(task_name)
] = None
db_k8scluster_update["_admin.{}.id".format(task_name)] = k8s_id
- db_k8scluster_update[
- "_admin.{}.created".format(task_name)
- ] = uninstall_sw
+ db_k8scluster_update["_admin.{}.created".format(task_name)] = (
+ uninstall_sw
+ )
db_k8scluster_update[
"_admin.{}.operationalState".format(task_name)
] = "ENABLED"
# update database
step = "Updating database for " + task_name
@@ -1347,13 +1347,13 @@
)
cluster_removed = await self.helm3_k8scluster.reset(
cluster_uuid=k8s_h3c_id, uninstall_sw=uninstall_sw
)
db_k8scluster_update["_admin.helm-chart-v3.id"] = None
- db_k8scluster_update[
- "_admin.helm-chart-v3.operationalState"
- ] = "DISABLED"
+ db_k8scluster_update["_admin.helm-chart-v3.operationalState"] = (
+ "DISABLED"
+ )
# Try to remove from cluster_inserted to clean old versions
if k8s_hc_id and cluster_removed:
step = "Removing k8scluster='{}' from k8srepos".format(k8scluster_id)
self.logger.debug(logging_text + step)
would reformat /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/vim_sdn.py
--- /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/ns.py 2024-01-26 09:50:35.648553+00:00
+++ /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/ns.py 2024-01-28 09:51:22.348622+00:00
@@ -671,13 +671,13 @@
if "status" not in vdur:
vdur["status"] = "ERROR"
vnfr_update["vdur.{}.status".format(vdu_index)] = "ERROR"
if error_text:
vdur["status-detailed"] = str(error_text)
- vnfr_update[
- "vdur.{}.status-detailed".format(vdu_index)
- ] = "ERROR"
+ vnfr_update["vdur.{}.status-detailed".format(vdu_index)] = (
+ "ERROR"
+ )
self.update_db_2("vnfrs", db_vnfr["_id"], vnfr_update)
except DbException as e:
self.logger.error("Cannot update vnf. {}".format(e))
def ns_update_vnfr(self, db_vnfrs, nsr_desc_RO):
@@ -875,13 +875,13 @@
target_vld["vim_info"],
(other_target_vim, param.replace("-", "_")),
vim_net,
)
else: # isinstance str
- target_vld["vim_info"][target_vim][
- param.replace("-", "_")
- ] = vld_params[param]
+ target_vld["vim_info"][target_vim][param.replace("-", "_")] = (
+ vld_params[param]
+ )
if vld_params.get("common_id"):
target_vld["common_id"] = vld_params.get("common_id")
# modify target["ns"]["vld"] with instantiation parameters to override vnf vim-account
def update_ns_vld_target(target, ns_params):
@@ -1141,13 +1141,13 @@
if "cidr" in ip_profile_source_data:
ip_profile_dest_data["subnet-address"] = ip_profile_source_data[
"cidr"
]
if "gateway-ip" in ip_profile_source_data:
- ip_profile_dest_data[
- "gateway-address"
- ] = ip_profile_source_data["gateway-ip"]
+ ip_profile_dest_data["gateway-address"] = (
+ ip_profile_source_data["gateway-ip"]
+ )
if "dhcp-enabled" in ip_profile_source_data:
ip_profile_dest_data["dhcp-params"] = {
"enabled": ip_profile_source_data["dhcp-enabled"]
}
@@ -1829,23 +1829,27 @@
# Get artifact path
if base_folder["pkg-dir"]:
artifact_path = "{}/{}/{}/{}".format(
base_folder["folder"],
base_folder["pkg-dir"],
- "charms"
- if vca_type
- in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
- else "helm-charts",
+ (
+ "charms"
+ if vca_type
+ in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
+ else "helm-charts"
+ ),
vca_name,
)
else:
artifact_path = "{}/Scripts/{}/{}/".format(
base_folder["folder"],
- "charms"
- if vca_type
- in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
- else "helm-charts",
+ (
+ "charms"
+ if vca_type
+ in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
+ else "helm-charts"
+ ),
vca_name,
)
self.logger.debug("Artifact path > {}".format(artifact_path))
@@ -2253,13 +2257,13 @@
:param other_update: Other required changes at database if provided, will be cleared
:return:
"""
try:
db_dict = other_update or {}
- db_dict[
- "_admin.nslcmop"
- ] = current_operation_id # for backward compatibility
+ db_dict["_admin.nslcmop"] = (
+ current_operation_id # for backward compatibility
+ )
db_dict["_admin.current-operation"] = current_operation_id
db_dict["_admin.operation-type"] = (
current_operation if current_operation != "IDLE" else None
)
db_dict["currentOperation"] = current_operation
@@ -2337,13 +2341,13 @@
db_path = "configurationStatus.{}.".format(vca_index)
db_dict = other_update or {}
if status:
db_dict[db_path + "status"] = status
if element_under_configuration:
- db_dict[
- db_path + "elementUnderConfiguration"
- ] = element_under_configuration
+ db_dict[db_path + "elementUnderConfiguration"] = (
+ element_under_configuration
+ )
if element_type:
db_dict[db_path + "elementType"] = element_type
self.update_db_2("nsrs", nsr_id, db_dict)
except DbException as e:
self.logger.warn(
@@ -3380,13 +3384,13 @@
# Mgmt service found, Obtain service ip
ip = service.get("external_ip", service.get("cluster_ip"))
if isinstance(ip, list) and len(ip) == 1:
ip = ip[0]
- vnfr_update_dict[
- "kdur.{}.ip-address".format(kdu_index)
- ] = ip
+ vnfr_update_dict["kdur.{}.ip-address".format(kdu_index)] = (
+ ip
+ )
# Check if must update also mgmt ip at the vnf
service_external_cp = mgmt_service.get(
"external-connection-point-ref"
)
@@ -3516,13 +3520,13 @@
k8s_credentials, reuse_cluster_uuid=cluster_id
)
db_k8scluster_update = {}
db_k8scluster_update["_admin.helm-chart-v3.error_msg"] = None
db_k8scluster_update["_admin.helm-chart-v3.id"] = k8s_id
- db_k8scluster_update[
- "_admin.helm-chart-v3.created"
- ] = uninstall_sw
+ db_k8scluster_update["_admin.helm-chart-v3.created"] = (
+ uninstall_sw
+ )
db_k8scluster_update[
"_admin.helm-chart-v3.operationalState"
] = "ENABLED"
self.update_db_2(
"k8sclusters", cluster_id, db_k8scluster_update
@@ -3891,14 +3895,13 @@
nsr_id,
nslcmop_id,
"instantiate_N2VC-{}".format(vca_index),
task_n2vc,
)
- task_instantiation_info[
- task_n2vc
- ] = self.task_name_deploy_vca + " {}.{}".format(
- member_vnf_index or "", vdu_id or ""
+ task_instantiation_info[task_n2vc] = (
+ self.task_name_deploy_vca
+ + " {}.{}".format(member_vnf_index or "", vdu_id or "")
)
@staticmethod
def _create_nslcmop(nsr_id, operation, params):
"""
@@ -4281,13 +4284,13 @@
self.logger.debug(logging_text + stage[2])
self.update_db_2("nsrs", nsr_id, db_nsr_update)
self._write_op_status(nslcmop_id, stage)
desc = await self.RO.delete("ns", ro_nsr_id)
ro_delete_action = desc["action_id"]
- db_nsr_update[
- "_admin.deployed.RO.nsr_delete_action_id"
- ] = ro_delete_action
+ db_nsr_update["_admin.deployed.RO.nsr_delete_action_id"] = (
+ ro_delete_action
+ )
db_nsr_update["_admin.deployed.RO.nsr_id"] = None
db_nsr_update["_admin.deployed.RO.nsr_status"] = "DELETED"
if ro_delete_action:
# wait until NS is deleted from VIM
stage[2] = "Waiting ns deleted from VIM."
@@ -4404,14 +4407,14 @@
for index, vnf_deployed in enumerate(nsr_deployed["RO"]["vnfd"]):
if not vnf_deployed or not vnf_deployed["id"]:
continue
try:
ro_vnfd_id = vnf_deployed["id"]
- stage[
- 2
- ] = "Deleting member_vnf_index={} ro_vnfd_id={} from RO.".format(
- vnf_deployed["member-vnf-index"], ro_vnfd_id
+ stage[2] = (
+ "Deleting member_vnf_index={} ro_vnfd_id={} from RO.".format(
+ vnf_deployed["member-vnf-index"], ro_vnfd_id
+ )
)
db_nsr_update["detailed-status"] = " ".join(stage)
self.update_db_2("nsrs", nsr_id, db_nsr_update)
self._write_op_status(nslcmop_id, stage)
await self.RO.delete("vnfd", ro_vnfd_id)
@@ -4421,13 +4424,13 @@
db_nsr_update["_admin.deployed.RO.vnfd.{}.id".format(index)] = None
except Exception as e:
if (
isinstance(e, ROclient.ROClientException) and e.http_code == 404
): # not found
- db_nsr_update[
- "_admin.deployed.RO.vnfd.{}.id".format(index)
- ] = None
+ db_nsr_update["_admin.deployed.RO.vnfd.{}.id".format(index)] = (
+ None
+ )
self.logger.debug(
logging_text
+ "ro_vnfd_id={} already deleted ".format(ro_vnfd_id)
)
elif (
@@ -4515,13 +4518,13 @@
for vnfr in db_vnfrs_list:
vnfd_id = vnfr["vnfd-id"]
if vnfd_id not in db_vnfds_from_id:
vnfd = self.db.get_one("vnfds", {"_id": vnfd_id})
db_vnfds_from_id[vnfd_id] = vnfd
- db_vnfds_from_member_index[
- vnfr["member-vnf-index-ref"]
- ] = db_vnfds_from_id[vnfd_id]
+ db_vnfds_from_member_index[vnfr["member-vnf-index-ref"]] = (
+ db_vnfds_from_id[vnfd_id]
+ )
# Destroy individual execution environments when there are terminating primitives.
# Rest of EE will be deleted at once
# TODO - check before calling _destroy_N2VC
# if not operation_params.get("skip_terminate_primitives"):#
@@ -4533,13 +4536,15 @@
for vca_index, vca in enumerate(get_iterable(nsr_deployed, "VCA")):
config_descriptor = None
vca_member_vnf_index = vca.get("member-vnf-index")
vca_id = self.get_vca_id(
- db_vnfrs_dict.get(vca_member_vnf_index)
- if vca_member_vnf_index
- else None,
+ (
+ db_vnfrs_dict.get(vca_member_vnf_index)
+ if vca_member_vnf_index
+ else None
+ ),
db_nsr,
)
if not vca or not vca.get("ee_id"):
continue
if not vca.get("member-vnf-index"):
@@ -4644,13 +4649,13 @@
+ "Unknown k8s deployment type {}".format(
kdu.get("k8scluster-type")
)
)
continue
- tasks_dict_info[
- task_delete_kdu_instance
- ] = "Terminating KDU '{}'".format(kdu.get("kdu-name"))
+ tasks_dict_info[task_delete_kdu_instance] = (
+ "Terminating KDU '{}'".format(kdu.get("kdu-name"))
+ )
# remove from RO
stage[1] = "Deleting ns from VIM."
if self.ro_config.ng:
task_delete_ro = asyncio.ensure_future(
@@ -5407,15 +5412,13 @@
logging_text + "Exit Exception {} {}".format(type(e).__name__, e),
exc_info=True,
)
finally:
if exc:
- db_nslcmop_update[
- "detailed-status"
- ] = (
- detailed_status
- ) = error_description_nslcmop = "FAILED {}: {}".format(step, exc)
+ db_nslcmop_update["detailed-status"] = detailed_status = (
+ error_description_nslcmop
+ ) = "FAILED {}: {}".format(step, exc)
nslcmop_operation_state = "FAILED"
if db_nsr:
self._write_ns_status(
nsr_id=nsr_id,
ns_state=db_nsr[
@@ -6238,15 +6241,13 @@
logging_text + "Exit Exception {} {}".format(type(e).__name__, e),
exc_info=True,
)
finally:
if exc:
- db_nslcmop_update[
- "detailed-status"
- ] = (
- detailed_status
- ) = error_description_nslcmop = "FAILED {}: {}".format(step, exc)
+ db_nslcmop_update["detailed-status"] = detailed_status = (
+ error_description_nslcmop
+ ) = "FAILED {}: {}".format(step, exc)
nslcmop_operation_state = "FAILED"
db_nsr_update["operational-status"] = old_operational_status
if db_nsr:
self._write_ns_status(
nsr_id=nsr_id,
@@ -6731,14 +6732,14 @@
and scaling_type == "SCALE_OUT"
):
vnf_config_primitive = scaling_config_action[
"vnf-config-primitive-name-ref"
]
- step = db_nslcmop_update[
- "detailed-status"
- ] = "executing pre-scale scaling-config-action '{}'".format(
- vnf_config_primitive
+ step = db_nslcmop_update["detailed-status"] = (
+ "executing pre-scale scaling-config-action '{}'".format(
+ vnf_config_primitive
+ )
)
# look for primitive
for config_primitive in (
get_configuration(db_vnfd, db_vnfd["id"]) or {}
@@ -6846,33 +6847,33 @@
# PRE-SCALE END
db_nsr_update[
"_admin.scaling-group.{}.nb-scale-op".format(admin_scale_index)
] = nb_scale_op
- db_nsr_update[
- "_admin.scaling-group.{}.time".format(admin_scale_index)
- ] = time()
+ db_nsr_update["_admin.scaling-group.{}.time".format(admin_scale_index)] = (
+ time()
+ )
# SCALE-IN VCA - BEGIN
if vca_scaling_info:
- step = db_nslcmop_update[
- "detailed-status"
- ] = "Deleting the execution environments"
+ step = db_nslcmop_update["detailed-status"] = (
+ "Deleting the execution environments"
+ )
scale_process = "VCA"
for vca_info in vca_scaling_info:
if vca_info["type"] == "delete" and not vca_info.get("osm_kdu_id"):
member_vnf_index = str(vca_info["member-vnf-index"])
self.logger.debug(
logging_text + "vdu info: {}".format(vca_info)
)
if vca_info.get("osm_vdu_id"):
vdu_id = vca_info["osm_vdu_id"]
vdu_index = int(vca_info["vdu_index"])
- stage[
- 1
- ] = "Scaling member_vnf_index={}, vdu_id={}, vdu_index={} ".format(
- member_vnf_index, vdu_id, vdu_index
+ stage[1] = (
+ "Scaling member_vnf_index={}, vdu_id={}, vdu_index={} ".format(
+ member_vnf_index, vdu_id, vdu_index
+ )
)
stage[2] = step = "Scaling in VCA"
self._write_op_status(op_id=nslcmop_id, stage=stage)
vca_update = db_nsr["_admin"]["deployed"]["VCA"]
config_update = db_nsr["configurationStatus"]
@@ -6980,13 +6981,13 @@
if db_nsr_update:
self.update_db_2("nsrs", nsr_id, db_nsr_update)
# SCALE-UP VCA - BEGIN
if vca_scaling_info:
- step = db_nslcmop_update[
- "detailed-status"
- ] = "Creating new execution environments"
+ step = db_nslcmop_update["detailed-status"] = (
+ "Creating new execution environments"
+ )
scale_process = "VCA"
for vca_info in vca_scaling_info:
if vca_info["type"] == "create" and not vca_info.get("osm_kdu_id"):
member_vnf_index = str(vca_info["member-vnf-index"])
self.logger.debug(
@@ -7044,14 +7045,14 @@
db_vnfr, vdu_id, vdu_count_index=vdu_index
)
if descriptor_config:
vdu_name = None
kdu_name = None
- stage[
- 1
- ] = "Scaling member_vnf_index={}, vdu_id={}, vdu_index={} ".format(
- member_vnf_index, vdu_id, vdu_index
+ stage[1] = (
+ "Scaling member_vnf_index={}, vdu_id={}, vdu_index={} ".format(
+ member_vnf_index, vdu_id, vdu_index
+ )
)
stage[2] = step = "Scaling out VCA"
self._write_op_status(op_id=nslcmop_id, stage=stage)
self._deploy_n2vc(
logging_text=logging_text
@@ -7093,14 +7094,14 @@
and scaling_type == "SCALE_OUT"
):
vnf_config_primitive = scaling_config_action[
"vnf-config-primitive-name-ref"
]
- step = db_nslcmop_update[
- "detailed-status"
- ] = "executing post-scale scaling-config-action '{}'".format(
- vnf_config_primitive
+ step = db_nslcmop_update["detailed-status"] = (
+ "executing post-scale scaling-config-action '{}'".format(
+ vnf_config_primitive
+ )
)
vnfr_params = {"VDU_SCALE_INFO": scaling_info}
if db_vnfr.get("additionalParamsForVnf"):
vnfr_params.update(db_vnfr["additionalParamsForVnf"])
@@ -7206,13 +7207,13 @@
raise LcmException(result_detail)
db_nsr_update["config-status"] = old_config_status
scale_process = None
# POST-SCALE END
- db_nsr_update[
- "detailed-status"
- ] = "" # "scaled {} {}".format(scaling_group, scaling_type)
+ db_nsr_update["detailed-status"] = (
+ "" # "scaled {} {}".format(scaling_group, scaling_type)
+ )
db_nsr_update["operational-status"] = (
"running"
if old_operational_status == "failed"
else old_operational_status
)
@@ -7254,27 +7255,27 @@
stage,
nslcmop_id,
nsr_id=nsr_id,
)
if exc:
- db_nslcmop_update[
- "detailed-status"
- ] = error_description_nslcmop = "FAILED {}: {}".format(step, exc)
+ db_nslcmop_update["detailed-status"] = error_description_nslcmop = (
+ "FAILED {}: {}".format(step, exc)
+ )
nslcmop_operation_state = "FAILED"
if db_nsr:
db_nsr_update["operational-status"] = old_operational_status
db_nsr_update["config-status"] = old_config_status
db_nsr_update["detailed-status"] = ""
if scale_process:
if "VCA" in scale_process:
db_nsr_update["config-status"] = "failed"
if "RO" in scale_process:
db_nsr_update["operational-status"] = "failed"
- db_nsr_update[
- "detailed-status"
- ] = "FAILED scaling nslcmop={} {}: {}".format(
- nslcmop_id, step, exc
+ db_nsr_update["detailed-status"] = (
+ "FAILED scaling nslcmop={} {}: {}".format(
+ nslcmop_id, step, exc
+ )
)
else:
error_description_nslcmop = None
nslcmop_operation_state = "COMPLETED"
db_nslcmop_update["detailed-status"] = "Done"
@@ -7947,20 +7948,20 @@
stage,
nslcmop_id,
nsr_id=nsr_id,
)
if exc:
- db_nslcmop_update[
- "detailed-status"
- ] = error_description_nslcmop = "FAILED {}: {}".format(step, exc)
+ db_nslcmop_update["detailed-status"] = error_description_nslcmop = (
+ "FAILED {}: {}".format(step, exc)
+ )
nslcmop_operation_state = "FAILED"
if db_nsr:
db_nsr_update["operational-status"] = old_operational_status
db_nsr_update["config-status"] = old_config_status
- db_nsr_update[
- "detailed-status"
- ] = "FAILED healing nslcmop={} {}: {}".format(nslcmop_id, step, exc)
+ db_nsr_update["detailed-status"] = (
+ "FAILED healing nslcmop={} {}: {}".format(nslcmop_id, step, exc)
+ )
for task, task_name in tasks_dict_info.items():
if not task.done() or task.cancelled() or task.exception():
if task_name.startswith(self.task_name_deploy_vca):
# A N2VC task is pending
db_nsr_update["config-status"] = "failed"
@@ -8248,14 +8249,13 @@
nsr_id,
nslcmop_id,
"instantiate_N2VC-{}".format(vca_index),
task_n2vc,
)
- task_instantiation_info[
- task_n2vc
- ] = self.task_name_deploy_vca + " {}.{}".format(
- member_vnf_index or "", vdu_id or ""
+ task_instantiation_info[task_n2vc] = (
+ self.task_name_deploy_vca
+ + " {}.{}".format(member_vnf_index or "", vdu_id or "")
)
async def heal_N2VC(
self,
logging_text,
@@ -8320,23 +8320,27 @@
# Get artifact path
if base_folder["pkg-dir"]:
artifact_path = "{}/{}/{}/{}".format(
base_folder["folder"],
base_folder["pkg-dir"],
- "charms"
- if vca_type
- in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
- else "helm-charts",
+ (
+ "charms"
+ if vca_type
+ in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
+ else "helm-charts"
+ ),
vca_name,
)
else:
artifact_path = "{}/Scripts/{}/{}/".format(
base_folder["folder"],
- "charms"
- if vca_type
- in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
- else "helm-charts",
+ (
+ "charms"
+ if vca_type
+ in ("native_charm", "lxc_proxy_charm", "k8s_proxy_charm")
+ else "helm-charts"
+ ),
vca_name,
)
self.logger.debug("Artifact path > {}".format(artifact_path))
would reformat /home/jenkins/workspace/LCM-stage_2-merge_v13.0/osm_lcm/ns.py
Oh no! 💥 💔 💥
6 files would be reformatted, 34 files would be left unchanged.
ERROR: InvocationError for command /tmp/.tox/black/bin/black --check --diff osm_lcm/ (exited with code 1)
=================================== log end ====================================
✖ FAIL black in 23.802 seconds
✔ OK safety in 49.296 seconds
safety create: /tmp/.tox/safety
safety installdeps: -r/home/jenkins/workspace/LCM-stage_2-merge_v13.0/requirements.txt, safety
safety develop-inst: /home/jenkins/workspace/LCM-stage_2-merge_v13.0
safety installed: aiohttp==3.7.4.post0,async-timeout==3.0.1,attrs==22.1.0,boltons==21.0.0,certifi==2023.11.17,chardet==4.0.0,charset-normalizer==3.3.2,checksumdir==1.2.0,click==8.1.7,config-man==0.0.4,dparse==0.6.3,face==22.0.0,glom==22.1.0,grpcio==1.50.0,grpcio-tools==1.48.1,grpclib==0.4.3,h2==4.1.0,hpack==4.0.0,hyperframe==6.0.1,idna==3.4,Jinja2==3.1.2,MarkupSafe==2.1.1,multidict==6.0.2,-e git+https://osm.etsi.org/gerrit/osm/LCM.git@c96cb8bb9cef9b19c51baefe6a67c22fa8a71830#egg=osm_lcm,packaging==21.3,protobuf==3.20.3,pydantic==1.10.2,pyparsing==3.1.1,PyYAML==5.4.1,requests==2.31.0,ruamel.yaml==0.18.5,ruamel.yaml.clib==0.2.8,safety==2.3.5,six==1.16.0,tomli==2.0.1,typing_extensions==4.4.0,urllib3==2.1.0,yarl==1.8.1
safety run-test-pre: PYTHONHASHSEED='2538429602'
safety run-test: commands[0] | - safety check --full-report
+==============================================================================+
/$$$$$$ /$$
/$$__ $$ | $$
/$$$$$$$ /$$$$$$ | $$ \__//$$$$$$ /$$$$$$ /$$ /$$
/$$_____/ |____ $$| $$$$ /$$__ $$|_ $$_/ | $$ | $$
| $$$$$$ /$$$$$$$| $$_/ | $$$$$$$$ | $$ | $$ | $$
\____ $$ /$$__ $$| $$ | $$_____/ | $$ /$$| $$ | $$
/$$$$$$$/| $$$$$$$| $$ | $$$$$$$ | $$$$/| $$$$$$$
|_______/ \_______/|__/ \_______/ \___/ \____ $$
/$$ | $$
| $$$$$$/
by pyup.io \______/
+==============================================================================+
[1mREPORT[0m
[31m[22m Safety is using PyUp's free open-source vulnerability database. This
data is 30 days old and limited. [0m
[31m For real-time enhanced vulnerability data, fix recommendations, severity
reporting, cybersecurity support, team and project policy management and more
sign up at https://pyup.io or email sales@pyup.io[0m
Safety [1mv2.3.5[0m is scanning for [1mVulnerabilities[0m[1m...[0m
[1m Scanning dependencies[0m in your [1menvironment:[0m
-> /home/jenkins/workspace/LCM-stage_2-merge_v13.0
-> /tmp/.tox/safety/lib/python3.8/site-packages
Using [1mnon-commercial database[0m
[1m Found and scanned 41 packages[0m
Timestamp [1m2024-01-28 09:51:48[0m
[1m 11[0m[1m vulnerabilities found[0m
[1m 0[0m[1m vulnerabilities ignored[0m
+==============================================================================+
[31m[1mVULNERABILITIES FOUND[0m
+==============================================================================+
[31m-> Vulnerability found in aiohttp version 3.7.4.post0
[0m[1m Vulnerability ID: [0m62583
[1m Affected spec: [0m<3.9.0
[1m ADVISORY: [0mAiohttp 3.9.0 includes a fix for CVE-2023-49082: Improper
validation makes it possible for an attacker to modify the HTTP request (e.g.
insert a new header) or even create a new HTTP request if the attacker
controls the HTTP method. The vulnerability occurs only if the attacker can
control the HTTP method (GET, POST etc.) of the request. If the attacker can
control the HTTP version of the request it will be able to modify the request
(request smuggling).https://github.com/aio-
libs/aiohttp/security/advisories/GHSA-qvrw-v9rv-5rjx
[1m CVE-2023-49082[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/62583/f17[0m
[31m-> Vulnerability found in aiohttp version 3.7.4.post0
[0m[1m Vulnerability ID: [0m62582
[1m Affected spec: [0m<3.9.0
[1m ADVISORY: [0mAiohttp 3.9.0 includes a fix for CVE-2023-49081: Improper
validation made it possible for an attacker to modify the HTTP request (e.g.
to insert a new header) or create a new HTTP request if the attacker controls
the HTTP version. The vulnerability only occurs if the attacker can control
the HTTP version of the request.https://github.com/aio-
libs/aiohttp/security/advisories/GHSA-q3qx-c6g2-7pw2
[1m CVE-2023-49081[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/62582/f17[0m
[31m-> Vulnerability found in aiohttp version 3.7.4.post0
[0m[1m Vulnerability ID: [0m59725
[1m Affected spec: [0m<=3.8.4
[1m ADVISORY: [0mAiohttp 3.8.5 includes a fix for CVE-2023-37276: Sending a
crafted HTTP request will cause the server to misinterpret one of the HTTP
header values leading to HTTP request smuggling.https://github.com/aio-libs/a
iohttp/commit/9337fb3f2ab2b5f38d7e98a194bde6f7e3d16c40https://github.com/aio-
libs/aiohttp/security/advisories/GHSA-45c4-8wx5-qw6w
[1m CVE-2023-37276[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/59725/f17[0m
[31m-> Vulnerability found in aiohttp version 3.7.4.post0
[0m[1m Vulnerability ID: [0m42692
[1m Affected spec: [0m<3.8.0
[1m ADVISORY: [0mAiohttp 3.8.0 adds validation of HTTP header keys and
values to prevent header injection.https://github.com/aio-
libs/aiohttp/issues/4818
[1m PVE-2021-42692[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/42692/f17[0m
[31m-> Vulnerability found in aiohttp version 3.7.4.post0
[0m[1m Vulnerability ID: [0m62327
[1m Affected spec: [0m<3.8.0
[1m ADVISORY: [0mAiohttp 3.8.0 includes a fix for CVE-2023-47641: Affected
versions of aiohttp have a security vulnerability regarding the inconsistent
interpretation of the http protocol. HTTP/1.1 is a persistent protocol, if
both Content-Length(CL) and Transfer-Encoding(TE) header values are present
it can lead to incorrect interpretation of two entities that parse the HTTP
and we can poison other sockets with this incorrect interpretation. A
possible Proof-of-Concept (POC) would be a configuration with a reverse
proxy(frontend) that accepts both CL and TE headers and aiohttp as backend.
As aiohttp parses anything with chunked, we can pass a chunked123 as TE, the
frontend entity will ignore this header and will parse Content-Length. The
impact of this vulnerability is that it is possible to bypass any proxy rule,
poisoning sockets to other users like passing Authentication Headers, also if
it is present an Open Redirect an attacker could combine it to redirect
random users to another website and log the request.https://github.com/aio-
libs/aiohttp/security/advisories/GHSA-xx9p-xxvh-7g8j
[1m CVE-2023-47641[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/62327/f17[0m
[31m-> Vulnerability found in aiohttp version 3.7.4.post0
[0m[1m Vulnerability ID: [0m62326
[1m Affected spec: [0m<3.8.6
[1m ADVISORY: [0mAiohttp 3.8.6 includes a fix for CVE-2023-47627: The HTTP
parser in AIOHTTP has numerous problems with header parsing, which could lead
to request smuggling. This parser is only used when AIOHTTP_NO_EXTENSIONS is
enabled (or not using a prebuilt wheel).https://github.com/aio-
libs/aiohttp/security/advisories/GHSA-gfw2-4jvh-wgfg
[1m CVE-2023-47627[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/62326/f17[0m
[31m-> Vulnerability found in grpcio version 1.50.0
[0m[1m Vulnerability ID: [0m61191
[1m Affected spec: [0m<1.53.2
[1m ADVISORY: [0mGrpcio 1.53.2, 1.54.3, 1.55.3 and 1.56.2 include a fix for
CVE-2023-4785: Lack of error handling in the TCP server in Google's gRPC
starting version 1.23 on posix-compatible platforms (ex. Linux) allows an
attacker to cause a denial of service by initiating a significant number of
connections with the server. Note that gRPC C++ Python, and Ruby are
affected, but gRPC Java, and Go are NOT
affected.https://github.com/grpc/grpc/pull/33656
[1m CVE-2023-4785[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/61191/f17[0m
[31m-> Vulnerability found in grpcio version 1.50.0
[0m[1m Vulnerability ID: [0m59869
[1m Affected spec: [0m<1.53.0
[1m ADVISORY: [0mGrpcio 1.53.0 includes a fix for a Connection Confusion
vulnerability. When gRPC HTTP2 stack raised a header size exceeded error, it
skipped parsing the rest of the HPACK frame. This caused any HPACK table
mutations to also be skipped, resulting in a desynchronization of HPACK
tables between sender and receiver. If leveraged, say, between a proxy and a
backend, this could lead to requests from the proxy being interpreted as
containing headers from different proxy clients - leading to an information
leak that can be used for privilege escalation or data
exfiltration.https://github.com/advisories/GHSA-cfgp-2977-2fmm
[1m CVE-2023-32731[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/59869/f17[0m
[31m-> Vulnerability found in grpcio version 1.50.0
[0m[1m Vulnerability ID: [0m59867
[1m Affected spec: [0m<1.53.0
[1m ADVISORY: [0mGrpcio 1.53.0 includes a fix for a Reachable Assertion
vulnerability. https://github.com/advisories/GHSA-6628-q6j9-w8vg
[1m CVE-2023-1428[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/59867/f17[0m
[31m-> Vulnerability found in grpcio version 1.50.0
[0m[1m Vulnerability ID: [0m59868
[1m Affected spec: [0m<1.53.0
[1m ADVISORY: [0mGrpcio 1.53.0 includes a fix for a Connection Termination
vulnerability. The prior versions contain a vulnerability whereby a client
can cause a termination of connection between a HTTP2 proxy and a gRPC
server: a base64 encoding error for -bin suffixed headers will result in a
disconnection by the gRPC server, but is typically allowed by HTTP2
proxies.https://github.com/advisories/GHSA-9hxf-ppjv-w6rq
[1m CVE-2023-32732[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/59868/f17[0m
[31m-> Vulnerability found in pydantic version 1.10.2
[0m[1m Vulnerability ID: [0m61416
[1m Affected spec: [0m<1.10.13
[1m ADVISORY: [0mPydantic 1.10.13 and 2.4.0 include a fix for a regular
expression denial of service vulnerability (REDoS).https://github.com/pydanti
c/pydantic/pull/7360https://github.com/pydantic/pydantic/pull/7673
[1m PVE-2023-61416[0m
[1m For more information, please visit
[0mhttps://data.safetycli.com/v/61416/f17[0m
Scan was completed. 11 vulnerabilities were found.
+==============================================================================+
[32m[1m REMEDIATIONS[0m
11 vulnerabilities were found in 3 packages. For detailed remediation & fix
recommendations, upgrade to a commercial license.
+==============================================================================+
[31m[22m Safety is using PyUp's free open-source vulnerability database. This
data is 30 days old and limited. [0m
[31m For real-time enhanced vulnerability data, fix recommendations, severity
reporting, cybersecurity support, team and project policy management and more
sign up at https://pyup.io or email sales@pyup.io[0m
+==============================================================================+[0m
✔ OK pylint in 2 minutes, 15.102 seconds
pylint create: /tmp/.tox/pylint
pylint installdeps: -r/home/jenkins/workspace/LCM-stage_2-merge_v13.0/requirements.txt, -r/home/jenkins/workspace/LCM-stage_2-merge_v13.0/requirements-dev.txt, -r/home/jenkins/workspace/LCM-stage_2-merge_v13.0/requirements-test.txt, pylint
pylint develop-inst: /home/jenkins/workspace/LCM-stage_2-merge_v13.0
pylint installed: aiohttp==3.7.4.post0,aiokafka==0.7.2,astroid==3.0.2,async-timeout==3.0.1,asynctest==0.13.0,attrs==22.1.0,bcrypt==4.0.1,boltons==21.0.0,cachetools==5.2.0,certifi==2022.9.24,cffi==1.15.1,chardet==4.0.0,charset-normalizer==2.1.1,checksumdir==1.2.0,config-man==0.0.4,coverage==6.5.0,cryptography==38.0.1,dataclasses==0.6,dill==0.3.8,face==22.0.0,glom==22.1.0,google-auth==2.12.0,grpcio==1.50.0,grpcio-tools==1.48.1,grpclib==0.4.3,h2==4.1.0,hpack==4.0.0,hyperframe==6.0.1,idna==3.4,isort==5.13.2,Jinja2==3.1.2,juju==3.0.0,jujubundlelib==0.5.7,kafka-python==2.0.2,kubernetes==24.2.0,macaroonbakery==1.3.1,MarkupSafe==2.1.1,mccabe==0.7.0,mock==4.0.3,motor==1.3.1,multidict==6.0.2,mypy-extensions==0.4.3,N2VC @ git+https://osm.etsi.org/gerrit/osm/N2VC.git@6eac0f117eca51b7970926f92f0b5a8abdd9ba4a,nose2==0.12.0,oauthlib==3.2.1,osm-common @ git+https://osm.etsi.org/gerrit/osm/common.git@303ffe4f33c7a0fcc6b5c267d402c0e7d44e5a57,-e git+https://osm.etsi.org/gerrit/osm/LCM.git@c96cb8bb9cef9b19c51baefe6a67c22fa8a71830#egg=osm_lcm,paramiko==2.11.0,platformdirs==4.1.0,protobuf==3.20.3,pyasn1==0.4.8,pyasn1-modules==0.2.8,pycparser==2.21,pycrypto==2.6.1,pydantic==1.10.2,pylint==3.0.3,pymacaroons==0.13.0,pymongo==3.12.3,PyNaCl==1.5.0,pyRFC3339==1.1,python-dateutil==2.8.2,pytz==2022.4,PyYAML==5.4.1,requests==2.28.1,requests-oauthlib==1.3.1,retrying-async==2.0.0,rsa==4.9,six==1.16.0,theblues==0.5.2,tomli==2.0.1,tomlkit==0.12.3,toposort==1.7,typing-inspect==0.8.0,typing_extensions==4.4.0,urllib3==1.26.12,websocket-client==1.4.1,websockets==7.0,yarl==1.8.1
pylint run-test-pre: PYTHONHASHSEED='727850233'
pylint run-test: commands[0] | - pylint -E osm_lcm
************* Module osm_lcm.ROclient
osm_lcm/ROclient.py:1271:25: E1101: Instance of 'ROClient' has no 'parse' member; maybe '_parse'? (no-member)
osm_lcm/ROclient.py:1314:25: E1101: Instance of 'ROClient' has no 'parse' member; maybe '_parse'? (no-member)
osm_lcm/ROclient.py:1366:32: E1120: No value for argument 'session' in method call (no-value-for-parameter)
osm_lcm/ROclient.py:1376:25: E1101: Instance of 'ROClient' has no 'get_datacenter' member; maybe '_get_datacenter'? (no-member)
************* Module osm_lcm.lcm
osm_lcm/lcm.py:482:28: E1101: Instance of 'Lcm' has no 'lcm_ns_tasks' member (no-member)
osm_lcm/lcm.py:544:28: E1101: Instance of 'Lcm' has no 'lcm_netslice_tasks' member (no-member)
************* Module osm_lcm.vim_sdn
osm_lcm/vim_sdn.py:1046:16: E1137: 'db_sdn' does not support item assignment (unsupported-assignment-operation)
osm_lcm/vim_sdn.py:1047:16: E1137: 'db_sdn' does not support item assignment (unsupported-assignment-operation)
************* Module osm_lcm.osm_config
osm_lcm/osm_config.py:16:0: E0611: No name 'BaseModel' in module 'pydantic' (no-name-in-module)
osm_lcm/osm_config.py:31:4: E0213: Method 'parse_services' should have "self" as first argument (no-self-argument)
************* Module osm_lcm.ns
osm_lcm/ns.py:1399:48: E1101: Instance of 'Exception' has no 'http_code' member (no-member)
osm_lcm/ns.py:1406:50: E1101: Instance of 'Exception' has no 'http_code' member (no-member)
osm_lcm/ns.py:1675:44: E1101: Instance of 'NgRoClient' has no 'create_action' member (no-member)
osm_lcm/ns.py:4284:29: E1121: Too many positional arguments for method call (too-many-function-args)
osm_lcm/ns.py:4307:33: E1101: Instance of 'NgRoClient' has no 'show' member (no-member)
osm_lcm/ns.py:4317:48: E1101: Instance of 'NgRoClient' has no 'check_action_status' member (no-member)
osm_lcm/ns.py:4347:62: E1101: Instance of 'Exception' has no 'http_code' member (no-member)
osm_lcm/ns.py:4356:62: E1101: Instance of 'Exception' has no 'http_code' member (no-member)
osm_lcm/ns.py:4377:22: E1121: Too many positional arguments for method call (too-many-function-args)
osm_lcm/ns.py:4384:66: E1101: Instance of 'Exception' has no 'http_code' member (no-member)
osm_lcm/ns.py:4391:66: E1101: Instance of 'Exception' has no 'http_code' member (no-member)
osm_lcm/ns.py:4417:26: E1121: Too many positional arguments for method call (too-many-function-args)
osm_lcm/ns.py:4424:70: E1101: Instance of 'Exception' has no 'http_code' member (no-member)
osm_lcm/ns.py:4434:70: E1101: Instance of 'Exception' has no 'http_code' member (no-member)
osm_lcm/ns.py:5398:71: E0601: Using variable 'step' before assignment (used-before-assignment)
************* Module osm_lcm.data_utils.database.vim_account
osm_lcm/data_utils/database/vim_account.py:32:4: E0213: Method 'get_vim_account_with_id' should have "self" as first argument (no-self-argument)
osm_lcm/data_utils/database/vim_account.py:37:4: E0211: Method 'initialize_db' has no argument (no-method-argument)
************* Module osm_lcm.data_utils.database.wim_account
osm_lcm/data_utils/database/wim_account.py:31:4: E0211: Method 'initialize_db' has no argument (no-method-argument)
osm_lcm/data_utils/database/wim_account.py:34:4: E0213: Method 'get_wim_account_with_id' should have "self" as first argument (no-self-argument)
osm_lcm/data_utils/database/wim_account.py:43:4: E0211: Method 'get_all_wim_accounts' has no argument (no-method-argument)
✔ OK cover in 2 minutes, 28.927 seconds
cover create: /tmp/.tox/cover
cover installdeps: -r/home/jenkins/workspace/LCM-stage_2-merge_v13.0/requirements.txt, -r/home/jenkins/workspace/LCM-stage_2-merge_v13.0/requirements-dev.txt, -r/home/jenkins/workspace/LCM-stage_2-merge_v13.0/requirements-test.txt
cover develop-inst: /home/jenkins/workspace/LCM-stage_2-merge_v13.0
cover installed: aiohttp==3.7.4.post0,aiokafka==0.7.2,async-timeout==3.0.1,asynctest==0.13.0,attrs==22.1.0,bcrypt==4.0.1,boltons==21.0.0,cachetools==5.2.0,certifi==2022.9.24,cffi==1.15.1,chardet==4.0.0,charset-normalizer==2.1.1,checksumdir==1.2.0,config-man==0.0.4,coverage==6.5.0,cryptography==38.0.1,dataclasses==0.6,face==22.0.0,glom==22.1.0,google-auth==2.12.0,grpcio==1.50.0,grpcio-tools==1.48.1,grpclib==0.4.3,h2==4.1.0,hpack==4.0.0,hyperframe==6.0.1,idna==3.4,Jinja2==3.1.2,juju==3.0.0,jujubundlelib==0.5.7,kafka-python==2.0.2,kubernetes==24.2.0,macaroonbakery==1.3.1,MarkupSafe==2.1.1,mock==4.0.3,motor==1.3.1,multidict==6.0.2,mypy-extensions==0.4.3,N2VC @ git+https://osm.etsi.org/gerrit/osm/N2VC.git@6eac0f117eca51b7970926f92f0b5a8abdd9ba4a,nose2==0.12.0,oauthlib==3.2.1,osm-common @ git+https://osm.etsi.org/gerrit/osm/common.git@303ffe4f33c7a0fcc6b5c267d402c0e7d44e5a57,-e git+https://osm.etsi.org/gerrit/osm/LCM.git@c96cb8bb9cef9b19c51baefe6a67c22fa8a71830#egg=osm_lcm,paramiko==2.11.0,protobuf==3.20.3,pyasn1==0.4.8,pyasn1-modules==0.2.8,pycparser==2.21,pycrypto==2.6.1,pydantic==1.10.2,pymacaroons==0.13.0,pymongo==3.12.3,PyNaCl==1.5.0,pyRFC3339==1.1,python-dateutil==2.8.2,pytz==2022.4,PyYAML==5.4.1,requests==2.28.1,requests-oauthlib==1.3.1,retrying-async==2.0.0,rsa==4.9,six==1.16.0,theblues==0.5.2,toposort==1.7,typing-inspect==0.8.0,typing_extensions==4.4.0,urllib3==1.26.12,websocket-client==1.4.1,websockets==7.0,yarl==1.8.1
cover run-test-pre: PYTHONHASHSEED='2190516959'
cover run-test: commands[0] | sh -c 'rm -f nosetests.xml'
cover run-test: commands[1] | coverage erase
cover run-test: commands[2] | nose2 -C --coverage osm_lcm
.ERROR:lcm.vca:Task vca_create=id Failed with exception: failed
ERROR:lcm.vca:Task vca_create=id Cannot update database: database exception failed
..ERROR:lcm.vca:Task vca_delete=id Failed with exception: failed deleting
ERROR:lcm.vca:Task vca_delete=id Cannot update database: database exception failed
..DEBUG:test_lcm_helm_conn:Initialize helm N2VC connector
DEBUG:test_lcm_helm_conn:initial vca_config: {'host': None, 'port': None, 'user': None, 'secret': None, 'cloud': None, 'k8s_cloud': None, 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': None, 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}
DEBUG:test_lcm_helm_conn:Initial retry time: 600
DEBUG:test_lcm_helm_conn:Retry time: 30
INFO:test_lcm_helm_conn:Helm N2VC connector initialized
INFO:test_lcm_helm_conn:create_execution_environment: namespace: testnamespace, artifact_path: helm_sample_charm, chart_model: {}, db_dict: helm_sample_charm, reuse_ee_id: None
DEBUG:test_lcm_helm_conn:install helm chart: /helm_sample_charm
.DEBUG:test_lcm_helm_conn:Initialize helm N2VC connector
DEBUG:test_lcm_helm_conn:initial vca_config: {'host': None, 'port': None, 'user': None, 'secret': None, 'cloud': None, 'k8s_cloud': None, 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': None, 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}
DEBUG:test_lcm_helm_conn:Initial retry time: 600
DEBUG:test_lcm_helm_conn:Retry time: 30
INFO:test_lcm_helm_conn:Helm N2VC connector initialized
INFO:test_lcm_helm_conn:ee_id: helm-v3:osm.helm_sample_charm_0001
INFO:test_lcm_helm_conn:ee_id: helm-v3:osm.helm_sample_charm_0001 deleted
.DEBUG:test_lcm_helm_conn:Initialize helm N2VC connector
DEBUG:test_lcm_helm_conn:initial vca_config: {'host': None, 'port': None, 'user': None, 'secret': None, 'cloud': None, 'k8s_cloud': None, 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': None, 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}
DEBUG:test_lcm_helm_conn:Initial retry time: 600
DEBUG:test_lcm_helm_conn:Retry time: 30
INFO:test_lcm_helm_conn:Helm N2VC connector initialized
DEBUG:test_lcm_helm_conn:Execute config primitive
INFO:test_lcm_helm_conn:exec primitive for ee_id : osm.helm_sample_charm_0001, primitive_name: config, params_dict: {'ssh-host-name': 'host1'}, db_dict: None
DEBUG:test_lcm_helm_conn:Executed config primitive ee_id_ osm.helm_sample_charm_0001, status: OK, message: CONFIG OK
.DEBUG:test_lcm_helm_conn:Initialize helm N2VC connector
DEBUG:test_lcm_helm_conn:initial vca_config: {'host': None, 'port': None, 'user': None, 'secret': None, 'cloud': None, 'k8s_cloud': None, 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': None, 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}
DEBUG:test_lcm_helm_conn:Initial retry time: 600
DEBUG:test_lcm_helm_conn:Retry time: 30
INFO:test_lcm_helm_conn:Helm N2VC connector initialized
INFO:test_lcm_helm_conn:exec primitive for ee_id : osm.helm_sample_charm_0001, primitive_name: sleep, params_dict: {}, db_dict: None
DEBUG:test_lcm_helm_conn:Executed primitive sleep ee_id_ osm.helm_sample_charm_0001, status: OK, message: test-ok
.DEBUG:test_lcm_helm_conn:Initialize helm N2VC connector
DEBUG:test_lcm_helm_conn:initial vca_config: {'host': None, 'port': None, 'user': None, 'secret': None, 'cloud': None, 'k8s_cloud': None, 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': None, 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}
DEBUG:test_lcm_helm_conn:Initial retry time: 600
DEBUG:test_lcm_helm_conn:Retry time: 30
INFO:test_lcm_helm_conn:Helm N2VC connector initialized
INFO:test_lcm_helm_conn:get_ee_ssh_public_key: ee_id: osm.helm_sample_charm_0001, db_dict: {}
...................................DEBUG:lcm.roclient:GET http://h//ns/v1/deploy/f48163a6-c807-47bc-9682-f72caef5af85/<MagicMock name='mock.vertical_scale().__getitem__()' id='139634154405504'>
.CRITICAL:lcm:Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
CRITICAL:lcm:starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
.2024-01-28T09:53:14 CRITICAL lcm lcm.py:96 Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
CRITICAL:lcm:Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
CRITICAL:lcm:starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
.2024-01-28T09:53:14 CRITICAL lcm lcm.py:96 Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
2024-01-28T09:53:14 CRITICAL lcm lcm.py:96 Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
CRITICAL:lcm:Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
CRITICAL:lcm:starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
.2024-01-28T09:53:14 CRITICAL lcm lcm.py:96 Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
2024-01-28T09:53:14 CRITICAL lcm lcm.py:96 Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
2024-01-28T09:53:14 CRITICAL lcm lcm.py:96 Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
CRITICAL:lcm:Loaded configuration:{'globalConfig': {'loglevel': 'DEBUG', 'logfile': None, 'nologging': False}, 'timeout': {'nsi_deploy': 7200, 'vca_on_error': 300, 'ns_deploy': 7200, 'ns_terminate': 1800, 'ns_heal': 1800, 'charm_delete': 600, 'primitive': 1800, 'ns_update': 1800, 'progress_primitive': 600, 'migrate': 1800, 'operate': 1800, 'verticalscale': 1800}, 'RO': {'host': 'ro', 'ng': True, 'port': 9090, 'uri': 'h', 'tenant': 'osm', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.roclient'}, 'VCA': {'host': 'vca', 'port': 17070, 'user': 'admin', 'secret': 'secret', 'cloud': 'localhost', 'k8s_cloud': 'k8scloud', 'helmpath': '/usr/local/bin/helm', 'helm3path': '/usr/local/bin/helm3', 'kubectlpath': '/usr/bin/kubectl', 'jujupath': '/usr/local/bin/juju', 'public_key': None, 'ca_cert': None, 'api_proxy': None, 'apt_mirror': None, 'eegrpcinittimeout': None, 'eegrpctimeout': None, 'eegrpc_tls_enforce': False, 'loglevel': 'DEBUG', 'logfile': None, 'ca_store': '/etc/ssl/certs/osm-ca.crt', 'kubectl_osm_namespace': 'osm', 'kubectl_osm_cluster_name': '_system-osm-k8s', 'helm_ee_service_port': 50050, 'helm_max_initial_retry_time': 600, 'helm_max_retry_time': 30, 'helm_ee_retry_delay': 10}, 'database': {'driver': 'memory', 'host': None, 'port': 27017, 'uri': None, 'name': 'osm', 'replicaset': None, 'user': None, 'password': None, 'commonkey': None, 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.db'}, 'storage': {'driver': 'local', 'path': '/tmp/storage', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.fs', 'collection': None, 'uri': None}, 'message': {'driver': 'local', 'path': '/tmp/kafka', 'host': 'kafka', 'port': 9092, 'loglevel': 'DEBUG', 'logfile': None, 'group_id': 'lcm-server', 'logger_name': 'lcm.msg'}, 'tsdb': {'driver': 'prometheus', 'path': '/tmp/prometheus', 'uri': 'http://prometheus:9090/', 'loglevel': 'DEBUG', 'logfile': None, 'logger_name': 'lcm.prometheus'}}
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
2024-01-28T09:53:14 CRITICAL lcm lcm.py:142 starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
CRITICAL:lcm:starting osm/lcm version 13.0.2+gc96cb8b 2020-04-08
2024-01-28T09:53:14 CRITICAL lcm lcm.py:638 unknown topic kafka and command 'ping'
2024-01-28T09:53:14 CRITICAL lcm lcm.py:638 unknown topic kafka and command 'ping'
2024-01-28T09:53:14 CRITICAL lcm lcm.py:638 unknown topic kafka and command 'ping'
2024-01-28T09:53:14 CRITICAL lcm lcm.py:638 unknown topic kafka and command 'ping'
CRITICAL:lcm:unknown topic kafka and command 'ping'
..........................
----------------------------------------------------------------------
Ran 74 tests in 36.245s
OK
Name Stmts Miss Cover
-----------------------------------------------------------------
osm_lcm/ROclient.py 824 759 8%
osm_lcm/__init__.py 7 2 71%
osm_lcm/data_utils/__init__.py 0 0 100%
osm_lcm/data_utils/database/__init__.py 0 0 100%
osm_lcm/data_utils/database/database.py 26 8 69%
osm_lcm/data_utils/database/vim_account.py 9 0 100%
osm_lcm/data_utils/database/wim_account.py 21 13 38%
osm_lcm/data_utils/dict_utils.py 7 2 71%
osm_lcm/data_utils/filesystem/__init__.py 0 0 100%
osm_lcm/data_utils/filesystem/filesystem.py 26 9 65%
osm_lcm/data_utils/lcm_config.py 156 4 97%
osm_lcm/data_utils/list_utils.py 5 0 100%
osm_lcm/data_utils/nsd.py 16 7 56%
osm_lcm/data_utils/nsr.py 17 4 76%
osm_lcm/data_utils/vca.py 97 42 57%
osm_lcm/data_utils/vim.py 28 13 54%
osm_lcm/data_utils/vnfd.py 78 41 47%
osm_lcm/data_utils/vnfr.py 41 13 68%
osm_lcm/data_utils/wim.py 72 55 24%
osm_lcm/frontend_grpc.py 20 6 70%
osm_lcm/frontend_pb2.py 27 0 100%
osm_lcm/lcm.py 497 363 27%
osm_lcm/lcm_hc.py 36 15 58%
osm_lcm/lcm_helm_conn.py 300 157 48%
osm_lcm/lcm_utils.py 303 157 48%
osm_lcm/netslice.py 440 425 3%
osm_lcm/ng_ro.py 205 151 26%
osm_lcm/ns.py 3629 2697 26%
osm_lcm/osm_config.py 21 0 100%
osm_lcm/prometheus.py 11 2 82%
osm_lcm/tests/test_db_descriptors.py 13 0 100%
osm_lcm/tests/test_lcm.py 58 1 98%
osm_lcm/tests/test_lcm_hc.py 43 0 100%
osm_lcm/tests/test_lcm_helm_conn.py 79 1 99%
osm_lcm/tests/test_lcm_utils.py 293 13 96%
osm_lcm/tests/test_ns.py 662 50 92%
osm_lcm/tests/test_osm_config.py 7 0 100%
osm_lcm/tests/test_prometheus.py 12 1 92%
osm_lcm/tests/test_vim_sdn.py 74 0 100%
osm_lcm/vim_sdn.py 934 823 12%
-----------------------------------------------------------------
TOTAL 9094 5834 36%
cover run-test: commands[3] | coverage report '--omit=*tests*'
Name Stmts Miss Cover
-----------------------------------------------------------------
osm_lcm/ROclient.py 824 759 8%
osm_lcm/__init__.py 7 2 71%
osm_lcm/data_utils/__init__.py 0 0 100%
osm_lcm/data_utils/database/__init__.py 0 0 100%
osm_lcm/data_utils/database/database.py 26 8 69%
osm_lcm/data_utils/database/vim_account.py 9 0 100%
osm_lcm/data_utils/database/wim_account.py 21 13 38%
osm_lcm/data_utils/dict_utils.py 7 2 71%
osm_lcm/data_utils/filesystem/__init__.py 0 0 100%
osm_lcm/data_utils/filesystem/filesystem.py 26 9 65%
osm_lcm/data_utils/lcm_config.py 156 4 97%
osm_lcm/data_utils/list_utils.py 5 0 100%
osm_lcm/data_utils/nsd.py 16 7 56%
osm_lcm/data_utils/nsr.py 17 4 76%
osm_lcm/data_utils/vca.py 97 42 57%
osm_lcm/data_utils/vim.py 28 13 54%
osm_lcm/data_utils/vnfd.py 78 41 47%
osm_lcm/data_utils/vnfr.py 41 13 68%
osm_lcm/data_utils/wim.py 72 55 24%
osm_lcm/frontend_grpc.py 20 6 70%
osm_lcm/frontend_pb2.py 27 0 100%
osm_lcm/lcm.py 497 363 27%
osm_lcm/lcm_hc.py 36 15 58%
osm_lcm/lcm_helm_conn.py 300 157 48%
osm_lcm/lcm_utils.py 303 157 48%
osm_lcm/netslice.py 440 425 3%
osm_lcm/ng_ro.py 205 151 26%
osm_lcm/ns.py 3629 2697 26%
osm_lcm/osm_config.py 21 0 100%
osm_lcm/prometheus.py 11 2 82%
osm_lcm/vim_sdn.py 934 823 12%
-----------------------------------------------------------------
TOTAL 7853 5768 27%
cover run-test: commands[4] | coverage html -d ./cover '--omit=*tests*'
Wrote HTML report to ./cover/index.html
cover run-test: commands[5] | coverage xml -o coverage.xml '--omit=*tests*'
Wrote XML report to coverage.xml
___________________________________ summary ____________________________________
ERROR: black: parallel child exit code 1
cover: commands succeeded
flake8: commands succeeded
pylint: commands succeeded
safety: commands succeeded