Standardize logging
This commit will standardize logging for all the code,
so that it follows the same format as LCM. The logging
for k8s_juju_conn.py is in commit 8638.
Change-Id: I0ea199018e96b167ecbf8fd1ab16e5065a2e2bdf
Signed-off-by: Dominik Fleischmann <dominik.fleischmann@canonical.com>
(cherry picked from commit f9bed35a0acf26a93eb2f6d0f146fa71579af74a)
diff --git a/n2vc/k8s_conn.py b/n2vc/k8s_conn.py
index 6ac8119..e49e497 100644
--- a/n2vc/k8s_conn.py
+++ b/n2vc/k8s_conn.py
@@ -50,13 +50,13 @@
# parent class
Loggable.__init__(self, log=log, log_to_console=True, prefix='\nK8S')
- self.info('Initializing generic K8S connector')
+ self.log.info('Initializing generic K8S connector')
# the database and update callback
self.db = db
self.on_update_db = on_update_db
- self.info('K8S generic connector initialized')
+ self.log.info('K8S generic connector initialized')
@abc.abstractmethod
async def init_env(
@@ -333,7 +333,7 @@
self.warning('No db_dict => No database write')
return False
- self.debug('status={}'.format(status))
+ self.log.debug('status={}'.format(status))
try:
@@ -366,5 +366,5 @@
return True
except Exception as e:
- self.info('Exception writing status to database: {}'.format(e))
+ self.log.info('Exception writing status to database: {}'.format(e))
return False
diff --git a/n2vc/k8s_helm_conn.py b/n2vc/k8s_helm_conn.py
index 496e320..ff1f331 100644
--- a/n2vc/k8s_helm_conn.py
+++ b/n2vc/k8s_helm_conn.py
@@ -67,7 +67,7 @@
on_update_db=on_update_db
)
- self.info('Initializing K8S Helm connector')
+ self.log.info('Initializing K8S Helm connector')
# random numbers for release name generation
random.seed(time.time())
@@ -84,7 +84,7 @@
self._check_file_exists(filename=helm_command, exception_if_not_exists=True)
# initialize helm client-only
- self.debug('Initializing helm client-only...')
+ self.log.debug('Initializing helm client-only...')
command = '{} init --client-only'.format(self._helm_command)
try:
asyncio.ensure_future(self._local_async_exec(command=command, raise_exception_on_error=False))
@@ -93,7 +93,7 @@
except Exception as e:
self.warning(msg='helm init failed (it was already initialized): {}'.format(e))
- self.info('K8S Helm connector initialized')
+ self.log.info('K8S Helm connector initialized')
async def init_env(
self,
@@ -117,7 +117,7 @@
if not cluster_uuid:
cluster_uuid = str(uuid4())
- self.debug('Initializing K8S environment. namespace: {}'.format(namespace))
+ self.log.debug('Initializing K8S environment. namespace: {}'.format(namespace))
# create config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -146,7 +146,7 @@
# helm init
n2vc_installed_sw = False
if not already_initialized:
- self.info('Initializing helm in client and server: {}'.format(cluster_uuid))
+ self.log.info('Initializing helm in client and server: {}'.format(cluster_uuid))
command = '{} --kubeconfig={} --tiller-namespace={} --home={} init'\
.format(self._helm_command, config_filename, namespace, helm_dir)
output, rc = await self._local_async_exec(command=command, raise_exception_on_error=True)
@@ -155,14 +155,14 @@
# check client helm installation
check_file = helm_dir + '/repository/repositories.yaml'
if not self._check_file_exists(filename=check_file, exception_if_not_exists=False):
- self.info('Initializing helm in client: {}'.format(cluster_uuid))
+ self.log.info('Initializing helm in client: {}'.format(cluster_uuid))
command = '{} --kubeconfig={} --tiller-namespace={} --home={} init --client-only'\
.format(self._helm_command, config_filename, namespace, helm_dir)
output, rc = await self._local_async_exec(command=command, raise_exception_on_error=True)
else:
- self.info('Helm client already initialized')
+ self.log.info('Helm client already initialized')
- self.info('Cluster initialized {}'.format(cluster_uuid))
+ self.log.info('Cluster initialized {}'.format(cluster_uuid))
return cluster_uuid, n2vc_installed_sw
@@ -174,7 +174,7 @@
repo_type: str = 'chart'
):
- self.debug('adding {} repository {}. URL: {}'.format(repo_type, name, url))
+ self.log.debug('adding {} repository {}. URL: {}'.format(repo_type, name, url))
# config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -182,13 +182,13 @@
# helm repo update
command = '{} --kubeconfig={} --home={} repo update'.format(self._helm_command, config_filename, helm_dir)
- self.debug('updating repo: {}'.format(command))
+ self.log.debug('updating repo: {}'.format(command))
await self._local_async_exec(command=command, raise_exception_on_error=False)
# helm repo add name url
command = '{} --kubeconfig={} --home={} repo add {} {}'\
.format(self._helm_command, config_filename, helm_dir, name, url)
- self.debug('adding repo: {}'.format(command))
+ self.log.debug('adding repo: {}'.format(command))
await self._local_async_exec(command=command, raise_exception_on_error=True)
async def repo_list(
@@ -201,7 +201,7 @@
:return: list of registered repositories: [ (name, url) .... ]
"""
- self.debug('list repositories for cluster {}'.format(cluster_uuid))
+ self.log.debug('list repositories for cluster {}'.format(cluster_uuid))
# config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -229,7 +229,7 @@
:return: True if successful
"""
- self.debug('list repositories for cluster {}'.format(cluster_uuid))
+ self.log.debug('list repositories for cluster {}'.format(cluster_uuid))
# config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -247,7 +247,7 @@
uninstall_sw: bool = False
) -> bool:
- self.debug('Resetting K8s environment. cluster uuid: {}'.format(cluster_uuid))
+ self.log.debug('Resetting K8s environment. cluster uuid: {}'.format(cluster_uuid))
# get kube and helm directories
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -261,19 +261,19 @@
try:
kdu_instance = r.get('Name')
chart = r.get('Chart')
- self.debug('Uninstalling {} -> {}'.format(chart, kdu_instance))
+ self.log.debug('Uninstalling {} -> {}'.format(chart, kdu_instance))
await self.uninstall(cluster_uuid=cluster_uuid, kdu_instance=kdu_instance)
except Exception as e:
- self.error('Error uninstalling release {}: {}'.format(kdu_instance, e))
+ self.log.error('Error uninstalling release {}: {}'.format(kdu_instance, e))
else:
msg = 'Cluster has releases and not force. Cannot reset K8s environment. Cluster uuid: {}'\
.format(cluster_uuid)
- self.error(msg)
+ self.log.error(msg)
raise K8sException(msg)
if uninstall_sw:
- self.debug('Uninstalling tiller from cluster {}'.format(cluster_uuid))
+ self.log.debug('Uninstalling tiller from cluster {}'.format(cluster_uuid))
# find namespace for tiller pod
command = '{} --kubeconfig={} get deployments --all-namespaces'\
@@ -290,31 +290,31 @@
pass
else:
msg = 'Tiller deployment not found in cluster {}'.format(cluster_uuid)
- self.error(msg)
+ self.log.error(msg)
- self.debug('namespace for tiller: {}'.format(namespace))
+ self.log.debug('namespace for tiller: {}'.format(namespace))
force_str = '--force'
if namespace:
# delete tiller deployment
- self.debug('Deleting tiller deployment for cluster {}, namespace {}'.format(cluster_uuid, namespace))
+ self.log.debug('Deleting tiller deployment for cluster {}, namespace {}'.format(cluster_uuid, namespace))
command = '{} --namespace {} --kubeconfig={} {} delete deployment tiller-deploy'\
.format(self.kubectl_command, namespace, config_filename, force_str)
await self._local_async_exec(command=command, raise_exception_on_error=False)
# uninstall tiller from cluster
- self.debug('Uninstalling tiller from cluster {}'.format(cluster_uuid))
+ self.log.debug('Uninstalling tiller from cluster {}'.format(cluster_uuid))
command = '{} --kubeconfig={} --home={} reset'\
.format(self._helm_command, config_filename, helm_dir)
- self.debug('resetting: {}'.format(command))
+ self.log.debug('resetting: {}'.format(command))
output, rc = await self._local_async_exec(command=command, raise_exception_on_error=True)
else:
- self.debug('namespace not found')
+ self.log.debug('namespace not found')
# delete cluster directory
dir = self.fs.path + '/' + cluster_uuid
- self.debug('Removing directory {}'.format(dir))
+ self.log.debug('Removing directory {}'.format(dir))
shutil.rmtree(dir, ignore_errors=True)
return True
@@ -330,7 +330,7 @@
kdu_name: str = None
):
- self.debug('installing {} in cluster {}'.format(kdu_model, cluster_uuid))
+ self.log.debug('installing {} in cluster {}'.format(kdu_model, cluster_uuid))
# config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -377,7 +377,7 @@
command = '{} install {} --output yaml --kubeconfig={} --home={} {} {} --name={} {} {}'\
.format(self._helm_command, atomic_str, config_filename, helm_dir,
params_str, timeout_str, kdu_instance, kdu_model, version_str)
- self.debug('installing: {}'.format(command))
+ self.log.debug('installing: {}'.format(command))
if atomic:
# exec helm in a task
@@ -424,10 +424,10 @@
if rc != 0:
msg = 'Error executing command: {}\nOutput: {}'.format(command, output)
- self.error(msg)
+ self.log.error(msg)
raise K8sException(msg)
- self.debug('Returning kdu_instance {}'.format(kdu_instance))
+ self.log.debug('Returning kdu_instance {}'.format(kdu_instance))
return kdu_instance
async def instances_list(
@@ -441,7 +441,7 @@
:return:
"""
- self.debug('list releases for cluster {}'.format(cluster_uuid))
+ self.log.debug('list releases for cluster {}'.format(cluster_uuid))
# config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -468,7 +468,7 @@
db_dict: dict = None
):
- self.debug('upgrading {} in cluster {}'.format(kdu_model, cluster_uuid))
+ self.log.debug('upgrading {} in cluster {}'.format(kdu_model, cluster_uuid))
# config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -499,7 +499,7 @@
command = '{} upgrade {} --output yaml --kubeconfig={} --home={} {} {} {} {} {}'\
.format(self._helm_command, atomic_str, config_filename, helm_dir,
params_str, timeout_str, kdu_instance, kdu_model, version_str)
- self.debug('upgrading: {}'.format(command))
+ self.log.debug('upgrading: {}'.format(command))
if atomic:
@@ -545,14 +545,14 @@
if rc != 0:
msg = 'Error executing command: {}\nOutput: {}'.format(command, output)
- self.error(msg)
+ self.log.error(msg)
raise K8sException(msg)
# return new revision number
instance = await self.get_instance_info(cluster_uuid=cluster_uuid, kdu_instance=kdu_instance)
if instance:
revision = int(instance.get('Revision'))
- self.debug('New revision: {}'.format(revision))
+ self.log.debug('New revision: {}'.format(revision))
return revision
else:
return 0
@@ -565,7 +565,7 @@
db_dict: dict = None
):
- self.debug('rollback kdu_instance {} to revision {} from cluster {}'
+ self.log.debug('rollback kdu_instance {} to revision {} from cluster {}'
.format(kdu_instance, revision, cluster_uuid))
# config filename
@@ -610,14 +610,14 @@
if rc != 0:
msg = 'Error executing command: {}\nOutput: {}'.format(command, output)
- self.error(msg)
+ self.log.error(msg)
raise K8sException(msg)
# return new revision number
instance = await self.get_instance_info(cluster_uuid=cluster_uuid, kdu_instance=kdu_instance)
if instance:
revision = int(instance.get('Revision'))
- self.debug('New revision: {}'.format(revision))
+ self.log.debug('New revision: {}'.format(revision))
return revision
else:
return 0
@@ -636,7 +636,7 @@
:return: True if successful
"""
- self.debug('uninstall kdu_instance {} from cluster {}'.format(kdu_instance, cluster_uuid))
+ self.log.debug('uninstall kdu_instance {} from cluster {}'.format(kdu_instance, cluster_uuid))
# config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -655,7 +655,7 @@
repo_url: str = None
) -> str:
- self.debug('inspect kdu_model {} from (optional) repo: {}'.format(kdu_model, repo_url))
+ self.log.debug('inspect kdu_model {} from (optional) repo: {}'.format(kdu_model, repo_url))
return await self._exec_inspect_comand(inspect_command='', kdu_model=kdu_model, repo_url=repo_url)
@@ -665,7 +665,7 @@
repo_url: str = None
) -> str:
- self.debug('inspect kdu_model values {} from (optional) repo: {}'.format(kdu_model, repo_url))
+ self.log.debug('inspect kdu_model values {} from (optional) repo: {}'.format(kdu_model, repo_url))
return await self._exec_inspect_comand(inspect_command='values', kdu_model=kdu_model, repo_url=repo_url)
@@ -675,7 +675,7 @@
repo_url: str = None
) -> str:
- self.debug('inspect kdu_model {} readme.md from repo: {}'.format(kdu_model, repo_url))
+ self.log.debug('inspect kdu_model {} readme.md from repo: {}'.format(kdu_model, repo_url))
return await self._exec_inspect_comand(inspect_command='readme', kdu_model=kdu_model, repo_url=repo_url)
@@ -695,7 +695,7 @@
async def synchronize_repos(self, cluster_uuid: str):
- self.debug("syncronize repos for cluster helm-id: {}",)
+ self.log.debug("syncronize repos for cluster helm-id: {}",)
try:
update_repos_timeout = 300 # max timeout to sync a single repos, more than this is too much
db_k8scluster = self.db.get_one("k8sclusters", {"_admin.helm-chart.id": cluster_uuid})
@@ -705,8 +705,8 @@
# elements that must be deleted
deleted_repo_list = []
added_repo_dict = {}
- self.debug("helm_chart_repos: {}".format(nbi_repo_list))
- self.debug("helm_charts_added: {}".format(cluster_repo_dict))
+ self.log.debug("helm_chart_repos: {}".format(nbi_repo_list))
+ self.log.debug("helm_charts_added: {}".format(cluster_repo_dict))
# obtain repos to add: registered by nbi but not added
repos_to_add = [repo for repo in nbi_repo_list if not cluster_repo_dict.get(repo)]
@@ -716,7 +716,7 @@
# delete repos: must delete first then add because there may be different repos with same name but
# different id and url
- self.debug("repos to delete: {}".format(repos_to_delete))
+ self.log.debug("repos to delete: {}".format(repos_to_delete))
for repo_id in repos_to_delete:
# try to delete repos
try:
@@ -730,25 +730,25 @@
deleted_repo_list.append(repo_id)
# add repos
- self.debug("repos to add: {}".format(repos_to_add))
+ self.log.debug("repos to add: {}".format(repos_to_add))
add_task_list = []
for repo_id in repos_to_add:
# obtain the repo data from the db
# if there is an error getting the repo in the database we will ignore this repo and continue
# because there is a possible race condition where the repo has been deleted while processing
db_repo = self.db.get_one("k8srepos", {"_id": repo_id})
- self.debug("obtained repo: id, {}, name: {}, url: {}".format(repo_id, db_repo["name"], db_repo["url"]))
+ self.log.debug("obtained repo: id, {}, name: {}, url: {}".format(repo_id, db_repo["name"], db_repo["url"]))
try:
repo_add_task = asyncio.ensure_future(self.repo_add(cluster_uuid=cluster_uuid,
name=db_repo["name"], url=db_repo["url"],
repo_type="chart"))
await asyncio.wait_for(repo_add_task, update_repos_timeout)
added_repo_dict[repo_id] = db_repo["name"]
- self.debug("added repo: id, {}, name: {}".format(repo_id, db_repo["name"]))
+ self.log.debug("added repo: id, {}, name: {}".format(repo_id, db_repo["name"]))
except Exception as e:
# deal with error adding repo, adding a repo that already exists does not raise any error
# will not raise error because a wrong repos added by anyone could prevent instantiating any ns
- self.error("Error adding repo id: {}, err_msg: {} ".format(repo_id, repr(e)))
+ self.log.error("Error adding repo id: {}, err_msg: {} ".format(repo_id, repr(e)))
return deleted_repo_list, added_repo_dict
@@ -756,7 +756,7 @@
raise K8sException("k8cluster with helm-id : {} not found".format(cluster_uuid))
except Exception as e:
- self.error("Error synchronizing repos: {}".format(str(e)))
+ self.log.error("Error synchronizing repos: {}".format(str(e)))
raise K8sException("Error synchronizing repos")
"""
@@ -793,7 +793,7 @@
return_text: bool = False
):
- self.debug('status of kdu_instance {}'.format(kdu_instance))
+ self.log.debug('status of kdu_instance {}'.format(kdu_instance))
# config filename
kube_dir, helm_dir, config_filename, cluster_dir = \
@@ -841,7 +841,7 @@
for instance in instances:
if instance.get('Name') == kdu_instance:
return instance
- self.debug('Instance {} not found'.format(kdu_instance))
+ self.log.debug('Instance {} not found'.format(kdu_instance))
return None
@staticmethod
@@ -895,8 +895,8 @@
await asyncio.sleep(check_every)
detailed_status = await self.status_kdu(cluster_uuid=cluster_uuid, kdu_instance=kdu_instance)
status = detailed_status.get('info').get('Description')
- self.debug('STATUS:\n{}'.format(status))
- self.debug('DETAILED STATUS:\n{}'.format(detailed_status))
+ self.log.debug('STATUS:\n{}'.format(status))
+ self.log.debug('DETAILED STATUS:\n{}'.format(detailed_status))
# write status to db
result = await self.write_app_status_to_db(
db_dict=db_dict,
@@ -904,13 +904,13 @@
detailed_status=str(detailed_status),
operation=operation)
if not result:
- self.info('Error writing in database. Task exiting...')
+ self.log.info('Error writing in database. Task exiting...')
return
except asyncio.CancelledError:
- self.debug('Task cancelled')
+ self.log.debug('Task cancelled')
return
except Exception as e:
- self.debug('_store_status exception: {}'.format(str(e)))
+ self.log.debug('_store_status exception: {}'.format(str(e)))
pass
finally:
if run_once:
@@ -957,7 +957,7 @@
current = int(parts[0])
total = int(parts[1])
if current < total:
- self.debug('NOT READY:\n {}'.format(line3))
+ self.log.debug('NOT READY:\n {}'.format(line3))
ready = False
line3 = resources[index]
index += 1
@@ -1085,31 +1085,31 @@
# base dir for cluster
cluster_dir = base + '/' + cluster_name
if create_if_not_exist and not os.path.exists(cluster_dir):
- self.debug('Creating dir {}'.format(cluster_dir))
+ self.log.debug('Creating dir {}'.format(cluster_dir))
os.makedirs(cluster_dir)
if not os.path.exists(cluster_dir):
msg = 'Base cluster dir {} does not exist'.format(cluster_dir)
- self.error(msg)
+ self.log.error(msg)
raise K8sException(msg)
# kube dir
kube_dir = cluster_dir + '/' + '.kube'
if create_if_not_exist and not os.path.exists(kube_dir):
- self.debug('Creating dir {}'.format(kube_dir))
+ self.log.debug('Creating dir {}'.format(kube_dir))
os.makedirs(kube_dir)
if not os.path.exists(kube_dir):
msg = 'Kube config dir {} does not exist'.format(kube_dir)
- self.error(msg)
+ self.log.error(msg)
raise K8sException(msg)
# helm home dir
helm_dir = cluster_dir + '/' + '.helm'
if create_if_not_exist and not os.path.exists(helm_dir):
- self.debug('Creating dir {}'.format(helm_dir))
+ self.log.debug('Creating dir {}'.format(helm_dir))
os.makedirs(helm_dir)
if not os.path.exists(helm_dir):
msg = 'Helm config dir {} does not exist'.format(helm_dir)
- self.error(msg)
+ self.log.error(msg)
raise K8sException(msg)
config_filename = kube_dir + '/config'
@@ -1127,13 +1127,13 @@
command: str
) -> (str, int):
command = K8sHelmConnector._remove_multiple_spaces(command)
- self.debug('Executing sync local command: {}'.format(command))
+ self.log.debug('Executing sync local command: {}'.format(command))
# raise exception if fails
output = ''
try:
output = subprocess.check_output(command, shell=True, universal_newlines=True)
return_code = 0
- self.debug(output)
+ self.log.debug(output)
except Exception as e:
return_code = 1
@@ -1148,7 +1148,7 @@
) -> (str, int):
command = K8sHelmConnector._remove_multiple_spaces(command)
- self.debug('Executing async local command: {}'.format(command))
+ self.log.debug('Executing async local command: {}'.format(command))
# split command
command = command.split(sep=' ')
@@ -1174,9 +1174,9 @@
# output = stderr.decode()
if return_code != 0 and show_error_log:
- self.debug('Return code (FAIL): {}\nOutput:\n{}'.format(return_code, output))
+ self.log.debug('Return code (FAIL): {}\nOutput:\n{}'.format(return_code, output))
else:
- self.debug('Return code: {}'.format(return_code))
+ self.log.debug('Return code: {}'.format(return_code))
if raise_exception_on_error and return_code != 0:
raise K8sException(output)
@@ -1193,20 +1193,20 @@
raise
except Exception as e:
msg = 'Exception executing command: {} -> {}'.format(command, e)
- self.error(msg)
+ self.log.error(msg)
if raise_exception_on_error:
raise K8sException(e) from e
else:
return '', -1
def _check_file_exists(self, filename: str, exception_if_not_exists: bool = False):
- self.debug('Checking if file {} exists...'.format(filename))
+ self.log.debug('Checking if file {} exists...'.format(filename))
if os.path.exists(filename):
return True
else:
msg = 'File {} does not exist'.format(filename)
if exception_if_not_exists:
- self.error(msg)
+ self.log.error(msg)
raise K8sException(msg)
diff --git a/n2vc/n2vc_conn.py b/n2vc/n2vc_conn.py
index 6ed9aec..8811d71 100644
--- a/n2vc/n2vc_conn.py
+++ b/n2vc/n2vc_conn.py
@@ -95,7 +95,7 @@
if fs is None:
raise N2VCBadArgumentsException('Argument fs is mandatory', ['fs'])
- self.info('url={}, username={}, vca_config={}'.format(url, username, vca_config))
+ self.log.info('url={}, username={}, vca_config={}'.format(url, username, vca_config))
# store arguments into self
self.db = db
@@ -399,10 +399,10 @@
entity_type: str
):
if not db_dict:
- self.debug('No db_dict => No database write')
+ self.log.debug('No db_dict => No database write')
return
- # self.debug('status={} / detailed-status={} / VCA-status={} / entity_type={}'
+ # self.log.debug('status={} / detailed-status={} / VCA-status={} / entity_type={}'
# .format(str(status.value), detailed_status, vca_status, entity_type))
try:
@@ -436,9 +436,9 @@
except DbException as e:
if e.http_code == HTTPStatus.NOT_FOUND:
- self.error('NOT_FOUND error: Exception writing status to database: {}'.format(e))
+ self.log.error('NOT_FOUND error: Exception writing status to database: {}'.format(e))
else:
- self.info('Exception writing status to database: {}'.format(e))
+ self.log.info('Exception writing status to database: {}'.format(e))
def juju_status_2_osm_status(type: str, status: str) -> N2VCDeploymentStatus:
diff --git a/n2vc/n2vc_juju_conn.py b/n2vc/n2vc_juju_conn.py
index e0f1824..4f0ee37 100644
--- a/n2vc/n2vc_juju_conn.py
+++ b/n2vc/n2vc_juju_conn.py
@@ -86,7 +86,7 @@
logging.getLogger('juju.client.connection').setLevel(logging.WARN)
logging.getLogger('model').setLevel(logging.WARN)
- self.info('Initializing N2VC juju connector...')
+ self.log.info('Initializing N2VC juju connector...')
"""
##############################################################
@@ -143,7 +143,7 @@
cacert,
)
except binascii.Error as e:
- self.debug("Caught binascii.Error: {}".format(e))
+ self.log.debug("Caught binascii.Error: {}".format(e))
raise N2VCInvalidCertificate(message="Invalid CA Certificate")
return cacert
@@ -154,7 +154,7 @@
if 'api_proxy' in vca_config:
self.api_proxy = vca_config['api_proxy']
- self.debug('api_proxy for native charms configured: {}'.format(self.api_proxy))
+ self.log.debug('api_proxy for native charms configured: {}'.format(self.api_proxy))
else:
self.warning('api_proxy is not configured. Support for native charms is disabled')
@@ -168,7 +168,7 @@
else:
self.apt_mirror = None
- self.debug('Arguments have been checked')
+ self.log.debug('Arguments have been checked')
# juju data
self.controller = None # it will be filled when connect to juju
@@ -181,11 +181,11 @@
# create juju pub key file in lcm container at ./local/share/juju/ssh/juju_id_rsa.pub
self._create_juju_public_key()
- self.info('N2VC juju connector initialized')
+ self.log.info('N2VC juju connector initialized')
async def get_status(self, namespace: str, yaml_format: bool = True):
- # self.info('Getting NS status. namespace: {}'.format(namespace))
+ # self.log.info('Getting NS status. namespace: {}'.format(namespace))
if not self._authenticated:
await self._juju_login()
@@ -195,7 +195,7 @@
model_name = ns_id
if model_name is None:
msg = 'Namespace {} not valid'.format(namespace)
- self.error(msg)
+ self.log.error(msg)
raise N2VCBadArgumentsException(msg, ['namespace'])
# get juju model (create model if needed)
@@ -217,7 +217,7 @@
total_timeout: float = None
) -> (str, dict):
- self.info('Creating execution environment. namespace: {}, reuse_ee_id: {}'.format(namespace, reuse_ee_id))
+ self.log.info('Creating execution environment. namespace: {}, reuse_ee_id: {}'.format(namespace, reuse_ee_id))
if not self._authenticated:
await self._juju_login()
@@ -232,7 +232,7 @@
# application name
application_name = self._get_application_name(namespace=namespace)
- self.debug('model name: {}, application name: {}, machine_id: {}'
+ self.log.debug('model name: {}, application name: {}, machine_id: {}'
.format(model_name, application_name, machine_id))
# create or reuse a new juju machine
@@ -247,7 +247,7 @@
)
except Exception as e:
message = 'Error creating machine on juju: {}'.format(e)
- self.error(message)
+ self.log.error(message)
raise N2VCException(message=message)
# id for the execution environment
@@ -256,13 +256,13 @@
application_name=application_name,
machine_id=str(machine.entity_id)
)
- self.debug('ee_id: {}'.format(ee_id))
+ self.log.debug('ee_id: {}'.format(ee_id))
# new machine credentials
credentials = dict()
credentials['hostname'] = machine.dns_name
- self.info('Execution environment created. ee_id: {}, credentials: {}'.format(ee_id, credentials))
+ self.log.info('Execution environment created. ee_id: {}, credentials: {}'.format(ee_id, credentials))
return ee_id, credentials
@@ -278,7 +278,7 @@
if not self._authenticated:
await self._juju_login()
- self.info('Registering execution environment. namespace={}, credentials={}'.format(namespace, credentials))
+ self.log.info('Registering execution environment. namespace={}, credentials={}'.format(namespace, credentials))
if credentials is None:
raise N2VCBadArgumentsException(message='credentials are mandatory', bad_args=['credentials'])
@@ -315,10 +315,10 @@
total_timeout=total_timeout
)
except Exception as e:
- self.error('Error registering machine: {}'.format(e))
+ self.log.error('Error registering machine: {}'.format(e))
raise N2VCException(message='Error registering machine on juju: {}'.format(e))
- self.info('Machine registered: {}'.format(machine_id))
+ self.log.info('Machine registered: {}'.format(machine_id))
# id for the execution environment
ee_id = N2VCJujuConnector._build_ee_id(
@@ -327,7 +327,7 @@
machine_id=str(machine_id)
)
- self.info('Execution environment registered. ee_id: {}'.format(ee_id))
+ self.log.info('Execution environment registered. ee_id: {}'.format(ee_id))
return ee_id
@@ -340,7 +340,7 @@
total_timeout: float = None
):
- self.info('Installing configuration sw on ee_id: {}, artifact path: {}, db_dict: {}'
+ self.log.info('Installing configuration sw on ee_id: {}, artifact path: {}, db_dict: {}'
.format(ee_id, artifact_path, db_dict))
if not self._authenticated:
@@ -356,7 +356,7 @@
try:
model_name, application_name, machine_id = N2VCJujuConnector._get_ee_id_components(ee_id=ee_id)
- self.debug('model: {}, application: {}, machine: {}'.format(model_name, application_name, machine_id))
+ self.log.debug('model: {}, application: {}, machine: {}'.format(model_name, application_name, machine_id))
except Exception as e:
raise N2VCBadArgumentsException(
message='ee_id={} is not a valid execution environment id'.format(ee_id),
@@ -390,7 +390,7 @@
except Exception as e:
raise N2VCException(message='Error desploying charm into ee={} : {}'.format(ee_id, e))
- self.info('Configuration sw installed')
+ self.log.info('Configuration sw installed')
async def get_ee_ssh_public__key(
self,
@@ -400,7 +400,7 @@
total_timeout: float = None
) -> str:
- self.info('Generating priv/pub key pair and get pub key on ee_id: {}, db_dict: {}'.format(ee_id, db_dict))
+ self.log.info('Generating priv/pub key pair and get pub key on ee_id: {}, db_dict: {}'.format(ee_id, db_dict))
if not self._authenticated:
await self._juju_login()
@@ -413,7 +413,7 @@
try:
model_name, application_name, machine_id = N2VCJujuConnector._get_ee_id_components(ee_id=ee_id)
- self.debug('model: {}, application: {}, machine: {}'.format(model_name, application_name, machine_id))
+ self.log.debug('model: {}, application: {}, machine: {}'.format(model_name, application_name, machine_id))
except Exception as e:
raise N2VCBadArgumentsException(
message='ee_id={} is not a valid execution environment id'.format(ee_id),
@@ -437,7 +437,7 @@
total_timeout=total_timeout
)
except Exception as e:
- self.info('Cannot execute action generate-ssh-key: {}\nContinuing...'.format(e))
+ self.log.info('Cannot execute action generate-ssh-key: {}\nContinuing...'.format(e))
# execute action: get-ssh-public-key
try:
@@ -451,7 +451,7 @@
)
except Exception as e:
msg = 'Cannot execute action get-ssh-public-key: {}\n'.format(e)
- self.info(msg)
+ self.log.info(msg)
raise e
# return public key if exists
@@ -465,25 +465,25 @@
endpoint_2: str
):
- self.debug('adding new relation between {} and {}, endpoints: {}, {}'
+ self.log.debug('adding new relation between {} and {}, endpoints: {}, {}'
.format(ee_id_1, ee_id_2, endpoint_1, endpoint_2))
# check arguments
if not ee_id_1:
message = 'EE 1 is mandatory'
- self.error(message)
+ self.log.error(message)
raise N2VCBadArgumentsException(message=message, bad_args=['ee_id_1'])
if not ee_id_2:
message = 'EE 2 is mandatory'
- self.error(message)
+ self.log.error(message)
raise N2VCBadArgumentsException(message=message, bad_args=['ee_id_2'])
if not endpoint_1:
message = 'endpoint 1 is mandatory'
- self.error(message)
+ self.log.error(message)
raise N2VCBadArgumentsException(message=message, bad_args=['endpoint_1'])
if not endpoint_2:
message = 'endpoint 2 is mandatory'
- self.error(message)
+ self.log.error(message)
raise N2VCBadArgumentsException(message=message, bad_args=['endpoint_2'])
if not self._authenticated:
@@ -496,7 +496,7 @@
# model must be the same
if model_1 != model_2:
message = 'EE models are not the same: {} vs {}'.format(ee_id_1, ee_id_2)
- self.error(message)
+ self.log.error(message)
raise N2VCBadArgumentsException(message=message, bad_args=['ee_id_1', 'ee_id_2'])
# add juju relations between two applications
@@ -510,7 +510,7 @@
)
except Exception as e:
message = 'Error adding relation between {} and {}'.format(ee_id_1, ee_id_2)
- self.error(message)
+ self.log.error(message)
raise N2VCException(message=message)
async def remove_relation(
@@ -519,7 +519,7 @@
if not self._authenticated:
await self._juju_login()
# TODO
- self.info('Method not implemented yet')
+ self.log.info('Method not implemented yet')
raise NotImplemented()
async def deregister_execution_environments(
@@ -528,7 +528,7 @@
if not self._authenticated:
await self._juju_login()
# TODO
- self.info('Method not implemented yet')
+ self.log.info('Method not implemented yet')
raise NotImplemented()
async def delete_namespace(
@@ -537,7 +537,7 @@
db_dict: dict = None,
total_timeout: float = None
):
- self.info('Deleting namespace={}'.format(namespace))
+ self.log.info('Deleting namespace={}'.format(namespace))
if not self._authenticated:
await self._juju_login()
@@ -558,7 +558,7 @@
else:
raise N2VCBadArgumentsException(message='only ns_id is permitted to delete yet', bad_args=['namespace'])
- self.info('Namespace {} deleted'.format(namespace))
+ self.log.info('Namespace {} deleted'.format(namespace))
async def delete_execution_environment(
self,
@@ -566,7 +566,7 @@
db_dict: dict = None,
total_timeout: float = None
):
- self.info('Deleting execution environment ee_id={}'.format(ee_id))
+ self.log.info('Deleting execution environment ee_id={}'.format(ee_id))
if not self._authenticated:
await self._juju_login()
@@ -595,7 +595,7 @@
raise N2VCException(message='Error deleting execution environment {} (machine {}) : {}'
.format(ee_id, machine_id, e))
- self.info('Execution environment {} deleted'.format(ee_id))
+ self.log.info('Execution environment {} deleted'.format(ee_id))
async def exec_primitive(
self,
@@ -607,7 +607,7 @@
total_timeout: float = None
) -> str:
- self.info('Executing primitive: {} on ee: {}, params: {}'.format(primitive_name, ee_id, params_dict))
+ self.log.info('Executing primitive: {} on ee: {}, params: {}'.format(primitive_name, ee_id, params_dict))
if not self._authenticated:
await self._juju_login()
@@ -640,7 +640,7 @@
total_timeout=total_timeout
)
except Exception as e:
- self.error('Error configuring juju application: {}'.format(e))
+ self.log.error('Error configuring juju application: {}'.format(e))
raise N2VCExecutionException(
message='Error configuring application into ee={} : {}'.format(ee_id, e),
primitive_name=primitive_name
@@ -662,14 +662,14 @@
else:
raise Exception('status is not completed: {}'.format(status))
except Exception as e:
- self.error('Error executing primitive {}: {}'.format(primitive_name, e))
+ self.log.error('Error executing primitive {}: {}'.format(primitive_name, e))
raise N2VCExecutionException(
message='Error executing primitive {} into ee={} : {}'.format(primitive_name, ee_id, e),
primitive_name=primitive_name
)
async def disconnect(self):
- self.info('closing juju N2VC...')
+ self.log.info('closing juju N2VC...')
await self._juju_logout()
"""
@@ -692,7 +692,7 @@
if not the_path[-1] == '.':
the_path = the_path + '.'
update_dict = {the_path + 'ee_id': ee_id}
- # self.debug('Writing ee_id to database: {}'.format(the_path))
+ # self.log.debug('Writing ee_id to database: {}'.format(the_path))
self.db.set_one(
table=the_table,
q_filter=the_filter,
@@ -700,7 +700,7 @@
fail_on_empty=True
)
except Exception as e:
- self.error('Error writing ee_id to database: {}'.format(e))
+ self.log.error('Error writing ee_id to database: {}'.format(e))
@staticmethod
def _build_ee_id(
@@ -781,7 +781,7 @@
total_timeout: float = None
) -> Machine:
- self.debug('creating machine in model: {}, existing machine id: {}'.format(model_name, machine_id))
+ self.log.debug('creating machine in model: {}, existing machine id: {}'.format(model_name, machine_id))
# get juju model and observer (create model if needed)
model = await self._juju_get_model(model_name=model_name)
@@ -790,15 +790,15 @@
# find machine id in model
machine = None
if machine_id is not None:
- self.debug('Finding existing machine id {} in model'.format(machine_id))
+ self.log.debug('Finding existing machine id {} in model'.format(machine_id))
# get juju existing machines in the model
existing_machines = await model.get_machines()
if machine_id in existing_machines:
- self.debug('Machine id {} found in model (reusing it)'.format(machine_id))
+ self.log.debug('Machine id {} found in model (reusing it)'.format(machine_id))
machine = model.machines[machine_id]
if machine is None:
- self.debug('Creating a new machine in juju...')
+ self.log.debug('Creating a new machine in juju...')
# machine does not exist, create it and wait for it
machine = await model.add_machine(
spec=None,
@@ -832,7 +832,7 @@
else:
- self.debug('Reusing old machine pending')
+ self.log.debug('Reusing old machine pending')
# register machine with observer
observer.register_machine(machine=machine, db_dict=db_dict)
@@ -843,7 +843,7 @@
progress_timeout=progress_timeout,
total_timeout=total_timeout)
- self.debug("Machine ready at " + str(machine.dns_name))
+ self.log.debug("Machine ready at " + str(machine.dns_name))
return machine
async def _juju_provision_machine(
@@ -859,10 +859,10 @@
if not self.api_proxy:
msg = 'Cannot provision machine: api_proxy is not defined'
- self.error(msg=msg)
+ self.log.error(msg=msg)
raise N2VCException(message=msg)
- self.debug('provisioning machine. model: {}, hostname: {}, username: {}'.format(model_name, hostname, username))
+ self.log.debug('provisioning machine. model: {}, hostname: {}, username: {}'.format(model_name, hostname, username))
if not self._authenticated:
await self._juju_login()
@@ -894,20 +894,20 @@
connection = model.connection()
# Submit the request.
- self.debug("Adding machine to model")
+ self.log.debug("Adding machine to model")
client_facade = client.ClientFacade.from_connection(connection)
results = await client_facade.AddMachines(params=[params])
error = results.machines[0].error
if error:
msg = "Error adding machine: {}}".format(error.message)
- self.error(msg=msg)
+ self.log.error(msg=msg)
raise ValueError(msg)
machine_id = results.machines[0].machine
# Need to run this after AddMachines has been called,
# as we need the machine_id
- self.debug("Installing Juju agent into machine {}".format(machine_id))
+ self.log.debug("Installing Juju agent into machine {}".format(machine_id))
asyncio.ensure_future(provisioner.install_agent(
connection=connection,
nonce=params.nonce,
@@ -920,28 +920,28 @@
for i in range(10):
machine_list = await model.get_machines()
if machine_id in machine_list:
- self.debug('Machine {} found in model!'.format(machine_id))
+ self.log.debug('Machine {} found in model!'.format(machine_id))
machine = model.machines.get(machine_id)
break
await asyncio.sleep(2)
if machine is None:
msg = 'Machine {} not found in model'.format(machine_id)
- self.error(msg=msg)
+ self.log.error(msg=msg)
raise Exception(msg)
# register machine with observer
observer.register_machine(machine=machine, db_dict=db_dict)
# wait for machine creation
- self.debug('waiting for provision finishes... {}'.format(machine_id))
+ self.log.debug('waiting for provision finishes... {}'.format(machine_id))
await observer.wait_for_machine(
machine_id=machine_id,
progress_timeout=progress_timeout,
total_timeout=total_timeout
)
- self.debug("Machine provisioned {}".format(machine_id))
+ self.log.debug("Machine provisioned {}".format(machine_id))
return machine_id
@@ -968,9 +968,9 @@
if application is None:
# application does not exist, create it and wait for it
- self.debug('deploying application {} to machine {}, model {}'
+ self.log.debug('deploying application {} to machine {}, model {}'
.format(application_name, machine_id, model_name))
- self.debug('charm: {}'.format(charm_path))
+ self.log.debug('charm: {}'.format(charm_path))
series = 'xenial'
# series = None
application = await model.deploy(
@@ -985,12 +985,12 @@
# register application with observer
observer.register_application(application=application, db_dict=db_dict)
- self.debug('waiting for application deployed... {}'.format(application.entity_id))
+ self.log.debug('waiting for application deployed... {}'.format(application.entity_id))
retries = await observer.wait_for_application(
application_id=application.entity_id,
progress_timeout=progress_timeout,
total_timeout=total_timeout)
- self.debug('application deployed')
+ self.log.debug('application deployed')
else:
@@ -998,12 +998,12 @@
observer.register_application(application=application, db_dict=db_dict)
# application already exists, but not finalised
- self.debug('application already exists, waiting for deployed...')
+ self.log.debug('application already exists, waiting for deployed...')
retries = await observer.wait_for_application(
application_id=application.entity_id,
progress_timeout=progress_timeout,
total_timeout=total_timeout)
- self.debug('application deployed')
+ self.log.debug('application deployed')
return application, retries
@@ -1028,7 +1028,7 @@
if unit is not None:
actions = await application.get_actions()
if action_name in actions:
- self.debug('executing action "{}" using params: {}'.format(action_name, kwargs))
+ self.log.debug('executing action "{}" using params: {}'.format(action_name, kwargs))
action = await unit.run_action(action_name, **kwargs)
# register action with observer
@@ -1038,7 +1038,7 @@
action_id=action.entity_id,
progress_timeout=progress_timeout,
total_timeout=total_timeout)
- self.debug('action completed with status: {}'.format(action.status))
+ self.log.debug('action completed with status: {}'.format(action.status))
output = await model.get_action_output(action_uuid=action.entity_id)
status = await model.get_action_status(uuid_or_prefix=action.entity_id)
if action.entity_id in status:
@@ -1065,15 +1065,15 @@
# get the application
application = await self._juju_get_application(model_name=model_name, application_name=application_name)
- self.debug('configuring the application {} -> {}'.format(application_name, config))
+ self.log.debug('configuring the application {} -> {}'.format(application_name, config))
res = await application.set_config(config)
- self.debug('application {} configured. res={}'.format(application_name, res))
+ self.log.debug('application {} configured. res={}'.format(application_name, res))
# Verify the config is set
new_conf = await application.get_config()
for key in config:
value = new_conf[key]['value']
- self.debug(' {} = {}'.format(key, value))
+ self.log.debug(' {} = {}'.format(key, value))
if config[key] != value:
raise N2VCException(
message='key {} is not configured correctly {} != {}'.format(key, config[key], new_conf[key])
@@ -1084,7 +1084,7 @@
actions = await application.get_actions()
if 'verify-ssh-credentials' not in actions:
msg = 'Action verify-ssh-credentials does not exist in application {}'.format(application_name)
- self.debug(msg=msg)
+ self.log.debug(msg=msg)
return False
# execute verify-credentials
@@ -1092,7 +1092,7 @@
retry_timeout = 15.0
for i in range(num_retries):
try:
- self.debug('Executing action verify-ssh-credentials...')
+ self.log.debug('Executing action verify-ssh-credentials...')
output, ok = await self._juju_execute_action(
model_name=model_name,
application_name=application_name,
@@ -1101,13 +1101,13 @@
progress_timeout=progress_timeout,
total_timeout=total_timeout
)
- self.debug('Result: {}, output: {}'.format(ok, output))
+ self.log.debug('Result: {}, output: {}'.format(ok, output))
return True
except Exception as e:
- self.debug('Error executing verify-ssh-credentials: {}. Retrying...'.format(e))
+ self.log.debug('Error executing verify-ssh-credentials: {}. Retrying...'.format(e))
await asyncio.sleep(retry_timeout)
else:
- self.error('Error executing verify-ssh-credentials after {} retries. '.format(num_retries))
+ self.log.error('Error executing verify-ssh-credentials after {} retries. '.format(num_retries))
return False
async def _juju_get_application(
@@ -1141,7 +1141,7 @@
return self.juju_models[model_name]
if self._creating_model:
- self.debug('Another coroutine is creating a model. Wait...')
+ self.log.debug('Another coroutine is creating a model. Wait...')
while self._creating_model:
# another coroutine is creating a model, wait
await asyncio.sleep(0.1)
@@ -1156,7 +1156,7 @@
model_list = await self.controller.list_models()
if model_name not in model_list:
- self.info('Model {} does not exist. Creating new model...'.format(model_name))
+ self.log.info('Model {} does not exist. Creating new model...'.format(model_name))
config_dict = {'authorized-keys': self.public_key}
if self.apt_mirror:
config_dict['apt-mirror'] = self.apt_mirror
@@ -1168,11 +1168,11 @@
model_name=model_name,
config=config_dict
)
- self.info('New model created, name={}'.format(model_name))
+ self.log.info('New model created, name={}'.format(model_name))
else:
- self.debug('Model already exists in juju. Getting model {}'.format(model_name))
+ self.log.debug('Model already exists in juju. Getting model {}'.format(model_name))
model = await self.controller.get_model(model_name)
- self.debug('Existing model in juju, name={}'.format(model_name))
+ self.log.debug('Existing model in juju, name={}'.format(model_name))
self.juju_models[model_name] = model
self.juju_observers[model_name] = JujuModelObserver(n2vc=self, model=model)
@@ -1180,7 +1180,7 @@
except Exception as e:
msg = 'Cannot get model {}. Exception: {}'.format(model_name, e)
- self.error(msg)
+ self.log.error(msg)
raise N2VCException(msg)
finally:
self._creating_model = False
@@ -1200,7 +1200,7 @@
r1 = '{}:{}'.format(application_name_1, relation_1)
r2 = '{}:{}'.format(application_name_2, relation_2)
- self.debug('adding relation: {} -> {}'.format(r1, r2))
+ self.log.debug('adding relation: {} -> {}'.format(r1, r2))
try:
await model.add_relation(relation1=r1, relation2=r2)
except JujuAPIError as e:
@@ -1219,7 +1219,7 @@
application_name: str
):
- self.debug('Destroying application {} in model {}'.format(application_name, model_name))
+ self.log.debug('Destroying application {} in model {}'.format(application_name, model_name))
# get juju model and observer
model = await self._juju_get_model(model_name=model_name)
@@ -1228,7 +1228,7 @@
if application:
await application.destroy()
else:
- self.debug('Application not found: {}'.format(application_name))
+ self.log.debug('Application not found: {}'.format(application_name))
async def _juju_destroy_machine(
self,
@@ -1237,7 +1237,7 @@
total_timeout: float = None
):
- self.debug('Destroying machine {} in model {}'.format(machine_id, model_name))
+ self.log.debug('Destroying machine {} in model {}'.format(machine_id, model_name))
if total_timeout is None:
total_timeout = 3600
@@ -1254,12 +1254,12 @@
# wait for machine removal
machines = await model.get_machines()
while machine_id in machines and time.time() < end:
- self.debug('Waiting for machine {} is destroyed'.format(machine_id))
+ self.log.debug('Waiting for machine {} is destroyed'.format(machine_id))
await asyncio.sleep(0.5)
machines = await model.get_machines()
- self.debug('Machine destroyed: {}'.format(machine_id))
+ self.log.debug('Machine destroyed: {}'.format(machine_id))
else:
- self.debug('Machine not found: {}'.format(machine_id))
+ self.log.debug('Machine not found: {}'.format(machine_id))
async def _juju_destroy_model(
self,
@@ -1267,7 +1267,7 @@
total_timeout: float = None
):
- self.debug('Destroying model {}'.format(model_name))
+ self.log.debug('Destroying model {}'.format(model_name))
if total_timeout is None:
total_timeout = 3600
@@ -1288,19 +1288,19 @@
self.juju_models[model_name] = None
self.juju_observers[model_name] = None
- self.debug('destroying model {}...'.format(model_name))
+ self.log.debug('destroying model {}...'.format(model_name))
await self.controller.destroy_model(uuid)
- self.debug('model destroy requested {}'.format(model_name))
+ self.log.debug('model destroy requested {}'.format(model_name))
# wait for model is completely destroyed
end = time.time() + total_timeout
while time.time() < end:
- self.debug('Waiting for model is destroyed...')
+ self.log.debug('Waiting for model is destroyed...')
try:
# await self.controller.get_model(uuid)
models = await self.controller.list_models()
if model_name not in models:
- self.debug('The model {} ({}) was destroyed'.format(model_name, uuid))
+ self.log.debug('The model {} ({}) was destroyed'.format(model_name, uuid))
return
except Exception as e:
pass
@@ -1326,7 +1326,7 @@
try:
self._connecting = True
- self.info(
+ self.log.info(
'connecting to juju controller: {} {}:{} ca_cert: {}'
.format(self.url, self.username, self.secret, '\n'+self.ca_cert if self.ca_cert else 'None'))
@@ -1340,10 +1340,10 @@
cacert=self.ca_cert
)
self._authenticated = True
- self.info('juju controller connected')
+ self.log.info('juju controller connected')
except Exception as e:
message = 'Exception connecting to juju: {}'.format(e)
- self.error(message)
+ self.log.error(message)
raise N2VCConnectionException(
message=message,
url=self.url
@@ -1361,10 +1361,10 @@
try:
await self._juju_disconnect_model(model_name)
except Exception as e:
- self.error('Error disconnecting model {} : {}'.format(model_name, e))
+ self.log.error('Error disconnecting model {} : {}'.format(model_name, e))
# continue with next model...
- self.info("Disconnecting controller")
+ self.log.info("Disconnecting controller")
try:
await self.controller.disconnect()
except Exception as e:
@@ -1372,13 +1372,13 @@
self.controller = None
self._authenticated = False
- self.info('disconnected')
+ self.log.info('disconnected')
async def _juju_disconnect_model(
self,
model_name: str
):
- self.debug("Disconnecting model {}".format(model_name))
+ self.log.debug("Disconnecting model {}".format(model_name))
if model_name in self.juju_models:
await self.juju_models[model_name].disconnect()
self.juju_models[model_name] = None
@@ -1404,15 +1404,15 @@
pk_path = "{}/.local/share/juju/ssh".format(os.path.expanduser('~'))
file_path = "{}/juju_id_rsa.pub".format(pk_path)
- self.debug('writing juju public key to file:\n{}\npublic key: {}'.format(file_path, self.public_key))
+ self.log.debug('writing juju public key to file:\n{}\npublic key: {}'.format(file_path, self.public_key))
if not os.path.exists(pk_path):
# create path and write file
os.makedirs(pk_path)
with open(file_path, 'w') as f:
- self.debug('Creating juju public key file: {}'.format(file_path))
+ self.log.debug('Creating juju public key file: {}'.format(file_path))
f.write(self.public_key)
else:
- self.debug('juju public key file already exists: {}'.format(file_path))
+ self.log.debug('juju public key file already exists: {}'.format(file_path))
@staticmethod
def _format_model_name(name: str) -> str: