2 # Copyright 2019 Telefonica Investigacion y Desarrollo, S.A.U.
3 # This file is part of OSM
6 # Licensed under the Apache License, Version 2.0 (the "License");
7 # you may not use this file except in compliance with the License.
8 # You may obtain a copy of the License at
10 # http://www.apache.org/licenses/LICENSE-2.0
12 # Unless required by applicable law or agreed to in writing, software
13 # distributed under the License is distributed on an "AS IS" BASIS,
14 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
16 # See the License for the specific language governing permissions and
17 # limitations under the License.
19 # For those usages not covered by the Apache License, Version 2.0 please
20 # contact with: nfvlabs@tid.es
29 from uuid
import uuid4
31 from n2vc
.k8s_conn
import K8sConnector
32 from n2vc
.exceptions
import K8sException
35 class K8sHelmConnector(K8sConnector
):
38 ##################################################################################################
39 ########################################## P U B L I C ###########################################
40 ##################################################################################################
47 kubectl_command
: str = '/usr/bin/kubectl',
48 helm_command
: str = '/usr/bin/helm',
54 :param fs: file system for kubernetes and helm configuration
55 :param db: database object to write current operation status
56 :param kubectl_command: path to kubectl executable
57 :param helm_command: path to helm executable
59 :param on_update_db: callback called when k8s connector updates database
63 K8sConnector
.__init
__(
67 on_update_db
=on_update_db
70 self
.log
.info('Initializing K8S Helm connector')
72 # random numbers for release name generation
73 random
.seed(time
.time())
78 # exception if kubectl is not installed
79 self
.kubectl_command
= kubectl_command
80 self
._check
_file
_exists
(filename
=kubectl_command
, exception_if_not_exists
=True)
82 # exception if helm is not installed
83 self
._helm
_command
= helm_command
84 self
._check
_file
_exists
(filename
=helm_command
, exception_if_not_exists
=True)
86 # initialize helm client-only
87 self
.log
.debug('Initializing helm client-only...')
88 command
= '{} init --client-only'.format(self
._helm
_command
)
90 asyncio
.ensure_future(self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False))
91 # loop = asyncio.get_event_loop()
92 # loop.run_until_complete(self._local_async_exec(command=command, raise_exception_on_error=False))
93 except Exception as e
:
94 self
.warning(msg
='helm init failed (it was already initialized): {}'.format(e
))
96 self
.log
.info('K8S Helm connector initialized')
101 namespace
: str = 'kube-system',
102 reuse_cluster_uuid
=None
105 It prepares a given K8s cluster environment to run Charts on both sides:
109 :param k8s_creds: credentials to access a given K8s cluster, i.e. a valid '.kube/config'
110 :param namespace: optional namespace to be used for helm. By default, 'kube-system' will be used
111 :param reuse_cluster_uuid: existing cluster uuid for reuse
112 :return: uuid of the K8s cluster and True if connector has installed some software in the cluster
113 (on error, an exception will be raised)
116 cluster_uuid
= reuse_cluster_uuid
118 cluster_uuid
= str(uuid4())
120 self
.log
.debug('Initializing K8S environment. namespace: {}'.format(namespace
))
122 # create config filename
123 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
124 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
125 f
= open(config_filename
, "w")
129 # check if tiller pod is up in cluster
130 command
= '{} --kubeconfig={} --namespace={} get deployments'\
131 .format(self
.kubectl_command
, config_filename
, namespace
)
132 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
134 output_table
= K8sHelmConnector
._output
_to
_table
(output
=output
)
136 # find 'tiller' pod in all pods
137 already_initialized
= False
139 for row
in output_table
:
140 if row
[0].startswith('tiller-deploy'):
141 already_initialized
= True
143 except Exception as e
:
147 n2vc_installed_sw
= False
148 if not already_initialized
:
149 self
.log
.info('Initializing helm in client and server: {}'.format(cluster_uuid
))
150 command
= '{} --kubeconfig={} --tiller-namespace={} --home={} init'\
151 .format(self
._helm
_command
, config_filename
, namespace
, helm_dir
)
152 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
153 n2vc_installed_sw
= True
155 # check client helm installation
156 check_file
= helm_dir
+ '/repository/repositories.yaml'
157 if not self
._check
_file
_exists
(filename
=check_file
, exception_if_not_exists
=False):
158 self
.log
.info('Initializing helm in client: {}'.format(cluster_uuid
))
159 command
= '{} --kubeconfig={} --tiller-namespace={} --home={} init --client-only'\
160 .format(self
._helm
_command
, config_filename
, namespace
, helm_dir
)
161 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
163 self
.log
.info('Helm client already initialized')
165 self
.log
.info('Cluster initialized {}'.format(cluster_uuid
))
167 return cluster_uuid
, n2vc_installed_sw
174 repo_type
: str = 'chart'
177 self
.log
.debug('adding {} repository {}. URL: {}'.format(repo_type
, name
, url
))
180 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
181 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
184 command
= '{} --kubeconfig={} --home={} repo update'.format(self
._helm
_command
, config_filename
, helm_dir
)
185 self
.log
.debug('updating repo: {}'.format(command
))
186 await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False)
188 # helm repo add name url
189 command
= '{} --kubeconfig={} --home={} repo add {} {}'\
190 .format(self
._helm
_command
, config_filename
, helm_dir
, name
, url
)
191 self
.log
.debug('adding repo: {}'.format(command
))
192 await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
199 Get the list of registered repositories
201 :return: list of registered repositories: [ (name, url) .... ]
204 self
.log
.debug('list repositories for cluster {}'.format(cluster_uuid
))
207 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
208 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
210 command
= '{} --kubeconfig={} --home={} repo list --output yaml'\
211 .format(self
._helm
_command
, config_filename
, helm_dir
)
213 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
214 if output
and len(output
) > 0:
215 return yaml
.load(output
, Loader
=yaml
.SafeLoader
)
219 async def repo_remove(
225 Remove a repository from OSM
227 :param cluster_uuid: the cluster
228 :param name: repo name in OSM
229 :return: True if successful
232 self
.log
.debug('list repositories for cluster {}'.format(cluster_uuid
))
235 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
236 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
238 command
= '{} --kubeconfig={} --home={} repo remove {}'\
239 .format(self
._helm
_command
, config_filename
, helm_dir
, name
)
241 await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
247 uninstall_sw
: bool = False
250 self
.log
.debug('Resetting K8s environment. cluster uuid: {}'.format(cluster_uuid
))
252 # get kube and helm directories
253 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
254 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=False)
256 # uninstall releases if needed
257 releases
= await self
.instances_list(cluster_uuid
=cluster_uuid
)
258 if len(releases
) > 0:
262 kdu_instance
= r
.get('Name')
263 chart
= r
.get('Chart')
264 self
.log
.debug('Uninstalling {} -> {}'.format(chart
, kdu_instance
))
265 await self
.uninstall(cluster_uuid
=cluster_uuid
, kdu_instance
=kdu_instance
)
266 except Exception as e
:
267 self
.log
.error('Error uninstalling release {}: {}'.format(kdu_instance
, e
))
269 msg
= 'Cluster has releases and not force. Cannot reset K8s environment. Cluster uuid: {}'\
270 .format(cluster_uuid
)
272 raise K8sException(msg
)
276 self
.log
.debug('Uninstalling tiller from cluster {}'.format(cluster_uuid
))
278 # find namespace for tiller pod
279 command
= '{} --kubeconfig={} get deployments --all-namespaces'\
280 .format(self
.kubectl_command
, config_filename
)
281 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False)
282 output_table
= K8sHelmConnector
._output
_to
_table
(output
=output
)
284 for r
in output_table
:
286 if 'tiller-deploy' in r
[1]:
289 except Exception as e
:
292 msg
= 'Tiller deployment not found in cluster {}'.format(cluster_uuid
)
295 self
.log
.debug('namespace for tiller: {}'.format(namespace
))
297 force_str
= '--force'
300 # delete tiller deployment
301 self
.log
.debug('Deleting tiller deployment for cluster {}, namespace {}'.format(cluster_uuid
, namespace
))
302 command
= '{} --namespace {} --kubeconfig={} {} delete deployment tiller-deploy'\
303 .format(self
.kubectl_command
, namespace
, config_filename
, force_str
)
304 await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False)
306 # uninstall tiller from cluster
307 self
.log
.debug('Uninstalling tiller from cluster {}'.format(cluster_uuid
))
308 command
= '{} --kubeconfig={} --home={} reset'\
309 .format(self
._helm
_command
, config_filename
, helm_dir
)
310 self
.log
.debug('resetting: {}'.format(command
))
311 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
313 self
.log
.debug('namespace not found')
315 # delete cluster directory
316 dir = self
.fs
.path
+ '/' + cluster_uuid
317 self
.log
.debug('Removing directory {}'.format(dir))
318 shutil
.rmtree(dir, ignore_errors
=True)
327 timeout
: float = 300,
329 db_dict
: dict = None,
333 self
.log
.debug('installing {} in cluster {}'.format(kdu_model
, cluster_uuid
))
336 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
337 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
340 # params_str = K8sHelmConnector._params_to_set_option(params)
341 params_str
, file_to_delete
= self
._params
_to
_file
_option
(cluster_uuid
=cluster_uuid
, params
=params
)
345 timeout_str
= '--timeout {}'.format(timeout
)
350 atomic_str
= '--atomic'
355 parts
= kdu_model
.split(sep
=':')
357 version_str
= '--version {}'.format(parts
[1])
360 # generate a name for the release. Then, check if already exists
362 while kdu_instance
is None:
363 kdu_instance
= K8sHelmConnector
._generate
_release
_name
(kdu_model
)
365 result
= await self
._status
_kdu
(
366 cluster_uuid
=cluster_uuid
,
367 kdu_instance
=kdu_instance
,
370 if result
is not None:
371 # instance already exists: generate a new one
377 command
= '{} install {} --output yaml --kubeconfig={} --home={} {} {} --name={} {} {}'\
378 .format(self
._helm
_command
, atomic_str
, config_filename
, helm_dir
,
379 params_str
, timeout_str
, kdu_instance
, kdu_model
, version_str
)
380 self
.log
.debug('installing: {}'.format(command
))
383 # exec helm in a task
384 exec_task
= asyncio
.ensure_future(
385 coro_or_future
=self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False)
388 # write status in another task
389 status_task
= asyncio
.ensure_future(
390 coro_or_future
=self
._store
_status
(
391 cluster_uuid
=cluster_uuid
,
392 kdu_instance
=kdu_instance
,
399 # wait for execution task
400 await asyncio
.wait([exec_task
])
405 output
, rc
= exec_task
.result()
409 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False)
411 # remove temporal values yaml file
413 os
.remove(file_to_delete
)
416 await self
._store
_status
(
417 cluster_uuid
=cluster_uuid
,
418 kdu_instance
=kdu_instance
,
426 msg
= 'Error executing command: {}\nOutput: {}'.format(command
, output
)
428 raise K8sException(msg
)
430 self
.log
.debug('Returning kdu_instance {}'.format(kdu_instance
))
433 async def instances_list(
438 returns a list of deployed releases in a cluster
440 :param cluster_uuid: the cluster
444 self
.log
.debug('list releases for cluster {}'.format(cluster_uuid
))
447 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
448 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
450 command
= '{} --kubeconfig={} --home={} list --output yaml'\
451 .format(self
._helm
_command
, config_filename
, helm_dir
)
453 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
455 if output
and len(output
) > 0:
456 return yaml
.load(output
, Loader
=yaml
.SafeLoader
).get('Releases')
464 kdu_model
: str = None,
466 timeout
: float = 300,
471 self
.log
.debug('upgrading {} in cluster {}'.format(kdu_model
, cluster_uuid
))
474 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
475 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
478 # params_str = K8sHelmConnector._params_to_set_option(params)
479 params_str
, file_to_delete
= self
._params
_to
_file
_option
(cluster_uuid
=cluster_uuid
, params
=params
)
483 timeout_str
= '--timeout {}'.format(timeout
)
488 atomic_str
= '--atomic'
492 if kdu_model
and ':' in kdu_model
:
493 parts
= kdu_model
.split(sep
=':')
495 version_str
= '--version {}'.format(parts
[1])
499 command
= '{} upgrade {} --output yaml --kubeconfig={} --home={} {} {} {} {} {}'\
500 .format(self
._helm
_command
, atomic_str
, config_filename
, helm_dir
,
501 params_str
, timeout_str
, kdu_instance
, kdu_model
, version_str
)
502 self
.log
.debug('upgrading: {}'.format(command
))
506 # exec helm in a task
507 exec_task
= asyncio
.ensure_future(
508 coro_or_future
=self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False)
510 # write status in another task
511 status_task
= asyncio
.ensure_future(
512 coro_or_future
=self
._store
_status
(
513 cluster_uuid
=cluster_uuid
,
514 kdu_instance
=kdu_instance
,
521 # wait for execution task
522 await asyncio
.wait([exec_task
])
526 output
, rc
= exec_task
.result()
530 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False)
532 # remove temporal values yaml file
534 os
.remove(file_to_delete
)
537 await self
._store
_status
(
538 cluster_uuid
=cluster_uuid
,
539 kdu_instance
=kdu_instance
,
547 msg
= 'Error executing command: {}\nOutput: {}'.format(command
, output
)
549 raise K8sException(msg
)
551 # return new revision number
552 instance
= await self
.get_instance_info(cluster_uuid
=cluster_uuid
, kdu_instance
=kdu_instance
)
554 revision
= int(instance
.get('Revision'))
555 self
.log
.debug('New revision: {}'.format(revision
))
568 self
.log
.debug('rollback kdu_instance {} to revision {} from cluster {}'
569 .format(kdu_instance
, revision
, cluster_uuid
))
572 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
573 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
575 command
= '{} rollback --kubeconfig={} --home={} {} {} --wait'\
576 .format(self
._helm
_command
, config_filename
, helm_dir
, kdu_instance
, revision
)
578 # exec helm in a task
579 exec_task
= asyncio
.ensure_future(
580 coro_or_future
=self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=False)
582 # write status in another task
583 status_task
= asyncio
.ensure_future(
584 coro_or_future
=self
._store
_status
(
585 cluster_uuid
=cluster_uuid
,
586 kdu_instance
=kdu_instance
,
588 operation
='rollback',
593 # wait for execution task
594 await asyncio
.wait([exec_task
])
599 output
, rc
= exec_task
.result()
602 await self
._store
_status
(
603 cluster_uuid
=cluster_uuid
,
604 kdu_instance
=kdu_instance
,
606 operation
='rollback',
612 msg
= 'Error executing command: {}\nOutput: {}'.format(command
, output
)
614 raise K8sException(msg
)
616 # return new revision number
617 instance
= await self
.get_instance_info(cluster_uuid
=cluster_uuid
, kdu_instance
=kdu_instance
)
619 revision
= int(instance
.get('Revision'))
620 self
.log
.debug('New revision: {}'.format(revision
))
631 Removes an existing KDU instance. It would implicitly use the `delete` call (this call would happen
632 after all _terminate-config-primitive_ of the VNF are invoked).
634 :param cluster_uuid: UUID of a K8s cluster known by OSM
635 :param kdu_instance: unique name for the KDU instance to be deleted
636 :return: True if successful
639 self
.log
.debug('uninstall kdu_instance {} from cluster {}'.format(kdu_instance
, cluster_uuid
))
642 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
643 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
645 command
= '{} --kubeconfig={} --home={} delete --purge {}'\
646 .format(self
._helm
_command
, config_filename
, helm_dir
, kdu_instance
)
648 output
, rc
= await self
._local
_async
_exec
(command
=command
, raise_exception_on_error
=True)
650 return self
._output
_to
_table
(output
)
652 async def inspect_kdu(
658 self
.log
.debug('inspect kdu_model {} from (optional) repo: {}'.format(kdu_model
, repo_url
))
660 return await self
._exec
_inspect
_comand
(inspect_command
='', kdu_model
=kdu_model
, repo_url
=repo_url
)
662 async def values_kdu(
668 self
.log
.debug('inspect kdu_model values {} from (optional) repo: {}'.format(kdu_model
, repo_url
))
670 return await self
._exec
_inspect
_comand
(inspect_command
='values', kdu_model
=kdu_model
, repo_url
=repo_url
)
678 self
.log
.debug('inspect kdu_model {} readme.md from repo: {}'.format(kdu_model
, repo_url
))
680 return await self
._exec
_inspect
_comand
(inspect_command
='readme', kdu_model
=kdu_model
, repo_url
=repo_url
)
682 async def status_kdu(
688 # call internal function
689 return await self
._status
_kdu
(
690 cluster_uuid
=cluster_uuid
,
691 kdu_instance
=kdu_instance
,
696 async def synchronize_repos(self
, cluster_uuid
: str):
698 self
.log
.debug("syncronize repos for cluster helm-id: {}",)
700 update_repos_timeout
= 300 # max timeout to sync a single repos, more than this is too much
701 db_k8scluster
= self
.db
.get_one("k8sclusters", {"_admin.helm-chart.id": cluster_uuid
})
703 nbi_repo_list
= db_k8scluster
.get("_admin").get("helm_chart_repos") or []
704 cluster_repo_dict
= db_k8scluster
.get("_admin").get("helm_charts_added") or {}
705 # elements that must be deleted
706 deleted_repo_list
= []
708 self
.log
.debug("helm_chart_repos: {}".format(nbi_repo_list
))
709 self
.log
.debug("helm_charts_added: {}".format(cluster_repo_dict
))
711 # obtain repos to add: registered by nbi but not added
712 repos_to_add
= [repo
for repo
in nbi_repo_list
if not cluster_repo_dict
.get(repo
)]
714 # obtain repos to delete: added by cluster but not in nbi list
715 repos_to_delete
= [repo
for repo
in cluster_repo_dict
.keys() if repo
not in nbi_repo_list
]
717 # delete repos: must delete first then add because there may be different repos with same name but
718 # different id and url
719 self
.log
.debug("repos to delete: {}".format(repos_to_delete
))
720 for repo_id
in repos_to_delete
:
721 # try to delete repos
723 repo_delete_task
= asyncio
.ensure_future(self
.repo_remove(cluster_uuid
=cluster_uuid
,
724 name
=cluster_repo_dict
[repo_id
]))
725 await asyncio
.wait_for(repo_delete_task
, update_repos_timeout
)
726 except Exception as e
:
727 self
.warning("Error deleting repo, id: {}, name: {}, err_msg: {}".format(repo_id
,
728 cluster_repo_dict
[repo_id
], str(e
)))
729 # always add to the list of to_delete if there is an error because if is not there deleting raises error
730 deleted_repo_list
.append(repo_id
)
733 self
.log
.debug("repos to add: {}".format(repos_to_add
))
735 for repo_id
in repos_to_add
:
736 # obtain the repo data from the db
737 # if there is an error getting the repo in the database we will ignore this repo and continue
738 # because there is a possible race condition where the repo has been deleted while processing
739 db_repo
= self
.db
.get_one("k8srepos", {"_id": repo_id
})
740 self
.log
.debug("obtained repo: id, {}, name: {}, url: {}".format(repo_id
, db_repo
["name"], db_repo
["url"]))
742 repo_add_task
= asyncio
.ensure_future(self
.repo_add(cluster_uuid
=cluster_uuid
,
743 name
=db_repo
["name"], url
=db_repo
["url"],
745 await asyncio
.wait_for(repo_add_task
, update_repos_timeout
)
746 added_repo_dict
[repo_id
] = db_repo
["name"]
747 self
.log
.debug("added repo: id, {}, name: {}".format(repo_id
, db_repo
["name"]))
748 except Exception as e
:
749 # deal with error adding repo, adding a repo that already exists does not raise any error
750 # will not raise error because a wrong repos added by anyone could prevent instantiating any ns
751 self
.log
.error("Error adding repo id: {}, err_msg: {} ".format(repo_id
, repr(e
)))
753 return deleted_repo_list
, added_repo_dict
755 else: # else db_k8scluster does not exist
756 raise K8sException("k8cluster with helm-id : {} not found".format(cluster_uuid
))
758 except Exception as e
:
759 self
.log
.error("Error synchronizing repos: {}".format(str(e
)))
760 raise K8sException("Error synchronizing repos")
763 ##################################################################################################
764 ########################################## P R I V A T E #########################################
765 ##################################################################################################
768 async def _exec_inspect_comand(
770 inspect_command
: str,
777 repo_str
= ' --repo {}'.format(repo_url
)
778 idx
= kdu_model
.find('/')
781 kdu_model
= kdu_model
[idx
:]
783 inspect_command
= '{} inspect {} {}{}'.format(self
._helm
_command
, inspect_command
, kdu_model
, repo_str
)
784 output
, rc
= await self
._local
_async
_exec
(command
=inspect_command
, encode_utf8
=True)
788 async def _status_kdu(
792 show_error_log
: bool = False,
793 return_text
: bool = False
796 self
.log
.debug('status of kdu_instance {}'.format(kdu_instance
))
799 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
800 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
802 command
= '{} --kubeconfig={} --home={} status {} --output yaml'\
803 .format(self
._helm
_command
, config_filename
, helm_dir
, kdu_instance
)
805 output
, rc
= await self
._local
_async
_exec
(
807 raise_exception_on_error
=True,
808 show_error_log
=show_error_log
817 data
= yaml
.load(output
, Loader
=yaml
.SafeLoader
)
819 # remove field 'notes'
821 del data
.get('info').get('status')['notes']
825 # parse field 'resources'
827 resources
= str(data
.get('info').get('status').get('resources'))
828 resource_table
= self
._output
_to
_table
(resources
)
829 data
.get('info').get('status')['resources'] = resource_table
830 except Exception as e
:
835 async def get_instance_info(
840 instances
= await self
.instances_list(cluster_uuid
=cluster_uuid
)
841 for instance
in instances
:
842 if instance
.get('Name') == kdu_instance
:
844 self
.log
.debug('Instance {} not found'.format(kdu_instance
))
848 def _generate_release_name(
851 # check embeded chart (file or dir)
852 if chart_name
.startswith('/'):
853 # extract file or directory name
854 chart_name
= chart_name
[chart_name
.rfind('/')+1:]
856 elif '://' in chart_name
:
857 # extract last portion of URL
858 chart_name
= chart_name
[chart_name
.rfind('/')+1:]
862 if c
.isalpha() or c
.isnumeric():
869 # if does not start with alpha character, prefix 'a'
870 if not name
[0].isalpha():
875 def get_random_number():
876 r
= random
.randrange(start
=1, stop
=99999999)
881 name
= name
+ get_random_number()
884 async def _store_status(
889 check_every
: float = 10,
890 db_dict
: dict = None,
891 run_once
: bool = False
895 await asyncio
.sleep(check_every
)
896 detailed_status
= await self
.status_kdu(cluster_uuid
=cluster_uuid
, kdu_instance
=kdu_instance
)
897 status
= detailed_status
.get('info').get('Description')
898 self
.log
.debug('STATUS:\n{}'.format(status
))
899 self
.log
.debug('DETAILED STATUS:\n{}'.format(detailed_status
))
901 result
= await self
.write_app_status_to_db(
904 detailed_status
=str(detailed_status
),
907 self
.log
.info('Error writing in database. Task exiting...')
909 except asyncio
.CancelledError
:
910 self
.log
.debug('Task cancelled')
912 except Exception as e
:
913 self
.log
.debug('_store_status exception: {}'.format(str(e
)))
919 async def _is_install_completed(
925 status
= await self
._status
_kdu
(cluster_uuid
=cluster_uuid
, kdu_instance
=kdu_instance
, return_text
=False)
927 # extract info.status.resources-> str
930 # NAME READY UP-TO-DATE AVAILABLE AGE
931 # halting-horse-mongodb 0/1 1 0 0s
932 # halting-petit-mongodb 1/1 1 0 0s
934 resources
= K8sHelmConnector
._get
_deep
(status
, ('info', 'status', 'resources'))
937 resources
= K8sHelmConnector
._output
_to
_table
(resources
)
939 num_lines
= len(resources
)
941 while index
< num_lines
:
943 line1
= resources
[index
]
945 # find '==>' in column 0
946 if line1
[0] == '==>':
947 line2
= resources
[index
]
949 # find READY in column 1
950 if line2
[1] == 'READY':
952 line3
= resources
[index
]
954 while len(line3
) > 1 and index
< num_lines
:
955 ready_value
= line3
[1]
956 parts
= ready_value
.split(sep
='/')
957 current
= int(parts
[0])
958 total
= int(parts
[1])
960 self
.log
.debug('NOT READY:\n {}'.format(line3
))
962 line3
= resources
[index
]
965 except Exception as e
:
971 def _get_deep(dictionary
: dict, members
: tuple):
976 value
= target
.get(m
)
981 except Exception as e
:
985 # find key:value in several lines
987 def _find_in_lines(p_lines
: list, p_key
: str) -> str:
990 if line
.startswith(p_key
+ ':'):
991 parts
= line
.split(':')
992 the_value
= parts
[1].strip()
994 except Exception as e
:
999 # params for use in -f file
1000 # returns values file option and filename (in order to delete it at the end)
1001 def _params_to_file_option(self
, cluster_uuid
: str, params
: dict) -> (str, str):
1003 if params
and len(params
) > 0:
1004 kube_dir
, helm_dir
, config_filename
, cluster_dir
= \
1005 self
._get
_paths
(cluster_name
=cluster_uuid
, create_if_not_exist
=True)
1007 def get_random_number():
1008 r
= random
.randrange(start
=1, stop
=99999999)
1016 value
= params
.get(key
)
1017 if '!!yaml' in str(value
):
1018 value
= yaml
.load(value
[7:])
1019 params2
[key
] = value
1021 values_file
= get_random_number() + '.yaml'
1022 with
open(values_file
, 'w') as stream
:
1023 yaml
.dump(params2
, stream
, indent
=4, default_flow_style
=False)
1025 return '-f {}'.format(values_file
), values_file
1029 # params for use in --set option
1031 def _params_to_set_option(params
: dict) -> str:
1033 if params
and len(params
) > 0:
1036 value
= params
.get(key
, None)
1037 if value
is not None:
1039 params_str
+= '--set '
1043 params_str
+= '{}={}'.format(key
, value
)
1047 def _output_to_lines(output
: str) -> list:
1048 output_lines
= list()
1049 lines
= output
.splitlines(keepends
=False)
1053 output_lines
.append(line
)
1057 def _output_to_table(output
: str) -> list:
1058 output_table
= list()
1059 lines
= output
.splitlines(keepends
=False)
1061 line
= line
.replace('\t', ' ')
1063 output_table
.append(line_list
)
1064 cells
= line
.split(sep
=' ')
1068 line_list
.append(cell
)
1071 def _get_paths(self
, cluster_name
: str, create_if_not_exist
: bool = False) -> (str, str, str, str):
1073 Returns kube and helm directories
1075 :param cluster_name:
1076 :param create_if_not_exist:
1077 :return: kube, helm directories, config filename and cluster dir.
1078 Raises exception if not exist and cannot create
1082 if base
.endswith("/") or base
.endswith("\\"):
1085 # base dir for cluster
1086 cluster_dir
= base
+ '/' + cluster_name
1087 if create_if_not_exist
and not os
.path
.exists(cluster_dir
):
1088 self
.log
.debug('Creating dir {}'.format(cluster_dir
))
1089 os
.makedirs(cluster_dir
)
1090 if not os
.path
.exists(cluster_dir
):
1091 msg
= 'Base cluster dir {} does not exist'.format(cluster_dir
)
1093 raise K8sException(msg
)
1096 kube_dir
= cluster_dir
+ '/' + '.kube'
1097 if create_if_not_exist
and not os
.path
.exists(kube_dir
):
1098 self
.log
.debug('Creating dir {}'.format(kube_dir
))
1099 os
.makedirs(kube_dir
)
1100 if not os
.path
.exists(kube_dir
):
1101 msg
= 'Kube config dir {} does not exist'.format(kube_dir
)
1103 raise K8sException(msg
)
1106 helm_dir
= cluster_dir
+ '/' + '.helm'
1107 if create_if_not_exist
and not os
.path
.exists(helm_dir
):
1108 self
.log
.debug('Creating dir {}'.format(helm_dir
))
1109 os
.makedirs(helm_dir
)
1110 if not os
.path
.exists(helm_dir
):
1111 msg
= 'Helm config dir {} does not exist'.format(helm_dir
)
1113 raise K8sException(msg
)
1115 config_filename
= kube_dir
+ '/config'
1116 return kube_dir
, helm_dir
, config_filename
, cluster_dir
1119 def _remove_multiple_spaces(str):
1122 str = str.replace(' ', ' ')
1129 command
= K8sHelmConnector
._remove
_multiple
_spaces
(command
)
1130 self
.log
.debug('Executing sync local command: {}'.format(command
))
1131 # raise exception if fails
1134 output
= subprocess
.check_output(command
, shell
=True, universal_newlines
=True)
1136 self
.log
.debug(output
)
1137 except Exception as e
:
1140 return output
, return_code
1142 async def _local_async_exec(
1145 raise_exception_on_error
: bool = False,
1146 show_error_log
: bool = True,
1147 encode_utf8
: bool = False
1150 command
= K8sHelmConnector
._remove
_multiple
_spaces
(command
)
1151 self
.log
.debug('Executing async local command: {}'.format(command
))
1154 command
= command
.split(sep
=' ')
1157 process
= await asyncio
.create_subprocess_exec(
1159 stdout
=asyncio
.subprocess
.PIPE
,
1160 stderr
=asyncio
.subprocess
.PIPE
1163 # wait for command terminate
1164 stdout
, stderr
= await process
.communicate()
1166 return_code
= process
.returncode
1170 output
= stdout
.decode('utf-8').strip()
1171 # output = stdout.decode()
1173 output
= stderr
.decode('utf-8').strip()
1174 # output = stderr.decode()
1176 if return_code
!= 0 and show_error_log
:
1177 self
.log
.debug('Return code (FAIL): {}\nOutput:\n{}'.format(return_code
, output
))
1179 self
.log
.debug('Return code: {}'.format(return_code
))
1181 if raise_exception_on_error
and return_code
!= 0:
1182 raise K8sException(output
)
1185 output
= output
.encode('utf-8').strip()
1186 output
= str(output
).replace('\\n', '\n')
1188 return output
, return_code
1190 except asyncio
.CancelledError
:
1192 except K8sException
:
1194 except Exception as e
:
1195 msg
= 'Exception executing command: {} -> {}'.format(command
, e
)
1197 if raise_exception_on_error
:
1198 raise K8sException(e
) from e
1202 def _check_file_exists(self
, filename
: str, exception_if_not_exists
: bool = False):
1203 self
.log
.debug('Checking if file {} exists...'.format(filename
))
1204 if os
.path
.exists(filename
):
1207 msg
= 'File {} does not exist'.format(filename
)
1208 if exception_if_not_exists
:
1210 raise K8sException(msg
)