1 # Copyright 2019 Canonical Ltd.
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
7 # http://www.apache.org/licenses/LICENSE-2.0
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
17 from .exceptions
import NotImplemented
21 # from juju.bundle import BundleHandler
22 from juju
.controller
import Controller
23 from juju
.model
import Model
24 from juju
.errors
import JujuAPIError
, JujuError
28 from n2vc
.k8s_conn
import K8sConnector
33 # from .vnf import N2VC
39 class K8sJujuConnector(K8sConnector
):
45 kubectl_command
: str = '/usr/bin/kubectl',
46 juju_command
: str = '/usr/bin/juju',
52 :param kubectl_command: path to kubectl executable
53 :param helm_command: path to helm executable
54 :param fs: file system for kubernetes and helm configuration
59 K8sConnector
.__init
__(
63 on_update_db
=on_update_db
,
67 self
.info('Initializing K8S Juju connector')
69 self
.authenticated
= False
71 self
.log
= logging
.getLogger(__name__
)
73 self
.juju_command
= juju_command
76 self
.info('K8S Juju connector initialized')
82 namespace
: str = 'kube-system',
83 reuse_cluster_uuid
: str = None,
86 It prepares a given K8s cluster environment to run Juju bundles.
88 :param k8s_creds: credentials to access a given K8s cluster, i.e. a valid '.kube/config'
89 :param namespace: optional namespace to be used for juju. By default, 'kube-system' will be used
90 :param reuse_cluster_uuid: existing cluster uuid for reuse
91 :return: uuid of the K8s cluster and True if connector has installed some software in the cluster
92 (on error, an exception will be raised)
97 Bootstrapping cannot be done, by design, through the API. We need to
104 1. Has the environment already been bootstrapped?
105 - Check the database to see if we have a record for this env
107 2. If this is a new env, create it
108 - Add the k8s cloud to Juju
110 - Record it in the database
112 3. Connect to the Juju controller for this cloud
115 # cluster_uuid = reuse_cluster_uuid
116 # if not cluster_uuid:
117 # cluster_uuid = str(uuid4())
119 ##################################################
120 # TODO: Pull info from db based on the namespace #
121 ##################################################
123 ###################################################
124 # TODO: Make it idempotent, calling add-k8s and #
125 # bootstrap whenever reuse_cluster_uuid is passed #
127 # `init_env` is called to initialize the K8s #
128 # cluster for juju. If this initialization fails, #
129 # it can be called again by LCM with the param #
130 # reuse_cluster_uuid, e.g. to try to fix it. #
131 ###################################################
133 if not reuse_cluster_uuid
:
134 # This is a new cluster, so bootstrap it
136 cluster_uuid
= str(uuid
.uuid4())
138 # Is a local k8s cluster?
139 localk8s
= self
.is_local_k8s(k8s_creds
)
141 # If the k8s is external, the juju controller needs a loadbalancer
142 loadbalancer
= False if localk8s
else True
144 # Name the new k8s cloud
145 k8s_cloud
= "k8s-{}".format(cluster_uuid
)
147 print("Adding k8s cloud {}".format(k8s_cloud
))
148 await self
.add_k8s(k8s_cloud
, k8s_creds
)
150 # Bootstrap Juju controller
151 print("Bootstrapping...")
152 await self
.bootstrap(k8s_cloud
, cluster_uuid
, loadbalancer
)
153 print("Bootstrap done.")
155 # Get the controller information
157 # Parse ~/.local/share/juju/controllers.yaml
158 # controllers.testing.api-endpoints|ca-cert|uuid
159 print("Getting controller endpoints")
160 with
open(os
.path
.expanduser(
161 "~/.local/share/juju/controllers.yaml"
163 controllers
= yaml
.load(f
, Loader
=yaml
.Loader
)
164 controller
= controllers
['controllers'][cluster_uuid
]
165 endpoints
= controller
['api-endpoints']
166 self
.juju_endpoint
= endpoints
[0]
167 self
.juju_ca_cert
= controller
['ca-cert']
169 # Parse ~/.local/share/juju/accounts
170 # controllers.testing.user|password
171 print("Getting accounts")
172 with
open(os
.path
.expanduser(
173 "~/.local/share/juju/accounts.yaml"
175 controllers
= yaml
.load(f
, Loader
=yaml
.Loader
)
176 controller
= controllers
['controllers'][cluster_uuid
]
178 self
.juju_user
= controller
['user']
179 self
.juju_secret
= controller
['password']
181 print("user: {}".format(self
.juju_user
))
182 print("secret: {}".format(self
.juju_secret
))
183 print("endpoint: {}".format(self
.juju_endpoint
))
184 print("ca-cert: {}".format(self
.juju_ca_cert
))
186 # raise Exception("EOL")
188 self
.juju_public_key
= None
191 'endpoint': self
.juju_endpoint
,
192 'username': self
.juju_user
,
193 'secret': self
.juju_secret
,
194 'cacert': self
.juju_ca_cert
,
195 'namespace': namespace
,
196 'loadbalancer': loadbalancer
,
199 # Store the cluster configuration so it
200 # can be used for subsequent calls
201 print("Setting config")
202 await self
.set_config(cluster_uuid
, config
)
205 # This is an existing cluster, so get its config
206 cluster_uuid
= reuse_cluster_uuid
208 config
= self
.get_config(cluster_uuid
)
210 self
.juju_endpoint
= config
['endpoint']
211 self
.juju_user
= config
['username']
212 self
.juju_secret
= config
['secret']
213 self
.juju_ca_cert
= config
['cacert']
214 self
.juju_public_key
= None
216 # Login to the k8s cluster
217 if not self
.authenticated
:
218 await self
.login(cluster_uuid
)
220 # We're creating a new cluster
221 print("Getting model {}".format(self
.get_namespace(cluster_uuid
), cluster_uuid
=cluster_uuid
))
222 model
= await self
.get_model(
223 self
.get_namespace(cluster_uuid
),
224 cluster_uuid
=cluster_uuid
227 # Disconnect from the model
228 if model
and model
.is_connected():
229 await model
.disconnect()
231 return cluster_uuid
, True
233 """Repo Management"""
240 raise NotImplemented()
242 async def repo_list(self
):
243 raise NotImplemented()
245 async def repo_remove(
249 raise NotImplemented()
256 uninstall_sw
: bool = False
260 Resets the Kubernetes cluster by removing the model that represents it.
262 :param cluster_uuid str: The UUID of the cluster to reset
263 :return: Returns True if successful or raises an exception.
267 if not self
.authenticated
:
268 await self
.login(cluster_uuid
)
270 if self
.controller
.is_connected():
272 namespace
= self
.get_namespace(cluster_uuid
)
273 if await self
.has_model(namespace
):
274 print("[reset] Destroying model")
275 await self
.controller
.destroy_model(
280 # Disconnect from the controller
281 print("[reset] Disconnecting controller")
282 await self
.controller
.disconnect()
284 # Destroy the controller (via CLI)
285 print("[reset] Destroying controller")
286 await self
.destroy_controller(cluster_uuid
)
288 print("[reset] Removing k8s cloud")
289 namespace
= self
.get_namespace(cluster_uuid
)
290 k8s_cloud
= "{}-k8s".format(namespace
)
291 await self
.remove_cloud(k8s_cloud
)
293 except Exception as ex
:
294 print("Caught exception during reset: {}".format(ex
))
303 timeout
: float = 300,
309 :param cluster_uuid str: The UUID of the cluster to install to
310 :param kdu_model str: The name or path of a bundle to install
311 :param atomic bool: If set, waits until the model is active and resets
312 the cluster on failure.
313 :param timeout int: The time, in seconds, to wait for the install
315 :param params dict: Key-value pairs of instantiation parameters
317 :return: If successful, returns ?
320 if not self
.authenticated
:
321 print("[install] Logging in to the controller")
322 await self
.login(cluster_uuid
)
325 # Get or create the model, based on the NS
327 model_name
= db_dict
["filter"]["_id"]
329 self
.log
.debug("Checking for model named {}".format(model_name
))
330 model
= await self
.get_model(model_name
, cluster_uuid
=cluster_uuid
)
332 # Create the new model
333 self
.log
.debug("Adding model: {}".format(model_name
))
334 model
= await self
.add_model(model_name
, cluster_uuid
=cluster_uuid
)
337 # TODO: Instantiation parameters
340 "Juju bundle that models the KDU, in any of the following ways:
341 - <juju-repo>/<juju-bundle>
342 - <juju-bundle folder under k8s_models folder in the package>
343 - <juju-bundle tgz file (w/ or w/o extension) under k8s_models folder in the package>
344 - <URL_where_to_fetch_juju_bundle>
348 if kdu_model
.startswith("cs:"):
350 elif kdu_model
.startswith("http"):
356 # if kdu_model.endswith(".tar.gz") or kdu_model.endswith(".tgz")
357 # Uncompress temporarily
358 # bundle = <uncompressed file>
362 # Raise named exception that the bundle could not be found
365 print("[install] deploying {}".format(bundle
))
366 await model
.deploy(bundle
)
368 # Get the application
370 # applications = model.applications
371 print("[install] Applications: {}".format(model
.applications
))
372 for name
in model
.applications
:
373 print("[install] Waiting for {} to settle".format(name
))
374 application
= model
.applications
[name
]
376 # It's not enough to wait for all units to be active;
377 # the application status needs to be active as well.
378 print("Waiting for all units to be active...")
379 await model
.block_until(
381 unit
.agent_status
== 'idle'
382 and application
.status
in ['active', 'unknown']
383 and unit
.workload_status
in [
385 ] for unit
in application
.units
389 print("All units active.")
391 except concurrent
.futures
._base
.TimeoutError
:
392 print("[install] Timeout exceeded; resetting cluster")
393 await self
.reset(cluster_uuid
)
396 # Wait for the application to be active
397 if model
.is_connected():
398 print("[install] Disconnecting model")
399 await model
.disconnect()
402 raise Exception("Unable to install")
404 async def instances_list(
409 returns a list of deployed releases in a cluster
411 :param cluster_uuid: the cluster
420 kdu_model
: str = None,
425 :param cluster_uuid str: The UUID of the cluster to upgrade
426 :param kdu_instance str: The unique name of the KDU instance
427 :param kdu_model str: The name or path of the bundle to upgrade to
428 :param params dict: Key-value pairs of instantiation parameters
430 :return: If successful, reference to the new revision number of the
434 # TODO: Loop through the bundle and upgrade each charm individually
437 The API doesn't have a concept of bundle upgrades, because there are
438 many possible changes: charm revision, disk, number of units, etc.
440 As such, we are only supporting a limited subset of upgrades. We'll
441 upgrade the charm revision but leave storage and scale untouched.
443 Scale changes should happen through OSM constructs, and changes to
444 storage would require a redeployment of the service, at least in this
447 namespace
= self
.get_namespace(cluster_uuid
)
448 model
= await self
.get_model(namespace
, cluster_uuid
=cluster_uuid
)
450 with
open(kdu_model
, 'r') as f
:
451 bundle
= yaml
.safe_load(f
)
455 'description': 'Test bundle',
456 'bundle': 'kubernetes',
459 'charm': 'cs:~charmed-osm/mariadb-k8s-20',
462 'password': 'manopw',
463 'root_password': 'osm4u',
466 'series': 'kubernetes'
471 # TODO: This should be returned in an agreed-upon format
472 for name
in bundle
['applications']:
473 print(model
.applications
)
474 application
= model
.applications
[name
]
477 path
= bundle
['applications'][name
]['charm']
480 await application
.upgrade_charm(switch
=path
)
481 except juju
.errors
.JujuError
as ex
:
482 if 'already running charm' in str(ex
):
483 # We're already running this version
486 await model
.disconnect()
489 raise NotImplemented()
500 :param cluster_uuid str: The UUID of the cluster to rollback
501 :param kdu_instance str: The unique name of the KDU instance
502 :param revision int: The revision to revert to. If omitted, rolls back
503 the previous upgrade.
505 :return: If successful, returns the revision of active KDU instance,
506 or raises an exception
508 raise NotImplemented()
516 """Uninstall a KDU instance
518 :param cluster_uuid str: The UUID of the cluster to uninstall
519 :param kdu_instance str: The unique name of the KDU instance
521 :return: Returns True if successful, or raises an exception
525 # Remove an application from the model
526 model
= await self
.get_model(self
.get_namespace(cluster_uuid
), cluster_uuid
=cluster_uuid
)
529 # Get the application
530 if kdu_instance
not in model
.applications
:
531 # TODO: Raise a named exception
532 raise Exception("Application not found.")
534 application
= model
.applications
[kdu_instance
]
536 # Destroy the application
537 await application
.destroy()
539 # TODO: Verify removal
545 async def inspect_kdu(
551 Inspects a bundle and returns a dictionary of config parameters and
552 their default values.
554 :param kdu_model str: The name or path of the bundle to inspect.
556 :return: If successful, returns a dictionary of available parameters
557 and their default values.
561 with
open(kdu_model
, 'r') as f
:
562 bundle
= yaml
.safe_load(f
)
566 'description': 'Test bundle',
567 'bundle': 'kubernetes',
570 'charm': 'cs:~charmed-osm/mariadb-k8s-20',
573 'password': 'manopw',
574 'root_password': 'osm4u',
577 'series': 'kubernetes'
582 # TODO: This should be returned in an agreed-upon format
583 kdu
= bundle
['applications']
593 If available, returns the README of the bundle.
595 :param kdu_model str: The name or path of a bundle
597 :return: If found, returns the contents of the README.
601 files
= ['README', 'README.txt', 'README.md']
602 path
= os
.path
.dirname(kdu_model
)
603 for file in os
.listdir(path
):
605 with
open(file, 'r') as f
:
611 async def status_kdu(
616 """Get the status of the KDU
618 Get the current status of the KDU instance.
620 :param cluster_uuid str: The UUID of the cluster
621 :param kdu_instance str: The unique id of the KDU instance
623 :return: Returns a dictionary containing namespace, state, resources,
628 model
= await self
.get_model(self
.get_namespace(cluster_uuid
), cluster_uuid
=cluster_uuid
)
630 # model = await self.get_model_by_uuid(cluster_uuid)
632 model_status
= await model
.get_status()
633 status
= model_status
.applications
635 for name
in model_status
.applications
:
636 application
= model_status
.applications
[name
]
638 'status': application
['status']['status']
641 if model
.is_connected():
642 await model
.disconnect()
652 """Add a k8s cloud to Juju
654 Adds a Kubernetes cloud to Juju, so it can be bootstrapped with a
657 :param cloud_name str: The name of the cloud to add.
658 :param credentials dict: A dictionary representing the output of
659 `kubectl config view --raw`.
661 :returns: True if successful, otherwise raises an exception.
664 cmd
= [self
.juju_command
, "add-k8s", "--local", cloud_name
]
667 process
= await asyncio
.create_subprocess_exec(
669 stdout
=asyncio
.subprocess
.PIPE
,
670 stderr
=asyncio
.subprocess
.PIPE
,
671 stdin
=asyncio
.subprocess
.PIPE
,
674 # Feed the process the credentials
675 process
.stdin
.write(credentials
.encode("utf-8"))
676 await process
.stdin
.drain()
677 process
.stdin
.close()
679 stdout
, stderr
= await process
.communicate()
681 return_code
= process
.returncode
683 print("add-k8s return code: {}".format(return_code
))
686 raise Exception(stderr
)
694 ) -> juju
.model
.Model
:
695 """Adds a model to the controller
697 Adds a new model to the Juju controller
699 :param model_name str: The name of the model to add.
700 :returns: The juju.model.Model object of the new model upon success or
703 if not self
.authenticated
:
704 await self
.login(cluster_uuid
)
706 self
.log
.debug("Adding model '{}' to cluster_uuid '{}'".format(model_name
, cluster_uuid
))
707 model
= await self
.controller
.add_model(
709 config
={'authorized-keys': self
.juju_public_key
}
719 """Bootstrap a Kubernetes controller
721 Bootstrap a Juju controller inside the Kubernetes cluster
723 :param cloud_name str: The name of the cloud.
724 :param cluster_uuid str: The UUID of the cluster to bootstrap.
725 :param loadbalancer bool: If the controller should use loadbalancer or not.
726 :returns: True upon success or raises an exception.
730 cmd
= [self
.juju_command
, "bootstrap", cloud_name
, cluster_uuid
]
733 For public clusters, specify that the controller service is using a LoadBalancer.
735 cmd
= [self
.juju_command
, "bootstrap", cloud_name
, cluster_uuid
, "--config", "controller-service-type=loadbalancer"]
737 print("Bootstrapping controller {} in cloud {}".format(
738 cluster_uuid
, cloud_name
741 process
= await asyncio
.create_subprocess_exec(
743 stdout
=asyncio
.subprocess
.PIPE
,
744 stderr
=asyncio
.subprocess
.PIPE
,
747 stdout
, stderr
= await process
.communicate()
749 return_code
= process
.returncode
753 if b
'already exists' not in stderr
:
754 raise Exception(stderr
)
758 async def destroy_controller(
762 """Destroy a Kubernetes controller
764 Destroy an existing Kubernetes controller.
766 :param cluster_uuid str: The UUID of the cluster to bootstrap.
767 :returns: True upon success or raises an exception.
771 "destroy-controller",
772 "--destroy-all-models",
778 process
= await asyncio
.create_subprocess_exec(
780 stdout
=asyncio
.subprocess
.PIPE
,
781 stderr
=asyncio
.subprocess
.PIPE
,
784 stdout
, stderr
= await process
.communicate()
786 return_code
= process
.returncode
790 if 'already exists' not in stderr
:
791 raise Exception(stderr
)
797 """Get the cluster configuration
799 Gets the configuration of the cluster
801 :param cluster_uuid str: The UUID of the cluster.
802 :return: A dict upon success, or raises an exception.
804 cluster_config
= "{}/{}.yaml".format(self
.fs
.path
, cluster_uuid
)
805 if os
.path
.exists(cluster_config
):
806 with
open(cluster_config
, 'r') as f
:
807 config
= yaml
.safe_load(f
.read())
811 "Unable to locate configuration for cluster {}".format(
820 ) -> juju
.model
.Model
:
821 """Get a model from the Juju Controller.
823 Note: Model objects returned must call disconnected() before it goes
826 :param model_name str: The name of the model to get
827 :return The juju.model.Model object if found, or None.
829 if not self
.authenticated
:
830 await self
.login(cluster_uuid
)
833 models
= await self
.controller
.list_models()
834 self
.log
.debug(models
)
835 if model_name
in models
:
836 self
.log
.debug("Found model: {}".format(model_name
))
837 model
= await self
.controller
.get_model(
846 """Get the namespace UUID
847 Gets the namespace's unique name
849 :param cluster_uuid str: The UUID of the cluster
850 :returns: The namespace UUID, or raises an exception
852 config
= self
.get_config(cluster_uuid
)
854 # Make sure the name is in the config
855 if 'namespace' not in config
:
856 raise Exception("Namespace not found.")
858 # TODO: We want to make sure this is unique to the cluster, in case
859 # the cluster is being reused.
860 # Consider pre/appending the cluster id to the namespace string
861 return config
['namespace']
867 """Check if a model exists in the controller
869 Checks to see if a model exists in the connected Juju controller.
871 :param model_name str: The name of the model
872 :return: A boolean indicating if the model exists
874 models
= await self
.controller
.list_models()
876 if model_name
in models
:
884 """Check if a cluster is local
886 Checks if a cluster is running in the local host
888 :param credentials dict: A dictionary containing the k8s credentials
889 :returns: A boolean if the cluster is running locally
891 creds
= yaml
.safe_load(credentials
)
892 if os
.getenv("OSMLCM_VCA_APIPROXY"):
893 host_ip
= os
.getenv("OSMLCM_VCA_APIPROXY")
895 if creds
and host_ip
:
896 for cluster
in creds
['clusters']:
897 if 'server' in cluster
['cluster']:
898 if host_ip
in cluster
['cluster']['server']:
903 async def login(self
, cluster_uuid
):
904 """Login to the Juju controller."""
906 if self
.authenticated
:
909 self
.connecting
= True
911 # Test: Make sure we have the credentials loaded
912 config
= self
.get_config(cluster_uuid
)
914 self
.juju_endpoint
= config
['endpoint']
915 self
.juju_user
= config
['username']
916 self
.juju_secret
= config
['secret']
917 self
.juju_ca_cert
= config
['cacert']
918 self
.juju_public_key
= None
920 self
.controller
= Controller()
924 "Connecting to controller... ws://{} as {}/{}".format(
931 await self
.controller
.connect(
932 endpoint
=self
.juju_endpoint
,
933 username
=self
.juju_user
,
934 password
=self
.juju_secret
,
935 cacert
=self
.juju_ca_cert
,
937 self
.authenticated
= True
938 self
.log
.debug("JujuApi: Logged into controller")
939 except Exception as ex
:
941 self
.log
.debug("Caught exception: {}".format(ex
))
944 self
.log
.fatal("VCA credentials not configured.")
945 self
.authenticated
= False
947 async def logout(self
):
948 """Logout of the Juju controller."""
950 if not self
.authenticated
:
953 for model
in self
.models
:
954 print("Logging out of model {}".format(model
))
955 await self
.models
[model
].disconnect()
958 self
.log
.debug("Disconnecting controller {}".format(
961 await self
.controller
.disconnect()
962 self
.controller
= None
964 self
.authenticated
= False
966 async def remove_cloud(
970 """Remove a k8s cloud from Juju
972 Removes a Kubernetes cloud from Juju.
974 :param cloud_name str: The name of the cloud to add.
976 :returns: True if successful, otherwise raises an exception.
979 # Remove the bootstrapped controller
980 cmd
= [self
.juju_command
, "remove-k8s", "--client", cloud_name
]
981 process
= await asyncio
.create_subprocess_exec(
983 stdout
=asyncio
.subprocess
.PIPE
,
984 stderr
=asyncio
.subprocess
.PIPE
,
987 stdout
, stderr
= await process
.communicate()
989 return_code
= process
.returncode
992 raise Exception(stderr
)
994 # Remove the cloud from the local config
995 cmd
= [self
.juju_command
, "remove-cloud", "--client", cloud_name
]
996 process
= await asyncio
.create_subprocess_exec(
998 stdout
=asyncio
.subprocess
.PIPE
,
999 stderr
=asyncio
.subprocess
.PIPE
,
1002 stdout
, stderr
= await process
.communicate()
1004 return_code
= process
.returncode
1007 raise Exception(stderr
)
1011 async def set_config(
1016 """Save the cluster configuration
1018 Saves the cluster information to the file store
1020 :param cluster_uuid str: The UUID of the cluster
1021 :param config dict: A dictionary containing the cluster configuration
1022 :returns: Boolean upon success or raises an exception.
1025 cluster_config
= "{}/{}.yaml".format(self
.fs
.path
, cluster_uuid
)
1026 if not os
.path
.exists(cluster_config
):
1027 print("Writing config to {}".format(cluster_config
))
1028 with
open(cluster_config
, 'w') as f
:
1029 f
.write(yaml
.dump(config
, Dumper
=yaml
.Dumper
))