--- /dev/null
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+venv
+.vscode
+build
+kafka-exporter.charm
+.coverage
+.stestr
+cover
--- /dev/null
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+---
+extends: default
+
+yaml-files:
+ - "*.yaml"
+ - "*.yml"
+ - ".yamllint"
+ignore: |
+ .tox
+ build/
--- /dev/null
+<!-- Copyright 2021 Canonical Ltd.
+
+Licensed under the Apache License, Version 2.0 (the "License"); you may
+not use this file except in compliance with the License. You may obtain
+a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+License for the specific language governing permissions and limitations
+under the License.
+
+For those usages not covered by the Apache License, Version 2.0 please
+contact: legal@canonical.com
+
+To get in touch with the maintainers, please contact:
+osm-charmers@lists.launchpad.net -->
+
+# Prometheus kafka exporter operator Charm for Kubernetes
+
+## Requirements
--- /dev/null
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+options:
+ ingress_whitelist_source_range:
+ type: string
+ description: |
+ A comma-separated list of CIDRs to store in the
+ ingress.kubernetes.io/whitelist-source-range annotation.
+
+ This can be used to lock down access to
+ Keystone based on source IP address.
+ default: ""
+ tls_secret_name:
+ type: string
+ description: TLS Secret name
+ default: ""
+ site_url:
+ type: string
+ description: Ingress URL
+ default: ""
--- /dev/null
+{
+ "annotations": {
+ "list": [
+ {
+ "builtIn": 1,
+ "datasource": "-- Grafana --",
+ "enable": true,
+ "hide": true,
+ "iconColor": "rgba(0, 211, 255, 1)",
+ "name": "Annotations & Alerts",
+ "type": "dashboard"
+ }
+ ]
+ },
+ "description": "Kafka resource usage and throughput",
+ "editable": true,
+ "gnetId": 7589,
+ "graphTooltip": 0,
+ "id": 10,
+ "iteration": 1578848023483,
+ "links": [],
+ "panels": [
+ {
+ "aliasColors": {},
+ "bars": false,
+ "dashLength": 10,
+ "dashes": false,
+ "datasource": "prometheus - Juju generated source",
+ "fill": 0,
+ "fillGradient": 0,
+ "gridPos": {
+ "h": 10,
+ "w": 10,
+ "x": 0,
+ "y": 0
+ },
+ "id": 14,
+ "legend": {
+ "alignAsTable": true,
+ "avg": false,
+ "current": true,
+ "max": true,
+ "min": false,
+ "rightSide": false,
+ "show": true,
+ "sideWidth": 480,
+ "sort": "max",
+ "sortDesc": true,
+ "total": false,
+ "values": true
+ },
+ "lines": true,
+ "linewidth": 1,
+ "links": [],
+ "nullPointMode": "connected",
+ "options": {
+ "dataLinks": []
+ },
+ "percentage": false,
+ "pointradius": 5,
+ "points": false,
+ "renderer": "flot",
+ "seriesOverrides": [],
+ "spaceLength": 10,
+ "stack": false,
+ "steppedLine": false,
+ "targets": [
+ {
+ "expr": "sum(kafka_topic_partition_current_offset - kafka_topic_partition_oldest_offset{instance=\"$instance\", topic=~\"$topic\"}) by (topic)",
+ "format": "time_series",
+ "intervalFactor": 1,
+ "legendFormat": "{{topic}}",
+ "refId": "B"
+ }
+ ],
+ "thresholds": [],
+ "timeFrom": null,
+ "timeRegions": [],
+ "timeShift": null,
+ "title": "Messages stored per topic",
+ "tooltip": {
+ "shared": true,
+ "sort": 0,
+ "value_type": "individual"
+ },
+ "type": "graph",
+ "xaxis": {
+ "buckets": null,
+ "mode": "time",
+ "name": null,
+ "show": true,
+ "values": []
+ },
+ "yaxes": [
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": "0",
+ "show": true
+ },
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": null,
+ "show": true
+ }
+ ],
+ "yaxis": {
+ "align": false,
+ "alignLevel": null
+ }
+ },
+ {
+ "aliasColors": {},
+ "bars": false,
+ "dashLength": 10,
+ "dashes": false,
+ "datasource": "prometheus - Juju generated source",
+ "fill": 0,
+ "fillGradient": 0,
+ "gridPos": {
+ "h": 10,
+ "w": 10,
+ "x": 10,
+ "y": 0
+ },
+ "id": 12,
+ "legend": {
+ "alignAsTable": true,
+ "avg": false,
+ "current": true,
+ "max": true,
+ "min": false,
+ "rightSide": false,
+ "show": true,
+ "sideWidth": 480,
+ "sort": "max",
+ "sortDesc": true,
+ "total": false,
+ "values": true
+ },
+ "lines": true,
+ "linewidth": 1,
+ "links": [],
+ "nullPointMode": "connected",
+ "options": {
+ "dataLinks": []
+ },
+ "percentage": false,
+ "pointradius": 5,
+ "points": false,
+ "renderer": "flot",
+ "seriesOverrides": [],
+ "spaceLength": 10,
+ "stack": false,
+ "steppedLine": false,
+ "targets": [
+ {
+ "expr": "sum(kafka_consumergroup_lag{instance=\"$instance\",topic=~\"$topic\"}) by (consumergroup, topic) ",
+ "format": "time_series",
+ "instant": false,
+ "interval": "",
+ "intervalFactor": 1,
+ "legendFormat": " {{topic}} ({{consumergroup}})",
+ "refId": "A"
+ }
+ ],
+ "thresholds": [],
+ "timeFrom": null,
+ "timeRegions": [],
+ "timeShift": null,
+ "title": "Lag by Consumer Group",
+ "tooltip": {
+ "shared": true,
+ "sort": 2,
+ "value_type": "individual"
+ },
+ "type": "graph",
+ "xaxis": {
+ "buckets": null,
+ "mode": "time",
+ "name": null,
+ "show": true,
+ "values": []
+ },
+ "yaxes": [
+ {
+ "format": "short",
+ "label": "",
+ "logBase": 1,
+ "max": null,
+ "min": "0",
+ "show": true
+ },
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": null,
+ "show": true
+ }
+ ],
+ "yaxis": {
+ "align": false,
+ "alignLevel": null
+ }
+ },
+ {
+ "aliasColors": {},
+ "bars": false,
+ "dashLength": 10,
+ "dashes": false,
+ "datasource": "prometheus - Juju generated source",
+ "fill": 0,
+ "fillGradient": 0,
+ "gridPos": {
+ "h": 10,
+ "w": 10,
+ "x": 0,
+ "y": 10
+ },
+ "id": 16,
+ "legend": {
+ "alignAsTable": true,
+ "avg": false,
+ "current": true,
+ "max": true,
+ "min": false,
+ "rightSide": false,
+ "show": true,
+ "sideWidth": 480,
+ "total": false,
+ "values": true
+ },
+ "lines": true,
+ "linewidth": 1,
+ "links": [],
+ "nullPointMode": "connected",
+ "options": {
+ "dataLinks": []
+ },
+ "percentage": false,
+ "pointradius": 5,
+ "points": false,
+ "renderer": "flot",
+ "seriesOverrides": [],
+ "spaceLength": 10,
+ "stack": false,
+ "steppedLine": false,
+ "targets": [
+ {
+ "expr": "sum(delta(kafka_topic_partition_current_offset{instance=~'$instance', topic=~\"$topic\"}[5m])/5) by (topic)",
+ "format": "time_series",
+ "intervalFactor": 1,
+ "legendFormat": "{{topic}}",
+ "refId": "A"
+ }
+ ],
+ "thresholds": [],
+ "timeFrom": null,
+ "timeRegions": [],
+ "timeShift": null,
+ "title": "Messages produced per minute",
+ "tooltip": {
+ "shared": true,
+ "sort": 0,
+ "value_type": "individual"
+ },
+ "type": "graph",
+ "xaxis": {
+ "buckets": null,
+ "mode": "time",
+ "name": null,
+ "show": true,
+ "values": []
+ },
+ "yaxes": [
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": null,
+ "show": true
+ },
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": null,
+ "show": true
+ }
+ ],
+ "yaxis": {
+ "align": false,
+ "alignLevel": null
+ }
+ },
+ {
+ "aliasColors": {},
+ "bars": false,
+ "dashLength": 10,
+ "dashes": false,
+ "datasource": "prometheus - Juju generated source",
+ "fill": 0,
+ "fillGradient": 0,
+ "gridPos": {
+ "h": 10,
+ "w": 10,
+ "x": 10,
+ "y": 10
+ },
+ "id": 18,
+ "legend": {
+ "alignAsTable": true,
+ "avg": false,
+ "current": true,
+ "max": true,
+ "min": false,
+ "rightSide": false,
+ "show": true,
+ "sideWidth": 480,
+ "sort": "current",
+ "sortDesc": true,
+ "total": false,
+ "values": true
+ },
+ "lines": true,
+ "linewidth": 1,
+ "links": [],
+ "nullPointMode": "connected",
+ "options": {
+ "dataLinks": []
+ },
+ "percentage": false,
+ "pointradius": 5,
+ "points": false,
+ "renderer": "flot",
+ "seriesOverrides": [],
+ "spaceLength": 10,
+ "stack": false,
+ "steppedLine": false,
+ "targets": [
+ {
+ "expr": "sum(delta(kafka_consumergroup_current_offset{instance=~'$instance',topic=~\"$topic\"}[5m])/5) by (consumergroup, topic)",
+ "format": "time_series",
+ "intervalFactor": 1,
+ "legendFormat": " {{topic}} ({{consumergroup}})",
+ "refId": "A"
+ }
+ ],
+ "thresholds": [],
+ "timeFrom": null,
+ "timeRegions": [],
+ "timeShift": null,
+ "title": "Messages consumed per minute",
+ "tooltip": {
+ "shared": true,
+ "sort": 0,
+ "value_type": "individual"
+ },
+ "type": "graph",
+ "xaxis": {
+ "buckets": null,
+ "mode": "time",
+ "name": null,
+ "show": true,
+ "values": []
+ },
+ "yaxes": [
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": null,
+ "show": true
+ },
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": null,
+ "show": true
+ }
+ ],
+ "yaxis": {
+ "align": false,
+ "alignLevel": null
+ }
+ },
+ {
+ "aliasColors": {},
+ "bars": true,
+ "dashLength": 10,
+ "dashes": false,
+ "datasource": "prometheus - Juju generated source",
+ "fill": 1,
+ "fillGradient": 0,
+ "gridPos": {
+ "h": 7,
+ "w": 20,
+ "x": 0,
+ "y": 20
+ },
+ "id": 8,
+ "legend": {
+ "alignAsTable": true,
+ "avg": false,
+ "current": true,
+ "max": false,
+ "min": false,
+ "rightSide": true,
+ "show": true,
+ "sideWidth": 420,
+ "total": false,
+ "values": true
+ },
+ "lines": false,
+ "linewidth": 1,
+ "links": [],
+ "nullPointMode": "null",
+ "options": {
+ "dataLinks": []
+ },
+ "percentage": false,
+ "pointradius": 5,
+ "points": false,
+ "renderer": "flot",
+ "seriesOverrides": [],
+ "spaceLength": 10,
+ "stack": false,
+ "steppedLine": false,
+ "targets": [
+ {
+ "expr": "sum by(topic) (kafka_topic_partitions{instance=\"$instance\",topic=~\"$topic\"})",
+ "format": "time_series",
+ "intervalFactor": 1,
+ "legendFormat": "{{topic}}",
+ "refId": "A"
+ }
+ ],
+ "thresholds": [],
+ "timeFrom": null,
+ "timeRegions": [],
+ "timeShift": null,
+ "title": "Partitions per Topic",
+ "tooltip": {
+ "shared": false,
+ "sort": 0,
+ "value_type": "individual"
+ },
+ "type": "graph",
+ "xaxis": {
+ "buckets": null,
+ "mode": "series",
+ "name": null,
+ "show": false,
+ "values": [
+ "current"
+ ]
+ },
+ "yaxes": [
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": null,
+ "show": true
+ },
+ {
+ "format": "short",
+ "label": null,
+ "logBase": 1,
+ "max": null,
+ "min": null,
+ "show": true
+ }
+ ],
+ "yaxis": {
+ "align": false,
+ "alignLevel": null
+ }
+ }
+ ],
+ "refresh": "5s",
+ "schemaVersion": 19,
+ "style": "dark",
+ "tags": [],
+ "templating": {
+ "list": [
+ {
+ "allValue": null,
+ "current": {
+ "text": "osm-kafka-exporter-service",
+ "value": "osm-kafka-exporter-service"
+ },
+ "datasource": "prometheus - Juju generated source",
+ "definition": "",
+ "hide": 0,
+ "includeAll": false,
+ "label": "Job",
+ "multi": false,
+ "name": "job",
+ "options": [],
+ "query": "label_values(kafka_consumergroup_current_offset, job)",
+ "refresh": 1,
+ "regex": "",
+ "skipUrlSync": false,
+ "sort": 0,
+ "tagValuesQuery": "",
+ "tags": [],
+ "tagsQuery": "",
+ "type": "query",
+ "useTags": false
+ },
+ {
+ "allValue": null,
+ "datasource": "prometheus - Juju generated source",
+ "definition": "",
+ "hide": 0,
+ "includeAll": false,
+ "label": "Instance",
+ "multi": false,
+ "name": "instance",
+ "options": [],
+ "query": "label_values(kafka_consumergroup_current_offset{job=~\"$job\"}, instance)",
+ "refresh": 1,
+ "regex": "",
+ "skipUrlSync": false,
+ "sort": 0,
+ "tagValuesQuery": "",
+ "tags": [],
+ "tagsQuery": "",
+ "type": "query",
+ "useTags": false
+ },
+ {
+ "allValue": null,
+ "current": {
+ "tags": [],
+ "text": "All",
+ "value": [
+ "$__all"
+ ]
+ },
+ "datasource": "prometheus - Juju generated source",
+ "definition": "",
+ "hide": 0,
+ "includeAll": true,
+ "label": "Topic",
+ "multi": true,
+ "name": "topic",
+ "options": [],
+ "query": "label_values(kafka_topic_partition_current_offset{instance='$instance',topic!='__consumer_offsets',topic!='--kafka'}, topic)",
+ "refresh": 1,
+ "regex": "",
+ "skipUrlSync": false,
+ "sort": 1,
+ "tagValuesQuery": "",
+ "tags": [],
+ "tagsQuery": "topic",
+ "type": "query",
+ "useTags": false
+ }
+ ]
+ },
+ "time": {
+ "from": "now-1h",
+ "to": "now"
+ },
+ "timepicker": {
+ "refresh_intervals": [
+ "5s",
+ "10s",
+ "30s",
+ "1m",
+ "5m",
+ "15m",
+ "30m",
+ "1h",
+ "2h",
+ "1d"
+ ],
+ "time_options": [
+ "5m",
+ "15m",
+ "1h",
+ "6h",
+ "12h",
+ "24h",
+ "2d",
+ "7d",
+ "30d"
+ ]
+ },
+ "timezone": "browser",
+ "title": "Kafka",
+ "uid": "jwPKIsniz",
+ "version": 2
+}
--- /dev/null
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+name: kafka-exporter
+summary: OSM Prometheus Kafka Exporter
+description: |
+ A CAAS charm to deploy OSM's Prometheus Kafka Exporter.
+series:
+ - kubernetes
+tags:
+ - kubernetes
+ - osm
+ - prometheus
+ - kafka-exporter
+min-juju-version: 2.8.0
+deployment:
+ type: stateless
+ service: cluster
+resources:
+ image:
+ type: oci-image
+ description: Image of kafka-exporter
+ upstream-source: "bitnami/kafka-exporter:latest"
+requires:
+ kafka:
+ interface: kafka
+provides:
+ prometheus-scrape:
+ interface: prometheus
+ grafana-dashboard:
+ interface: grafana-dashboard
--- /dev/null
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+ops
+git+https://github.com/juju-solutions/resource-oci-image/@c5778285d332edf3d9a538f9d0c06154b7ec1b0b#egg=oci-image
--- /dev/null
+#!/usr/bin/env python3
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+import logging
+from pathlib import Path
+from typing import Dict, List, NoReturn
+from urllib.parse import urlparse
+
+from ops.charm import CharmBase
+from ops.framework import EventBase, StoredState
+from ops.main import main
+from ops.model import ActiveStatus, Application, BlockedStatus, MaintenanceStatus, Unit
+from oci_image import OCIImageResource, OCIImageResourceError
+
+from pod_spec import make_pod_spec
+
+logger = logging.getLogger(__name__)
+
+KAFKA_EXPORTER_PORT = 9308
+
+
+class RelationsMissing(Exception):
+ def __init__(self, missing_relations: List):
+ self.message = ""
+ if missing_relations and isinstance(missing_relations, list):
+ self.message += f'Waiting for {", ".join(missing_relations)} relation'
+ if "," in self.message:
+ self.message += "s"
+
+
+class RelationDefinition:
+ def __init__(self, relation_name: str, keys: List, source_type):
+ if source_type != Application and source_type != Unit:
+ raise TypeError(
+ "source_type should be ops.model.Application or ops.model.Unit"
+ )
+ self.relation_name = relation_name
+ self.keys = keys
+ self.source_type = source_type
+
+
+def check_missing_relation_data(
+ data: Dict,
+ expected_relations_data: List[RelationDefinition],
+):
+ missing_relations = []
+ for relation_data in expected_relations_data:
+ if not all(
+ f"{relation_data.relation_name}_{k}" in data for k in relation_data.keys
+ ):
+ missing_relations.append(relation_data.relation_name)
+ if missing_relations:
+ raise RelationsMissing(missing_relations)
+
+
+def get_relation_data(
+ charm: CharmBase,
+ relation_data: RelationDefinition,
+) -> Dict:
+ data = {}
+ relation = charm.model.get_relation(relation_data.relation_name)
+ if relation:
+ self_app_unit = (
+ charm.app if relation_data.source_type == Application else charm.unit
+ )
+ expected_type = relation_data.source_type
+ for app_unit in relation.data:
+ if app_unit != self_app_unit and isinstance(app_unit, expected_type):
+ if all(k in relation.data[app_unit] for k in relation_data.keys):
+ for k in relation_data.keys:
+ data[f"{relation_data.relation_name}_{k}"] = relation.data[
+ app_unit
+ ].get(k)
+ break
+ return data
+
+
+class KafkaExporterCharm(CharmBase):
+ """Kafka Exporter Charm."""
+
+ state = StoredState()
+
+ def __init__(self, *args) -> NoReturn:
+ """Kafka Exporter Charm constructor."""
+ super().__init__(*args)
+
+ # Internal state initialization
+ self.state.set_default(pod_spec=None)
+
+ self.port = KAFKA_EXPORTER_PORT
+ self.image = OCIImageResource(self, "image")
+
+ # Registering regular events
+ self.framework.observe(self.on.start, self.configure_pod)
+ self.framework.observe(self.on.config_changed, self.configure_pod)
+
+ # Registering required relation events
+ self.framework.observe(self.on.kafka_relation_changed, self.configure_pod)
+
+ # Registering required relation departed events
+ self.framework.observe(self.on.kafka_relation_departed, self.configure_pod)
+
+ # Registering provided relation events
+ self.framework.observe(
+ self.on.prometheus_scrape_relation_joined, self._publish_scrape_info
+ )
+ self.framework.observe(
+ self.on.grafana_dashboard_relation_joined, self._publish_dashboard_info
+ )
+
+ def _publish_scrape_info(self, event: EventBase) -> NoReturn:
+ """Publishes scrape information.
+
+ Args:
+ event (EventBase): Exporter relation event.
+ """
+ rel_data = {
+ "hostname": urlparse(self.model.config["site_url"]).hostname
+ if self.model.config["site_url"]
+ else self.model.app.name,
+ "port": "80" if self.model.config["site_url"] else str(KAFKA_EXPORTER_PORT),
+ "metrics_path": "/metrics",
+ "scrape_interval": "30s",
+ "scrape_timeout": "15s",
+ }
+ for k, v in rel_data.items():
+ event.relation.data[self.unit][k] = v
+
+ def _publish_dashboard_info(self, event: EventBase) -> NoReturn:
+ """Publishes dashboard information.
+
+ Args:
+ event (EventBase): Exporter relation event.
+ """
+ rel_data = {
+ "name": "osm-kafka",
+ "dashboard": Path("files/kafka_exporter_dashboard.json").read_text(),
+ }
+ for k, v in rel_data.items():
+ event.relation.data[self.unit][k] = v
+
+ @property
+ def relations_requirements(self):
+ return [RelationDefinition("kafka", ["host", "port"], Unit)]
+
+ def get_relation_state(self):
+ relation_state = {}
+ for relation_requirements in self.relations_requirements:
+ data = get_relation_data(self, relation_requirements)
+ relation_state = {**relation_state, **data}
+ check_missing_relation_data(relation_state, self.relations_requirements)
+ return relation_state
+
+ def configure_pod(self, _=None) -> NoReturn:
+ """Assemble the pod spec and apply it, if possible.
+
+ Args:
+ event (EventBase): Hook or Relation event that started the
+ function.
+ """
+ if not self.unit.is_leader():
+ self.unit.status = ActiveStatus("ready")
+ return
+
+ relation_state = None
+ try:
+ relation_state = self.get_relation_state()
+ except RelationsMissing as exc:
+ logger.exception("Relation missing error")
+ self.unit.status = BlockedStatus(exc.message)
+ return
+
+ self.unit.status = MaintenanceStatus("Assembling pod spec")
+
+ # Fetch image information
+ try:
+ self.unit.status = MaintenanceStatus("Fetching image information")
+ image_info = self.image.fetch()
+ except OCIImageResourceError:
+ self.unit.status = BlockedStatus("Error fetching image information")
+ return
+
+ try:
+ pod_spec = make_pod_spec(
+ image_info,
+ self.model.config,
+ relation_state,
+ self.model.app.name,
+ self.port,
+ )
+ except ValueError as exc:
+ logger.exception("Config/Relation data validation error")
+ self.unit.status = BlockedStatus(str(exc))
+ return
+
+ if self.state.pod_spec != pod_spec:
+ self.model.pod.set_spec(pod_spec)
+ self.state.pod_spec = pod_spec
+
+ self.unit.status = ActiveStatus("ready")
+
+
+if __name__ == "__main__":
+ main(KafkaExporterCharm)
--- /dev/null
+#!/usr/bin/env python3
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+import logging
+from ipaddress import ip_network
+from typing import Any, Dict, List
+from urllib.parse import urlparse
+
+logger = logging.getLogger(__name__)
+
+
+def _validate_ip_network(network: str) -> bool:
+ """Validate IP network.
+
+ Args:
+ network (str): IP network range.
+
+ Returns:
+ bool: True if valid, false otherwise.
+ """
+ if not network:
+ return True
+
+ try:
+ ip_network(network)
+ except ValueError:
+ return False
+
+ return True
+
+
+def _validate_data(config_data: Dict[str, Any], relation_data: Dict[str, Any]) -> bool:
+ """Validates passed information.
+
+ Args:
+ config_data (Dict[str, Any]): configuration information.
+ relation_data (Dict[str, Any]): relation information
+
+ Raises:
+ ValueError: when config and/or relation data is not valid.
+ """
+ config_validators = {
+ "site_url": lambda value, _: isinstance(value, str)
+ if value is not None
+ else True,
+ "ingress_whitelist_source_range": lambda value, _: _validate_ip_network(value),
+ "tls_secret_name": lambda value, _: isinstance(value, str)
+ if value is not None
+ else True,
+ }
+ relation_validators = {
+ "kafka_host": lambda value, _: isinstance(value, str) and len(value) > 0,
+ "kafka_port": lambda value, _: isinstance(value, str)
+ and len(value) > 0
+ and int(value) > 0,
+ }
+ problems = []
+
+ for key, validator in config_validators.items():
+ valid = validator(config_data.get(key), config_data)
+
+ if not valid:
+ problems.append(key)
+
+ for key, validator in relation_validators.items():
+ valid = validator(relation_data.get(key), relation_data)
+
+ if not valid:
+ problems.append(key)
+
+ if len(problems) > 0:
+ raise ValueError("Errors found in: {}".format(", ".join(problems)))
+
+ return True
+
+
+def _make_pod_ports(port: int) -> List[Dict[str, Any]]:
+ """Generate pod ports details.
+
+ Args:
+ port (int): port to expose.
+
+ Returns:
+ List[Dict[str, Any]]: pod port details.
+ """
+ return [{"name": "kafka-exporter", "containerPort": port, "protocol": "TCP"}]
+
+
+def _make_pod_envconfig(
+ config: Dict[str, Any], relation_state: Dict[str, Any]
+) -> Dict[str, Any]:
+ """Generate pod environment configuration.
+
+ Args:
+ config (Dict[str, Any]): configuration information.
+ relation_state (Dict[str, Any]): relation state information.
+
+ Returns:
+ Dict[str, Any]: pod environment configuration.
+ """
+ envconfig = {}
+
+ return envconfig
+
+
+def _make_pod_ingress_resources(
+ config: Dict[str, Any], app_name: str, port: int
+) -> List[Dict[str, Any]]:
+ """Generate pod ingress resources.
+
+ Args:
+ config (Dict[str, Any]): configuration information.
+ app_name (str): application name.
+ port (int): port to expose.
+
+ Returns:
+ List[Dict[str, Any]]: pod ingress resources.
+ """
+ site_url = config.get("site_url")
+
+ if not site_url:
+ return
+
+ parsed = urlparse(site_url)
+
+ if not parsed.scheme.startswith("http"):
+ return
+
+ ingress_whitelist_source_range = config["ingress_whitelist_source_range"]
+
+ annotations = {}
+
+ if ingress_whitelist_source_range:
+ annotations[
+ "nginx.ingress.kubernetes.io/whitelist-source-range"
+ ] = ingress_whitelist_source_range
+
+ ingress_spec_tls = None
+
+ if parsed.scheme == "https":
+ ingress_spec_tls = [{"hosts": [parsed.hostname]}]
+ tls_secret_name = config["tls_secret_name"]
+ if tls_secret_name:
+ ingress_spec_tls[0]["secretName"] = tls_secret_name
+ else:
+ annotations["nginx.ingress.kubernetes.io/ssl-redirect"] = "false"
+
+ ingress = {
+ "name": "{}-ingress".format(app_name),
+ "annotations": annotations,
+ "spec": {
+ "rules": [
+ {
+ "host": parsed.hostname,
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": app_name,
+ "servicePort": port,
+ },
+ }
+ ]
+ },
+ }
+ ]
+ },
+ }
+ if ingress_spec_tls:
+ ingress["spec"]["tls"] = ingress_spec_tls
+
+ return [ingress]
+
+
+def _make_readiness_probe(port: int) -> Dict[str, Any]:
+ """Generate readiness probe.
+
+ Args:
+ port (int): service port.
+
+ Returns:
+ Dict[str, Any]: readiness probe.
+ """
+ return {
+ "httpGet": {
+ "path": "/api/health",
+ "port": port,
+ },
+ "initialDelaySeconds": 10,
+ "periodSeconds": 10,
+ "timeoutSeconds": 5,
+ "successThreshold": 1,
+ "failureThreshold": 3,
+ }
+
+
+def _make_liveness_probe(port: int) -> Dict[str, Any]:
+ """Generate liveness probe.
+
+ Args:
+ port (int): service port.
+
+ Returns:
+ Dict[str, Any]: liveness probe.
+ """
+ return {
+ "httpGet": {
+ "path": "/api/health",
+ "port": port,
+ },
+ "initialDelaySeconds": 60,
+ "timeoutSeconds": 30,
+ "failureThreshold": 10,
+ }
+
+
+def _make_pod_command(relation: Dict[str, Any]) -> List[str]:
+ """Generate the startup command.
+
+ Args:
+ relation (Dict[str, Any]): Relation information.
+
+ Returns:
+ List[str]: command to startup the process.
+ """
+ command = [
+ "kafka_exporter",
+ "--kafka.server={}:{}".format(
+ relation.get("kafka_host"), relation.get("kafka_port")
+ ),
+ ]
+
+ return command
+
+
+def make_pod_spec(
+ image_info: Dict[str, str],
+ config: Dict[str, Any],
+ relation_state: Dict[str, Any],
+ app_name: str = "kafka-exporter",
+ port: int = 9308,
+) -> Dict[str, Any]:
+ """Generate the pod spec information.
+
+ Args:
+ image_info (Dict[str, str]): Object provided by
+ OCIImageResource("image").fetch().
+ config (Dict[str, Any]): Configuration information.
+ relation_state (Dict[str, Any]): Relation state information.
+ app_name (str, optional): Application name. Defaults to "ro".
+ port (int, optional): Port for the container. Defaults to 9090.
+
+ Returns:
+ Dict[str, Any]: Pod spec dictionary for the charm.
+ """
+ if not image_info:
+ return None
+
+ _validate_data(config, relation_state)
+
+ ports = _make_pod_ports(port)
+ env_config = _make_pod_envconfig(config, relation_state)
+ readiness_probe = _make_readiness_probe(port)
+ liveness_probe = _make_liveness_probe(port)
+ ingress_resources = _make_pod_ingress_resources(config, app_name, port)
+ command = _make_pod_command(relation_state)
+
+ return {
+ "version": 3,
+ "containers": [
+ {
+ "name": app_name,
+ "imageDetails": image_info,
+ "imagePullPolicy": "Always",
+ "ports": ports,
+ "envConfig": env_config,
+ "command": command,
+ "kubernetes": {
+ "readinessProbe": readiness_probe,
+ "livenessProbe": liveness_probe,
+ },
+ }
+ ],
+ "kubernetesResources": {
+ "ingressResources": ingress_resources or [],
+ },
+ }
--- /dev/null
+#!/usr/bin/env python3
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+"""Init mocking for unit tests."""
+
+import sys
+import mock
+
+sys.path.append("src")
+
+oci_image = mock.MagicMock()
+sys.modules["oci_image"] = oci_image
--- /dev/null
+#!/usr/bin/env python3
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+from typing import NoReturn
+import unittest
+
+from ops.model import BlockedStatus
+from ops.testing import Harness
+
+from charm import KafkaExporterCharm
+
+
+class TestCharm(unittest.TestCase):
+ """Kafka Exporter Charm unit tests."""
+
+ def setUp(self) -> NoReturn:
+ """Test setup"""
+ self.harness = Harness(KafkaExporterCharm)
+ self.harness.set_leader(is_leader=True)
+ self.harness.begin()
+
+ def test_on_start_without_relations(self) -> NoReturn:
+ """Test installation without any relation."""
+ self.harness.charm.on.start.emit()
+
+ # Verifying status
+ self.assertIsInstance(self.harness.charm.unit.status, BlockedStatus)
+
+ # Verifying status message
+ self.assertGreater(len(self.harness.charm.unit.status.message), 0)
+ self.assertTrue(
+ self.harness.charm.unit.status.message.startswith("Waiting for ")
+ )
+ self.assertIn("kafka", self.harness.charm.unit.status.message)
+ self.assertTrue(self.harness.charm.unit.status.message.endswith(" relation"))
+
+ def test_on_start_with_relations_without_http(self) -> NoReturn:
+ """Test deployment."""
+ expected_result = {
+ "version": 3,
+ "containers": [
+ {
+ "name": "kafka-exporter",
+ "imageDetails": self.harness.charm.image.fetch(),
+ "imagePullPolicy": "Always",
+ "ports": [
+ {
+ "name": "kafka-exporter",
+ "containerPort": 9308,
+ "protocol": "TCP",
+ }
+ ],
+ "envConfig": {},
+ "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
+ "kubernetes": {
+ "readinessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": 9308,
+ },
+ "initialDelaySeconds": 10,
+ "periodSeconds": 10,
+ "timeoutSeconds": 5,
+ "successThreshold": 1,
+ "failureThreshold": 3,
+ },
+ "livenessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": 9308,
+ },
+ "initialDelaySeconds": 60,
+ "timeoutSeconds": 30,
+ "failureThreshold": 10,
+ },
+ },
+ },
+ ],
+ "kubernetesResources": {"ingressResources": []},
+ }
+
+ self.harness.charm.on.start.emit()
+
+ # Initializing the kafka relation
+ relation_id = self.harness.add_relation("kafka", "kafka")
+ self.harness.add_relation_unit(relation_id, "kafka/0")
+ self.harness.update_relation_data(
+ relation_id,
+ "kafka/0",
+ {
+ "host": "kafka",
+ "port": "9090",
+ },
+ )
+
+ # Verifying status
+ self.assertNotIsInstance(self.harness.charm.unit.status, BlockedStatus)
+
+ pod_spec, _ = self.harness.get_pod_spec()
+
+ self.assertDictEqual(expected_result, pod_spec)
+
+ def test_ingress_resources_with_http(self) -> NoReturn:
+ """Test ingress resources with HTTP."""
+ expected_result = {
+ "version": 3,
+ "containers": [
+ {
+ "name": "kafka-exporter",
+ "imageDetails": self.harness.charm.image.fetch(),
+ "imagePullPolicy": "Always",
+ "ports": [
+ {
+ "name": "kafka-exporter",
+ "containerPort": 9308,
+ "protocol": "TCP",
+ }
+ ],
+ "envConfig": {},
+ "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
+ "kubernetes": {
+ "readinessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": 9308,
+ },
+ "initialDelaySeconds": 10,
+ "periodSeconds": 10,
+ "timeoutSeconds": 5,
+ "successThreshold": 1,
+ "failureThreshold": 3,
+ },
+ "livenessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": 9308,
+ },
+ "initialDelaySeconds": 60,
+ "timeoutSeconds": 30,
+ "failureThreshold": 10,
+ },
+ },
+ },
+ ],
+ "kubernetesResources": {
+ "ingressResources": [
+ {
+ "name": "kafka-exporter-ingress",
+ "annotations": {
+ "nginx.ingress.kubernetes.io/ssl-redirect": "false",
+ },
+ "spec": {
+ "rules": [
+ {
+ "host": "kafka-exporter",
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": "kafka-exporter",
+ "servicePort": 9308,
+ },
+ }
+ ]
+ },
+ }
+ ]
+ },
+ }
+ ],
+ },
+ }
+
+ self.harness.charm.on.start.emit()
+
+ # Initializing the kafka relation
+ relation_id = self.harness.add_relation("kafka", "kafka")
+ self.harness.add_relation_unit(relation_id, "kafka/0")
+ self.harness.update_relation_data(
+ relation_id,
+ "kafka/0",
+ {
+ "host": "kafka",
+ "port": "9090",
+ },
+ )
+
+ self.harness.update_config({"site_url": "http://kafka-exporter"})
+
+ pod_spec, _ = self.harness.get_pod_spec()
+
+ self.assertDictEqual(expected_result, pod_spec)
+
+ def test_ingress_resources_with_https(self) -> NoReturn:
+ """Test ingress resources with HTTPS."""
+ expected_result = {
+ "version": 3,
+ "containers": [
+ {
+ "name": "kafka-exporter",
+ "imageDetails": self.harness.charm.image.fetch(),
+ "imagePullPolicy": "Always",
+ "ports": [
+ {
+ "name": "kafka-exporter",
+ "containerPort": 9308,
+ "protocol": "TCP",
+ }
+ ],
+ "envConfig": {},
+ "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
+ "kubernetes": {
+ "readinessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": 9308,
+ },
+ "initialDelaySeconds": 10,
+ "periodSeconds": 10,
+ "timeoutSeconds": 5,
+ "successThreshold": 1,
+ "failureThreshold": 3,
+ },
+ "livenessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": 9308,
+ },
+ "initialDelaySeconds": 60,
+ "timeoutSeconds": 30,
+ "failureThreshold": 10,
+ },
+ },
+ },
+ ],
+ "kubernetesResources": {
+ "ingressResources": [
+ {
+ "name": "kafka-exporter-ingress",
+ "annotations": {},
+ "spec": {
+ "rules": [
+ {
+ "host": "kafka-exporter",
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": "kafka-exporter",
+ "servicePort": 9308,
+ },
+ }
+ ]
+ },
+ }
+ ],
+ "tls": [
+ {
+ "hosts": ["kafka-exporter"],
+ "secretName": "kafka-exporter",
+ }
+ ],
+ },
+ }
+ ],
+ },
+ }
+
+ self.harness.charm.on.start.emit()
+
+ # Initializing the kafka relation
+ relation_id = self.harness.add_relation("kafka", "kafka")
+ self.harness.add_relation_unit(relation_id, "kafka/0")
+ self.harness.update_relation_data(
+ relation_id,
+ "kafka/0",
+ {
+ "host": "kafka",
+ "port": "9090",
+ },
+ )
+
+ self.harness.update_config(
+ {
+ "site_url": "https://kafka-exporter",
+ "tls_secret_name": "kafka-exporter",
+ }
+ )
+
+ pod_spec, _ = self.harness.get_pod_spec()
+
+ self.assertDictEqual(expected_result, pod_spec)
+
+ def test_ingress_resources_with_https_and_ingress_whitelist(self) -> NoReturn:
+ """Test ingress resources with HTTPS and ingress whitelist."""
+ expected_result = {
+ "version": 3,
+ "containers": [
+ {
+ "name": "kafka-exporter",
+ "imageDetails": self.harness.charm.image.fetch(),
+ "imagePullPolicy": "Always",
+ "ports": [
+ {
+ "name": "kafka-exporter",
+ "containerPort": 9308,
+ "protocol": "TCP",
+ }
+ ],
+ "envConfig": {},
+ "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
+ "kubernetes": {
+ "readinessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": 9308,
+ },
+ "initialDelaySeconds": 10,
+ "periodSeconds": 10,
+ "timeoutSeconds": 5,
+ "successThreshold": 1,
+ "failureThreshold": 3,
+ },
+ "livenessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": 9308,
+ },
+ "initialDelaySeconds": 60,
+ "timeoutSeconds": 30,
+ "failureThreshold": 10,
+ },
+ },
+ },
+ ],
+ "kubernetesResources": {
+ "ingressResources": [
+ {
+ "name": "kafka-exporter-ingress",
+ "annotations": {
+ "nginx.ingress.kubernetes.io/whitelist-source-range": "0.0.0.0/0",
+ },
+ "spec": {
+ "rules": [
+ {
+ "host": "kafka-exporter",
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": "kafka-exporter",
+ "servicePort": 9308,
+ },
+ }
+ ]
+ },
+ }
+ ],
+ "tls": [
+ {
+ "hosts": ["kafka-exporter"],
+ "secretName": "kafka-exporter",
+ }
+ ],
+ },
+ }
+ ],
+ },
+ }
+
+ self.harness.charm.on.start.emit()
+
+ # Initializing the kafka relation
+ relation_id = self.harness.add_relation("kafka", "kafka")
+ self.harness.add_relation_unit(relation_id, "kafka/0")
+ self.harness.update_relation_data(
+ relation_id,
+ "kafka/0",
+ {
+ "host": "kafka",
+ "port": "9090",
+ },
+ )
+
+ self.harness.update_config(
+ {
+ "site_url": "https://kafka-exporter",
+ "tls_secret_name": "kafka-exporter",
+ "ingress_whitelist_source_range": "0.0.0.0/0",
+ }
+ )
+
+ pod_spec, _ = self.harness.get_pod_spec()
+
+ self.assertDictEqual(expected_result, pod_spec)
+
+ def test_on_kafka_unit_relation_changed(self) -> NoReturn:
+ """Test to see if kafka relation is updated."""
+ self.harness.charm.on.start.emit()
+
+ relation_id = self.harness.add_relation("kafka", "kafka")
+ self.harness.add_relation_unit(relation_id, "kafka/0")
+ self.harness.update_relation_data(
+ relation_id,
+ "kafka/0",
+ {
+ "host": "kafka",
+ "port": "9090",
+ },
+ )
+
+ # Verifying status
+ self.assertNotIsInstance(self.harness.charm.unit.status, BlockedStatus)
+
+ def test_publish_target_info(self) -> NoReturn:
+ """Test to see if target relation is updated."""
+ expected_result = {
+ "hostname": "kafka-exporter",
+ "port": "9308",
+ "metrics_path": "/metrics",
+ "scrape_interval": "30s",
+ "scrape_timeout": "15s",
+ }
+
+ self.harness.charm.on.start.emit()
+
+ relation_id = self.harness.add_relation("prometheus-scrape", "prometheus")
+ self.harness.add_relation_unit(relation_id, "prometheus/0")
+ relation_data = self.harness.get_relation_data(relation_id, "kafka-exporter/0")
+
+ self.assertDictEqual(expected_result, relation_data)
+
+ def test_publish_target_info_with_site_url(self) -> NoReturn:
+ """Test to see if target relation is updated."""
+ expected_result = {
+ "hostname": "kafka-exporter-osm",
+ "port": "80",
+ "metrics_path": "/metrics",
+ "scrape_interval": "30s",
+ "scrape_timeout": "15s",
+ }
+
+ self.harness.charm.on.start.emit()
+
+ self.harness.update_config({"site_url": "http://kafka-exporter-osm"})
+
+ relation_id = self.harness.add_relation("prometheus-scrape", "prometheus")
+ self.harness.add_relation_unit(relation_id, "prometheus/0")
+ relation_data = self.harness.get_relation_data(relation_id, "kafka-exporter/0")
+
+ self.assertDictEqual(expected_result, relation_data)
+
+ def test_publish_dashboard_info(self) -> NoReturn:
+ """Test to see if dashboard relation is updated."""
+ self.harness.charm.on.start.emit()
+
+ relation_id = self.harness.add_relation("grafana-dashboard", "grafana")
+ self.harness.add_relation_unit(relation_id, "grafana/0")
+ relation_data = self.harness.get_relation_data(relation_id, "kafka-exporter/0")
+
+ self.assertTrue("dashboard" in relation_data)
+ self.assertTrue(len(relation_data["dashboard"]) > 0)
+
+
+if __name__ == "__main__":
+ unittest.main()
--- /dev/null
+#!/usr/bin/env python3
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+from typing import NoReturn
+import unittest
+
+import pod_spec
+
+
+class TestPodSpec(unittest.TestCase):
+ """Pod spec unit tests."""
+
+ def test_make_pod_ports(self) -> NoReturn:
+ """Testing make pod ports."""
+ port = 9308
+
+ expected_result = [
+ {
+ "name": "kafka-exporter",
+ "containerPort": port,
+ "protocol": "TCP",
+ }
+ ]
+
+ pod_ports = pod_spec._make_pod_ports(port)
+
+ self.assertListEqual(expected_result, pod_ports)
+
+ def test_make_pod_envconfig(self) -> NoReturn:
+ """Teting make pod envconfig."""
+ config = {}
+ relation_state = {}
+
+ expected_result = {}
+
+ pod_envconfig = pod_spec._make_pod_envconfig(config, relation_state)
+
+ self.assertDictEqual(expected_result, pod_envconfig)
+
+ def test_make_pod_ingress_resources_without_site_url(self) -> NoReturn:
+ """Testing make pod ingress resources without site_url."""
+ config = {"site_url": ""}
+ app_name = "kafka-exporter"
+ port = 9308
+
+ pod_ingress_resources = pod_spec._make_pod_ingress_resources(
+ config, app_name, port
+ )
+
+ self.assertIsNone(pod_ingress_resources)
+
+ def test_make_pod_ingress_resources(self) -> NoReturn:
+ """Testing make pod ingress resources."""
+ config = {
+ "site_url": "http://kafka-exporter",
+ "ingress_whitelist_source_range": "",
+ }
+ app_name = "kafka-exporter"
+ port = 9308
+
+ expected_result = [
+ {
+ "name": f"{app_name}-ingress",
+ "annotations": {
+ "nginx.ingress.kubernetes.io/ssl-redirect": "false",
+ },
+ "spec": {
+ "rules": [
+ {
+ "host": app_name,
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": app_name,
+ "servicePort": port,
+ },
+ }
+ ]
+ },
+ }
+ ]
+ },
+ }
+ ]
+
+ pod_ingress_resources = pod_spec._make_pod_ingress_resources(
+ config, app_name, port
+ )
+
+ self.assertListEqual(expected_result, pod_ingress_resources)
+
+ def test_make_pod_ingress_resources_with_whitelist_source_range(self) -> NoReturn:
+ """Testing make pod ingress resources with whitelist_source_range."""
+ config = {
+ "site_url": "http://kafka-exporter",
+ "ingress_whitelist_source_range": "0.0.0.0/0",
+ }
+ app_name = "kafka-exporter"
+ port = 9308
+
+ expected_result = [
+ {
+ "name": f"{app_name}-ingress",
+ "annotations": {
+ "nginx.ingress.kubernetes.io/ssl-redirect": "false",
+ "nginx.ingress.kubernetes.io/whitelist-source-range": config[
+ "ingress_whitelist_source_range"
+ ],
+ },
+ "spec": {
+ "rules": [
+ {
+ "host": app_name,
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": app_name,
+ "servicePort": port,
+ },
+ }
+ ]
+ },
+ }
+ ]
+ },
+ }
+ ]
+
+ pod_ingress_resources = pod_spec._make_pod_ingress_resources(
+ config, app_name, port
+ )
+
+ self.assertListEqual(expected_result, pod_ingress_resources)
+
+ def test_make_pod_ingress_resources_with_https(self) -> NoReturn:
+ """Testing make pod ingress resources with HTTPs."""
+ config = {
+ "site_url": "https://kafka-exporter",
+ "max_file_size": 0,
+ "ingress_whitelist_source_range": "",
+ "tls_secret_name": "",
+ }
+ app_name = "kafka-exporter"
+ port = 9308
+
+ expected_result = [
+ {
+ "name": f"{app_name}-ingress",
+ "annotations": {},
+ "spec": {
+ "rules": [
+ {
+ "host": app_name,
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": app_name,
+ "servicePort": port,
+ },
+ }
+ ]
+ },
+ }
+ ],
+ "tls": [{"hosts": [app_name]}],
+ },
+ }
+ ]
+
+ pod_ingress_resources = pod_spec._make_pod_ingress_resources(
+ config, app_name, port
+ )
+
+ self.assertListEqual(expected_result, pod_ingress_resources)
+
+ def test_make_pod_ingress_resources_with_https_tls_secret_name(self) -> NoReturn:
+ """Testing make pod ingress resources with HTTPs and TLS secret name."""
+ config = {
+ "site_url": "https://kafka-exporter",
+ "max_file_size": 0,
+ "ingress_whitelist_source_range": "",
+ "tls_secret_name": "secret_name",
+ }
+ app_name = "kafka-exporter"
+ port = 9308
+
+ expected_result = [
+ {
+ "name": f"{app_name}-ingress",
+ "annotations": {},
+ "spec": {
+ "rules": [
+ {
+ "host": app_name,
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": app_name,
+ "servicePort": port,
+ },
+ }
+ ]
+ },
+ }
+ ],
+ "tls": [
+ {"hosts": [app_name], "secretName": config["tls_secret_name"]}
+ ],
+ },
+ }
+ ]
+
+ pod_ingress_resources = pod_spec._make_pod_ingress_resources(
+ config, app_name, port
+ )
+
+ self.assertListEqual(expected_result, pod_ingress_resources)
+
+ def test_make_readiness_probe(self) -> NoReturn:
+ """Testing make readiness probe."""
+ port = 9308
+
+ expected_result = {
+ "httpGet": {
+ "path": "/api/health",
+ "port": port,
+ },
+ "initialDelaySeconds": 10,
+ "periodSeconds": 10,
+ "timeoutSeconds": 5,
+ "successThreshold": 1,
+ "failureThreshold": 3,
+ }
+
+ readiness_probe = pod_spec._make_readiness_probe(port)
+
+ self.assertDictEqual(expected_result, readiness_probe)
+
+ def test_make_liveness_probe(self) -> NoReturn:
+ """Testing make liveness probe."""
+ port = 9308
+
+ expected_result = {
+ "httpGet": {
+ "path": "/api/health",
+ "port": port,
+ },
+ "initialDelaySeconds": 60,
+ "timeoutSeconds": 30,
+ "failureThreshold": 10,
+ }
+
+ liveness_probe = pod_spec._make_liveness_probe(port)
+
+ self.assertDictEqual(expected_result, liveness_probe)
+
+ def test_make_pod_command(self) -> NoReturn:
+ """Testing make pod command."""
+ relation = {
+ "kakfa_host": "kafka",
+ "kafka_port": "9090",
+ }
+
+ expected_result = [
+ "kafka-exporter",
+ "--kafka.server={}:{}".format(
+ relation.get("kafka_host"), relation.get("kafka_port")
+ ),
+ ]
+
+ pod_envconfig = pod_spec._make_pod_command(relation)
+
+ self.assertListEqual(expected_result, pod_envconfig)
+
+ def test_make_pod_spec(self) -> NoReturn:
+ """Testing make pod spec."""
+ image_info = {"upstream-source": "bitnami/kafka-exporter:latest"}
+ config = {
+ "site_url": "",
+ }
+ relation_state = {
+ "kafka_host": "kafka",
+ "kafka_port": "9090",
+ }
+ app_name = "kafka-exporter"
+ port = 9308
+
+ expected_result = {
+ "version": 3,
+ "containers": [
+ {
+ "name": app_name,
+ "imageDetails": image_info,
+ "imagePullPolicy": "Always",
+ "ports": [
+ {
+ "name": app_name,
+ "containerPort": port,
+ "protocol": "TCP",
+ }
+ ],
+ "envConfig": {},
+ "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
+ "kubernetes": {
+ "readinessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": port,
+ },
+ "initialDelaySeconds": 10,
+ "periodSeconds": 10,
+ "timeoutSeconds": 5,
+ "successThreshold": 1,
+ "failureThreshold": 3,
+ },
+ "livenessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": port,
+ },
+ "initialDelaySeconds": 60,
+ "timeoutSeconds": 30,
+ "failureThreshold": 10,
+ },
+ },
+ }
+ ],
+ "kubernetesResources": {"ingressResources": []},
+ }
+
+ spec = pod_spec.make_pod_spec(
+ image_info, config, relation_state, app_name, port
+ )
+
+ self.assertDictEqual(expected_result, spec)
+
+ def test_make_pod_spec_with_ingress(self) -> NoReturn:
+ """Testing make pod spec."""
+ image_info = {"upstream-source": "bitnami/kafka-exporter:latest"}
+ config = {
+ "site_url": "https://kafka-exporter",
+ "tls_secret_name": "kafka-exporter",
+ "max_file_size": 0,
+ "ingress_whitelist_source_range": "0.0.0.0/0",
+ }
+ relation_state = {
+ "kafka_host": "kafka",
+ "kafka_port": "9090",
+ }
+ app_name = "kafka-exporter"
+ port = 9308
+
+ expected_result = {
+ "version": 3,
+ "containers": [
+ {
+ "name": app_name,
+ "imageDetails": image_info,
+ "imagePullPolicy": "Always",
+ "ports": [
+ {
+ "name": app_name,
+ "containerPort": port,
+ "protocol": "TCP",
+ }
+ ],
+ "envConfig": {},
+ "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
+ "kubernetes": {
+ "readinessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": port,
+ },
+ "initialDelaySeconds": 10,
+ "periodSeconds": 10,
+ "timeoutSeconds": 5,
+ "successThreshold": 1,
+ "failureThreshold": 3,
+ },
+ "livenessProbe": {
+ "httpGet": {
+ "path": "/api/health",
+ "port": port,
+ },
+ "initialDelaySeconds": 60,
+ "timeoutSeconds": 30,
+ "failureThreshold": 10,
+ },
+ },
+ }
+ ],
+ "kubernetesResources": {
+ "ingressResources": [
+ {
+ "name": "{}-ingress".format(app_name),
+ "annotations": {
+ "nginx.ingress.kubernetes.io/whitelist-source-range": config.get(
+ "ingress_whitelist_source_range"
+ ),
+ },
+ "spec": {
+ "rules": [
+ {
+ "host": app_name,
+ "http": {
+ "paths": [
+ {
+ "path": "/",
+ "backend": {
+ "serviceName": app_name,
+ "servicePort": port,
+ },
+ }
+ ]
+ },
+ }
+ ],
+ "tls": [
+ {
+ "hosts": [app_name],
+ "secretName": config.get("tls_secret_name"),
+ }
+ ],
+ },
+ }
+ ],
+ },
+ }
+
+ spec = pod_spec.make_pod_spec(
+ image_info, config, relation_state, app_name, port
+ )
+
+ self.assertDictEqual(expected_result, spec)
+
+ def test_make_pod_spec_without_image_info(self) -> NoReturn:
+ """Testing make pod spec without image_info."""
+ image_info = None
+ config = {
+ "site_url": "",
+ }
+ relation_state = {
+ "kafka_host": "kafka",
+ "kafka_port": "9090",
+ }
+ app_name = "kafka-exporter"
+ port = 9308
+
+ spec = pod_spec.make_pod_spec(
+ image_info, config, relation_state, app_name, port
+ )
+
+ self.assertIsNone(spec)
+
+ def test_make_pod_spec_without_relation_state(self) -> NoReturn:
+ """Testing make pod spec without relation_state."""
+ image_info = {"upstream-source": "bitnami/kafka-exporter:latest"}
+ config = {
+ "site_url": "",
+ }
+ relation_state = {}
+ app_name = "kafka-exporter"
+ port = 9308
+
+ with self.assertRaises(ValueError):
+ pod_spec.make_pod_spec(image_info, config, relation_state, app_name, port)
+
+
+if __name__ == "__main__":
+ unittest.main()
--- /dev/null
+# Copyright 2021 Canonical Ltd.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+#
+# For those usages not covered by the Apache License, Version 2.0 please
+# contact: legal@canonical.com
+#
+# To get in touch with the maintainers, please contact:
+# osm-charmers@lists.launchpad.net
+##
+
+[tox]
+skipsdist = True
+envlist = unit, lint
+sitepackages = False
+skip_missing_interpreters = False
+
+[testenv]
+basepython = python3
+setenv =
+ PYTHONHASHSEED=0
+ PYTHONPATH = {toxinidir}/src
+ CHARM_NAME = kafka-exporter
+
+[testenv:build]
+passenv=HTTP_PROXY HTTPS_PROXY NO_PROXY
+whitelist_externals =
+ charmcraft
+ rm
+ unzip
+commands =
+ rm -rf release kafka-exporter.charm
+ charmcraft build
+ unzip kafka-exporter.charm -d release
+
+[testenv:unit]
+commands =
+ coverage erase
+ stestr run --slowest --test-path=./tests --top-dir=./
+ coverage combine
+ coverage html -d cover
+ coverage xml -o cover/coverage.xml
+ coverage report
+deps =
+ coverage
+ stestr
+ mock
+ ops
+setenv =
+ {[testenv]setenv}
+ PYTHON=coverage run
+
+[testenv:lint]
+deps =
+ black
+ yamllint
+ flake8
+commands =
+ black --check --diff . --exclude "build/|.tox/|mod/|lib/"
+ yamllint .
+ flake8 . --max-line-length=100 --ignore="E501,W503,W504,F722" --exclude "build/ .tox/ mod/ lib/"
+
+[coverage:run]
+branch = True
+concurrency = multiprocessing
+parallel = True
+source =
+ .
+omit =
+ .tox/*
+ tests/*
+++ /dev/null
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-venv
-.vscode
-build
-prometheus-kafka-exporter.charm
-.coverage
-.stestr
-cover
+++ /dev/null
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
----
-extends: default
-
-yaml-files:
- - "*.yaml"
- - "*.yml"
- - ".yamllint"
-ignore: |
- .tox
- build/
+++ /dev/null
-<!-- Copyright 2021 Canonical Ltd.
-
-Licensed under the Apache License, Version 2.0 (the "License"); you may
-not use this file except in compliance with the License. You may obtain
-a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-License for the specific language governing permissions and limitations
-under the License.
-
-For those usages not covered by the Apache License, Version 2.0 please
-contact: legal@canonical.com
-
-To get in touch with the maintainers, please contact:
-osm-charmers@lists.launchpad.net -->
-
-# Prometheus kafka exporter operator Charm for Kubernetes
-
-## Requirements
+++ /dev/null
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-options:
- ingress_whitelist_source_range:
- type: string
- description: |
- A comma-separated list of CIDRs to store in the
- ingress.kubernetes.io/whitelist-source-range annotation.
-
- This can be used to lock down access to
- Keystone based on source IP address.
- default: ""
- tls_secret_name:
- type: string
- description: TLS Secret name
- default: ""
- site_url:
- type: string
- description: Ingress URL
- default: ""
+++ /dev/null
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
----
-annotations:
- list:
- - builtIn: 1
- datasource: '-- Grafana --'
- enable: true
- hide: true
- iconColor: 'rgba(0, 211, 255, 1)'
- name: Annotations & Alerts
- type: dashboard
-description: Kafka resource usage and throughput
-editable: true
-gnetId: 7589
-graphTooltip: 0
-id: 10
-iteration: 1578848023483
-links: []
-panels:
- - aliasColors: {}
- bars: false
- dashLength: 10
- dashes: false
- datasource: Prometheus
- fill: 0
- fillGradient: 0
- gridPos:
- h: 10
- w: 10
- x: 0
- 'y': 0
- id: 14
- legend:
- alignAsTable: true
- avg: false
- current: true
- max: true
- min: false
- rightSide: false
- show: true
- sideWidth: 480
- sort: max
- sortDesc: true
- total: false
- values: true
- lines: true
- linewidth: 1
- links: []
- nullPointMode: connected
- options:
- dataLinks: []
- percentage: false
- pointradius: 5
- points: false
- renderer: flot
- seriesOverrides: []
- spaceLength: 10
- stack: false
- steppedLine: false
- targets:
- - expr: >-
- sum(kafka_topic_partition_current_offset -
- kafka_topic_partition_oldest_offset{instance="$instance",
- topic=~"$topic"}) by (topic)
- format: time_series
- intervalFactor: 1
- legendFormat: '{{topic}}'
- refId: B
- thresholds: []
- timeFrom: null
- timeRegions: []
- timeShift: null
- title: Messages stored per topic
- tooltip:
- shared: true
- sort: 0
- value_type: individual
- type: graph
- xaxis:
- buckets: null
- mode: time
- name: null
- show: true
- values: []
- yaxes:
- - format: short
- label: null
- logBase: 1
- max: null
- min: '0'
- show: true
- - format: short
- label: null
- logBase: 1
- max: null
- min: null
- show: true
- yaxis:
- align: false
- alignLevel: null
- - aliasColors: {}
- bars: false
- dashLength: 10
- dashes: false
- datasource: Prometheus
- fill: 0
- fillGradient: 0
- gridPos:
- h: 10
- w: 10
- x: 10
- 'y': 0
- id: 12
- legend:
- alignAsTable: true
- avg: false
- current: true
- max: true
- min: false
- rightSide: false
- show: true
- sideWidth: 480
- sort: max
- sortDesc: true
- total: false
- values: true
- lines: true
- linewidth: 1
- links: []
- nullPointMode: connected
- options:
- dataLinks: []
- percentage: false
- pointradius: 5
- points: false
- renderer: flot
- seriesOverrides: []
- spaceLength: 10
- stack: false
- steppedLine: false
- targets:
- - expr: >-
- sum(kafka_consumergroup_lag{instance="$instance",topic=~"$topic"}) by
- (consumergroup, topic)
- format: time_series
- instant: false
- interval: ''
- intervalFactor: 1
- legendFormat: ' {{topic}} ({{consumergroup}})'
- refId: A
- thresholds: []
- timeFrom: null
- timeRegions: []
- timeShift: null
- title: Lag by Consumer Group
- tooltip:
- shared: true
- sort: 2
- value_type: individual
- type: graph
- xaxis:
- buckets: null
- mode: time
- name: null
- show: true
- values: []
- yaxes:
- - format: short
- label: ''
- logBase: 1
- max: null
- min: '0'
- show: true
- - format: short
- label: null
- logBase: 1
- max: null
- min: null
- show: true
- yaxis:
- align: false
- alignLevel: null
- - aliasColors: {}
- bars: false
- dashLength: 10
- dashes: false
- datasource: Prometheus
- fill: 0
- fillGradient: 0
- gridPos:
- h: 10
- w: 10
- x: 0
- 'y': 10
- id: 16
- legend:
- alignAsTable: true
- avg: false
- current: true
- max: true
- min: false
- rightSide: false
- show: true
- sideWidth: 480
- total: false
- values: true
- lines: true
- linewidth: 1
- links: []
- nullPointMode: connected
- options:
- dataLinks: []
- percentage: false
- pointradius: 5
- points: false
- renderer: flot
- seriesOverrides: []
- spaceLength: 10
- stack: false
- steppedLine: false
- targets:
- - expr: >-
- sum(delta(kafka_topic_partition_current_offset{instance=~'$instance',
- topic=~"$topic"}[5m])/5) by (topic)
- format: time_series
- intervalFactor: 1
- legendFormat: '{{topic}}'
- refId: A
- thresholds: []
- timeFrom: null
- timeRegions: []
- timeShift: null
- title: Messages produced per minute
- tooltip:
- shared: true
- sort: 0
- value_type: individual
- type: graph
- xaxis:
- buckets: null
- mode: time
- name: null
- show: true
- values: []
- yaxes:
- - format: short
- label: null
- logBase: 1
- max: null
- min: null
- show: true
- - format: short
- label: null
- logBase: 1
- max: null
- min: null
- show: true
- yaxis:
- align: false
- alignLevel: null
- - aliasColors: {}
- bars: false
- dashLength: 10
- dashes: false
- datasource: Prometheus
- fill: 0
- fillGradient: 0
- gridPos:
- h: 10
- w: 10
- x: 10
- 'y': 10
- id: 18
- legend:
- alignAsTable: true
- avg: false
- current: true
- max: true
- min: false
- rightSide: false
- show: true
- sideWidth: 480
- sort: current
- sortDesc: true
- total: false
- values: true
- lines: true
- linewidth: 1
- links: []
- nullPointMode: connected
- options:
- dataLinks: []
- percentage: false
- pointradius: 5
- points: false
- renderer: flot
- seriesOverrides: []
- spaceLength: 10
- stack: false
- steppedLine: false
- targets:
- - expr: >-
- sum(delta(kafka_consumergroup_current_offset{instance=~'$instance',topic=~"$topic"}[5m])/5)
- by (consumergroup, topic)
- format: time_series
- intervalFactor: 1
- legendFormat: ' {{topic}} ({{consumergroup}})'
- refId: A
- thresholds: []
- timeFrom: null
- timeRegions: []
- timeShift: null
- title: Messages consumed per minute
- tooltip:
- shared: true
- sort: 0
- value_type: individual
- type: graph
- xaxis:
- buckets: null
- mode: time
- name: null
- show: true
- values: []
- yaxes:
- - format: short
- label: null
- logBase: 1
- max: null
- min: null
- show: true
- - format: short
- label: null
- logBase: 1
- max: null
- min: null
- show: true
- yaxis:
- align: false
- alignLevel: null
- - aliasColors: {}
- bars: true
- dashLength: 10
- dashes: false
- datasource: Prometheus
- fill: 1
- fillGradient: 0
- gridPos:
- h: 7
- w: 20
- x: 0
- 'y': 20
- id: 8
- legend:
- alignAsTable: true
- avg: false
- current: true
- max: false
- min: false
- rightSide: true
- show: true
- sideWidth: 420
- total: false
- values: true
- lines: false
- linewidth: 1
- links: []
- nullPointMode: 'null'
- options:
- dataLinks: []
- percentage: false
- pointradius: 5
- points: false
- renderer: flot
- seriesOverrides: []
- spaceLength: 10
- stack: false
- steppedLine: false
- targets:
- - expr: >-
- sum by(topic)
- (kafka_topic_partitions{instance="$instance",topic=~"$topic"})
- format: time_series
- intervalFactor: 1
- legendFormat: '{{topic}}'
- refId: A
- thresholds: []
- timeFrom: null
- timeRegions: []
- timeShift: null
- title: Partitions per Topic
- tooltip:
- shared: false
- sort: 0
- value_type: individual
- type: graph
- xaxis:
- buckets: null
- mode: series
- name: null
- show: false
- values:
- - current
- yaxes:
- - format: short
- label: null
- logBase: 1
- max: null
- min: null
- show: true
- - format: short
- label: null
- logBase: 1
- max: null
- min: null
- show: true
- yaxis:
- align: false
- alignLevel: null
-refresh: 5s
-schemaVersion: 19
-style: dark
-tags: []
-templating:
- list:
- - allValue: null
- current:
- text: osm-kafka-exporter-service
- value: osm-kafka-exporter-service
- datasource: Prometheus
- definition: ''
- hide: 0
- includeAll: false
- label: Job
- multi: false
- name: job
- options: []
- query: 'label_values(kafka_consumergroup_current_offset, job)'
- refresh: 1
- regex: ''
- skipUrlSync: false
- sort: 0
- tagValuesQuery: ''
- tags: []
- tagsQuery: ''
- type: query
- useTags: false
- - allValue: null
- datasource: Prometheus
- definition: ''
- hide: 0
- includeAll: false
- label: Instance
- multi: false
- name: instance
- options: []
- query: >-
- 'label_values(kafka_consumergroup_current_offset{job=~"$job"},'
- 'instance)'
- refresh: 1
- regex: ''
- skipUrlSync: false
- sort: 0
- tagValuesQuery: ''
- tags: []
- tagsQuery: ''
- type: query
- useTags: false
- - allValue: null
- current:
- tags: []
- text: All
- value:
- - $__all
- datasource: Prometheus
- definition: ''
- hide: 0
- includeAll: true
- label: Topic
- multi: true
- name: topic
- options: []
- query: >-
- label_values(kafka_topic_partition_current_offset{instance='$instance',topic!='__consumer_offsets',topic!='--kafka'},
- topic)
- refresh: 1
- regex: ''
- skipUrlSync: false
- sort: 1
- tagValuesQuery: ''
- tags: []
- tagsQuery: topic
- type: query
- useTags: false
-time:
- from: now-1h
- to: now
-timepicker:
- refresh_intervals:
- - 5s
- - 10s
- - 30s
- - 1m
- - 5m
- - 15m
- - 30m
- - 1h
- - 2h
- - 1d
- time_options:
- - 5m
- - 15m
- - 1h
- - 6h
- - 12h
- - 24h
- - 2d
- - 7d
- - 30d
-timezone: browser
-title: Kafka
-uid: jwPKIsniz
-version: 2
+++ /dev/null
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-name: prometheus-kafka-exporter
-summary: OSM Prometheus Kafka Exporter
-description: |
- A CAAS charm to deploy OSM's Prometheus Kafka Exporter.
-series:
- - kubernetes
-tags:
- - kubernetes
- - osm
- - prometheus
- - kafka-exporter
-min-juju-version: 2.8.0
-deployment:
- type: stateless
- service: cluster
-resources:
- image:
- type: oci-image
- description: Image of kafka-exporter
- upstream-source: "bitnami/kafka-exporter:latest"
-requires:
- kafka:
- interface: kafka
-provides:
- prometheus-target:
- interface: http
- grafana-dashboard:
- interface: grafana-dashboard
+++ /dev/null
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-ops
-git+https://github.com/juju-solutions/resource-oci-image/@c5778285d332edf3d9a538f9d0c06154b7ec1b0b#egg=oci-image
+++ /dev/null
-#!/usr/bin/env python3
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-import logging
-from pathlib import Path
-from typing import Dict, List, NoReturn
-
-from ops.charm import CharmBase
-from ops.framework import EventBase, StoredState
-from ops.main import main
-from ops.model import ActiveStatus, Application, BlockedStatus, MaintenanceStatus, Unit
-from oci_image import OCIImageResource, OCIImageResourceError
-
-from pod_spec import make_pod_spec
-
-logger = logging.getLogger(__name__)
-
-PROMETHEUS_KAFKA_EXPORTER_PORT = 9308
-
-
-class RelationsMissing(Exception):
- def __init__(self, missing_relations: List):
- self.message = ""
- if missing_relations and isinstance(missing_relations, list):
- self.message += f'Waiting for {", ".join(missing_relations)} relation'
- if "," in self.message:
- self.message += "s"
-
-
-class RelationDefinition:
- def __init__(self, relation_name: str, keys: List, source_type):
- if source_type != Application and source_type != Unit:
- raise TypeError(
- "source_type should be ops.model.Application or ops.model.Unit"
- )
- self.relation_name = relation_name
- self.keys = keys
- self.source_type = source_type
-
-
-def check_missing_relation_data(
- data: Dict,
- expected_relations_data: List[RelationDefinition],
-):
- missing_relations = []
- for relation_data in expected_relations_data:
- if not all(
- f"{relation_data.relation_name}_{k}" in data for k in relation_data.keys
- ):
- missing_relations.append(relation_data.relation_name)
- if missing_relations:
- raise RelationsMissing(missing_relations)
-
-
-def get_relation_data(
- charm: CharmBase,
- relation_data: RelationDefinition,
-) -> Dict:
- data = {}
- relation = charm.model.get_relation(relation_data.relation_name)
- if relation:
- self_app_unit = (
- charm.app if relation_data.source_type == Application else charm.unit
- )
- expected_type = relation_data.source_type
- for app_unit in relation.data:
- if app_unit != self_app_unit and isinstance(app_unit, expected_type):
- if all(k in relation.data[app_unit] for k in relation_data.keys):
- for k in relation_data.keys:
- data[f"{relation_data.relation_name}_{k}"] = relation.data[
- app_unit
- ].get(k)
- break
- return data
-
-
-class PrometheusKafkaExporterCharm(CharmBase):
- """Prometheus Kafka Exporter Charm."""
-
- state = StoredState()
-
- def __init__(self, *args) -> NoReturn:
- """Prometheus Kafka Exporter Charm constructor."""
- super().__init__(*args)
-
- # Internal state initialization
- self.state.set_default(pod_spec=None)
-
- self.port = PROMETHEUS_KAFKA_EXPORTER_PORT
- self.image = OCIImageResource(self, "image")
-
- # Registering regular events
- self.framework.observe(self.on.start, self.configure_pod)
- self.framework.observe(self.on.config_changed, self.configure_pod)
-
- # Registering required relation events
- self.framework.observe(self.on.kafka_relation_changed, self.configure_pod)
-
- # Registering required relation departed events
- self.framework.observe(self.on.kafka_relation_departed, self.configure_pod)
-
- # Registering provided relation events
- self.framework.observe(
- self.on.prometheus_target_relation_joined, self._publish_target_info
- )
- self.framework.observe(
- self.on.grafana_dashboard_relation_joined, self._publish_dashboard_info
- )
-
- def _publish_target_info(self, event: EventBase) -> NoReturn:
- """Publishes target information.
-
- Args:
- event (EventBase): Exporter relation event.
- """
- rel_data = {
- "hostname": self.model.app.name,
- "port": str(PROMETHEUS_KAFKA_EXPORTER_PORT),
- }
- for k, v in rel_data.items():
- event.relation.data[self.unit][k] = v
-
- def _publish_dashboard_info(self, event: EventBase) -> NoReturn:
- """Publishes dashboard information.
-
- Args:
- event (EventBase): Exporter relation event.
- """
- rel_data = {
- "dashboard": Path("files/kafka_exporter_dashboard.yaml").read_text(),
- }
- for k, v in rel_data.items():
- event.relation.data[self.unit][k] = v
-
- @property
- def relations_requirements(self):
- return [RelationDefinition("kafka", ["host", "port"], Unit)]
-
- def get_relation_state(self):
- relation_state = {}
- for relation_requirements in self.relations_requirements:
- data = get_relation_data(self, relation_requirements)
- relation_state = {**relation_state, **data}
- check_missing_relation_data(relation_state, self.relations_requirements)
- return relation_state
-
- def configure_pod(self, _=None) -> NoReturn:
- """Assemble the pod spec and apply it, if possible.
-
- Args:
- event (EventBase): Hook or Relation event that started the
- function.
- """
- if not self.unit.is_leader():
- self.unit.status = ActiveStatus("ready")
- return
-
- relation_state = None
- try:
- relation_state = self.get_relation_state()
- except RelationsMissing as exc:
- logger.exception("Relation missing error")
- self.unit.status = BlockedStatus(exc.message)
- return
-
- self.unit.status = MaintenanceStatus("Assembling pod spec")
-
- # Fetch image information
- try:
- self.unit.status = MaintenanceStatus("Fetching image information")
- image_info = self.image.fetch()
- except OCIImageResourceError:
- self.unit.status = BlockedStatus("Error fetching image information")
- return
-
- try:
- pod_spec = make_pod_spec(
- image_info,
- self.model.config,
- relation_state,
- self.model.app.name,
- self.port,
- )
- except ValueError as exc:
- logger.exception("Config/Relation data validation error")
- self.unit.status = BlockedStatus(str(exc))
- return
-
- if self.state.pod_spec != pod_spec:
- self.model.pod.set_spec(pod_spec)
- self.state.pod_spec = pod_spec
-
- self.unit.status = ActiveStatus("ready")
-
-
-if __name__ == "__main__":
- main(PrometheusKafkaExporterCharm)
+++ /dev/null
-#!/usr/bin/env python3
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-import logging
-from ipaddress import ip_network
-from typing import Any, Dict, List
-from urllib.parse import urlparse
-
-logger = logging.getLogger(__name__)
-
-
-def _validate_ip_network(network: str) -> bool:
- """Validate IP network.
-
- Args:
- network (str): IP network range.
-
- Returns:
- bool: True if valid, false otherwise.
- """
- if not network:
- return True
-
- try:
- ip_network(network)
- except ValueError:
- return False
-
- return True
-
-
-def _validate_data(config_data: Dict[str, Any], relation_data: Dict[str, Any]) -> bool:
- """Validates passed information.
-
- Args:
- config_data (Dict[str, Any]): configuration information.
- relation_data (Dict[str, Any]): relation information
-
- Raises:
- ValueError: when config and/or relation data is not valid.
- """
- config_validators = {
- "site_url": lambda value, _: isinstance(value, str)
- if value is not None
- else True,
- "ingress_whitelist_source_range": lambda value, _: _validate_ip_network(value),
- "tls_secret_name": lambda value, _: isinstance(value, str)
- if value is not None
- else True,
- }
- relation_validators = {
- "kafka_host": lambda value, _: isinstance(value, str) and len(value) > 0,
- "kafka_port": lambda value, _: isinstance(value, str)
- and len(value) > 0
- and int(value) > 0,
- }
- problems = []
-
- for key, validator in config_validators.items():
- valid = validator(config_data.get(key), config_data)
-
- if not valid:
- problems.append(key)
-
- for key, validator in relation_validators.items():
- valid = validator(relation_data.get(key), relation_data)
-
- if not valid:
- problems.append(key)
-
- if len(problems) > 0:
- raise ValueError("Errors found in: {}".format(", ".join(problems)))
-
- return True
-
-
-def _make_pod_ports(port: int) -> List[Dict[str, Any]]:
- """Generate pod ports details.
-
- Args:
- port (int): port to expose.
-
- Returns:
- List[Dict[str, Any]]: pod port details.
- """
- return [
- {"name": "prometheus-kafka-exporter", "containerPort": port, "protocol": "TCP"}
- ]
-
-
-def _make_pod_envconfig(
- config: Dict[str, Any], relation_state: Dict[str, Any]
-) -> Dict[str, Any]:
- """Generate pod environment configuration.
-
- Args:
- config (Dict[str, Any]): configuration information.
- relation_state (Dict[str, Any]): relation state information.
-
- Returns:
- Dict[str, Any]: pod environment configuration.
- """
- envconfig = {}
-
- return envconfig
-
-
-def _make_pod_ingress_resources(
- config: Dict[str, Any], app_name: str, port: int
-) -> List[Dict[str, Any]]:
- """Generate pod ingress resources.
-
- Args:
- config (Dict[str, Any]): configuration information.
- app_name (str): application name.
- port (int): port to expose.
-
- Returns:
- List[Dict[str, Any]]: pod ingress resources.
- """
- site_url = config.get("site_url")
-
- if not site_url:
- return
-
- parsed = urlparse(site_url)
-
- if not parsed.scheme.startswith("http"):
- return
-
- ingress_whitelist_source_range = config["ingress_whitelist_source_range"]
-
- annotations = {}
-
- if ingress_whitelist_source_range:
- annotations[
- "nginx.ingress.kubernetes.io/whitelist-source-range"
- ] = ingress_whitelist_source_range
-
- ingress_spec_tls = None
-
- if parsed.scheme == "https":
- ingress_spec_tls = [{"hosts": [parsed.hostname]}]
- tls_secret_name = config["tls_secret_name"]
- if tls_secret_name:
- ingress_spec_tls[0]["secretName"] = tls_secret_name
- else:
- annotations["nginx.ingress.kubernetes.io/ssl-redirect"] = "false"
-
- ingress = {
- "name": "{}-ingress".format(app_name),
- "annotations": annotations,
- "spec": {
- "rules": [
- {
- "host": parsed.hostname,
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": app_name,
- "servicePort": port,
- },
- }
- ]
- },
- }
- ]
- },
- }
- if ingress_spec_tls:
- ingress["spec"]["tls"] = ingress_spec_tls
-
- return [ingress]
-
-
-def _make_readiness_probe(port: int) -> Dict[str, Any]:
- """Generate readiness probe.
-
- Args:
- port (int): service port.
-
- Returns:
- Dict[str, Any]: readiness probe.
- """
- return {
- "httpGet": {
- "path": "/api/health",
- "port": port,
- },
- "initialDelaySeconds": 10,
- "periodSeconds": 10,
- "timeoutSeconds": 5,
- "successThreshold": 1,
- "failureThreshold": 3,
- }
-
-
-def _make_liveness_probe(port: int) -> Dict[str, Any]:
- """Generate liveness probe.
-
- Args:
- port (int): service port.
-
- Returns:
- Dict[str, Any]: liveness probe.
- """
- return {
- "httpGet": {
- "path": "/api/health",
- "port": port,
- },
- "initialDelaySeconds": 60,
- "timeoutSeconds": 30,
- "failureThreshold": 10,
- }
-
-
-def _make_pod_command(relation: Dict[str, Any]) -> List[str]:
- """Generate the startup command.
-
- Args:
- relation (Dict[str, Any]): Relation information.
-
- Returns:
- List[str]: command to startup the process.
- """
- command = [
- "kafka-exporter",
- "--kafka.server={}:{}".format(
- relation.get("kafka_host"), relation.get("kafka_port")
- ),
- ]
-
- return command
-
-
-def make_pod_spec(
- image_info: Dict[str, str],
- config: Dict[str, Any],
- relation_state: Dict[str, Any],
- app_name: str = "prometheus-kafka-exporter",
- port: int = 9308,
-) -> Dict[str, Any]:
- """Generate the pod spec information.
-
- Args:
- image_info (Dict[str, str]): Object provided by
- OCIImageResource("image").fetch().
- config (Dict[str, Any]): Configuration information.
- relation_state (Dict[str, Any]): Relation state information.
- app_name (str, optional): Application name. Defaults to "ro".
- port (int, optional): Port for the container. Defaults to 9090.
-
- Returns:
- Dict[str, Any]: Pod spec dictionary for the charm.
- """
- if not image_info:
- return None
-
- _validate_data(config, relation_state)
-
- ports = _make_pod_ports(port)
- env_config = _make_pod_envconfig(config, relation_state)
- readiness_probe = _make_readiness_probe(port)
- liveness_probe = _make_liveness_probe(port)
- ingress_resources = _make_pod_ingress_resources(config, app_name, port)
- command = _make_pod_command(relation_state)
-
- return {
- "version": 3,
- "containers": [
- {
- "name": app_name,
- "imageDetails": image_info,
- "imagePullPolicy": "Always",
- "ports": ports,
- "envConfig": env_config,
- "command": command,
- "kubernetes": {
- "readinessProbe": readiness_probe,
- "livenessProbe": liveness_probe,
- },
- }
- ],
- "kubernetesResources": {
- "ingressResources": ingress_resources or [],
- },
- }
+++ /dev/null
-#!/usr/bin/env python3
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-"""Init mocking for unit tests."""
-
-import sys
-import mock
-
-sys.path.append("src")
-
-oci_image = mock.MagicMock()
-sys.modules["oci_image"] = oci_image
+++ /dev/null
-#!/usr/bin/env python3
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-from typing import NoReturn
-import unittest
-
-from ops.model import BlockedStatus
-from ops.testing import Harness
-
-from charm import PrometheusKafkaExporterCharm
-
-
-class TestCharm(unittest.TestCase):
- """Prometheus Kafka Exporter Charm unit tests."""
-
- def setUp(self) -> NoReturn:
- """Test setup"""
- self.harness = Harness(PrometheusKafkaExporterCharm)
- self.harness.set_leader(is_leader=True)
- self.harness.begin()
-
- def test_on_start_without_relations(self) -> NoReturn:
- """Test installation without any relation."""
- self.harness.charm.on.start.emit()
-
- # Verifying status
- self.assertIsInstance(self.harness.charm.unit.status, BlockedStatus)
-
- # Verifying status message
- self.assertGreater(len(self.harness.charm.unit.status.message), 0)
- self.assertTrue(
- self.harness.charm.unit.status.message.startswith("Waiting for ")
- )
- self.assertIn("kafka", self.harness.charm.unit.status.message)
- self.assertTrue(self.harness.charm.unit.status.message.endswith(" relation"))
-
- def test_on_start_with_relations_without_http(self) -> NoReturn:
- """Test deployment."""
- expected_result = {
- "version": 3,
- "containers": [
- {
- "name": "prometheus-kafka-exporter",
- "imageDetails": self.harness.charm.image.fetch(),
- "imagePullPolicy": "Always",
- "ports": [
- {
- "name": "prometheus-kafka-exporter",
- "containerPort": 9308,
- "protocol": "TCP",
- }
- ],
- "envConfig": {},
- "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
- "kubernetes": {
- "readinessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": 9308,
- },
- "initialDelaySeconds": 10,
- "periodSeconds": 10,
- "timeoutSeconds": 5,
- "successThreshold": 1,
- "failureThreshold": 3,
- },
- "livenessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": 9308,
- },
- "initialDelaySeconds": 60,
- "timeoutSeconds": 30,
- "failureThreshold": 10,
- },
- },
- },
- ],
- "kubernetesResources": {"ingressResources": []},
- }
-
- self.harness.charm.on.start.emit()
-
- # Initializing the kafka relation
- relation_id = self.harness.add_relation("kafka", "kafka")
- self.harness.add_relation_unit(relation_id, "kafka/0")
- self.harness.update_relation_data(
- relation_id,
- "kafka/0",
- {
- "host": "kafka",
- "port": "9090",
- },
- )
-
- # Verifying status
- self.assertNotIsInstance(self.harness.charm.unit.status, BlockedStatus)
-
- pod_spec, _ = self.harness.get_pod_spec()
-
- self.assertDictEqual(expected_result, pod_spec)
-
- def test_ingress_resources_with_http(self) -> NoReturn:
- """Test ingress resources with HTTP."""
- expected_result = {
- "version": 3,
- "containers": [
- {
- "name": "prometheus-kafka-exporter",
- "imageDetails": self.harness.charm.image.fetch(),
- "imagePullPolicy": "Always",
- "ports": [
- {
- "name": "prometheus-kafka-exporter",
- "containerPort": 9308,
- "protocol": "TCP",
- }
- ],
- "envConfig": {},
- "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
- "kubernetes": {
- "readinessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": 9308,
- },
- "initialDelaySeconds": 10,
- "periodSeconds": 10,
- "timeoutSeconds": 5,
- "successThreshold": 1,
- "failureThreshold": 3,
- },
- "livenessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": 9308,
- },
- "initialDelaySeconds": 60,
- "timeoutSeconds": 30,
- "failureThreshold": 10,
- },
- },
- },
- ],
- "kubernetesResources": {
- "ingressResources": [
- {
- "name": "prometheus-kafka-exporter-ingress",
- "annotations": {
- "nginx.ingress.kubernetes.io/ssl-redirect": "false",
- },
- "spec": {
- "rules": [
- {
- "host": "prometheus-kafka-exporter",
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": "prometheus-kafka-exporter",
- "servicePort": 9308,
- },
- }
- ]
- },
- }
- ]
- },
- }
- ],
- },
- }
-
- self.harness.charm.on.start.emit()
-
- # Initializing the kafka relation
- relation_id = self.harness.add_relation("kafka", "kafka")
- self.harness.add_relation_unit(relation_id, "kafka/0")
- self.harness.update_relation_data(
- relation_id,
- "kafka/0",
- {
- "host": "kafka",
- "port": "9090",
- },
- )
-
- self.harness.update_config({"site_url": "http://prometheus-kafka-exporter"})
-
- pod_spec, _ = self.harness.get_pod_spec()
-
- self.assertDictEqual(expected_result, pod_spec)
-
- def test_ingress_resources_with_https(self) -> NoReturn:
- """Test ingress resources with HTTPS."""
- expected_result = {
- "version": 3,
- "containers": [
- {
- "name": "prometheus-kafka-exporter",
- "imageDetails": self.harness.charm.image.fetch(),
- "imagePullPolicy": "Always",
- "ports": [
- {
- "name": "prometheus-kafka-exporter",
- "containerPort": 9308,
- "protocol": "TCP",
- }
- ],
- "envConfig": {},
- "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
- "kubernetes": {
- "readinessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": 9308,
- },
- "initialDelaySeconds": 10,
- "periodSeconds": 10,
- "timeoutSeconds": 5,
- "successThreshold": 1,
- "failureThreshold": 3,
- },
- "livenessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": 9308,
- },
- "initialDelaySeconds": 60,
- "timeoutSeconds": 30,
- "failureThreshold": 10,
- },
- },
- },
- ],
- "kubernetesResources": {
- "ingressResources": [
- {
- "name": "prometheus-kafka-exporter-ingress",
- "annotations": {},
- "spec": {
- "rules": [
- {
- "host": "prometheus-kafka-exporter",
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": "prometheus-kafka-exporter",
- "servicePort": 9308,
- },
- }
- ]
- },
- }
- ],
- "tls": [
- {
- "hosts": ["prometheus-kafka-exporter"],
- "secretName": "prometheus-kafka-exporter",
- }
- ],
- },
- }
- ],
- },
- }
-
- self.harness.charm.on.start.emit()
-
- # Initializing the kafka relation
- relation_id = self.harness.add_relation("kafka", "kafka")
- self.harness.add_relation_unit(relation_id, "kafka/0")
- self.harness.update_relation_data(
- relation_id,
- "kafka/0",
- {
- "host": "kafka",
- "port": "9090",
- },
- )
-
- self.harness.update_config(
- {
- "site_url": "https://prometheus-kafka-exporter",
- "tls_secret_name": "prometheus-kafka-exporter",
- }
- )
-
- pod_spec, _ = self.harness.get_pod_spec()
-
- self.assertDictEqual(expected_result, pod_spec)
-
- def test_ingress_resources_with_https_and_ingress_whitelist(self) -> NoReturn:
- """Test ingress resources with HTTPS and ingress whitelist."""
- expected_result = {
- "version": 3,
- "containers": [
- {
- "name": "prometheus-kafka-exporter",
- "imageDetails": self.harness.charm.image.fetch(),
- "imagePullPolicy": "Always",
- "ports": [
- {
- "name": "prometheus-kafka-exporter",
- "containerPort": 9308,
- "protocol": "TCP",
- }
- ],
- "envConfig": {},
- "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
- "kubernetes": {
- "readinessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": 9308,
- },
- "initialDelaySeconds": 10,
- "periodSeconds": 10,
- "timeoutSeconds": 5,
- "successThreshold": 1,
- "failureThreshold": 3,
- },
- "livenessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": 9308,
- },
- "initialDelaySeconds": 60,
- "timeoutSeconds": 30,
- "failureThreshold": 10,
- },
- },
- },
- ],
- "kubernetesResources": {
- "ingressResources": [
- {
- "name": "prometheus-kafka-exporter-ingress",
- "annotations": {
- "nginx.ingress.kubernetes.io/whitelist-source-range": "0.0.0.0/0",
- },
- "spec": {
- "rules": [
- {
- "host": "prometheus-kafka-exporter",
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": "prometheus-kafka-exporter",
- "servicePort": 9308,
- },
- }
- ]
- },
- }
- ],
- "tls": [
- {
- "hosts": ["prometheus-kafka-exporter"],
- "secretName": "prometheus-kafka-exporter",
- }
- ],
- },
- }
- ],
- },
- }
-
- self.harness.charm.on.start.emit()
-
- # Initializing the kafka relation
- relation_id = self.harness.add_relation("kafka", "kafka")
- self.harness.add_relation_unit(relation_id, "kafka/0")
- self.harness.update_relation_data(
- relation_id,
- "kafka/0",
- {
- "host": "kafka",
- "port": "9090",
- },
- )
-
- self.harness.update_config(
- {
- "site_url": "https://prometheus-kafka-exporter",
- "tls_secret_name": "prometheus-kafka-exporter",
- "ingress_whitelist_source_range": "0.0.0.0/0",
- }
- )
-
- pod_spec, _ = self.harness.get_pod_spec()
-
- self.assertDictEqual(expected_result, pod_spec)
-
- def test_on_kafka_unit_relation_changed(self) -> NoReturn:
- """Test to see if kafka relation is updated."""
- self.harness.charm.on.start.emit()
-
- relation_id = self.harness.add_relation("kafka", "kafka")
- self.harness.add_relation_unit(relation_id, "kafka/0")
- self.harness.update_relation_data(
- relation_id,
- "kafka/0",
- {
- "host": "kafka",
- "port": "9090",
- },
- )
-
- # Verifying status
- self.assertNotIsInstance(self.harness.charm.unit.status, BlockedStatus)
-
- def test_publish_target_info(self) -> NoReturn:
- """Test to see if target relation is updated."""
- expected_result = {
- "hostname": "prometheus-kafka-exporter",
- "port": "9308",
- }
-
- self.harness.charm.on.start.emit()
-
- relation_id = self.harness.add_relation("prometheus-target", "prometheus")
- self.harness.add_relation_unit(relation_id, "prometheus/0")
- relation_data = self.harness.get_relation_data(
- relation_id, "prometheus-kafka-exporter/0"
- )
-
- self.assertDictEqual(expected_result, relation_data)
-
- def test_publish_dashboard_info(self) -> NoReturn:
- """Test to see if dashboard relation is updated."""
- self.harness.charm.on.start.emit()
-
- relation_id = self.harness.add_relation("grafana-dashboard", "grafana")
- self.harness.add_relation_unit(relation_id, "grafana/0")
- relation_data = self.harness.get_relation_data(
- relation_id, "prometheus-kafka-exporter/0"
- )
-
- self.assertTrue("dashboard" in relation_data)
- self.assertTrue(len(relation_data["dashboard"]) > 0)
-
-
-if __name__ == "__main__":
- unittest.main()
+++ /dev/null
-#!/usr/bin/env python3
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-from typing import NoReturn
-import unittest
-
-import pod_spec
-
-
-class TestPodSpec(unittest.TestCase):
- """Pod spec unit tests."""
-
- def test_make_pod_ports(self) -> NoReturn:
- """Testing make pod ports."""
- port = 9308
-
- expected_result = [
- {
- "name": "prometheus-kafka-exporter",
- "containerPort": port,
- "protocol": "TCP",
- }
- ]
-
- pod_ports = pod_spec._make_pod_ports(port)
-
- self.assertListEqual(expected_result, pod_ports)
-
- def test_make_pod_envconfig(self) -> NoReturn:
- """Teting make pod envconfig."""
- config = {}
- relation_state = {}
-
- expected_result = {}
-
- pod_envconfig = pod_spec._make_pod_envconfig(config, relation_state)
-
- self.assertDictEqual(expected_result, pod_envconfig)
-
- def test_make_pod_ingress_resources_without_site_url(self) -> NoReturn:
- """Testing make pod ingress resources without site_url."""
- config = {"site_url": ""}
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- pod_ingress_resources = pod_spec._make_pod_ingress_resources(
- config, app_name, port
- )
-
- self.assertIsNone(pod_ingress_resources)
-
- def test_make_pod_ingress_resources(self) -> NoReturn:
- """Testing make pod ingress resources."""
- config = {
- "site_url": "http://prometheus-kafka-exporter",
- "ingress_whitelist_source_range": "",
- }
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- expected_result = [
- {
- "name": f"{app_name}-ingress",
- "annotations": {
- "nginx.ingress.kubernetes.io/ssl-redirect": "false",
- },
- "spec": {
- "rules": [
- {
- "host": app_name,
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": app_name,
- "servicePort": port,
- },
- }
- ]
- },
- }
- ]
- },
- }
- ]
-
- pod_ingress_resources = pod_spec._make_pod_ingress_resources(
- config, app_name, port
- )
-
- self.assertListEqual(expected_result, pod_ingress_resources)
-
- def test_make_pod_ingress_resources_with_whitelist_source_range(self) -> NoReturn:
- """Testing make pod ingress resources with whitelist_source_range."""
- config = {
- "site_url": "http://prometheus-kafka-exporter",
- "ingress_whitelist_source_range": "0.0.0.0/0",
- }
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- expected_result = [
- {
- "name": f"{app_name}-ingress",
- "annotations": {
- "nginx.ingress.kubernetes.io/ssl-redirect": "false",
- "nginx.ingress.kubernetes.io/whitelist-source-range": config[
- "ingress_whitelist_source_range"
- ],
- },
- "spec": {
- "rules": [
- {
- "host": app_name,
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": app_name,
- "servicePort": port,
- },
- }
- ]
- },
- }
- ]
- },
- }
- ]
-
- pod_ingress_resources = pod_spec._make_pod_ingress_resources(
- config, app_name, port
- )
-
- self.assertListEqual(expected_result, pod_ingress_resources)
-
- def test_make_pod_ingress_resources_with_https(self) -> NoReturn:
- """Testing make pod ingress resources with HTTPs."""
- config = {
- "site_url": "https://prometheus-kafka-exporter",
- "max_file_size": 0,
- "ingress_whitelist_source_range": "",
- "tls_secret_name": "",
- }
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- expected_result = [
- {
- "name": f"{app_name}-ingress",
- "annotations": {},
- "spec": {
- "rules": [
- {
- "host": app_name,
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": app_name,
- "servicePort": port,
- },
- }
- ]
- },
- }
- ],
- "tls": [{"hosts": [app_name]}],
- },
- }
- ]
-
- pod_ingress_resources = pod_spec._make_pod_ingress_resources(
- config, app_name, port
- )
-
- self.assertListEqual(expected_result, pod_ingress_resources)
-
- def test_make_pod_ingress_resources_with_https_tls_secret_name(self) -> NoReturn:
- """Testing make pod ingress resources with HTTPs and TLS secret name."""
- config = {
- "site_url": "https://prometheus-kafka-exporter",
- "max_file_size": 0,
- "ingress_whitelist_source_range": "",
- "tls_secret_name": "secret_name",
- }
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- expected_result = [
- {
- "name": f"{app_name}-ingress",
- "annotations": {},
- "spec": {
- "rules": [
- {
- "host": app_name,
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": app_name,
- "servicePort": port,
- },
- }
- ]
- },
- }
- ],
- "tls": [
- {"hosts": [app_name], "secretName": config["tls_secret_name"]}
- ],
- },
- }
- ]
-
- pod_ingress_resources = pod_spec._make_pod_ingress_resources(
- config, app_name, port
- )
-
- self.assertListEqual(expected_result, pod_ingress_resources)
-
- def test_make_readiness_probe(self) -> NoReturn:
- """Testing make readiness probe."""
- port = 9308
-
- expected_result = {
- "httpGet": {
- "path": "/api/health",
- "port": port,
- },
- "initialDelaySeconds": 10,
- "periodSeconds": 10,
- "timeoutSeconds": 5,
- "successThreshold": 1,
- "failureThreshold": 3,
- }
-
- readiness_probe = pod_spec._make_readiness_probe(port)
-
- self.assertDictEqual(expected_result, readiness_probe)
-
- def test_make_liveness_probe(self) -> NoReturn:
- """Testing make liveness probe."""
- port = 9308
-
- expected_result = {
- "httpGet": {
- "path": "/api/health",
- "port": port,
- },
- "initialDelaySeconds": 60,
- "timeoutSeconds": 30,
- "failureThreshold": 10,
- }
-
- liveness_probe = pod_spec._make_liveness_probe(port)
-
- self.assertDictEqual(expected_result, liveness_probe)
-
- def test_make_pod_command(self) -> NoReturn:
- """Testing make pod command."""
- relation = {
- "kakfa_host": "kafka",
- "kafka_port": "9090",
- }
-
- expected_result = [
- "kafka-exporter",
- "--kafka.server={}:{}".format(
- relation.get("kafka_host"), relation.get("kafka_port")
- ),
- ]
-
- pod_envconfig = pod_spec._make_pod_command(relation)
-
- self.assertListEqual(expected_result, pod_envconfig)
-
- def test_make_pod_spec(self) -> NoReturn:
- """Testing make pod spec."""
- image_info = {"upstream-source": "bitnami/kafka-exporter:latest"}
- config = {
- "site_url": "",
- }
- relation_state = {
- "kafka_host": "kafka",
- "kafka_port": "9090",
- }
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- expected_result = {
- "version": 3,
- "containers": [
- {
- "name": app_name,
- "imageDetails": image_info,
- "imagePullPolicy": "Always",
- "ports": [
- {
- "name": app_name,
- "containerPort": port,
- "protocol": "TCP",
- }
- ],
- "envConfig": {},
- "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
- "kubernetes": {
- "readinessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": port,
- },
- "initialDelaySeconds": 10,
- "periodSeconds": 10,
- "timeoutSeconds": 5,
- "successThreshold": 1,
- "failureThreshold": 3,
- },
- "livenessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": port,
- },
- "initialDelaySeconds": 60,
- "timeoutSeconds": 30,
- "failureThreshold": 10,
- },
- },
- }
- ],
- "kubernetesResources": {"ingressResources": []},
- }
-
- spec = pod_spec.make_pod_spec(
- image_info, config, relation_state, app_name, port
- )
-
- self.assertDictEqual(expected_result, spec)
-
- def test_make_pod_spec_with_ingress(self) -> NoReturn:
- """Testing make pod spec."""
- image_info = {"upstream-source": "bitnami/kafka-exporter:latest"}
- config = {
- "site_url": "https://prometheus-kafka-exporter",
- "tls_secret_name": "prometheus-kafka-exporter",
- "max_file_size": 0,
- "ingress_whitelist_source_range": "0.0.0.0/0",
- }
- relation_state = {
- "kafka_host": "kafka",
- "kafka_port": "9090",
- }
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- expected_result = {
- "version": 3,
- "containers": [
- {
- "name": app_name,
- "imageDetails": image_info,
- "imagePullPolicy": "Always",
- "ports": [
- {
- "name": app_name,
- "containerPort": port,
- "protocol": "TCP",
- }
- ],
- "envConfig": {},
- "command": ["kafka-exporter", "--kafka.server=kafka:9090"],
- "kubernetes": {
- "readinessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": port,
- },
- "initialDelaySeconds": 10,
- "periodSeconds": 10,
- "timeoutSeconds": 5,
- "successThreshold": 1,
- "failureThreshold": 3,
- },
- "livenessProbe": {
- "httpGet": {
- "path": "/api/health",
- "port": port,
- },
- "initialDelaySeconds": 60,
- "timeoutSeconds": 30,
- "failureThreshold": 10,
- },
- },
- }
- ],
- "kubernetesResources": {
- "ingressResources": [
- {
- "name": "{}-ingress".format(app_name),
- "annotations": {
- "nginx.ingress.kubernetes.io/whitelist-source-range": config.get(
- "ingress_whitelist_source_range"
- ),
- },
- "spec": {
- "rules": [
- {
- "host": app_name,
- "http": {
- "paths": [
- {
- "path": "/",
- "backend": {
- "serviceName": app_name,
- "servicePort": port,
- },
- }
- ]
- },
- }
- ],
- "tls": [
- {
- "hosts": [app_name],
- "secretName": config.get("tls_secret_name"),
- }
- ],
- },
- }
- ],
- },
- }
-
- spec = pod_spec.make_pod_spec(
- image_info, config, relation_state, app_name, port
- )
-
- self.assertDictEqual(expected_result, spec)
-
- def test_make_pod_spec_without_image_info(self) -> NoReturn:
- """Testing make pod spec without image_info."""
- image_info = None
- config = {
- "site_url": "",
- }
- relation_state = {
- "kafka_host": "kafka",
- "kafka_port": "9090",
- }
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- spec = pod_spec.make_pod_spec(
- image_info, config, relation_state, app_name, port
- )
-
- self.assertIsNone(spec)
-
- def test_make_pod_spec_without_relation_state(self) -> NoReturn:
- """Testing make pod spec without relation_state."""
- image_info = {"upstream-source": "bitnami/kafka-exporter:latest"}
- config = {
- "site_url": "",
- }
- relation_state = {}
- app_name = "prometheus-kafka-exporter"
- port = 9308
-
- with self.assertRaises(ValueError):
- pod_spec.make_pod_spec(image_info, config, relation_state, app_name, port)
-
-
-if __name__ == "__main__":
- unittest.main()
+++ /dev/null
-# Copyright 2021 Canonical Ltd.
-#
-# Licensed under the Apache License, Version 2.0 (the "License"); you may
-# not use this file except in compliance with the License. You may obtain
-# a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
-# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
-# License for the specific language governing permissions and limitations
-# under the License.
-#
-# For those usages not covered by the Apache License, Version 2.0 please
-# contact: legal@canonical.com
-#
-# To get in touch with the maintainers, please contact:
-# osm-charmers@lists.launchpad.net
-##
-
-[tox]
-skipsdist = True
-envlist = unit, lint
-sitepackages = False
-skip_missing_interpreters = False
-
-[testenv]
-basepython = python3
-setenv =
- PYTHONHASHSEED=0
- PYTHONPATH = {toxinidir}/src
- CHARM_NAME = prometheus-kafka-exporter
-
-[testenv:build]
-passenv=HTTP_PROXY HTTPS_PROXY NO_PROXY
-whitelist_externals =
- charmcraft
- rm
- unzip
-commands =
- rm -rf release grafana.charm
- charmcraft build
- unzip prometheus-kafka-exporter.charm -d release
-
-[testenv:unit]
-commands =
- coverage erase
- stestr run --slowest --test-path=./tests --top-dir=./
- coverage combine
- coverage html -d cover
- coverage xml -o cover/coverage.xml
- coverage report
-deps =
- coverage
- stestr
- mock
- ops
-setenv =
- {[testenv]setenv}
- PYTHON=coverage run
-
-[testenv:lint]
-deps =
- black
- yamllint
- flake8
-commands =
- black --check --diff . --exclude "build/|.tox/|mod/|lib/"
- yamllint .
- flake8 . --max-line-length=100 --ignore="E501,W503,W504,F722" --exclude "build/ .tox/ mod/ lib/"
-
-[coverage:run]
-branch = True
-concurrency = multiprocessing
-parallel = True
-source =
- .
-omit =
- .tox/*
- tests/*