X-Git-Url: https://osm.etsi.org/gitweb/?a=blobdiff_plain;f=installers%2Fhelm%2Fosm%2Fvalues.yaml;h=5bfc94ef90d0490f0dc277bfb92edf2b5fa6e44d;hb=7870d172322963e22c1cf04558768d3d8103fffa;hp=65c4406aa7588c8d763f76d6b6b96c14012aeb8a;hpb=8080e4b32d3dc5e66ea85a503a4daa79df3098d0;p=osm%2Fdevops.git diff --git a/installers/helm/osm/values.yaml b/installers/helm/osm/values.yaml index 65c4406a..5bfc94ef 100644 --- a/installers/helm/osm/values.yaml +++ b/installers/helm/osm/values.yaml @@ -20,12 +20,13 @@ global: replicaCount: 1 + logLevel: INFO image: repositoryBase: docker.io/opensourcemano pullPolicy: IfNotPresent # Overrides the image tag whose default is the chart appVersion. - tag: "13" + tag: "testing-daily" imagePullSecrets: [] nameOverride: "" @@ -64,16 +65,111 @@ global: oldServiceAssurance: false +airflow: + enabled: true + certauth: enabled: true grafana: enabled: true + fullnameOverride: "grafana" + service: + type: NodePort + nodePort: 3000 + port: 3000 + extraVolumes: + - emptyDir: {} + name: sc-dashboard-volume-k8s + - emptyDir: {} + name: sc-dashboard-volume-osm + extraVolumeMounts: + - name: sc-dashboard-volume-k8s + mountPath: "/tmp/dashboards/Kubernetes Cluster" + - name: sc-dashboard-volume-osm + mountPath: "/tmp/dashboards/Open Source MANO" + rbac: + extraClusterRoleRules: + - apiGroups: + - "" + resources: + - configmaps + - secrets + verbs: + - get + - watch + - list + datasources: + datasource.yaml: + apiVersion: 1 + datasources: + - name: osm_prometheus + type: prometheus + url: http://prometheus:9090 + access: proxy + allowUiUpdates: true + isDefault: true + - name: Prometheus + type: prometheus + url: http://osm-monitoring-kube-promet-prometheus.monitoring:9090 + access: proxy + allowUiUpdates: true + isDefault: false + dashboardProviders: + provider.yaml: + apiVersion: 1 + providers: + - name: 'Kubernetes Cluster' + orgId: 1 + folder: 'Kubernetes Cluster' + type: file + disableDeletion: false + options: + path: '/tmp/dashboards/Kubernetes Cluster' + - name: 'Open Source MANO' + orgId: 1 + folder: 'Open Source MANO' + type: file + disableDeletion: false + options: + path: '/tmp/dashboards/Open Source MANO' + sidecar: + dashboards: + enabled: true + folder: "/tmp/dashboards/" + resource: both + searchNamespace: osm + extraMounts: + - name: sc-dashboard-volume-k8s + mountPath: "/tmp/dashboards/Kubernetes Cluster" + - name: sc-dashboard-volume-osm + mountPath: "/tmp/dashboards/Open Source MANO" + extraInitContainers: + - env: + - name: METHOD + value: LIST + - name: LABEL + value: grafana_datasource + - name: FOLDER + value: /etc/grafana/provisioning/datasources + - name: RESOURCE + value: both + image: kiwigrid/k8s-sidecar:1.15.6 + imagePullPolicy: IfNotPresent + name: grafana-sc-datasources + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/grafana/provisioning/datasources + name: config kafka: enabled: true - service: - port: 9092 + listeners: + client: + protocol: "PLAINTEXT" + fullnameOverride: "kafka" # replicaCount: 1 keystone: @@ -82,16 +178,17 @@ keystone: port: 5000 image: {} # repository: opensourcemano/keystone - # tag: "13" + # tag: "testing-daily" # replicaCount: 1 useOsmSecret: true # secretName: "keystone-secret" lcm: enabled: true + logLevel: DEBUG image: {} # repository: opensourcemano/lcm - # tag: "13" + # tag: "testing-daily" # replicaCount: 1 useOsmSecret: true # secretName: "lcm-secret" @@ -104,35 +201,31 @@ lcm: mon: enabled: true - service: - port: 8662 + # logLevel: DEBUG image: {} # repository: opensourcemano/mon - # tag: "13" + # tag: "testing-daily" # replicaCount: 1 useOsmSecret: true # secretName: "mon-secret" config: {} - # OS_NOTIFIER_URI: "http://DEFAULT_IP:8662" mysql: enabled: true - # dbHostPath: "/var/lib/osm/osm" - service: - port: 3306 - # replicaCount: 1 - useOsmSecret: true - # secretName: "ro-db-secret" + image: + tag: "8.1-debian-11" + fullnameOverride: "mysql" nbi: enabled: true + logLevel: DEBUG service: type: NodePort port: 9999 nodePort: 9999 image: {} # repository: opensourcemano/nbi - # tag: "13" + # tag: "testing-daily" # replicaCount: 1 useOsmSecret: true # secretName: "nbi-secret" @@ -145,50 +238,255 @@ ngui: nodePort: 80 image: {} # repository: opensourcemano/ng-ui - # tag: "13" + # tag: "testing-daily" # replicaCount: 1 # pla module is disabled by default unless global.oldServiceAssurance and pla.enabled are set to true pla: enabled: false + # logLevel: DEBUG image: {} # repository: opensourcemano/pla - # tag: "13" + # tag: "testing-daily" # replicaCount: 1 # pol module is disabled by default unless global.oldServiceAssurance and pol.enabled are set to true pol: enabled: true + # logLevel: DEBUG image: {} # repository: opensourcemano/pol - # tag: "13" + # tag: "testing-daily" # replicaCount: 1 useOsmSecret: true # secretName: "pol-secret" prometheus: enabled: true - service: - type: NodePort - nodePort: 9091 - # replicaCount: 1 - sidecarImage: {} - # repository: opensourcemano/ro - # tag: "13" + configmapReload: + prometheus: + enabled: true + prometheus-node-exporter: + enabled: false + kube-state-metrics: + enabled: false + server: + extraFlags: + - web.enable-lifecycle + statefulSet: + enabled: true + configPath: /etc/prometheus/prometheus.yml + fullnameOverride: "prometheus" + service: + servicePort: 9090 + type: NodePort + nodePort: 9091 + sidecarContainers: + prometheus-config-sidecar: + securityContext: + # readOnlyRootFilesystem: true + allowPrivilegeEscalation: false + # runAsNonRoot: true + image: opensourcemano/prometheus:testing-daily + imagePullPolicy: IfNotPresent + command: ["/bin/sh", "-c"] + args: ["sleep 50; python -u ./app.py"] + volumeMounts: + - name: prom-config + mountPath: /etc/prometheus + - name: prom-config-base + mountPath: /etc/prometheus_base + resources: + limits: + memory: 1024Mi + requests: + memory: 128Mi + envFrom: + - configMapRef: + name: osm-prometheus-sidecar-configmap + extraInitContainers: + - name: prometheus-init-config + image: busybox + command: ["/bin/sh", "-c"] + #args: [' sleep 100000 '] + args: ['if [ ! -f "/etc/prometheus/prometheus.yml" ]; then cp /config/prometheus.yml /etc/prometheus; fi; cp /config/osm_metric_rules.yml /config/osm_alert_rules.yml /etc/prometheus'] + volumeMounts: + - name: prom-config + mountPath: /etc/prometheus + - name: prom-config-base + mountPath: /config + extraVolumeMounts: + - name: prom-config + mountPath: /etc/prometheus + extraVolumes: + - name: prom-config + emptyDir: {} + - name: prom-config-base + projected: + sources: + - configMap: + name: osm-prom-configmap + items: + - key: prometheus.yml + path: prometheus.yml + - configMap: + name: osm-prometheus-recordingrules-configmap + items: + - key: osm_metric_rules.yml + path: osm_metric_rules.yml + - configMap: + name: osm-prometheus-alertingrules-configmap + items: + - key: osm_alert_rules.yml + path: osm_alert_rules.yml + extraManifests: + - | + apiVersion: v1 + kind: ConfigMap + metadata: + name: osm-prometheus-alertingrules-configmap + data: + osm_alert_rules.yml: | + groups: + - name: osm_alert_rules + rules: + - alert: vdu_down + expr: vm_status_extended != 1 + for: 3m + annotations: + summary: "VDU {{ "{{" }} $labels.vm_id }} in VIM {{ "{{" }} $labels.vim_id }} is down" + description: "VDU {{ "{{" }} $labels.vm_id }} in VIM {{ "{{" }} $labels.vim_id }} has been down for more than 3 minutes. NS instance id is {{ "{{" }} $labels.ns_id }}" + - | + apiVersion: v1 + kind: ConfigMap + metadata: + name: osm-prometheus-recordingrules-configmap + data: + osm_metric_rules.yml: | + groups: + - name: osm_metric_rules + rules: + - record: vm_status_extended + expr: (last_over_time(vm_status[1m]) * on (vm_id, vim_id) group_left(ns_id, vnf_id, vdu_id, project_id, job, vdu_name, vnf_member_index) last_over_time(ns_topology[1m])) or (last_over_time(ns_topology[1m]) * -1) + labels: + job: osm_prometheus + - record: vnf_status + expr: (0 * (count (vm_status_extended==0) by (ns_id, vnf_id)>=0)) or (min by (ns_id, vnf_id) (vm_status_extended)) + labels: + job: osm_prometheus + - record: ns_status + expr: (0 * (count (vm_status_extended==0) by (ns_id)>=0)) or (min by (ns_id) (vm_status_extended)) + labels: + job: osm_prometheus + - | + apiVersion: v1 + kind: ConfigMap + metadata: + name: osm-prometheus-sidecar-configmap + data: + MONGODB_URL: "mongodb://mongodb-k8s:27017/?replicaSet=rs0" + PROMETHEUS_CONFIG_FILE: "/etc/prometheus/prometheus.yml" + PROMETHEUS_BASE_CONFIG_FILE: "/etc/prometheus_base/prometheus.yml" + TARGET_DATABASE: "osm" + PROMETHEUS_URL: "http://prometheus:9090" + - | + apiVersion: v1 + kind: ConfigMap + metadata: + name: osm-prom-configmap + data: + prometheus.yml: | + global: + scrape_interval: 15s # Set the scrape interval to every 15 seconds. Default is every 1 minute. + evaluation_interval: 15s # Evaluate rules every 15 seconds. The default is every 1 minute. + # scrape_timeout is set to the global default (10s). + # Alertmanager configuration + alerting: + alertmanagers: + - static_configs: + - targets: + - alertmanager:9093 + rule_files: + - "osm_metric_rules.yml" + - "osm_alert_rules.yml" + scrape_configs: + - job_name: 'mon_exporter' + static_configs: + - targets: ['mon:8000'] + - job_name: pushgateway + honor_labels: true + scrape_interval: 30s + static_configs: + - targets: + - pushgateway-prometheus-pushgateway:9091 + - job_name: prometheus + static_configs: + - targets: + - localhost:9090 + - job_name: node + static_configs: + - targets: ['prometheus-node-exporter:9100'] + alertmanager: + fullnameOverride: "alertmanager" + extraArgs: + log.level: debug + service: + type: NodePort + nodePort: 9093 + port: 9093 + config: + receivers: + - name: default-receiver + - name: vdu-webhook + webhook_configs: + - url: http://webhook-translator:9998/vdu_down + - name: scaleout-webhook + webhook_configs: + - url: http://webhook-translator:9998/scaleout_vdu + - name: scalein-webhook + webhook_configs: + - url: http://webhook-translator:9998/scalein_vdu + - name: alarm-webhook + webhook_configs: + - url: http://webhook-translator:9998/vdu_alarm + route: + group_wait: 10s + group_interval: 2m + receiver: default-receiver + routes: + - receiver: vdu-webhook + repeat_interval: 15m + matchers: + - alertname = "vdu_down" + - receiver: 'scaleout-webhook' + repeat_interval: 5m + matchers: + - alertname =~ "^scaleout_.*" + - receiver: 'scalein-webhook' + repeat_interval: 5m + matchers: + - alertname =~ "^scalein_.*" + - receiver: 'alarm-webhook' + repeat_interval: 5m + matchers: + - alertname =~ "^vdu_alarm_.*" + prometheus-pushgateway: + fullnameOverride: "pushgateway-prometheus-pushgateway" ro: enabled: true + # logLevel: DEBUG service: port: 9090 image: {} # repository: opensourcemano/ro - # tag: "13" + # tag: "testing-daily" # replicaCount: 1 useOsmSecret: true # secretName: "ro-secret" vca: - enabled: true + enabled: false # host: "" # secret: "" # cacert: "" @@ -199,13 +497,8 @@ webhookTranslator: service: type: NodePort nodePort: 9998 + port: 9998 image: {} # repository: opensourcemano/webhook - # tag: "13" - # replicaCount: 1 - -zookeeper: - enabled: true - service: - port: 2181 + # tag: "testing-daily" # replicaCount: 1