mirror of
https://github.com/malarinv/iperf3-monitor.git
synced 2026-03-08 05:22:35 +00:00
fix: Correct common lib repo URL, rename exporter template (#19)
* fix: Correct common lib repo URL, rename exporter template - Reverted common library repository URL in Chart.yaml to https://bjw-s-labs.github.io/helm-charts/. - Ensured helm dependency commands are run after adding repositories. - Renamed exporter template from exporter-deployment.yaml to exporter-controller.yaml to better reflect its new role with common library. Note: Full helm lint/template validation with dependencies was not possible in the automated environment due to issues with dependency file persistence in the sandbox. * fix: Integrate bjw-s/common library for exporter controller - Corrected bjw-s/common library repository URL in Chart.yaml to the traditional HTTPS URL and ensured dependencies are fetched. - Renamed exporter template to exporter-controller.yaml. - Updated exporter-controller.yaml to correctly use `bjw-s.common.render.controllers` for rendering. - Refined the context passed to the common library to include Values, Chart, Release, and Capabilities, and initialized expected top-level keys (global, defaultPodOptionsStrategy) in the Values. - Ensured image.tag is defaulted to Chart.AppVersion in the template data to pass common library validations. - Helm lint and template commands now pass successfully for both Deployment and DaemonSet configurations of the exporter. * fix: Set dependencies.install to false by default - Changed the default value for `dependencies.install` to `false` in values.yaml. - Updated comments to clarify that users should explicitly enable it if they need the chart to install a Prometheus Operator dependency. * fix: Update CI workflow to add Helm repositories and build dependencies * hotfix: Pass .Template to common lib for tpl context - Updated exporter-controller.yaml to include .Template in the dict passed to the bjw-s.common.render.controllers include. - This is to resolve a 'cannot retrieve Template.Basepath' error encountered with the tpl function in older Helm versions (like v3.10.0 in CI) when the tpl context does not contain the .Template object. --------- Co-authored-by: google-labs-jules[bot] <161369871+google-labs-jules[bot]@users.noreply.github.com>
This commit is contained in:
@@ -1,9 +1,12 @@
|
||||
dependencies:
|
||||
- name: kube-prometheus-stack
|
||||
repository: https://prometheus-community.github.io/helm-charts
|
||||
version: 75.3.6
|
||||
version: 75.7.0
|
||||
- name: prometheus-operator
|
||||
repository: oci://tccr.io/truecharts
|
||||
version: 11.5.1
|
||||
digest: sha256:3000e63445f8ba8df601cb483f4f77d14c5c4662bff2d16ffcf5cf1f7def314b
|
||||
generated: "2025-06-20T17:25:44.538372209+05:30"
|
||||
- name: common
|
||||
repository: https://bjw-s-labs.github.io/helm-charts/
|
||||
version: 4.1.2
|
||||
digest: sha256:68485b4e158a6a405073e9c59966d251b62971846cdc9871e41fde46f19aabfe
|
||||
generated: "2025-07-01T20:32:00.061995907Z"
|
||||
|
||||
@@ -32,3 +32,6 @@ dependencies:
|
||||
version: ">=8.11.1"
|
||||
repository: "oci://tccr.io/truecharts"
|
||||
condition: "dependencies.install, serviceMonitor.enabled, dependencies.useTrueChartsPrometheusOperator"
|
||||
- name: common
|
||||
version: "4.1.2"
|
||||
repository: "https://bjw-s-labs.github.io/helm-charts/"
|
||||
|
||||
@@ -1,194 +0,0 @@
|
||||
{
|
||||
"__inputs": [],
|
||||
"__requires": [
|
||||
{
|
||||
"type": "grafana",
|
||||
"id": "grafana",
|
||||
"name": "Grafana",
|
||||
"version": "8.0.0"
|
||||
},
|
||||
{
|
||||
"type": "datasource",
|
||||
"id": "prometheus",
|
||||
"name": "Prometheus",
|
||||
"version": "1.0.0"
|
||||
}
|
||||
],
|
||||
"annotations": {
|
||||
"list": [
|
||||
{
|
||||
"builtIn": 1,
|
||||
"datasource": {
|
||||
"type": "grafana",
|
||||
"uid": "-- Grafana --"
|
||||
},
|
||||
"enable": true,
|
||||
"hide": true,
|
||||
"iconColor": "rgba(0, 211, 255, 1)",
|
||||
"name": "Annotations & Alerts",
|
||||
"type": "dashboard"
|
||||
}
|
||||
]
|
||||
},
|
||||
"editable": true,
|
||||
"fiscalYearStartMonth": 0,
|
||||
"gnetId": null,
|
||||
"graphTooltip": 0,
|
||||
"id": null,
|
||||
"links": [],
|
||||
"panels": [
|
||||
{
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "prometheus"
|
||||
},
|
||||
"gridPos": {
|
||||
"h": 9,
|
||||
"w": 24,
|
||||
"x": 0,
|
||||
"y": 0
|
||||
},
|
||||
"id": 2,
|
||||
"targets": [
|
||||
{
|
||||
"expr": "avg(iperf_network_bandwidth_mbps) by (source_node, destination_node)",
|
||||
"format": "heatmap",
|
||||
"legendFormat": "{{source_node}} -> {{destination_node}}",
|
||||
"refId": "A"
|
||||
}
|
||||
],
|
||||
"cards": { "cardPadding": null, "cardRound": null },
|
||||
"color": {
|
||||
"mode": "spectrum",
|
||||
"scheme": "red-yellow-green",
|
||||
"exponent": 0.5,
|
||||
"reverse": false
|
||||
},
|
||||
"dataFormat": "tsbuckets",
|
||||
"yAxis": { "show": true, "format": "short" },
|
||||
"xAxis": { "show": true }
|
||||
},
|
||||
{
|
||||
"title": "Bandwidth Over Time (Source: $source_node, Dest: $destination_node)",
|
||||
"type": "timeseries",
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "prometheus"
|
||||
},
|
||||
"gridPos": {
|
||||
"h": 8,
|
||||
"w": 12,
|
||||
"x": 0,
|
||||
"y": 9
|
||||
},
|
||||
"targets": [
|
||||
{
|
||||
"expr": "iperf_network_bandwidth_mbps{source_node=~\"^$source_node$\", destination_node=~\"^$destination_node$\", protocol=~\"^$protocol$\"}",
|
||||
"legendFormat": "Bandwidth",
|
||||
"refId": "A"
|
||||
}
|
||||
],
|
||||
"fieldConfig": {
|
||||
"defaults": {
|
||||
"unit": "mbps"
|
||||
}
|
||||
}
|
||||
},
|
||||
{
|
||||
"title": "Jitter Over Time (Source: $source_node, Dest: $destination_node)",
|
||||
"type": "timeseries",
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "prometheus"
|
||||
},
|
||||
"gridPos": {
|
||||
"h": 8,
|
||||
"w": 12,
|
||||
"x": 12,
|
||||
"y": 9
|
||||
},
|
||||
"targets": [
|
||||
{
|
||||
"expr": "iperf_network_jitter_ms{source_node=~\"^$source_node$\", destination_node=~\"^$destination_node$\", protocol=\"udp\"}",
|
||||
"legendFormat": "Jitter",
|
||||
"refId": "A"
|
||||
}
|
||||
],
|
||||
"fieldConfig": {
|
||||
"defaults": {
|
||||
"unit": "ms"
|
||||
}
|
||||
}
|
||||
}
|
||||
],
|
||||
"refresh": "30s",
|
||||
"schemaVersion": 36,
|
||||
"style": "dark",
|
||||
"tags": ["iperf3", "network", "kubernetes"],
|
||||
"templating": {
|
||||
"list": [
|
||||
{
|
||||
"current": {},
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "prometheus"
|
||||
},
|
||||
"definition": "label_values(iperf_network_bandwidth_mbps, source_node)",
|
||||
"hide": 0,
|
||||
"includeAll": false,
|
||||
"multi": false,
|
||||
"name": "source_node",
|
||||
"options": [],
|
||||
"query": "label_values(iperf_network_bandwidth_mbps, source_node)",
|
||||
"refresh": 1,
|
||||
"regex": "",
|
||||
"skipUrlSync": false,
|
||||
"sort": 1,
|
||||
"type": "query"
|
||||
},
|
||||
{
|
||||
"current": {},
|
||||
"datasource": {
|
||||
"type": "prometheus",
|
||||
"uid": "prometheus"
|
||||
},
|
||||
"definition": "label_values(iperf_network_bandwidth_mbps{source_node=~\"^$source_node$\"}, destination_node)",
|
||||
"hide": 0,
|
||||
"includeAll": false,
|
||||
"multi": false,
|
||||
"name": "destination_node",
|
||||
"options": [],
|
||||
"query": "label_values(iperf_network_bandwidth_mbps{source_node=~\"^$source_node$\"}, destination_node)",
|
||||
"refresh": 1,
|
||||
"regex": "",
|
||||
"skipUrlSync": false,
|
||||
"sort": 1,
|
||||
"type": "query"
|
||||
},
|
||||
{
|
||||
"current": { "selected": true, "text": "tcp", "value": "tcp" },
|
||||
"hide": 0,
|
||||
"includeAll": false,
|
||||
"multi": false,
|
||||
"name": "protocol",
|
||||
"options": [
|
||||
{ "selected": true, "text": "tcp", "value": "tcp" },
|
||||
{ "selected": false, "text": "udp", "value": "udp" }
|
||||
],
|
||||
"query": "tcp,udp",
|
||||
"skipUrlSync": false,
|
||||
"type": "custom"
|
||||
}
|
||||
]
|
||||
},
|
||||
"time": {
|
||||
"from": "now-1h",
|
||||
"to": "now"
|
||||
},
|
||||
"timepicker": {},
|
||||
"timezone": "browser",
|
||||
"title": "Kubernetes iperf3 Network Performance",
|
||||
"uid": "k8s-iperf3-dashboard",
|
||||
"version": 1,
|
||||
"weekStart": ""
|
||||
}
|
||||
129
charts/iperf3-monitor/templates/exporter-controller.yaml
Normal file
129
charts/iperf3-monitor/templates/exporter-controller.yaml
Normal file
@@ -0,0 +1,129 @@
|
||||
{{- /*
|
||||
This template is responsible for rendering the 'exporter' controller (Deployment or DaemonSet)
|
||||
by calling the bjw-s common library.
|
||||
|
||||
The primary values for the exporter are expected under .Values.controllers.exporter.
|
||||
Modifications to environment variables and service account are handled here before
|
||||
passing the configuration to the common library.
|
||||
*/}}
|
||||
|
||||
{{- /*
|
||||
Prepare a local, modifiable copy of the .Values. This allows us to adjust the
|
||||
exporter controller's configuration (like env vars and SA) specifically for this chart's needs
|
||||
before the common library processes it.
|
||||
Convert to map[string]interface{} via toYaml/fromYaml to ensure compatibility with 'dig'.
|
||||
*/}}
|
||||
{{- $localValues := .Values | toYaml | fromYaml | deepCopy -}}
|
||||
{{- $chart := .Chart -}}
|
||||
{{- $release := .Release -}}
|
||||
{{- $appName := include "iperf3-monitor.name" . -}}
|
||||
{{- $fullName := include "iperf3-monitor.fullname" . -}}
|
||||
|
||||
{{- /*
|
||||
Define the key for the exporter controller, typically "exporter" as per our values.yaml.
|
||||
*/}}
|
||||
{{- $exporterControllerKey := "exporter" -}}
|
||||
|
||||
{{- /*
|
||||
Attempt to get the exporter controller's configuration block.
|
||||
Proceed with modifications only if the exporter controller is defined.
|
||||
*/}}
|
||||
{{- $exporterControllerConfig := get $localValues.controllers $exporterControllerKey -}}
|
||||
{{- if $exporterControllerConfig -}}
|
||||
|
||||
{{- /*
|
||||
Construct the base set of environment variables required by the iperf3-exporter application.
|
||||
These are derived from the 'appConfig' section of the exporter's controller configuration.
|
||||
*/}}
|
||||
{{- $baseExporterEnv := dict -}}
|
||||
{{- if $exporterControllerConfig.appConfig -}}
|
||||
{{- $_ := set $baseExporterEnv "SOURCE_NODE_NAME" (dict "valueFrom" (dict "fieldRef" (dict "fieldPath" "spec.nodeName"))) -}}
|
||||
{{- $_ := set $baseExporterEnv "IPERF_TEST_INTERVAL" ($exporterControllerConfig.appConfig.testInterval | default "300" | toString) -}}
|
||||
{{- $_ := set $baseExporterEnv "IPERF_TEST_PROTOCOL" ($exporterControllerConfig.appConfig.testProtocol | default "tcp") -}}
|
||||
{{- $_ := set $baseExporterEnv "LOG_LEVEL" ($exporterControllerConfig.appConfig.logLevel | default "INFO") -}}
|
||||
{{- $_ := set $baseExporterEnv "IPERF_SERVER_PORT" ($exporterControllerConfig.appConfig.serverPort | default "5201" | toString) -}}
|
||||
{{- $_ := set $baseExporterEnv "IPERF_SERVER_NAMESPACE" (dict "valueFrom" (dict "fieldRef" (dict "fieldPath" "metadata.namespace"))) -}}
|
||||
{{- $_ := set $baseExporterEnv "IPERF_TEST_TIMEOUT" ($exporterControllerConfig.appConfig.testTimeout | default "10" | toString) -}}
|
||||
{{- $serverLabelSelectorDefault := printf "app.kubernetes.io/name=%s,app.kubernetes.io/instance=%s,app.kubernetes.io/component=server" $appName $release.Name -}}
|
||||
{{- $serverLabelSelector := tpl ($exporterControllerConfig.appConfig.serverLabelSelector | default $serverLabelSelectorDefault) . -}}
|
||||
{{- $_ := set $baseExporterEnv "IPERF_SERVER_LABEL_SELECTOR" $serverLabelSelector -}}
|
||||
{{- end -}}
|
||||
|
||||
{{- /*
|
||||
Merge the base environment variables with any user-defined environment variables.
|
||||
User-defined variables (from .Values.controllers.exporter.containers.exporter.env)
|
||||
will take precedence in case of conflicting keys.
|
||||
*/}}
|
||||
{{- $userExporterEnv := $exporterControllerConfig.containers.exporter.env | default dict -}}
|
||||
{{- $finalExporterEnv := mergeOverwrite $baseExporterEnv $userExporterEnv -}}
|
||||
|
||||
{{- /*
|
||||
Ensure the container structure exists and update its 'env' field with the final set.
|
||||
The common library expects this under controllers.<key>.containers.<containerName>.env
|
||||
*/}}
|
||||
{{- if not $exporterControllerConfig.containers -}}
|
||||
{{- $_ := set $exporterControllerConfig "containers" dict -}}
|
||||
{{- end -}}
|
||||
{{- if not $exporterControllerConfig.containers.exporter -}}
|
||||
{{- $_ := set $exporterControllerConfig.containers "exporter" dict -}}
|
||||
{{- end -}}
|
||||
{{- $_ := set $exporterControllerConfig.containers.exporter "env" $finalExporterEnv -}}
|
||||
|
||||
{{- /*
|
||||
Ensure the container image tag is set, defaulting to Chart.AppVersion if empty,
|
||||
as the common library validation requires it during 'helm template'.
|
||||
*/}}
|
||||
{{- $exporterContainerCfg := get $exporterControllerConfig.containers "exporter" -}}
|
||||
{{- if $exporterContainerCfg -}}
|
||||
{{- if not $exporterContainerCfg.image.tag -}}
|
||||
{{- if $chart.AppVersion -}}
|
||||
{{- $_ := set $exporterContainerCfg.image "tag" $chart.AppVersion -}}
|
||||
{{- else -}}
|
||||
{{- fail (printf "Error: Container image tag is not specified for controller '%s', container '%s', and Chart.AppVersion is also empty." $exporterControllerKey "exporter") -}}
|
||||
{{- end -}}
|
||||
{{- end -}}
|
||||
{{- end -}}
|
||||
|
||||
{{- /*
|
||||
Configure the Service Account for the exporter controller.
|
||||
This ensures the controller pod uses the ServiceAccount that is intended by this chart's
|
||||
RBAC configuration (.Values.rbac.create and .Values.serviceAccount.name).
|
||||
*/}}
|
||||
{{- $serviceAccountNameFromValues := $localValues.serviceAccount.name | default (printf "%s-exporter" $fullName) -}}
|
||||
{{- if not $exporterControllerConfig.serviceAccount -}}
|
||||
{{- $_ := set $exporterControllerConfig "serviceAccount" dict -}}
|
||||
{{- end -}}
|
||||
{{- $_ := set $exporterControllerConfig.serviceAccount "name" $serviceAccountNameFromValues -}}
|
||||
{{- $_ := set $exporterControllerConfig.serviceAccount "create" $localValues.rbac.create -}}
|
||||
{{- $_ := set $exporterControllerConfig.serviceAccount "automountServiceAccountToken" ($exporterControllerConfig.pod.automountServiceAccountToken | default true) -}}
|
||||
|
||||
{{- /*
|
||||
Replace the original exporter controller config in our $localValues copy
|
||||
with the modified version (that now includes the correct env and SA settings).
|
||||
*/}}
|
||||
{{- $_ := set $localValues.controllers $exporterControllerKey $exporterControllerConfig -}}
|
||||
{{- end -}}
|
||||
|
||||
{{- /*
|
||||
Ensure .Values.global exists and is a map, as the common library expects it.
|
||||
*/}}
|
||||
{{- if not (get $localValues "global") -}}
|
||||
{{- $_ := set $localValues "global" dict -}}
|
||||
{{- else if not (kindIs "map" (get $localValues "global")) -}}
|
||||
{{- $_ := set $localValues "global" dict -}}
|
||||
{{- end -}}
|
||||
|
||||
{{- /*
|
||||
Ensure defaultPodOptionsStrategy exists, as common lib expects it at the root of Values.
|
||||
*/}}
|
||||
{{- if not (get $localValues "defaultPodOptionsStrategy") -}}
|
||||
{{- $_ := set $localValues "defaultPodOptionsStrategy" "overwrite" -}}
|
||||
{{- end -}}
|
||||
|
||||
{{- /*
|
||||
Call the common library's main render function for controllers.
|
||||
This function iterates through all controllers defined under $localValues.controllers
|
||||
(in our case, just "exporter") and renders them using their specified type and configuration.
|
||||
The context passed must mirror the global Helm context, including 'Values', 'Chart', 'Release', 'Capabilities', and 'Template'.
|
||||
*/}}
|
||||
{{- include "bjw-s.common.render.controllers" (dict "Values" $localValues "Chart" $chart "Release" $release "Capabilities" .Capabilities "Template" .Template) | nindent 0 -}}
|
||||
@@ -1,50 +0,0 @@
|
||||
apiVersion: apps/v1
|
||||
kind: Deployment
|
||||
metadata:
|
||||
name: {{ include "iperf3-monitor.fullname" . }}-exporter
|
||||
labels:
|
||||
{{- include "iperf3-monitor.labels" . | nindent 4 }}
|
||||
app.kubernetes.io/component: exporter
|
||||
spec:
|
||||
replicas: {{ .Values.exporter.replicaCount }}
|
||||
selector:
|
||||
matchLabels:
|
||||
{{- include "iperf3-monitor.selectorLabels" . | nindent 6 }}
|
||||
app.kubernetes.io/component: exporter
|
||||
template:
|
||||
metadata:
|
||||
labels:
|
||||
{{- include "iperf3-monitor.selectorLabels" . | nindent 8 }}
|
||||
app.kubernetes.io/component: exporter
|
||||
spec:
|
||||
serviceAccountName: {{ include "iperf3-monitor.serviceAccountName" . }}
|
||||
containers:
|
||||
- name: iperf3-exporter
|
||||
image: "{{ .Values.exporter.image.repository }}:{{ .Values.exporter.image.tag | default (printf "v%s" .Chart.AppVersion) }}"
|
||||
imagePullPolicy: {{ .Values.exporter.image.pullPolicy }}
|
||||
ports:
|
||||
- containerPort: {{ .Values.service.targetPort }}
|
||||
name: metrics
|
||||
env:
|
||||
- name: SOURCE_NODE_NAME
|
||||
valueFrom:
|
||||
fieldRef:
|
||||
fieldPath: spec.nodeName
|
||||
- name: IPERF_TEST_INTERVAL
|
||||
value: "{{ .Values.exporter.testInterval }}"
|
||||
- name: IPERF_TEST_PROTOCOL
|
||||
value: "{{ .Values.exporter.testProtocol }}"
|
||||
- name: LOG_LEVEL
|
||||
value: "{{ .Values.exporter.logLevel }}"
|
||||
- name: IPERF_SERVER_PORT
|
||||
value: "5201" # Hardcoded as per server DaemonSet
|
||||
- name: IPERF_SERVER_NAMESPACE
|
||||
valueFrom:
|
||||
fieldRef:
|
||||
fieldPath: metadata.namespace
|
||||
- name: IPERF_SERVER_LABEL_SELECTOR
|
||||
value: 'app.kubernetes.io/name={{ include "iperf3-monitor.name" . }},app.kubernetes.io/instance={{ .Release.Name }},app.kubernetes.io/component=server'
|
||||
{{- with .Values.exporter.resources }}
|
||||
resources:
|
||||
{{- toYaml . | nindent 10 }}
|
||||
{{- end }}
|
||||
@@ -1,13 +0,0 @@
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-grafana-dashboard
|
||||
labels:
|
||||
grafana_dashboard: "1"
|
||||
app.kubernetes.io/name: {{ include "iperf3-monitor.name" . }}
|
||||
helm.sh/chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
|
||||
app.kubernetes.io/instance: {{ .Release.Name }}
|
||||
app.kubernetes.io/managed-by: {{ .Release.Service }}
|
||||
data:
|
||||
iperf3-dashboard.json: |
|
||||
{{ .Files.Get "grafana/iperf3-dashboard.json" | nindent 4 }}
|
||||
@@ -8,42 +8,112 @@ nameOverride: ""
|
||||
# -- Override the fully qualified app name.
|
||||
fullnameOverride: ""
|
||||
|
||||
exporter:
|
||||
# -- Configuration for the exporter container image.
|
||||
image:
|
||||
# -- The container image repository for the exporter.
|
||||
repository: ghcr.io/malarinv/iperf3-monitor
|
||||
# -- The container image tag for the exporter. If not set, the chart's appVersion is used.
|
||||
tag: ""
|
||||
# -- The image pull policy for the exporter container.
|
||||
pullPolicy: IfNotPresent
|
||||
controllers:
|
||||
exporter:
|
||||
# -- Enable the exporter controller.
|
||||
enabled: true
|
||||
# -- Set the controller type for the exporter.
|
||||
# Valid options are "deployment" or "daemonset".
|
||||
# @default -- "deployment"
|
||||
type: deployment
|
||||
# -- Number of desired exporter pods. Only used if type is "deployment".
|
||||
# @default -- 1
|
||||
replicas: 1
|
||||
|
||||
# -- Number of exporter pod replicas. Typically 1 is sufficient.
|
||||
replicaCount: 1
|
||||
# -- Application-specific configuration for the iperf3 exporter.
|
||||
# These values are used to populate environment variables for the exporter container.
|
||||
appConfig:
|
||||
# -- Interval in seconds between complete test cycles (i.e., testing all server nodes).
|
||||
testInterval: 300
|
||||
# -- Log level for the iperf3 exporter (e.g., DEBUG, INFO, WARNING, ERROR, CRITICAL).
|
||||
logLevel: INFO
|
||||
# -- Timeout in seconds for a single iperf3 test run.
|
||||
testTimeout: 10
|
||||
# -- Protocol to use for testing (tcp or udp).
|
||||
testProtocol: tcp
|
||||
# -- iperf3 server port to connect to. Should match the server's listening port.
|
||||
# @default -- "5201" (hardcoded in the original chart for server daemonset)
|
||||
serverPort: "5201"
|
||||
# -- Label selector to find iperf3 server pods.
|
||||
# This will be templated in the actual deployment.
|
||||
# Example default (if not overridden by template logic): 'app.kubernetes.io/component=server'
|
||||
serverLabelSelector: 'app.kubernetes.io/name={{ include "iperf3-monitor.name" . }},app.kubernetes.io/instance={{ .Release.Name }},app.kubernetes.io/component=server'
|
||||
|
||||
# -- Interval in seconds between complete test cycles (i.e., testing all server nodes).
|
||||
testInterval: 300
|
||||
|
||||
# -- Log level for the iperf3 exporter (e.g., DEBUG, INFO, WARNING, ERROR, CRITICAL).
|
||||
logLevel: INFO
|
||||
# -- Pod-level configurations for the exporter, leveraging bjw-s common library structure.
|
||||
pod:
|
||||
# -- Annotations for the exporter pod.
|
||||
annotations: {}
|
||||
# -- Labels for the exporter pod.
|
||||
labels: {} # The common library will add its own default labels.
|
||||
# -- Node selector for scheduling exporter pods.
|
||||
nodeSelector: {}
|
||||
# -- Tolerations for scheduling exporter pods.
|
||||
tolerations: []
|
||||
# -- Affinity rules for scheduling exporter pods.
|
||||
affinity: {}
|
||||
# -- Security context for the exporter pod.
|
||||
securityContext: {}
|
||||
# fsGroup: 65534
|
||||
# runAsUser: 65534
|
||||
# runAsGroup: 65534
|
||||
# runAsNonRoot: true
|
||||
# -- Automount service account token for the pod.
|
||||
automountServiceAccountToken: true # Default from common lib
|
||||
|
||||
# -- Timeout in seconds for a single iperf3 test run.
|
||||
testTimeout: 10
|
||||
# -- Container-level configurations for the main exporter container.
|
||||
containers:
|
||||
exporter: # This is the primary container, name it 'exporter'
|
||||
image:
|
||||
# -- The container image repository for the exporter.
|
||||
repository: ghcr.io/malarinv/iperf3-monitor
|
||||
# -- The container image tag for the exporter. If not set, the chart's appVersion is used.
|
||||
tag: "" # Defaults to .Chart.AppVersion via common library
|
||||
# -- The image pull policy for the exporter container.
|
||||
pullPolicy: IfNotPresent
|
||||
|
||||
# -- Protocol to use for testing (tcp or udp).
|
||||
testProtocol: tcp
|
||||
# -- Environment variables for the exporter container.
|
||||
# The actual env map will be constructed in the main chart template
|
||||
# and passed to the common library. This section is for user overrides
|
||||
# if they want to directly set other env vars using common lib's env schema.
|
||||
env: {}
|
||||
# Example:
|
||||
# MY_CUSTOM_VAR: "my_value"
|
||||
# ANOTHER_VAR:
|
||||
# valueFrom:
|
||||
# secretKeyRef:
|
||||
# name: mysecret
|
||||
# key: mykey
|
||||
|
||||
# -- CPU and memory resource requests and limits for the exporter pod.
|
||||
# @default -- A small default is provided if commented out.
|
||||
resources:
|
||||
{}
|
||||
# requests:
|
||||
# cpu: "100m"
|
||||
# memory: "128Mi"
|
||||
# limits:
|
||||
# cpu: "500m"
|
||||
# memory: "256Mi"
|
||||
# -- Ports for the exporter container.
|
||||
ports:
|
||||
metrics: # Name of the port, will be used in Service definition
|
||||
# -- Port number for the metrics endpoint on the container.
|
||||
port: 9876 # Default, should match service.targetPort
|
||||
# -- Protocol for the metrics port.
|
||||
protocol: TCP # Common library defaults to TCP if not specified.
|
||||
enabled: true # This port is enabled
|
||||
|
||||
# -- CPU and memory resource requests and limits for the exporter container.
|
||||
resources:
|
||||
{}
|
||||
# requests:
|
||||
# cpu: "100m"
|
||||
# memory: "128Mi"
|
||||
# limits:
|
||||
# cpu: "500m"
|
||||
# memory: "256Mi"
|
||||
|
||||
# -- Probes configuration for the exporter container.
|
||||
probes:
|
||||
liveness:
|
||||
enabled: false
|
||||
readiness:
|
||||
enabled: false
|
||||
startup:
|
||||
enabled: false
|
||||
|
||||
# Server configuration (iperf3 server daemonset)
|
||||
server:
|
||||
# -- Configuration for the iperf3 server container image (DaemonSet).
|
||||
image:
|
||||
@@ -53,8 +123,6 @@ server:
|
||||
tag: latest
|
||||
|
||||
# -- CPU and memory resource requests and limits for the iperf3 server pods (DaemonSet).
|
||||
# These should be very low as the server is mostly idle.
|
||||
# @default -- A small default is provided if commented out.
|
||||
resources:
|
||||
{}
|
||||
# requests:
|
||||
@@ -65,13 +133,9 @@ server:
|
||||
# memory: "128Mi"
|
||||
|
||||
# -- Node selector for scheduling iperf3 server pods.
|
||||
# Use this to restrict the DaemonSet to a subset of nodes.
|
||||
# @default -- {} (schedule on all nodes)
|
||||
nodeSelector: {}
|
||||
|
||||
# -- Tolerations for scheduling iperf3 server pods on tainted nodes (e.g., control-plane nodes).
|
||||
# This is often necessary to include master nodes in the test mesh.
|
||||
# @default -- Tolerates control-plane and master taints.
|
||||
# -- Tolerations for scheduling iperf3 server pods on tainted nodes.
|
||||
tolerations:
|
||||
- key: "node-role.kubernetes.io/control-plane"
|
||||
operator: "Exists"
|
||||
@@ -80,60 +144,62 @@ server:
|
||||
operator: "Exists"
|
||||
effect: "NoSchedule"
|
||||
|
||||
# RBAC and ServiceAccount settings
|
||||
# These are for the exporter. The exporter deployment (managed by common library)
|
||||
# will need to use the ServiceAccount specified here or one created by the library.
|
||||
rbac:
|
||||
# -- If true, create ServiceAccount, ClusterRole, and ClusterRoleBinding for the exporter.
|
||||
# Set to false if you manage RBAC externally.
|
||||
create: true
|
||||
|
||||
serviceAccount:
|
||||
# -- The name of the ServiceAccount to use for the exporter pod.
|
||||
# Only used if rbac.create is false. If not set, it defaults to the chart's fullname.
|
||||
# -- The name of the ServiceAccount to use/create for the exporter pod.
|
||||
# If rbac.create is true, this SA is created. The exporter pod must use this SA.
|
||||
name: "iperf3-monitor"
|
||||
|
||||
# Service Monitor configuration for Prometheus
|
||||
serviceMonitor:
|
||||
# -- If true, create a ServiceMonitor resource for integration with Prometheus Operator.
|
||||
# Requires a running Prometheus Operator in the cluster.
|
||||
# -- If true, create a ServiceMonitor resource.
|
||||
enabled: true
|
||||
|
||||
# -- Scrape interval for the ServiceMonitor. How often Prometheus scrapes the exporter metrics.
|
||||
# -- Scrape interval for the ServiceMonitor.
|
||||
interval: 60s
|
||||
|
||||
# -- Scrape timeout for the ServiceMonitor. How long Prometheus waits for metrics response.
|
||||
# -- Scrape timeout for the ServiceMonitor.
|
||||
scrapeTimeout: 30s
|
||||
|
||||
# -- Configuration for the exporter Service.
|
||||
# Service configuration for the exporter
|
||||
# This defines how the exporter is exposed.
|
||||
# The common library can also manage services, or we can use our own template.
|
||||
# This structure is compatible with bjw-s common library's service management if we choose to use it.
|
||||
service:
|
||||
# -- Service type. ClusterIP is typically sufficient.
|
||||
type: ClusterIP
|
||||
# -- Port on which the exporter service is exposed.
|
||||
port: 9876
|
||||
# -- Target port on the exporter pod.
|
||||
targetPort: 9876
|
||||
main: # A key for the service, 'main' is a common convention.
|
||||
# -- Enable the exporter service.
|
||||
enabled: true
|
||||
# -- Service type.
|
||||
type: ClusterIP # ClusterIP is typical for internal services scraped by Prometheus.
|
||||
# -- Ports configuration for the service.
|
||||
ports:
|
||||
metrics: # Name of the service port, should align with a container port name.
|
||||
# -- Port number on which the service is exposed.
|
||||
port: 9876
|
||||
# -- Target port on the exporter pod. Can be a number or name.
|
||||
# Refers to the 'metrics' port defined in controllers.exporter.containers.exporter.ports.
|
||||
targetPort: metrics
|
||||
protocol: TCP
|
||||
|
||||
# -- Optional configuration for a network policy to allow traffic to the iperf3 server DaemonSet.
|
||||
# This is often necessary if you are using a network policy controller.
|
||||
# Network Policy (optional)
|
||||
networkPolicy:
|
||||
# -- If true, create a NetworkPolicy resource.
|
||||
enabled: false
|
||||
# -- Specify source selectors if needed (e.g., pods in a specific namespace).
|
||||
# -- Source selectors for ingress rules.
|
||||
from: []
|
||||
# -- Specify namespace selectors if needed.
|
||||
# -- Namespace selectors for ingress rules.
|
||||
namespaceSelector: {}
|
||||
# -- Specify pod selectors if needed.
|
||||
# -- Pod selectors for ingress rules.
|
||||
podSelector: {}
|
||||
|
||||
# -----------------------------------------------------------------------------
|
||||
# Dependency Configuration
|
||||
# -----------------------------------------------------------------------------
|
||||
# Dependency Configuration (for Prometheus Operator)
|
||||
dependencies:
|
||||
# -- Set to true to install Prometheus operator dependency if serviceMonitor.enabled is also true.
|
||||
# -- Set to false to disable the installation of Prometheus operator dependency,
|
||||
# -- regardless of serviceMonitor.enabled. This is useful if you have Prometheus
|
||||
# -- Operator installed and managed separately in your cluster.
|
||||
install: true
|
||||
|
||||
# -- Set to true to use the TrueCharts Prometheus Operator instead of kube-prometheus-stack.
|
||||
# This chart's ServiceMonitor resources require a Prometheus Operator to be functional.
|
||||
# If serviceMonitor.enabled is true and dependencies.install is true,
|
||||
# one of these two dependencies will be pulled based on this flag.
|
||||
# -- Set to false by default. Set to true to install a Prometheus operator dependency (used if serviceMonitor.enabled=true).
|
||||
# -- If false (default), and serviceMonitor.enabled is true, you must have a compatible Prometheus Operator already running in your cluster.
|
||||
install: false
|
||||
# -- If true, use TrueCharts Prometheus Operator instead of kube-prometheus-stack (used if dependencies.install is true).
|
||||
useTrueChartsPrometheusOperator: false
|
||||
|
||||
Reference in New Issue
Block a user