1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243 |
- rbac:
- create: true
- podSecurityPolicy:
- enabled: false
- imagePullSecrets: []
- # - name: "image-pull-secret"
- ## Define serviceAccount names for components. Defaults to component's fully qualified name.
- ##
- serviceAccounts:
- server:
- create: true
- name: ""
- annotations: {}
- ## Monitors ConfigMap changes and POSTs to a URL
- ## Ref: https://github.com/prometheus-operator/prometheus-operator/tree/main/cmd/prometheus-config-reloader
- ##
- configmapReload:
- prometheus:
- ## If false, the configmap-reload container will not be deployed
- ##
- enabled: true
- ## configmap-reload container name
- ##
- name: configmap-reload
- ## configmap-reload container image
- ##
- image:
- repository: quay.io/prometheus-operator/prometheus-config-reloader
- tag: v0.63.0
- # When digest is set to a non-empty value, images will be pulled by digest (regardless of tag value).
- digest: ""
- pullPolicy: IfNotPresent
- # containerPort: 9533
- ## Additional configmap-reload container arguments
- ##
- extraArgs: {}
- ## Additional configmap-reload volume directories
- ##
- extraVolumeDirs: []
- ## Additional configmap-reload mounts
- ##
- extraConfigmapMounts: []
- # - name: prometheus-alerts
- # mountPath: /etc/alerts.d
- # subPath: ""
- # configMap: prometheus-alerts
- # readOnly: true
- ## Security context to be added to configmap-reload container
- containerSecurityContext: {}
- ## configmap-reload resource requests and limits
- ## Ref: http://kubernetes.io/docs/user-guide/compute-resources/
- ##
- resources:
- limits:
- cpu: 4
- memory: 8Gi
- server:
- ## Prometheus server container name
- ##
- name: server
- ## Use a ClusterRole (and ClusterRoleBinding)
- ## - If set to false - we define a RoleBinding in the defined namespaces ONLY
- ##
- ## NB: because we need a Role with nonResourceURL's ("/metrics") - you must get someone with Cluster-admin privileges to define this role for you, before running with this setting enabled.
- ## This makes prometheus work - for users who do not have ClusterAdmin privs, but wants prometheus to operate on their own namespaces, instead of clusterwide.
- ##
- ## You MUST also set namespaces to the ones you have access to and want monitored by Prometheus.
- ##
- # useExistingClusterRoleName: nameofclusterrole
- ## namespaces to monitor (instead of monitoring all - clusterwide). Needed if you want to run without Cluster-admin privileges.
- # namespaces:
- # - yournamespace
- # sidecarContainers - add more containers to prometheus server
- # Key/Value where Key is the sidecar `- name: <Key>`
- # Example:
- # sidecarContainers:
- # webserver:
- # image: nginx
- sidecarContainers: {}
- # sidecarTemplateValues - context to be used in template for sidecarContainers
- # Example:
- # sidecarTemplateValues: *your-custom-globals
- # sidecarContainers:
- # webserver: |-
- # {{ include "webserver-container-template" . }}
- # Template for `webserver-container-template` might looks like this:
- # image: "{{ .Values.server.sidecarTemplateValues.repository }}:{{ .Values.server.sidecarTemplateValues.tag }}"
- # ...
- #
- sidecarTemplateValues: {}
- ## Prometheus server container image
- ##
- image:
- repository: quay.io/prometheus/prometheus
- # if not set appVersion field from Chart.yaml is used
- tag: ""
- # When digest is set to a non-empty value, images will be pulled by digest (regardless of tag value).
- digest: ""
- pullPolicy: IfNotPresent
- ## prometheus server priorityClassName
- ##
- priorityClassName: ""
- ## EnableServiceLinks indicates whether information about services should be injected
- ## into pod's environment variables, matching the syntax of Docker links.
- ## WARNING: the field is unsupported and will be skipped in K8s prior to v1.13.0.
- ##
- enableServiceLinks: true
- ## The URL prefix at which the container can be accessed. Useful in the case the '-web.external-url' includes a slug
- ## so that the various internal URLs are still able to access as they are in the default case.
- ## (Optional)
- prefixURL: ""
- ## External URL which can access prometheus
- ## Maybe same with Ingress host name
- baseURL: ""
- ## Additional server container environment variables
- ##
- ## You specify this manually like you would a raw deployment manifest.
- ## This means you can bind in environment variables from secrets.
- ##
- ## e.g. static environment variable:
- ## - name: DEMO_GREETING
- ## value: "Hello from the environment"
- ##
- ## e.g. secret environment variable:
- ## - name: USERNAME
- ## valueFrom:
- ## secretKeyRef:
- ## name: mysecret
- ## key: username
- env: []
- # List of flags to override default parameters, e.g:
- # - --enable-feature=agent
- # - --storage.agent.retention.max-time=30m
- defaultFlagsOverride: []
- extraFlags:
- - web.enable-lifecycle
- - web.enable-remote-write-receiver
- ## web.enable-admin-api flag controls access to the administrative HTTP API which includes functionality such as
- ## deleting time series. This is disabled by default.
- # - web.enable-admin-api
- ##
- ## storage.tsdb.no-lockfile flag controls BD locking
- # - storage.tsdb.no-lockfile
- ##
- ## storage.tsdb.wal-compression flag enables compression of the write-ahead log (WAL)
- # - storage.tsdb.wal-compression
- ## Path to a configuration file on prometheus server container FS
- configPath: /etc/config/prometheus.yml
- ### The data directory used by prometheus to set --storage.tsdb.path
- ### When empty server.persistentVolume.mountPath is used instead
- storagePath: ""
- global:
- ## How frequently to scrape targets by default
- ##
- scrape_interval: 30s
- ## How long until a scrape request times out
- ##
- scrape_timeout: 10s
- ## How frequently to evaluate rules
- ##
- evaluation_interval: 1m
- ## https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_write
- ##
- remoteWrite: []
- ## https://prometheus.io/docs/prometheus/latest/configuration/configuration/#remote_read
- ##
- remoteRead: []
- ## Custom HTTP headers for Liveness/Readiness/Startup Probe
- ##
- ## Useful for providing HTTP Basic Auth to healthchecks
- probeHeaders: []
- # - name: "Authorization"
- # value: "Bearer ABCDEabcde12345"
- ## Additional Prometheus server container arguments
- ##
- extraArgs: {}
- ## Additional InitContainers to initialize the pod
- ##
- extraInitContainers: []
- ## Additional Prometheus server Volume mounts
- ##
- extraVolumeMounts: []
- ## Additional Prometheus server Volumes
- ##
- extraVolumes: []
- ## Additional Prometheus server hostPath mounts
- ##
- extraHostPathMounts: []
- # - name: certs-dir
- # mountPath: /etc/kubernetes/certs
- # subPath: ""
- # hostPath: /etc/kubernetes/certs
- # readOnly: true
- extraConfigmapMounts: []
- # - name: certs-configmap
- # mountPath: /prometheus
- # subPath: ""
- # configMap: certs-configmap
- # readOnly: true
- ## Additional Prometheus server Secret mounts
- # Defines additional mounts with secrets. Secrets must be manually created in the namespace.
- extraSecretMounts: []
- # - name: secret-files
- # mountPath: /etc/secrets
- # subPath: ""
- # secretName: prom-secret-files
- # readOnly: true
- ## ConfigMap override where fullname is {{.Release.Name}}-{{.Values.server.configMapOverrideName}}
- ## Defining configMapOverrideName will cause templates/server-configmap.yaml
- ## to NOT generate a ConfigMap resource
- ##
- configMapOverrideName: ""
- ## Extra labels for Prometheus server ConfigMap (ConfigMap that holds serverFiles)
- extraConfigmapLabels: {}
- ingress:
- ## If true, Prometheus server Ingress will be created
- ##
- enabled: true
- # For Kubernetes >= 1.18 you should specify the ingress-controller via the field ingressClassName
- # See https://kubernetes.io/blog/2020/04/02/improvements-to-the-ingress-api-in-kubernetes-1.18/#specifying-the-class-of-an-ingress
- ingressClassName: nginx
- ## Prometheus server Ingress annotations
- ##
- annotations: {}
- # kubernetes.io/ingress.class: nginx
- # kubernetes.io/tls-acme: 'true'
- ## Prometheus server Ingress additional labels
- ##
- extraLabels: {}
- ## Prometheus server Ingress hostnames with optional path
- ## Must be provided if Ingress is enabled
- ##
- hosts:
- - prometheus.cestong.com.cn
- # - domain.com/prometheus
- path: /
- # pathType is only for k8s >= 1.18
- pathType: Prefix
- ## Extra paths to prepend to every host configuration. This is useful when working with annotation based services.
- extraPaths: []
- # - path: /*
- # backend:
- # serviceName: ssl-redirect
- # servicePort: use-annotation
- ## Prometheus server Ingress TLS configuration
- ## Secrets must be manually created in the namespace
- ##
- tls: []
- # - secretName: prometheus-server-tls
- # hosts:
- # - prometheus.domain.com
- ## Server Deployment Strategy type
- strategy:
- type: Recreate
- ## hostAliases allows adding entries to /etc/hosts inside the containers
- hostAliases: []
- # - ip: "127.0.0.1"
- # hostnames:
- # - "example.com"
- ## Node tolerations for server scheduling to nodes with taints
- ## Ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/
- ##
- tolerations: []
- # - key: "key"
- # operator: "Equal|Exists"
- # value: "value"
- # effect: "NoSchedule|PreferNoSchedule|NoExecute(1.6 only)"
- ## Node labels for Prometheus server pod assignment
- ## Ref: https://kubernetes.io/docs/user-guide/node-selection/
- ##
- nodeSelector: {}
- ## Pod affinity
- ##
- affinity: {}
- ## PodDisruptionBudget settings
- ## ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/
- ##
- podDisruptionBudget:
- enabled: false
- maxUnavailable: 1
- ## Use an alternate scheduler, e.g. "stork".
- ## ref: https://kubernetes.io/docs/tasks/administer-cluster/configure-multiple-schedulers/
- ##
- # schedulerName:
- persistentVolume:
- ## If true, Prometheus server will create/use a Persistent Volume Claim
- ## If false, use emptyDir
- ##
- enabled: true
- ## Prometheus server data Persistent Volume access modes
- ## Must match those of existing PV or dynamic provisioner
- ## Ref: http://kubernetes.io/docs/user-guide/persistent-volumes/
- ##
- accessModes:
- - ReadWriteOnce
- ## Prometheus server data Persistent Volume labels
- ##
- labels: {}
- ## Prometheus server data Persistent Volume annotations
- ##
- annotations: {}
- ## Prometheus server data Persistent Volume existing claim name
- ## Requires server.persistentVolume.enabled: true
- ## If defined, PVC must be created manually before volume will be bound
- existingClaim: ""
- ## Prometheus server data Persistent Volume mount root path
- ##
- mountPath: /data
- ## Prometheus server data Persistent Volume size
- ##
- size: 8Gi
- ## Prometheus server data Persistent Volume Storage Class
- ## If defined, storageClassName: <storageClass>
- ## If set to "-", storageClassName: "", which disables dynamic provisioning
- ## If undefined (the default) or set to null, no storageClassName spec is
- ## set, choosing the default provisioner. (gp2 on AWS, standard on
- ## GKE, AWS & OpenStack)
- ##
- storageClass: "openebs-hostpath"
- ## Prometheus server data Persistent Volume Binding Mode
- ## If defined, volumeBindingMode: <volumeBindingMode>
- ## If undefined (the default) or set to null, no volumeBindingMode spec is
- ## set, choosing the default mode.
- ##
- # volumeBindingMode: ""
- ## Subdirectory of Prometheus server data Persistent Volume to mount
- ## Useful if the volume's root directory is not empty
- ##
- subPath: ""
- ## Persistent Volume Claim Selector
- ## Useful if Persistent Volumes have been provisioned in advance
- ## Ref: https://kubernetes.io/docs/concepts/storage/persistent-volumes/#selector
- ##
- # selector:
- # matchLabels:
- # release: "stable"
- # matchExpressions:
- # - { key: environment, operator: In, values: [ dev ] }
- ## Persistent Volume Name
- ## Useful if Persistent Volumes have been provisioned in advance and you want to use a specific one
- ##
- # volumeName: ""
- emptyDir:
- ## Prometheus server emptyDir volume size limit
- ##
- sizeLimit: ""
- ## Annotations to be added to Prometheus server pods
- ##
- podAnnotations: {}
- # iam.amazonaws.com/role: prometheus
- ## Labels to be added to Prometheus server pods
- ##
- podLabels: {}
- ## Prometheus AlertManager configuration
- ##
- alertmanagers:
- - static_configs:
- - targets: ['prometheus-alertmanager:9093']
- ## Specify if a Pod Security Policy for node-exporter must be created
- ## Ref: https://kubernetes.io/docs/concepts/policy/pod-security-policy/
- ##
- podSecurityPolicy:
- annotations: {}
- ## Specify pod annotations
- ## Ref: https://kubernetes.io/docs/concepts/policy/pod-security-policy/#apparmor
- ## Ref: https://kubernetes.io/docs/concepts/policy/pod-security-policy/#seccomp
- ## Ref: https://kubernetes.io/docs/concepts/policy/pod-security-policy/#sysctl
- ##
- # seccomp.security.alpha.kubernetes.io/allowedProfileNames: '*'
- # seccomp.security.alpha.kubernetes.io/defaultProfileName: 'docker/default'
- # apparmor.security.beta.kubernetes.io/defaultProfileName: 'runtime/default'
- ## Use a StatefulSet if replicaCount needs to be greater than 1 (see below)
- ##
- replicaCount: 1
- ## Annotations to be added to deployment
- ##
- deploymentAnnotations: {}
- statefulSet:
- ## If true, use a statefulset instead of a deployment for pod management.
- ## This allows to scale replicas to more than 1 pod
- ##
- enabled: false
- annotations: {}
- labels: {}
- podManagementPolicy: OrderedReady
- ## Alertmanager headless service to use for the statefulset
- ##
- headless:
- annotations: {}
- labels: {}
- servicePort: 80
- ## Enable gRPC port on service to allow auto discovery with thanos-querier
- gRPC:
- enabled: false
- servicePort: 10901
- # nodePort: 10901
- ## Prometheus server readiness and liveness probe initial delay and timeout
- ## Ref: https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/
- ##
- tcpSocketProbeEnabled: false
- probeScheme: HTTP
- readinessProbeInitialDelay: 30
- readinessProbePeriodSeconds: 5
- readinessProbeTimeout: 4
- readinessProbeFailureThreshold: 3
- readinessProbeSuccessThreshold: 1
- livenessProbeInitialDelay: 30
- livenessProbePeriodSeconds: 15
- livenessProbeTimeout: 10
- livenessProbeFailureThreshold: 3
- livenessProbeSuccessThreshold: 1
- startupProbe:
- enabled: false
- periodSeconds: 5
- failureThreshold: 30
- timeoutSeconds: 10
- ## Prometheus server resource requests and limits
- ## Ref: http://kubernetes.io/docs/user-guide/compute-resources/
- ##
- resources:
- limits:
- cpu: 1
- memory: 4Gi
- requests:
- cpu: 500m
- memory: 2Gi
- # Required for use in managed kubernetes clusters (such as AWS EKS) with custom CNI (such as calico),
- # because control-plane managed by AWS cannot communicate with pods' IP CIDR and admission webhooks are not working
- ##
- hostNetwork: false
- # When hostNetwork is enabled, this will set to ClusterFirstWithHostNet automatically
- dnsPolicy: ClusterFirst
- # Use hostPort
- # hostPort: 9090
- ## Vertical Pod Autoscaler config
- ## Ref: https://github.com/kubernetes/autoscaler/tree/master/vertical-pod-autoscaler
- verticalAutoscaler:
- ## If true a VPA object will be created for the controller (either StatefulSet or Deployemnt, based on above configs)
- enabled: false
- # updateMode: "Auto"
- # containerPolicies:
- # - containerName: 'prometheus-server'
- # Custom DNS configuration to be added to prometheus server pods
- dnsConfig: {}
- # nameservers:
- # - 1.2.3.4
- # searches:
- # - ns1.svc.cluster-domain.example
- # - my.dns.search.suffix
- # options:
- # - name: ndots
- # value: "2"
- # - name: edns0
- ## Security context to be added to server pods
- ##
- securityContext:
- runAsUser: 65534
- runAsNonRoot: true
- runAsGroup: 65534
- fsGroup: 65534
- ## Security context to be added to server container
- ##
- containerSecurityContext: {}
- service:
- ## If false, no Service will be created for the Prometheus server
- ##
- enabled: true
- annotations: {}
- labels: {}
- clusterIP: ""
- ## List of IP addresses at which the Prometheus server service is available
- ## Ref: https://kubernetes.io/docs/user-guide/services/#external-ips
- ##
- externalIPs: []
- loadBalancerIP: ""
- loadBalancerSourceRanges: []
- servicePort: 80
- sessionAffinity: None
- type: ClusterIP
- ## Enable gRPC port on service to allow auto discovery with thanos-querier
- gRPC:
- enabled: false
- servicePort: 10901
- # nodePort: 10901
- ## If using a statefulSet (statefulSet.enabled=true), configure the
- ## service to connect to a specific replica to have a consistent view
- ## of the data.
- statefulsetReplica:
- enabled: false
- replica: 0
- ## Prometheus server pod termination grace period
- ##
- terminationGracePeriodSeconds: 300
- ## Prometheus data retention period (default if not specified is 15 days)
- ##
- retention: "365d"
- ## Prometheus server ConfigMap entries for rule files (allow prometheus labels interpolation)
- ruleFiles: {}
- ## Prometheus server ConfigMap entries
- ##
- serverFiles:
- ## Alerts configuration
- ## Ref: https://prometheus.io/docs/prometheus/latest/configuration/alerting_rules/
- alerting_rules.yml:
- groups:
- - name: 主机状态报警
- rules:
- - alert: 主机cpu使用报警
- expr: sum(avg without (cpu)(irate(node_cpu_seconds_total{mode!='idle'}[5m]))) by (instance) > 0.85
- for: 1m
- labels:
- severity: page
- annotations:
- summary: "Instance {{ $labels.instance }} CPU usgae high"
- description: "{{ $labels.instance }} CPU usage above 85% (current value: {{ $value }})"
- - alert: 主机内存使用报警
- expr: (node_memory_MemTotal_bytes - node_memory_MemAvailable_bytes)/node_memory_MemTotal_bytes > 0.85
- for: 1m
- labels:
- severity: page
- annotations:
- summary: "Instance {{ $labels.instance }} MEM usgae high"
- description: "{{ $labels.instance }} MEM usage above 85% (current value: {{ $value }})"
- # groups:
- # - name: Instances
- # rules:
- # - alert: InstanceDown
- # expr: up == 0
- # for: 5m
- # labels:
- # severity: page
- # annotations:
- # description: '{{ $labels.instance }} of job {{ $labels.job }} has been down for more than 5 minutes.'
- # summary: 'Instance {{ $labels.instance }} down'
- ## DEPRECATED DEFAULT VALUE, unless explicitly naming your files, please use alerting_rules.yml
- alerts: {}
- ## Records configuration
- ## Ref: https://prometheus.io/docs/prometheus/latest/configuration/recording_rules/
- recording_rules.yml: {}
- ## DEPRECATED DEFAULT VALUE, unless explicitly naming your files, please use recording_rules.yml
- rules: {}
- prometheus.yml:
- rule_files:
- - /etc/config/recording_rules.yml
- - /etc/config/alerting_rules.yml
- ## Below two files are DEPRECATED will be removed from this default values file
- - /etc/config/rules
- - /etc/config/alerts
- scrape_configs:
- - job_name: prometheus
- static_configs:
- - targets:
- - "localhost:9090"
- # A scrape configuration for running Prometheus on a Kubernetes cluster.
- # This uses separate scrape configs for cluster components (i.e. API server, node)
- # and services to allow each to use different authentication configs.
- #
- # Kubernetes labels will be added as Prometheus labels on metrics via the
- # `labelmap` relabeling action.
- # Scrape config for API servers.
- #
- # Kubernetes exposes API servers as endpoints to the default/kubernetes
- # service so this uses `endpoints` role and uses relabelling to only keep
- # the endpoints associated with the default/kubernetes service using the
- # default named port `https`. This works for single API server deployments as
- # well as HA API server deployments.
- - job_name: 'kubernetes-apiservers'
- kubernetes_sd_configs:
- - role: endpoints
- # Default to scraping over https. If required, just disable this or change to
- # `http`.
- scheme: https
- # This TLS & bearer token file config is used to connect to the actual scrape
- # endpoints for cluster components. This is separate to discovery auth
- # configuration because discovery & scraping are two separate concerns in
- # Prometheus. The discovery auth config is automatic if Prometheus runs inside
- # the cluster. Otherwise, more config options have to be provided within the
- # <kubernetes_sd_config>.
- tls_config:
- ca_file: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
- # If your node certificates are self-signed or use a different CA to the
- # master CA, then disable certificate verification below. Note that
- # certificate verification is an integral part of a secure infrastructure
- # so this should only be disabled in a controlled environment. You can
- # disable certificate verification by uncommenting the line below.
- #
- insecure_skip_verify: true
- bearer_token_file: /var/run/secrets/kubernetes.io/serviceaccount/token
- # Keep only the default/kubernetes service endpoints for the https port. This
- # will add targets for each API server which Kubernetes adds an endpoint to
- # the default/kubernetes service.
- relabel_configs:
- - source_labels: [__meta_kubernetes_namespace, __meta_kubernetes_service_name, __meta_kubernetes_endpoint_port_name]
- action: keep
- regex: default;kubernetes;https
- - job_name: 'kubernetes-nodes'
- # Default to scraping over https. If required, just disable this or change to
- # `http`.
- scheme: https
- # This TLS & bearer token file config is used to connect to the actual scrape
- # endpoints for cluster components. This is separate to discovery auth
- # configuration because discovery & scraping are two separate concerns in
- # Prometheus. The discovery auth config is automatic if Prometheus runs inside
- # the cluster. Otherwise, more config options have to be provided within the
- # <kubernetes_sd_config>.
- tls_config:
- ca_file: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
- # If your node certificates are self-signed or use a different CA to the
- # master CA, then disable certificate verification below. Note that
- # certificate verification is an integral part of a secure infrastructure
- # so this should only be disabled in a controlled environment. You can
- # disable certificate verification by uncommenting the line below.
- #
- insecure_skip_verify: true
- bearer_token_file: /var/run/secrets/kubernetes.io/serviceaccount/token
- kubernetes_sd_configs:
- - role: node
- relabel_configs:
- - action: labelmap
- regex: __meta_kubernetes_node_label_(.+)
- - target_label: __address__
- replacement: kubernetes.default.svc:443
- - source_labels: [__meta_kubernetes_node_name]
- regex: (.+)
- target_label: __metrics_path__
- replacement: /api/v1/nodes/$1/proxy/metrics
- - job_name: 'kubernetes-nodes-cadvisor'
- # Default to scraping over https. If required, just disable this or change to
- # `http`.
- scheme: https
- # This TLS & bearer token file config is used to connect to the actual scrape
- # endpoints for cluster components. This is separate to discovery auth
- # configuration because discovery & scraping are two separate concerns in
- # Prometheus. The discovery auth config is automatic if Prometheus runs inside
- # the cluster. Otherwise, more config options have to be provided within the
- # <kubernetes_sd_config>.
- tls_config:
- ca_file: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt
- # If your node certificates are self-signed or use a different CA to the
- # master CA, then disable certificate verification below. Note that
- # certificate verification is an integral part of a secure infrastructure
- # so this should only be disabled in a controlled environment. You can
- # disable certificate verification by uncommenting the line below.
- #
- insecure_skip_verify: true
- bearer_token_file: /var/run/secrets/kubernetes.io/serviceaccount/token
- kubernetes_sd_configs:
- - role: node
- # This configuration will work only on kubelet 1.7.3+
- # As the scrape endpoints for cAdvisor have changed
- # if you are using older version you need to change the replacement to
- # replacement: /api/v1/nodes/$1:4194/proxy/metrics
- # more info here https://github.com/coreos/prometheus-operator/issues/633
- relabel_configs:
- - action: labelmap
- regex: __meta_kubernetes_node_label_(.+)
- - target_label: __address__
- replacement: kubernetes.default.svc:443
- - source_labels: [__meta_kubernetes_node_name]
- regex: (.+)
- target_label: __metrics_path__
- replacement: /api/v1/nodes/$1/proxy/metrics/cadvisor
- # Metric relabel configs to apply to samples before ingestion.
- # [Metric Relabeling](https://prometheus.io/docs/prometheus/latest/configuration/configuration/#metric_relabel_configs)
- # metric_relabel_configs:
- # - action: labeldrop
- # regex: (kubernetes_io_hostname|failure_domain_beta_kubernetes_io_region|beta_kubernetes_io_os|beta_kubernetes_io_arch|beta_kubernetes_io_instance_type|failure_domain_beta_kubernetes_io_zone)
- # Scrape config for service endpoints.
- #
- # The relabeling allows the actual service scrape endpoint to be configured
- # via the following annotations:
- #
- # * `prometheus.io/scrape`: Only scrape services that have a value of
- # `true`, except if `prometheus.io/scrape-slow` is set to `true` as well.
- # * `prometheus.io/scheme`: If the metrics endpoint is secured then you will need
- # to set this to `https` & most likely set the `tls_config` of the scrape config.
- # * `prometheus.io/path`: If the metrics path is not `/metrics` override this.
- # * `prometheus.io/port`: If the metrics are exposed on a different port to the
- # service then set this appropriately.
- # * `prometheus.io/param_<parameter>`: If the metrics endpoint uses parameters
- # then you can set any parameter
- - job_name: 'kubernetes-service-endpoints'
- honor_labels: true
- kubernetes_sd_configs:
- - role: endpoints
- relabel_configs:
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_scrape]
- action: keep
- regex: true
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_scrape_slow]
- action: drop
- regex: true
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_scheme]
- action: replace
- target_label: __scheme__
- regex: (https?)
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_path]
- action: replace
- target_label: __metrics_path__
- regex: (.+)
- - source_labels: [__address__, __meta_kubernetes_service_annotation_prometheus_io_port]
- action: replace
- target_label: __address__
- regex: (.+?)(?::\d+)?;(\d+)
- replacement: $1:$2
- - action: labelmap
- regex: __meta_kubernetes_service_annotation_prometheus_io_param_(.+)
- replacement: __param_$1
- - action: labelmap
- regex: __meta_kubernetes_service_label_(.+)
- - source_labels: [__meta_kubernetes_namespace]
- action: replace
- target_label: namespace
- - source_labels: [__meta_kubernetes_service_name]
- action: replace
- target_label: service
- - source_labels: [__meta_kubernetes_pod_node_name]
- action: replace
- target_label: node
- # Scrape config for slow service endpoints; same as above, but with a larger
- # timeout and a larger interval
- #
- # The relabeling allows the actual service scrape endpoint to be configured
- # via the following annotations:
- #
- # * `prometheus.io/scrape-slow`: Only scrape services that have a value of `true`
- # * `prometheus.io/scheme`: If the metrics endpoint is secured then you will need
- # to set this to `https` & most likely set the `tls_config` of the scrape config.
- # * `prometheus.io/path`: If the metrics path is not `/metrics` override this.
- # * `prometheus.io/port`: If the metrics are exposed on a different port to the
- # service then set this appropriately.
- # * `prometheus.io/param_<parameter>`: If the metrics endpoint uses parameters
- # then you can set any parameter
- - job_name: 'kubernetes-service-endpoints-slow'
- honor_labels: true
- scrape_interval: 5m
- scrape_timeout: 30s
- kubernetes_sd_configs:
- - role: endpoints
- relabel_configs:
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_scrape_slow]
- action: keep
- regex: true
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_scheme]
- action: replace
- target_label: __scheme__
- regex: (https?)
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_path]
- action: replace
- target_label: __metrics_path__
- regex: (.+)
- - source_labels: [__address__, __meta_kubernetes_service_annotation_prometheus_io_port]
- action: replace
- target_label: __address__
- regex: (.+?)(?::\d+)?;(\d+)
- replacement: $1:$2
- - action: labelmap
- regex: __meta_kubernetes_service_annotation_prometheus_io_param_(.+)
- replacement: __param_$1
- - action: labelmap
- regex: __meta_kubernetes_service_label_(.+)
- - source_labels: [__meta_kubernetes_namespace]
- action: replace
- target_label: namespace
- - source_labels: [__meta_kubernetes_service_name]
- action: replace
- target_label: service
- - source_labels: [__meta_kubernetes_pod_node_name]
- action: replace
- target_label: node
- - job_name: 'prometheus-pushgateway'
- honor_labels: true
- kubernetes_sd_configs:
- - role: service
- relabel_configs:
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_probe]
- action: keep
- regex: pushgateway
- # Example scrape config for probing services via the Blackbox Exporter.
- #
- # The relabeling allows the actual service scrape endpoint to be configured
- # via the following annotations:
- #
- # * `prometheus.io/probe`: Only probe services that have a value of `true`
- - job_name: 'kubernetes-services'
- honor_labels: true
- metrics_path: /probe
- params:
- module: [http_2xx]
- kubernetes_sd_configs:
- - role: service
- relabel_configs:
- - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_probe]
- action: keep
- regex: true
- - source_labels: [__address__]
- target_label: __param_target
- - target_label: __address__
- replacement: blackbox
- - source_labels: [__param_target]
- target_label: instance
- - action: labelmap
- regex: __meta_kubernetes_service_label_(.+)
- - source_labels: [__meta_kubernetes_namespace]
- target_label: namespace
- - source_labels: [__meta_kubernetes_service_name]
- target_label: service
- # Example scrape config for pods
- #
- # The relabeling allows the actual pod scrape endpoint to be configured via the
- # following annotations:
- #
- # * `prometheus.io/scrape`: Only scrape pods that have a value of `true`,
- # except if `prometheus.io/scrape-slow` is set to `true` as well.
- # * `prometheus.io/scheme`: If the metrics endpoint is secured then you will need
- # to set this to `https` & most likely set the `tls_config` of the scrape config.
- # * `prometheus.io/path`: If the metrics path is not `/metrics` override this.
- # * `prometheus.io/port`: Scrape the pod on the indicated port instead of the default of `9102`.
- - job_name: 'kubernetes-pods'
- honor_labels: true
- kubernetes_sd_configs:
- - role: pod
- relabel_configs:
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scrape]
- action: keep
- regex: true
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scrape_slow]
- action: drop
- regex: true
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scheme]
- action: replace
- regex: (https?)
- target_label: __scheme__
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_path]
- action: replace
- target_label: __metrics_path__
- regex: (.+)
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_port, __meta_kubernetes_pod_ip]
- action: replace
- regex: (\d+);(([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4})
- replacement: '[$2]:$1'
- target_label: __address__
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_port, __meta_kubernetes_pod_ip]
- action: replace
- regex: (\d+);((([0-9]+?)(\.|$)){4})
- replacement: $2:$1
- target_label: __address__
- - action: labelmap
- regex: __meta_kubernetes_pod_annotation_prometheus_io_param_(.+)
- replacement: __param_$1
- - action: labelmap
- regex: __meta_kubernetes_pod_label_(.+)
- - source_labels: [__meta_kubernetes_namespace]
- action: replace
- target_label: namespace
- - source_labels: [__meta_kubernetes_pod_name]
- action: replace
- target_label: pod
- - source_labels: [__meta_kubernetes_pod_phase]
- regex: Pending|Succeeded|Failed|Completed
- action: drop
- # Example Scrape config for pods which should be scraped slower. An useful example
- # would be stackriver-exporter which queries an API on every scrape of the pod
- #
- # The relabeling allows the actual pod scrape endpoint to be configured via the
- # following annotations:
- #
- # * `prometheus.io/scrape-slow`: Only scrape pods that have a value of `true`
- # * `prometheus.io/scheme`: If the metrics endpoint is secured then you will need
- # to set this to `https` & most likely set the `tls_config` of the scrape config.
- # * `prometheus.io/path`: If the metrics path is not `/metrics` override this.
- # * `prometheus.io/port`: Scrape the pod on the indicated port instead of the default of `9102`.
- - job_name: 'kubernetes-pods-slow'
- honor_labels: true
- scrape_interval: 5m
- scrape_timeout: 30s
- # kubernetes_sd_configs:
- # - role: pod
- relabel_configs:
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scrape_slow]
- action: keep
- regex: true
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scheme]
- action: replace
- regex: (https?)
- target_label: __scheme__
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_path]
- action: replace
- target_label: __metrics_path__
- regex: (.+)
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_port, __meta_kubernetes_pod_ip]
- action: replace
- regex: (\d+);(([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4})
- replacement: '[$2]:$1'
- target_label: __address__
- - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_port, __meta_kubernetes_pod_ip]
- action: replace
- regex: (\d+);((([0-9]+?)(\.|$)){4})
- replacement: $2:$1
- target_label: __address__
- - action: labelmap
- regex: __meta_kubernetes_pod_annotation_prometheus_io_param_(.+)
- replacement: __param_$1
- - action: labelmap
- regex: __meta_kubernetes_pod_label_(.+)
- - source_labels: [__meta_kubernetes_namespace]
- action: replace
- target_label: namespace
- - source_labels: [__meta_kubernetes_pod_name]
- action: replace
- target_label: pod
- - source_labels: [__meta_kubernetes_pod_phase]
- regex: Pending|Succeeded|Failed|Completed
- action: drop
- - job_name: 'mysql'
- static_configs:
- - targets:
- - 'mysql-metrics.observe.svc.cluster.local:9104'
- - job_name: 'clickhouse'
- static_configs:
- - targets:
- - 'clickhouse.observe.svc.cluster.local:8001'
- - job_name: 'otel-collector'
- static_configs:
- - targets:
- - 'otel-collector-opentelemetry-collector:8889'
- # 20240924 ADD ob-event metric
- - job_name: 'ob-event'
- static_configs:
- - targets:
- - 'ob-event.observe.svc.cluster.local'
- # adds additional scrape configs to prometheus.yml
- # must be a string so you have to add a | after extraScrapeConfigs:
- # example adds prometheus-blackbox-exporter scrape config
- #extraScrapeConfigs: ""
- #extraScrapeConfigs: |
- # - job_name: 'otel-collector'
- # honor_labels: true # 添加这个job后, 会向匹配到的数据添加job=otel-collector标签, 设置为true则不覆盖原来的标签, 否则原来的job将会被改成exported_job
- # static_configs:
- # - targets:
- # - 'otel-collector-opentelemetry-collector:8889'
- # relabel_configs:
- # - regex: (app_kubernetes_io_instance|app_kubernetes_io_name|component|controller_revision_hash|pod|pod_template_generation)
- # action: labeldrop
- # - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scrape_slow]
- # action: keep
- # regex: true
- # - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scheme]
- # action: replace
- # regex: (https?)
- # target_label: __scheme__
- # - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_path]
- # action: replace
- # target_label: __metrics_path__
- # regex: (.+)
- # - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_port, __meta_kubernetes_pod_ip]
- # action: replace
- # regex: (\d+);(([A-Fa-f0-9]{1,4}::?){1,7}[A-Fa-f0-9]{1,4})
- # replacement: '[$2]:$1'
- # target_label: __address__
- # - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_port, __meta_kubernetes_pod_ip]
- # action: replace
- # regex: (\d+);((([0-9]+?)(\.|$)){4})
- # replacement: $2:$1
- # target_label: __address__
- # - action: labelmap
- # regex: __meta_kubernetes_pod_annotation_prometheus_io_param_(.+)
- # replacement: __param_$1
- # - action: labelmap
- # regex: __meta_kubernetes_pod_label_(.+)
- # - source_labels: [__meta_kubernetes_namespace]
- # action: replace
- # target_label: namespace
- # - source_labels: [__meta_kubernetes_pod_name]
- # action: replace
- # target_label: pod
- # - source_labels: [__meta_kubernetes_pod_phase]
- # regex: Pending|Succeeded|Failed|Completed
- # action: drop
- # - job_name: 'prometheus-blackbox-exporter'
- # metrics_path: /probe
- # params:
- # module: [http_2xx]
- # static_configs:
- # - targets:
- # - https://example.com
- # relabel_configs:
- # - source_labels: [__address__]
- # target_label: __param_target
- # - source_labels: [__param_target]
- # target_label: instance
- # - target_label: __address__
- # replacement: prometheus-blackbox-exporter:9115
- # Adds option to add alert_relabel_configs to avoid duplicate alerts in alertmanager
- # useful in H/A prometheus with different external labels but the same alerts
- alertRelabelConfigs: {}
- # alert_relabel_configs:
- # - source_labels: [dc]
- # regex: (.+)\d+
- # target_label: dc
- networkPolicy:
- ## Enable creation of NetworkPolicy resources.
- ##
- enabled: false
- # Force namespace of namespaced resources
- forceNamespace: ""
- # Extra manifests to deploy as an array
- extraManifests: []
- # - apiVersion: v1
- # kind: ConfigMap
- # metadata:
- # labels:
- # name: prometheus-extra
- # data:
- # extra-data: "value"
- # Configuration of subcharts defined in Chart.yaml
- ## alertmanager sub-chart configurable values
- ## Please see https://github.com/prometheus-community/helm-charts/tree/main/charts/alertmanager
- ##
- alertmanager:
- ## If false, alertmanager will not be installed
- ##
- enabled: true
- persistence:
- size: 2Gi
- podSecurityContext:
- runAsUser: 65534
- runAsNonRoot: true
- runAsGroup: 65534
- fsGroup: 65534
- resources:
- limits:
- cpu: 100m
- memory: 256Mi
- requests:
- cpu: 100m
- memory: 128Mi
- ## kube-state-metrics sub-chart configurable values
- ## Please see https://github.com/prometheus-community/helm-charts/tree/main/charts/kube-state-metrics
- ##
- kube-state-metrics:
- ## If false, kube-state-metrics sub-chart will not be installed
- ##
- enabled: true
- resources:
- limits:
- cpu: 100m
- memory: 256Mi
- requests:
- cpu: 100m
- memory: 128Mi
- ## promtheus-node-exporter sub-chart configurable values
- ## Please see https://github.com/prometheus-community/helm-charts/tree/main/charts/prometheus-node-exporter
- ##
- prometheus-node-exporter:
- ## If false, node-exporter will not be installed
- ##
- enabled: false
- rbac:
- pspEnabled: false
- containerSecurityContext:
- allowPrivilegeEscalation: false
- resources:
- limits:
- cpu: 100m
- memory: 1024Mi
- requests:
- cpu: 100m
- memory: 256Mi
- ## pprometheus-pushgateway sub-chart configurable values
- ## Please see https://github.com/prometheus-community/helm-charts/tree/main/charts/prometheus-pushgateway
- ##
- prometheus-pushgateway:
- ## If false, pushgateway will not be installed
- ##
- enabled: true
- # Optional service annotations
- serviceAnnotations:
- prometheus.io/probe: pushgateway
- resources:
- limits:
- cpu: 100m
- memory: 256Mi
- requests:
- cpu: 100m
- memory: 128Mi
|