ts=2026-04-02T16:57:14.866Z caller=main.go:617 level=info msg="Starting Prometheus Server" mode=server version="(version=2.51.2, branch=HEAD, revision=b4c0ab52c3e9b940ab803581ddae9b3d9a452337)" ts=2026-04-02T16:57:14.866Z caller=main.go:622 level=info build_context="(go=go1.22.2, platform=linux/amd64, user=root@b63f02a423d9, date=20240410-14:05:54, tags=netgo,builtinassets,stringlabels)" ts=2026-04-02T16:57:14.866Z caller=main.go:623 level=info host_details="(Linux 5.15.0-174-generic #184-Ubuntu SMP Fri Mar 13 18:41:50 UTC 2026 x86_64 prometheus-kube-prometheus-stack-prometheus-0 (none))" ts=2026-04-02T16:57:14.866Z caller=main.go:624 level=info fd_limits="(soft=1048576, hard=1048576)" ts=2026-04-02T16:57:14.866Z caller=main.go:625 level=info vm_limits="(soft=unlimited, hard=unlimited)" ts=2026-04-02T16:57:14.870Z caller=web.go:568 level=info component=web msg="Start listening for connections" address=0.0.0.0:9090 ts=2026-04-02T16:57:14.870Z caller=main.go:1129 level=info msg="Starting TSDB ..." ts=2026-04-02T16:57:14.873Z caller=tls_config.go:313 level=info component=web msg="Listening on" address=[::]:9090 ts=2026-04-02T16:57:14.873Z caller=tls_config.go:352 level=info component=web msg="TLS is disabled." http2=false address=[::]:9090 ts=2026-04-02T16:57:14.878Z caller=head.go:616 level=info component=tsdb msg="Replaying on-disk memory mappable chunks if any" ts=2026-04-02T16:57:14.878Z caller=head.go:698 level=info component=tsdb msg="On-disk memory mappable chunks replay completed" duration=5.13µs ts=2026-04-02T16:57:14.878Z caller=head.go:706 level=info component=tsdb msg="Replaying WAL, this may take a while" ts=2026-04-02T16:57:14.886Z caller=head.go:778 level=info component=tsdb msg="WAL segment loaded" segment=0 maxSegment=0 ts=2026-04-02T16:57:14.886Z caller=head.go:815 level=info component=tsdb msg="WAL replay completed" checkpoint_replay_duration=43.112µs wal_replay_duration=8.658004ms wbl_replay_duration=210ns total_replay_duration=8.744127ms ts=2026-04-02T16:57:14.891Z caller=main.go:1150 level=info fs_type=EXT4_SUPER_MAGIC ts=2026-04-02T16:57:14.892Z caller=main.go:1153 level=info msg="TSDB started" ts=2026-04-02T16:57:14.892Z caller=main.go:1335 level=info msg="Loading configuration file" filename=/etc/prometheus/config_out/prometheus.env.yaml ts=2026-04-02T16:57:14.906Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=podMonitor/monitoring/openstack-database-exporter/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:14.906Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/kube-prometheus-stack-kube-scheduler/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:14.906Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/kube-prometheus-stack-prometheus/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:14.906Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/memcached/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:14.907Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/ingress-nginx-controller/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:14.907Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/keycloak/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:14.907Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/kube-prometheus-stack-apiserver/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:14.907Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=podMonitor/monitoring/ipmi-exporter/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:14.907Z caller=kubernetes.go:331 level=info component="discovery manager notify" discovery=kubernetes config=config-0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.042Z caller=main.go:1372 level=info msg="Completed loading of configuration file" filename=/etc/prometheus/config_out/prometheus.env.yaml totalDuration=150.857956ms db_storage=1.54µs remote_storage=1.89µs web_handler=680ns query_engine=1.26µs scrape=289.509µs scrape_sd=1.689174ms notify=19.621µs notify_sd=5.014138ms rules=130.087799ms tracing=21.56µs ts=2026-04-02T16:57:15.043Z caller=main.go:1114 level=info msg="Server is ready to receive web requests." ts=2026-04-02T16:57:15.043Z caller=main.go:1335 level=info msg="Loading configuration file" filename=/etc/prometheus/config_out/prometheus.env.yaml ts=2026-04-02T16:57:15.043Z caller=manager.go:163 level=info component="rule manager" msg="Starting rule manager..." ts=2026-04-02T16:57:15.062Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/kube-prometheus-stack-apiserver/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.062Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/kube-prometheus-stack-alertmanager/1 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.062Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/kube-prometheus-stack-kubelet/1 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.063Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=podMonitor/monitoring/rabbitmq/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.064Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=podMonitor/monitoring/ipmi-exporter/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.066Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/coredns/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.067Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/ingress-nginx-controller/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.067Z caller=kubernetes.go:331 level=info component="discovery manager scrape" discovery=kubernetes config=serviceMonitor/monitoring/keycloak/0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.068Z caller=kubernetes.go:331 level=info component="discovery manager notify" discovery=kubernetes config=config-0 msg="Using pod service account via in-cluster config" ts=2026-04-02T16:57:15.260Z caller=main.go:1372 level=info msg="Completed loading of configuration file" filename=/etc/prometheus/config_out/prometheus.env.yaml totalDuration=217.533617ms db_storage=2.47µs remote_storage=2.67µs web_handler=1.29µs query_engine=1.32µs scrape=165.416µs scrape_sd=6.444414ms notify=32.711µs notify_sd=488.465µs rules=191.926517ms tracing=10.09µs ts=2026-04-02T16:57:20.044Z caller=manager.go:171 level=error component="scrape manager" msg="error creating new scrape pool" err="error creating HTTP client: unable to load specified CA cert /certs/ca.crt: open /certs/ca.crt: no such file or directory" scrape_pool=serviceMonitor/monitoring/kube-prometheus-stack-prometheus-node-exporter/0 ts=2026-04-02T17:38:04.398Z caller=group.go:549 level=warn name=cluster:namespace:pod_memory:active:kube_pod_container_resource_limits index=4 component="rule manager" file=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0/monitoring-kube-prometheus-stack-k8s.rules.container-resource-58713baa-89cd-4ec3-b9eb-62ee3a40c6cf.yaml group=k8s.rules.container_resource msg="Error on ingesting out-of-order result from rule evaluation" numDropped=16 ts=2026-04-02T17:38:04.399Z caller=group.go:549 level=warn name=namespace_memory:kube_pod_container_resource_limits:sum index=5 component="rule manager" file=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0/monitoring-kube-prometheus-stack-k8s.rules.container-resource-58713baa-89cd-4ec3-b9eb-62ee3a40c6cf.yaml group=k8s.rules.container_resource msg="Error on ingesting out-of-order result from rule evaluation" numDropped=7 ts=2026-04-02T17:39:35.285Z caller=group.go:549 level=warn name=cluster:namespace:pod_memory:active:kube_pod_container_resource_limits index=4 component="rule manager" file=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0/monitoring-kube-prometheus-stack-k8s.rules.container-resource-58713baa-89cd-4ec3-b9eb-62ee3a40c6cf.yaml group=k8s.rules.container_resource msg="Error on ingesting out-of-order result from rule evaluation" numDropped=16 ts=2026-04-02T17:39:35.287Z caller=group.go:549 level=warn name=namespace_memory:kube_pod_container_resource_limits:sum index=5 component="rule manager" file=/etc/prometheus/rules/prometheus-kube-prometheus-stack-prometheus-rulefiles-0/monitoring-kube-prometheus-stack-k8s.rules.container-resource-58713baa-89cd-4ec3-b9eb-62ee3a40c6cf.yaml group=k8s.rules.container_resource msg="Error on ingesting out-of-order result from rule evaluation" numDropped=7