Targets


monitoring/kube-prometheus-stack-alertmanager/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.125.20:9093/metrics
up endpoint="web" instance="10.233.125.20:9093" job="kube-prometheus-stack-alertmanager" namespace="monitoring" pod="alertmanager-kube-prometheus-stack-alertmanager-0" service="kube-prometheus-stack-alertmanager" 24.232s ago 3.151ms

monitoring/kube-prometheus-stack-apiserver/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.66.12:6443/metrics
up endpoint="https" instance="192.168.66.12:6443" job="apiserver" namespace="default" service="kubernetes" 11.674s ago 149.9ms
https://192.168.66.13:6443/metrics
up endpoint="https" instance="192.168.66.13:6443" job="apiserver" namespace="default" service="kubernetes" 28.045s ago 81.7ms
https://192.168.66.14:6443/metrics
up endpoint="https" instance="192.168.66.14:6443" job="apiserver" namespace="default" service="kubernetes" 6.82s ago 64.77ms

monitoring/kube-prometheus-stack-coredns/0 (2/2 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.108.39:9153/metrics
up endpoint="http-metrics" instance="10.233.108.39:9153" job="coredns" namespace="kube-system" pod="coredns-74c9d4d795-ttrdx" service="kube-prometheus-stack-coredns" 19.955s ago 3.54ms
http://10.233.86.46:9153/metrics
up endpoint="http-metrics" instance="10.233.86.46:9153" job="coredns" namespace="kube-system" pod="coredns-74c9d4d795-4ffx9" service="kube-prometheus-stack-coredns" 18.56s ago 3.608ms

monitoring/kube-prometheus-stack-grafana/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.125.19:3000/metrics
up endpoint="service" instance="10.233.125.19:3000" job="kube-prometheus-stack-grafana" namespace="monitoring" pod="kube-prometheus-stack-grafana-6b7cb99c84-l2ch6" service="kube-prometheus-stack-grafana" 24.481s ago 3.127ms

monitoring/kube-prometheus-stack-kube-controller-manager/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://192.168.66.12:10252/metrics
up endpoint="http-metrics" instance="192.168.66.12:10252" job="kube-controller-manager" namespace="kube-system" pod="kube-controller-manager-preprod-master-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kube-controller-manager" 1.69s ago 3.484ms
http://192.168.66.13:10252/metrics
up endpoint="http-metrics" instance="192.168.66.13:10252" job="kube-controller-manager" namespace="kube-system" pod="kube-controller-manager-preprod-master-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kube-controller-manager" 19.079s ago 17.83ms
http://192.168.66.14:10252/metrics
up endpoint="http-metrics" instance="192.168.66.14:10252" job="kube-controller-manager" namespace="kube-system" pod="kube-controller-manager-preprod-master-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kube-controller-manager" 9.009s ago 2.568ms

monitoring/kube-prometheus-stack-kube-etcd/0 (0/0 up)

Endpoint State Labels Last Scrape Scrape Duration Error

monitoring/kube-prometheus-stack-kube-proxy/0 (0/11 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://192.168.66.11:10249/metrics
down endpoint="http-metrics" instance="192.168.66.11:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-rfb9h" service="kube-prometheus-stack-kube-proxy" 13.074s ago 708.4us Get "http://192.168.66.11:10249/metrics": dial tcp 192.168.66.11:10249: connect: connection refused
http://192.168.66.12:10249/metrics
down endpoint="http-metrics" instance="192.168.66.12:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-vgj7t" service="kube-prometheus-stack-kube-proxy" 10.08s ago 656.6us Get "http://192.168.66.12:10249/metrics": dial tcp 192.168.66.12:10249: connect: connection refused
http://192.168.66.13:10249/metrics
down endpoint="http-metrics" instance="192.168.66.13:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-fv5pw" service="kube-prometheus-stack-kube-proxy" 11.433s ago 607.7us Get "http://192.168.66.13:10249/metrics": dial tcp 192.168.66.13:10249: connect: connection refused
http://192.168.66.14:10249/metrics
down endpoint="http-metrics" instance="192.168.66.14:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-mn48v" service="kube-prometheus-stack-kube-proxy" 9.407s ago 651.5us Get "http://192.168.66.14:10249/metrics": dial tcp 192.168.66.14:10249: connect: connection refused
http://192.168.66.16:10249/metrics
down endpoint="http-metrics" instance="192.168.66.16:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-287wc" service="kube-prometheus-stack-kube-proxy" 11.325s ago 423.3us Get "http://192.168.66.16:10249/metrics": dial tcp 192.168.66.16:10249: connect: connection refused
http://192.168.66.18:10249/metrics
down endpoint="http-metrics" instance="192.168.66.18:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-fpwhw" service="kube-prometheus-stack-kube-proxy" 1.507s ago 655.4us Get "http://192.168.66.18:10249/metrics": dial tcp 192.168.66.18:10249: connect: connection refused
http://192.168.66.21:10249/metrics
down endpoint="http-metrics" instance="192.168.66.21:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-262mg" service="kube-prometheus-stack-kube-proxy" 9.547s ago 726.9us Get "http://192.168.66.21:10249/metrics": dial tcp 192.168.66.21:10249: connect: connection refused
http://192.168.66.22:10249/metrics
down endpoint="http-metrics" instance="192.168.66.22:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-gwv5n" service="kube-prometheus-stack-kube-proxy" 21.523s ago 675.5us Get "http://192.168.66.22:10249/metrics": dial tcp 192.168.66.22:10249: connect: connection refused
http://192.168.66.24:10249/metrics
down endpoint="http-metrics" instance="192.168.66.24:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-bl779" service="kube-prometheus-stack-kube-proxy" 17.192s ago 815.9us Get "http://192.168.66.24:10249/metrics": dial tcp 192.168.66.24:10249: connect: connection refused
http://192.168.66.25:10249/metrics
down endpoint="http-metrics" instance="192.168.66.25:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-jxv7h" service="kube-prometheus-stack-kube-proxy" 2.754s ago 671.4us Get "http://192.168.66.25:10249/metrics": dial tcp 192.168.66.25:10249: connect: connection refused
http://192.168.66.26:10249/metrics
down endpoint="http-metrics" instance="192.168.66.26:10249" job="kube-proxy" namespace="kube-system" pod="kube-proxy-77gzh" service="kube-prometheus-stack-kube-proxy" 20.437s ago 746.4us Get "http://192.168.66.26:10249/metrics": dial tcp 192.168.66.26:10249: connect: connection refused

monitoring/kube-prometheus-stack-kube-scheduler/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://192.168.66.12:10251/metrics
up endpoint="http-metrics" instance="192.168.66.12:10251" job="kube-scheduler" namespace="kube-system" pod="kube-scheduler-preprod-master-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kube-scheduler" 29.172s ago 3.783ms
http://192.168.66.13:10251/metrics
up endpoint="http-metrics" instance="192.168.66.13:10251" job="kube-scheduler" namespace="kube-system" pod="kube-scheduler-preprod-master-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kube-scheduler" 17.717s ago 4.369ms
http://192.168.66.14:10251/metrics
up endpoint="http-metrics" instance="192.168.66.14:10251" job="kube-scheduler" namespace="kube-system" pod="kube-scheduler-preprod-master-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kube-scheduler" 26.015s ago 3.927ms

monitoring/kube-prometheus-stack-kube-state-metrics/0 (0/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.125.21:8080/metrics
down endpoint="http" instance="10.233.125.21:8080" job="kube-state-metrics" namespace="monitoring" pod="kube-prometheus-stack-kube-state-metrics-857d997b65-r44sj" service="kube-prometheus-stack-kube-state-metrics" 1.82s ago 373.9us Get "http://10.233.125.21:8080/metrics": dial tcp 10.233.125.21:8080: connect: connection refused

monitoring/kube-prometheus-stack-kubelet/0 (11/11 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.66.11:10250/metrics
up endpoint="https-metrics" instance="192.168.66.11:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-worker-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 4.86s ago 22.04ms
https://192.168.66.12:10250/metrics
up endpoint="https-metrics" instance="192.168.66.12:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-master-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 9.036s ago 13.29ms
https://192.168.66.13:10250/metrics
up endpoint="https-metrics" instance="192.168.66.13:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-master-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 29.123s ago 8.176ms
https://192.168.66.14:10250/metrics
up endpoint="https-metrics" instance="192.168.66.14:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-master-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 16.057s ago 7.85ms
https://192.168.66.16:10250/metrics
up endpoint="https-metrics" instance="192.168.66.16:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-worker-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 7.845s ago 19.76ms
https://192.168.66.18:10250/metrics
up endpoint="https-metrics" instance="192.168.66.18:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-worker-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 23.853s ago 9.458ms
https://192.168.66.21:10250/metrics
up endpoint="https-metrics" instance="192.168.66.21:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-worker-4.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 13.455s ago 16.72ms
https://192.168.66.22:10250/metrics
up endpoint="https-metrics" instance="192.168.66.22:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-worker-5.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 27.878s ago 9.548ms
https://192.168.66.24:10250/metrics
up endpoint="https-metrics" instance="192.168.66.24:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-ingress-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 20.726s ago 16.79ms
https://192.168.66.25:10250/metrics
up endpoint="https-metrics" instance="192.168.66.25:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-worker-6.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 3.135s ago 41.56ms
https://192.168.66.26:10250/metrics
up endpoint="https-metrics" instance="192.168.66.26:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="preprod-ingress-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 13.014s ago 34.73ms

monitoring/kube-prometheus-stack-kubelet/1 (11/11 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.66.11:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.11:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-worker-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 21.732s ago 246.1ms
https://192.168.66.12:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.12:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-master-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 26.26s ago 46.18ms
https://192.168.66.13:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.13:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-master-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 15.028s ago 39.24ms
https://192.168.66.14:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.14:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-master-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 6.474s ago 82.63ms
https://192.168.66.16:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.16:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-worker-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 6.538s ago 63.8ms
https://192.168.66.18:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.18:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-worker-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 15.354s ago 82.07ms
https://192.168.66.21:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.21:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-worker-4.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 28.49s ago 83.75ms
https://192.168.66.22:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.22:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-worker-5.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 22.595s ago 121.3ms
https://192.168.66.24:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.24:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-ingress-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 17.934s ago 76.3ms
https://192.168.66.25:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.25:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-worker-6.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 6.44s ago 241.3ms
https://192.168.66.26:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.66.26:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="preprod-ingress-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 6.275s ago 54.67ms

monitoring/kube-prometheus-stack-kubelet/2 (11/11 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.66.11:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.11:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-worker-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 27.717s ago 3.075ms
https://192.168.66.12:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.12:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-master-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 25.621s ago 3.113ms
https://192.168.66.13:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.13:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-master-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 22.673s ago 1.732ms
https://192.168.66.14:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.14:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-master-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 12.563s ago 1.301ms
https://192.168.66.16:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.16:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-worker-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 17.41s ago 1.162ms
https://192.168.66.18:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.18:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-worker-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 4.912s ago 1.566ms
https://192.168.66.21:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.21:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-worker-4.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 5.052s ago 1.505ms
https://192.168.66.22:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.22:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-worker-5.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 6.242s ago 2.003ms
https://192.168.66.24:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.24:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-ingress-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 21.399s ago 1.441ms
https://192.168.66.25:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.25:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-worker-6.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 28.03s ago 2.737ms
https://192.168.66.26:10250/metrics/probes
up endpoint="https-metrics" instance="192.168.66.26:10250" job="kubelet" metrics_path="/metrics/probes" namespace="kube-system" node="preprod-ingress-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 14.059s ago 1.746ms

monitoring/kube-prometheus-stack-kubelet/3 (11/11 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.66.11:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.11:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-worker-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 28.909s ago 24.2ms
https://192.168.66.12:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.12:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-master-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 8.351s ago 4.18ms
https://192.168.66.13:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.13:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-master-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 484ms ago 1.869ms
https://192.168.66.14:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.14:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-master-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 18.905s ago 7.554ms
https://192.168.66.16:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.16:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-worker-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 19.414s ago 10.7ms
https://192.168.66.18:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.18:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-worker-3.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 6.217s ago 10.97ms
https://192.168.66.21:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.21:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-worker-4.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 23.129s ago 14.57ms
https://192.168.66.22:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.22:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-worker-5.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 24.541s ago 17.29ms
https://192.168.66.24:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.24:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-ingress-2.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 1.101s ago 2.065ms
https://192.168.66.25:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.25:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-worker-6.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 8.212s ago 4.242ms
https://192.168.66.26:10250/metrics/resource/v1alpha1
up endpoint="https-metrics" instance="192.168.66.26:10250" job="kubelet" metrics_path="/metrics/resource/v1alpha1" namespace="kube-system" node="preprod-ingress-1.preprod.pmc.vas-stream.ru" service="kube-prometheus-stack-kubelet" 26.079s ago 1.901ms

monitoring/kube-prometheus-stack-node-exporter/0 (11/11 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://192.168.66.11:9100/metrics
up endpoint="metrics" instance="192.168.66.11:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-5vj9n" service="kube-prometheus-stack-prometheus-node-exporter" 5.636s ago 38.96ms
http://192.168.66.12:9100/metrics
up endpoint="metrics" instance="192.168.66.12:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-5glnz" service="kube-prometheus-stack-prometheus-node-exporter" 2.411s ago 22.59ms
http://192.168.66.13:9100/metrics
up endpoint="metrics" instance="192.168.66.13:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-j5wrx" service="kube-prometheus-stack-prometheus-node-exporter" 5.689s ago 22.11ms
http://192.168.66.14:9100/metrics
up endpoint="metrics" instance="192.168.66.14:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-vgtxj" service="kube-prometheus-stack-prometheus-node-exporter" 3.884s ago 26.77ms
http://192.168.66.16:9100/metrics
up endpoint="metrics" instance="192.168.66.16:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-fbthg" service="kube-prometheus-stack-prometheus-node-exporter" 18.507s ago 23.84ms
http://192.168.66.18:9100/metrics
up endpoint="metrics" instance="192.168.66.18:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-rknlq" service="kube-prometheus-stack-prometheus-node-exporter" 2.814s ago 23.17ms
http://192.168.66.21:9100/metrics
up endpoint="metrics" instance="192.168.66.21:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-cvnh6" service="kube-prometheus-stack-prometheus-node-exporter" 18.417s ago 31.98ms
http://192.168.66.22:9100/metrics
up endpoint="metrics" instance="192.168.66.22:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-fh7dg" service="kube-prometheus-stack-prometheus-node-exporter" 5.476s ago 36.61ms
http://192.168.66.24:9100/metrics
up endpoint="metrics" instance="192.168.66.24:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-gs8gk" service="kube-prometheus-stack-prometheus-node-exporter" 6.41s ago 29.12ms
http://192.168.66.25:9100/metrics
up endpoint="metrics" instance="192.168.66.25:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-92w7w" service="kube-prometheus-stack-prometheus-node-exporter" 29.699s ago 44.87ms
http://192.168.66.26:9100/metrics
up endpoint="metrics" instance="192.168.66.26:9100" job="node-exporter" namespace="monitoring" pod="kube-prometheus-stack-prometheus-node-exporter-7p564" service="kube-prometheus-stack-prometheus-node-exporter" 8.687s ago 26.1ms

monitoring/kube-prometheus-stack-operator/0 (0/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.125.18:8080/metrics
down endpoint="http" instance="10.233.125.18:8080" job="kube-prometheus-stack-operator" namespace="monitoring" pod="kube-prometheus-stack-operator-5b48747dff-jpzxx" service="kube-prometheus-stack-operator" 15.331s ago 456.9us Get "http://10.233.125.18:8080/metrics": dial tcp 10.233.125.18:8080: connect: connection refused

monitoring/kube-prometheus-stack-prometheus/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.125.23:9090/metrics
up endpoint="web" instance="10.233.125.23:9090" job="kube-prometheus-stack-prometheus" namespace="monitoring" pod="prometheus-kube-prometheus-stack-prometheus-0" service="kube-prometheus-stack-prometheus" 20.816s ago 9.935ms

monitoring/partner-metrics-exporter/0 (0/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.111.251:9913/metrics
down endpoint="metrics" instance="10.233.111.251:9913" job="pmc-partner-api" namespace="pmc-production" pod="pmc-partner-api-68dd884bb6-2spn9" service="pmc-partner-api" tier="pmc-metrics-api" 17.593s ago 812us Get "http://10.233.111.251:9913/metrics": dial tcp 10.233.111.251:9913: connect: connection refused

monitoring/rabbit-metrics-exporter/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.110.210:15672/api/metrics
up endpoint="http" instance="10.233.110.210:15672" job="rabbitmq-internal" namespace="default" pod="rabbitmq-0" service="rabbitmq-internal" tier="pmc-rabbit" 27.067s ago 265.2ms
http://10.233.111.210:15672/api/metrics
up endpoint="http" instance="10.233.111.210:15672" job="rabbitmq-internal" namespace="default" pod="rabbitmq-1" service="rabbitmq-internal" tier="pmc-rabbit" 26.502s ago 305.3ms
http://10.233.126.129:15672/api/metrics
up endpoint="http" instance="10.233.126.129:15672" job="rabbitmq-internal" namespace="default" pod="rabbitmq-2" service="rabbitmq-internal" tier="pmc-rabbit" 27.171s ago 283.7ms

monitoring/statistic-metrics-exporter/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.233.110.65:8080/metrics
up endpoint="http" instance="10.233.110.65:8080" job="pmc-statistic-metrics" namespace="pmc-production" pod="pmc-statistic-metrics-7469b6f694-tx7l5" service="pmc-statistic-metrics" tier="pmc-statistic" 18.141s ago 25.9ms
http://10.233.111.189:8080/metrics
up endpoint="http" instance="10.233.111.189:8080" job="pmc-statistic-metrics" namespace="pmc-production" pod="pmc-statistic-metrics-7469b6f694-qr68r" service="pmc-statistic-metrics" tier="pmc-statistic" 14.567s ago 43.43ms
http://10.233.115.200:8080/metrics
up endpoint="http" instance="10.233.115.200:8080" job="pmc-statistic-metrics" namespace="pmc-production" pod="pmc-statistic-metrics-7469b6f694-bmpsn" service="pmc-statistic-metrics" tier="pmc-statistic" 23.085s ago 31.58ms