今天有个节点磁盘io异常,想查看具体是因为哪个service引起的,然后发现pod没有磁盘io相关的监控指标,发现是把磁盘相关的指标丢弃掉,是为什么啊?
# kubectl get servicemonitor -n kubesphere-monitoring-system kubelet -o yaml
...
spec:
endpoints:
- bearerTokenFile: /var/run/secrets/kubernetes.io/serviceaccount/token
honorLabels: true
interval: 1m
metricRelabelings:
- action: keep
regex: kubelet_node_name|kubelet_running_container_count|kubelet_running_pod_count|kubelet_volume_stats.*|kubelet_pleg_relist_duration_seconds_.+
sourceLabels:
- __name__
port: https-metrics
relabelings:
- action: labeldrop
regex: (service|endpoint)
scheme: https
tlsConfig:
insecureSkipVerify: true
- bearerTokenFile: /var/run/secrets/kubernetes.io/serviceaccount/token
honorLabels: true
interval: 1m
metricRelabelings:
# 除了这些指标其他都丢弃掉,磁盘的也不要?
- action: keep
regex: container_cpu_usage_seconds_total|container_memory_usage_bytes|container_memory_cache|container_network_.+_bytes_total|container_memory_working_set_bytes|container_cpu_cfs_.*periods_total
sourceLabels:
- __name__
path: /metrics/cadvisor
port: https-metrics
...