HhnsvetuK零S
这种情况排查的话,应该从哪入手呢
这种情况排查的话,应该从哪入手呢
容器内日志无报错信息,现在就是不清楚为什么这个会不停的自己启停,是什么事件触发的
补充一下详细信息 这是什么pod 具体作用是啥 是谁创建的
I0430 08:13:41.229132 1 scale_down.go:836] Scale-down: removing node ip-172-22-12-44.us-west-2.compute.internal, utilization: {0.3005181347150259 0.22000377567955187 0 cpu 0.3005181347150259}, pods to reschedule: ai-game/aigame-abtest-68cc5d6fd8-6zql6
I0430 08:13:41.230119 1 event.go:255] Event(v1.ObjectReference{Kind:"ConfigMap", Namespace:"ai-game", Name:"cluster-autoscaler-status", UID:"668d3962-5e57-4a06-8322-0148f5ed9d74", APIVersion:"v1", ResourceVersion:"166462868", FieldPath:""}): type: 'Normal' reason: 'ScaleDown' Scale-down: removing node ip-172-22-12-44.us-west-2.compute.internal, utilization: {0.3005181347150259 0.22000377567955187 0 cpu 0.3005181347150259}, pods to reschedule: ai-game/aigame-abtest-68cc5d6fd8-6zql6
I0430 08:13:41.247446 1 delete.go:102] Successfully added ToBeDeletedTaint on node ip-172-22-12-44.us-west-2.compute.internal
I0430 08:13:41.248742 1 event.go:255] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"ip-172-22-12-44.us-west-2.compute.internal", UID:"353daf10-0792-4608-b0ac-7df50eb0c0e9", APIVersion:"v1", ResourceVersion:"166462883", FieldPath:""}): type: 'Normal' reason: 'ScaleDown' marked the node as toBeDeleted/unschedulable
I0430 08:13:41.248826 1 event.go:255] Event(v1.ObjectReference{Kind:"Pod", Namespace:"ai-game", Name:"aigame-abtest-68cc5d6fd8-6zql6", UID:"d520213d-3740-4270-9c16-a5599df4ddce", APIVersion:"v1", ResourceVersion:"166458333", FieldPath:""}): type: 'Normal' reason: 'ScaleDown' deleting pod for node scale down
检查发现是k8s 调度的,这个有什么办法可以避免吗
环境是aws eks, nodegroup 配置了 ASG ,现在好像就是因为资源空余比较多,这个容器在两个物理节点反复横跳