问题描述
我的k8s 1.12.8
集群(通过kops创建)已经运行了6个月以上。最近,某件事导致主节点上的kube-scheduler
和kube-controller-manager
都死亡并重新启动:
SyncLoop (PLEG): "kube-controller-manager-ip-x-x-x-x.z.compute.internal_kube-system(abc123)",event: &pleg.PodLifecycleEvent{ID:"abc123",Type:"ContainerDied",Data:"def456"}
hostname for pod:"kube-controller-manager-ip-x-x-x-x.z.compute.internal" was longer than 63. Truncated hostname to :"kube-controller-manager-ip-x-x-x-x.z.compute.inter"
SyncLoop (PLEG): "kube-scheduler-ip-x-x-x-x.z.compute.internal_kube-system(hij678)",event: &pleg.PodLifecycleEvent{ID:"hij678",Data:"890klm"}
SyncLoop (PLEG): "kube-controller-manager-ip-x-x-x-x.eu-west-2.compute.internal_kube-system(abc123)",Type:"ContainerStarted",Data:"def345"}
SyncLoop (container unhealthy): "kube-scheduler-ip-x-x-x-x.z.compute.internal_kube-system(hjk678)"
SyncLoop (PLEG): "kube-scheduler-ip-x-x-x-x.z.compute.internal_kube-system(ghj567)",event: &pleg.PodLifecycleEvent{ID:"ghj567",Data:"hjk768"}
自kube-scheduler
和kube-controller-manager
重新启动以来,kubelet完全无法获取或更新任何节点状态:
Error updating node status,will retry: failed to patch status "{"status":{"$setElementOrder/conditions":[{"type":"NetworkUnavailable"},{"type":"OutOfDisk"},{"type":"MemoryPressure"},{"type":"DiskPressure"},{"type":"PIDPressure"},{"type":"Ready"}],"conditions":[{"lastHeartbeatTime":"2020-08-12T09:22:08Z","type":"OutOfDisk"},{"lastHeartbeatTime":"2020-08-12T09:22:08Z","type":"MemoryPressure"},"type":"DiskPressure"},"type":"PIDPressure"},"type":"Ready"}]}}" for node "ip-172-20-60-88.eu-west-2.compute.internal": Patch https://127.0.0.1/api/v1/nodes/ip-172-20-60-88.eu-west-2.compute.internal/status?timeout=10s: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
Error updating node status,will retry: error getting node "ip-x-x-x-x.z.compute.internal": Get https://127.0.0.1/api/v1/nodes/ip-x-x-x-x.z.compute.internal?timeout=10s: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
Error updating node status,will retry: error getting node "ip-x-x-x-x.z.compute.internal": Get https://127.0.0.1/api/v1/nodes/ip-x-x-x-x.z.compute.internal?timeout=10s: context deadline exceeded
Error updating node status,will retry: error getting node "ip-x-x-x-x.z.compute.internal": Get https://127.0.0.1/api/v1/nodes/ip-x-x-x-x.z.compute.internal?timeout=10s: context deadline exceeded (Client.Timeout exceeded while awaiting headers)
Unable to update node status: update node status exceeds retry count
在此状态下,群集完全无法执行任何更新。
- 是什么会导致主节点失去与诸如 这个吗?
- 是第一个日志输出中的第二行'被截断 主机名.. 可能是问题的根源?
- 我该如何进一步 诊断是什么真正导致获取/更新节点操作 失败?
解决方法
我记得kubernetes将主机名限制为少于64个字符。这次是否有主机名更新的情况? 如果是这样,那么使用此文档来重建kubelet配置会很好 https://kubernetes.io/docs/tasks/administer-cluster/reconfigure-kubelet/