问题描述
在Google容器引擎群集(GKE)上,有时会看到一个pod(或更多)未启动并查看其事件,我可以看到以下内容:
如果我等 - 重新尝试。
如果我删除了pod,并允许它们通过部署的副本集重新创建,它将正常启动。
行为是不一致的。
Kubernetes版本1.7.6和1.7.8
任何想法?
我可以在:
我们正在调查影响Google Container Engine(GKE)群集的问题,在Docker崩溃后或在节点上重新启动,豆荚不能被安排。
这个问题被认为影响了所有的GKE clust运行Kubernetes v1.6.11,v1.7.8和v1.8.1。
我们的工程团队建议:如果节点在版本v1.6.11上,请将您的节点降级到v1 .6.10。如果节点在版本v1.7.8上,请将您的节点降级到v1.7.6。如果节点在v1.8.1上,请将您的节点降级到v1.7.6。
工程小组还在。这些解决方法适用于无法降级节点的客户。
On a Google Container Engine cluster (GKE), I see sometimes a pod (or more) not starting and looking in its events, I can see the following
If I wait - it just keeps re-trying.
If I delete the pod, and allow them to be recreated by the Deployment's Replica Set, it will start properly.
The behavior is inconsistent.
Kubernetes versions 1.7.6 and 1.7.8
Any ideas?
I can see following message posted in Google Cloud Status Dashboard:
"We are investigating an issue affecting Google Container Engine (GKE) clusters where after docker crashes or is restarted on a node, pods are unable to be scheduled.
The issue is believed to be affecting all GKE clusters running Kubernetes v1.6.11, v1.7.8 and v1.8.1.
Our Engineering Team suggests: If nodes are on release v1.6.11, please downgrade your nodes to v1.6.10. If nodes are on release v1.7.8, please downgrade your nodes to v1.7.6. If nodes are on v1.8.1, please downgrade your nodes to v1.7.6.
Alternative workarounds are also provided by the Engineering team in this doc . These workarounds are applicable to the customers that are unable to downgrade their nodes."
这篇关于Kubernetes豆荚在“Pod沙箱改变,它将被杀死并重新创建”时失败。的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持!