如何在minikube上修复Cilium CrashLoopBackOff?

huangapple go评论129阅读模式
英文:

How to fix Cilium CrashLoopBackOff on minikube?

问题

I installed Meshery with Helm.
我使用Helm安装了Meshery。

All pods are ok except the Cilium.
除了Cilium之外,所有的Pod都正常。

Logs show:
日志显示:

I checked events also.
我还检查了事件。

I am adding logs from troubled pod.
我正在添加有问题的Pod的日志。

The Nginx pod shows network problems.
Nginx Pod显示有网络问题。

How to fix this issue?
如何解决这个问题?

英文:

I installed Meshery with Helm.

All pods are ok except the Cilium

  1. meshery-app-mesh-7d8cb57c85-cfk7j 1/1 Running 0 63m
  2. meshery-cilium-69b4f7d6c5-t9w55 0/1 CrashLoopBackOff 16 (3m16s ago) 63m
  3. meshery-consul-67669c97cc-hxvlz 1/1 Running 0 64m
  4. meshery-istio-687c7ff5bc-bn8mq 1/1 Running 0 63m
  5. meshery-kuma-66f947b988-887hm 1/1 Running 0 63m
  6. meshery-linkerd-854975dcb9-rsgtl 1/1 Running 0 63m
  7. meshery-nginx-sm-55675f57b5-xfpgz 1/1 Running 0 63m

Logs show

  1. level=info msg="Registering workloads with Meshery Server for version v1.14.0-snapshot.4" app=cilium-adapter
  2. panic: interface conversion: error is *fs.PathError, not *errors.Error
  3. goroutine 42 [running]:
  4. github.com/layer5io/meshkit/errors.GetCode({0x278e020?, 0xc0004f7e30?})

I checked events also

  1. kubectl get events --field-selector involvedObject.name=meshery-cilium-69b4f7d6c5-t9w55
  2. LAST SEEN TYPE REASON OBJECT MESSAGE
  3. 17m Normal Pulling pod/meshery-cilium-69b4f7d6c5-t9w55 Pulling image "layer5/meshery-cilium:stable-latest"
  4. 2m44s Warning BackOff pod/meshery-cilium-69b4f7d6c5-t9w55 Back-off restarting failed container meshery-cilium in pod meshery-cilium-69b4f7d6c5-t9w55_default(d7ccd0e8-27e5-4f40-89bc-f8a6dc8fa25a)

I am adding logs from troubled pod
Events:

  1. Type Reason Age From Message
  2. ---- ------ ---- ---- -------
  3. Normal Pulling 4h25m (x55 over 8h) kubelet Pulling image "layer5/meshery-cilium:stable-latest"
  4. Warning BackOff 3h45m (x1340 over 8h) kubelet Back-off restarting failed container meshery-cilium in pod meshery-cilium-69b4f7d6c5-t9w55_default(d7ccd0e8-27e5-4f40-89bc-f8a6dc8fa25a)
  5. Warning FailedMount 21m (x6 over 21m) kubelet MountVolume.SetUp failed for volume "kube-api-access-tcdml" : object "default"/"kube-root-ca.crt" not registered
  6. Warning NetworkNotReady 20m (x19 over 21m) kubelet network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
  7. Warning BackOff 72s (x82 over 19m) kubelet Back-off restarting failed container meshery-cilium in pod meshery-cilium-69b4f7d6c5-t9w55_default(d7ccd0e8-27e5-4f40-89bc-f8a6dc8fa25a)

The Nginx pod shows network problems

  1. Events:
  2. Type Reason Age From Message
  3. ---- ------ ---- ---- -------
  4. Warning NodeNotReady 22m node-controller Node is not ready
  5. Warning NetworkNotReady 22m (x18 over 23m) kubelet network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized
  6. Warning FailedMount 22m (x7 over 23m) kubelet MountVolume.SetUp failed for volume "kube-api-access-lscv2" : object "default"/"kube-root-ca.crt" not registered

How to fix this issue?

答案1

得分: 1

根据此博客由Patrick Londa所述,您的cilium Pod因“Back-off restarting failed container”错误而发生CrashLoopBack off错误,这意味着在Kubernetes启动容器后,您的容器突然终止。<br><br>1. 这是由于活动激增导致的临时资源超载而发生的。修复此问题的方法是调整periodSeconds或timeoutSeconds,以为应用程序提供更长的响应时间窗口。<br><br>2. 有时,这是由于内存资源不足引起的。您可以通过在容器的资源清单中更改**“resources:limits”**来增加内存限制,如博客中所示。<br><br>请验证上述步骤,并告诉我是否解决了问题。<br>

英文:

As per this Blog by Patrick Londa, your cilium Pod is getting CrashLoopBack off error due to the &quot;Back-off restarting failed container&quot;, this means that your container suddenly terminated after Kubernetes started it. <br><br>1. This occurs due to the temporary resource overload as a result of a spike in activity. The solution to fix this is to adjust periodSeconds or timeoutSeconds to give the application a longer window of time to respond. <br><br>2. Sometimes this happens due to insufficient memory resources. You can increase the memory limit by changing the "resources:limits" in the Container's resource manifest as shown in the blog. <br><br>Verify the above steps and let me know if this fixes the issue.<br>

huangapple
  • 本文由 发表于 2023年7月13日 13:13:24
  • 转载请务必保留本文链接:https://go.coder-hub.com/76676124.html
匿名

发表评论

匿名网友

:?: :razz: :sad: :evil: :!: :smile: :oops: :grin: :eek: :shock: :???: :cool: :lol: :mad: :twisted: :roll: :wink: :idea: :arrow: :neutral: :cry: :mrgreen:

确定