Failed to start ContainerManager
See original GitHub issueAll my pods end up stuck to Pending
kubectl get -n kube-system pods
NAME READY STATUS RESTARTS AGE
calico-kube-controllers-847c8c99d-t4kqm 0/1 Pending 0 164m
calico-node-rmc8z 0/1 Pending 0 164m
coredns-86f78bb79c-gr5bf 0/1 Pending 0 138m
hostpath-provisioner-5c65fbdb4f-bb698 0/1 Pending 0 137m
metrics-server-8bbfb4bdb-vnlcl 0/1 Pending 0 136m
kubernetes-dashboard-7ffd448895-69nf7 0/1 Pending 0 136m
dashboard-metrics-scraper-6c4568dc68-xxh2c 0/1 Pending 0 136m
Because the default node does not seem available
kubectl get events --all-namespaces | grep -i kubernetes-dashboard-7ffd448895-69nf7
kube-system 10s Warning FailedScheduling pod/kubernetes-dashboard-7ffd448895-69nf7 0/1 nodes are available: 1 node(s) had taint {node.kubernetes.io/not-ready: }, that the pod didn't tolerate.
It seems stuck in a weird loop
kubectl describe nodes
.....
Normal Starting 37s kubelet, bifrost Starting kubelet.
Warning InvalidDiskCapacity 37s kubelet, bifrost invalid capacity 0 on image filesystem
Normal NodeHasSufficientMemory 37s kubelet, bifrost Node bifrost status is now: NodeHasSufficientMemory
Normal Starting 30s kubelet, bifrost Starting kubelet.
Warning InvalidDiskCapacity 30s kubelet, bifrost invalid capacity 0 on image filesystem
Normal NodeHasSufficientPID 30s kubelet, bifrost Node bifrost status is now: NodeHasSufficientPID
Normal NodeHasNoDiskPressure 30s kubelet, bifrost Node bifrost status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientMemory 30s kubelet, bifrost Node bifrost status is now: NodeHasSufficientMemory
Normal Starting 24s kubelet, bifrost Starting kubelet.
.....
And I wonder whether has something to do with what I found in journalctl
microk8s.daemon-kubelet[1001446]: E0922 12:22:28.089018 1001446 kubelet.go:1765] skipping pod synchronization - container runtime status check may not have completed yet
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.091302 1001446 kubelet_node_status.go:70] Attempting to register node bifrost
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.102758 1001446 kubelet_node_status.go:108] Node bifrost was previously registered
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.102861 1001446 kubelet_node_status.go:73] Successfully registered node bifrost
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.103699 1001446 cpu_manager.go:184] [cpumanager] starting with none policy
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.103734 1001446 cpu_manager.go:185] [cpumanager] reconciling every 10s
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.103766 1001446 state_mem.go:36] [cpumanager] initializing new in-memory state store
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.103994 1001446 state_mem.go:88] [cpumanager] updated default cpuset: ""
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.104003 1001446 state_mem.go:96] [cpumanager] updated cpuset assignments: "map[]"
microk8s.daemon-kubelet[1001446]: I0922 12:22:28.104012 1001446 policy_none.go:43] [cpumanager] none policy: Start
microk8s.daemon-kubelet[1001446]: F0922 12:22:28.104037 1001446 kubelet.go:1296] Failed to start ContainerManager failed to get rootfs info: failed to get device for dir "/var/snap/microk8s/common/var/lib/kubelet": could not find device with major: 0, minor: 28 in cached partitions map
microk8s.daemon-kubelet[1001446]: goroutine 337 [running]:
microk8s.daemon-kubelet[1001446]: k8s.io/kubernetes/vendor/k8s.io/klog/v2.stacks(0xc000010001, 0xc00060d800, 0xfd, 0xfd)
microk8s.daemon-kubelet[1001446]: /build/microk8s/parts/k8s-binaries/build/go/src/github.com/kubernetes/kubernetes/_output/local/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:996 +0xb9
microk8s.daemon-kubelet[1001446]: k8s.io/kubernetes/vendor/k8s.io/klog/v2.(*loggingT).output(0x6ef8140, 0xc000000003, 0x0, 0x0, 0xc0005c0070, 0x6b48993, 0xa, 0x510, 0x0)
.....
How could I fix this? (If that’s the reason why it fails)
Issue Analytics
- State:
- Created 3 years ago
- Comments:11 (2 by maintainers)
Top Results From Across the Web
kubernetes - kubelet error: Failed to start ContainerManager ...
kubelet error: Failed to start ContainerManager failed to initialise top level QOS containers: root container /kubepods doesn't exist.
Read more >docker + crio: kubelet "Failed to start ContainerManager failed ...
Failed to start ContainerManager failed to initialise top level QOS containers kubernetes#43856. additional context: i've noticed that minikube (v1.20.0-beta.0, ...
Read more >1508040 – kubelet does not recover when "failed to start ...
Description of problem: The kubelet did not initialize modules properly, and it appeared the container manager failed to initialize or detect that the ......
Read more >Failed to start ContainerManager || Failed to create cgroup
OpenAPI AggregationController: action for item k8s_internal_local_delegation_chain_0000000000: Nothing (removed from the queue).
Read more >Kubernetes failing to start: failed to build map of initial containers
Starting today, k3s is failing to start with the following error: "Failed to start ContainerManager" err="failed to build map of initial ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
@zar3bski It looks like this is a feature gate [1] setup. You will need to configure kubelet with the respective feature gate [2] (
--feature-gates=....
). The kubelet arguments are placed in/var/snap/microk8s/current/args/kubelet
and after editing that file you will need to restart MicroK8s withmicrok8s.stop; microk8s.start;
.[1] https://kubernetes.io/docs/reference/command-line-tools-reference/feature-gates/ [2] https://kubernetes.io/docs/reference/command-line-tools-reference/kubelet/
https://github.com/kubernetes/kubernetes/pull/111513, https://github.com/kubernetes/enhancements/issues/361