question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Pods not assigned to nodes in autoscaling worker-pool

See original GitHub issue

Hi,

Thanks for your effort you are putting into this project! I was able to get pangeo running on a GCS cluster. I did pretty much everything as described in the tutorial and now my hub is running and I can log in and open a notebook. Now the problem: When I generate the dask-kubernetes-cluster with

from dask_kubernetes import KubeCluster
cluster = KubeCluster.from_yaml('../worker-template.yaml')
cluster.scale_up(5)

I just get 2 workers and as long as I wait it won’t become more. I can perform computations on those worker-pods though! When I check the pods in the cluster I can see, that the missing pods are not started due to missing cpu/memory resources. When I check the worker-pool, I can see that the resources are already allocated. So it seems, as if the autoscaling works, but the pods are somehow not distributed to the nodes?

Would be thankful for a hint!

Issue Analytics

  • State:closed
  • Created 5 years ago
  • Comments:5 (2 by maintainers)

github_iconTop GitHub Comments

1reaction
h4gencommented, Nov 1, 2018

@jacobtomlinson Thank you for your answer! I thought that autoscaling was enabled, because as you can see the events said, that the upscaling was triggered. Nevertheless I recreated the cluster and now it works like charme! Thank you very much for your help!

0reactions
jacobtomlinsoncommented, Nov 1, 2018

The key bit here is 0/3 nodes are available: 3 Insufficient cpu, 3 Insufficient memory..

This means you do not have enough cpu and memory on your compute nodes in your cluster to fulfill the request. The way to handle this is to scale you Kubernetes cluster up. If you are using GKE there should be an option to auto scale based on demand.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Deploying apps to worker pools where autoscaling is enabled
To limit a pod deployment to a specific worker pool that is managed by the cluster autoscaler, use a combination of labels and...
Read more >
Google Kubernetes: worker pool not scaling down to zero
My default-pool is a cheap g1-small to run the system pods (those never go away so the default pool can't autoscale to 0,...
Read more >
About cluster autoscaling | Google Kubernetes Engine (GKE)
If Pods are unschedulable because there are not enough nodes in the node pool, cluster autoscaler adds nodes, up to the maximum size...
Read more >
Chapter 1. Post-installation cluster tasks OpenShift Container ...
Create a default node selector so that pods without a node selector are assigned a subset of nodes to be deployed on, for...
Read more >
ReplicationController - Kubernetes
For example, your pods are re-created on a node after disruptive maintenance such as a kernel upgrade. For this reason, you should use...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found