-
Notifications
You must be signed in to change notification settings - Fork 28
Pending status for cow-job #202
Comments
Have you figured it out? Having the same problem of It seems like they aren't being assigned to the virtual-kubelets, despite ensuring the virtual kubelets have both labels:
|
Hi, yeah, you need to change the nodeSelector at your cow-job config file to one of the existing nodes (e.g., slurm-qpod3-cn01-cpn). |
@pisarukv I really appreciate the response. I assume you're referring to the "virtual-kubelet" node? The slurmjob's pod still isn't being assigned to any node, even after adding If it's not too much trouble, would you mind showing me the output for the following commands? Thanks again. |
Yes, I'm referring to the virtual-kubelet nodes. |
Many thanks! I think my issue may stem from the fact that I was running the k8s master and slurm master(with slurmctld) as the same node. I've set up a separate test env from out dev environment and got it working. Thanks again! |
I was trying to run the cow-job after setup environments by the following command:
vagrant up && vagrant ssh k8s-master
kubectl apply -f examples/cow.yaml
but when I run
kubectl get pods
my cow-job is "Pending":NAME READY STATUS RESTARTS AGE
cow-job 0/1 Pending 0 13s
wlm-operator-ffddd8795-lz98t 1/1 Running 0 16m
The text was updated successfully, but these errors were encountered: