You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
What does it mean that the workload is stuck? Do you mean the pods get created by cannot get scheduled?
If so then one option is to use waitForPodsReady. You can configure it to deactivate the workload after a couple of attempts. Then, you could have a small script to GC Deactivated workloads.
Unfortunately not, in our case it is only detected if the workload is not executed for a day. Then all events are already gone. My assumption is that maybe the workload is updated, but the unsuspending of the job fails because of an api server error. Is this possible?
woehrl01
changed the title
Stuck workload is not cleanedup/corretly handled
Stuck workload is not cleaned up/correctly handled
Feb 12, 2025
What happened:
In our cluster we irregularly find stuck workloads which are multiple days old.
What you expected to happen:
That the workload is scheduled or garbage collected
How to reproduce it (as minimally and precisely as possible):
Anything else we need to know?:
This is the workload which is "stuck":
Environment:
kubectl version
): 1.31git describe --tags --dirty --always
): v0.10.1cat /etc/os-release
):uname -a
):The text was updated successfully, but these errors were encountered: