-
Notifications
You must be signed in to change notification settings - Fork 706
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to restart the training JOB when one training process fails in cluster environment to recover the training? #2269
Comments
The current situation is that when a Pod encounters an exception, it can be automatically restarted. However, as for multi-node training, all Pods need to be restarted. How do I configure this YAML file? Thanks! |
The pytorchJOB YAML configuration file I used: |
I believe that the v2 Training Operator will solve this by utilizing JobSet Failure Policies. I'm unsure of the path for v1 (@tenzen-y @andreyvelich?). |
@kubeflow/wg-training-leads @kuizhiqing can we leverage the RestartPolicy API in V1 to restart all replica's pods in case of failure ? Yes, with V2, we can use FailurePolicy within JobSet. |
hi @kevinsummer219 , and how to use JobSet, I am sorry that I cannot find the doc for this @andreyvelich |
Hi @ltm920716, we are working on Kubeflow Training V2 API where that would be possible: https://github.com/kubeflow/training-operator/tree/master/docs/proposals/2170-kubeflow-training-v2 You can find the |
ok,very kind of you! |
What you would like to be added?
I use pytorchJOB.
Why is this needed?
I think, that is valid idea. So in such restart policy our controller should re-create all PyTorchJob's pods in case of single Pod failure.
Love this feature?
Give it a 👍 We prioritize the features with most 👍
The text was updated successfully, but these errors were encountered: