Skip to content
This repository has been archived by the owner on Sep 19, 2022. It is now read-only.

whether multi-gpu-per-pod setup be supported in PytorchJob #331

Open
tingweiwu opened this issue Apr 25, 2021 · 1 comment
Open

whether multi-gpu-per-pod setup be supported in PytorchJob #331

tingweiwu opened this issue Apr 25, 2021 · 1 comment

Comments

@tingweiwu
Copy link

tingweiwu commented Apr 25, 2021

If there is 2GPU per node, how to set the Worker spec In the PytorchJob
1 replicas with 2GPU per pod or 2 replicas with only 1GPU per pod?

I've seen similar issues: #219 , but there is no clear instrunctions on whether multi-gpu-per-pod setup be supported in PytorchJob ?

is pytorch-operator designed for 1-gpu-per-pod setup even through there is multi-gpu on the same node?

will multi-gpu-per-pod setup be supported ?

@tingweiwu tingweiwu changed the title confusion about muti-gpu per one pod how to submit the PytorchJob when there is multi-gpus on the same node Apr 27, 2021
@tingweiwu tingweiwu changed the title how to submit the PytorchJob when there is multi-gpus on the same node whether multi-gpu-per-pod setup be supported in PytorchJob Apr 27, 2021
@wallarug
Copy link

Hey @tingweiwu ,

Did you ever get this sorted? I am struggling with the same issue.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants