Multi-GPUs DDP - How the dataset is distributed accross the GPUs #13342
Answered
by
akihironitta
KevinCrp
asked this question in
DDP / multi-GPU / multi-node
-
Hi, I'm using some GPUs and the Distributed-Data-Parallel strategy and I want to know how the global dataset is split across all GPUs. Is it split iteratively:
Is it randomly distributed? Is it another way? I looked at the DistributedSampler class, but I didn't find the answer. |
Beta Was this translation helpful? Give feedback.
Answered by
akihironitta
Jun 21, 2022
Replies: 1 comment
-
I believe this line in PyTorch code explains it all: indices = indices[self.rank:self.total_size:self.num_replicas] |
Beta Was this translation helpful? Give feedback.
0 replies
Answer selected by
KevinCrp
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I believe this line in PyTorch code explains it all: