Skip to content

Effective batch size in DDP #13165

Discussion options

You must be logged in to vote

I have solved my problem and find out that the answer is: each gpu get #batch_size batches. If you have batch_size of 2 and 2 gpus are utilized, each gpu gets 2 batches and 4 batches in total are feed into a forward pass.

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@SagiPolaczek
Comment options

@Ki-Zhang
Comment options

Answer selected by pengzhangzhi
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
3 participants