Reputation: 2585
For the normal BatchNorm
, the least batch size per GPU is 2.
I wonder if I use the SyncBatchNorm
, can I use batch_size=1
for every GPU with more than a single GPU?
I.e, the total_batch_size
is more than 1 but batch_size_per_gpu
is 1.
I would appreciate answers for any deep learning framework, pytorch, tensorflow, mxnet, etc
Upvotes: 1
Views: 295
Reputation: 2585
For PyTorch, using batch_size_per_gpu=1
and more than one GPU is fine.
Upvotes: 1