how is batch normalization layer working in multi-gpu mode?
217 views
Skip to first unread message
Etienne Perot
unread,
Feb 23, 2016, 10:47:06 AM2/23/16
Reply to author
Sign in to reply to author
Forward
Sign in to forward
Delete
You do not have permission to delete messages in this group
Copy link
Report message
Show original message
Either email addresses are anonymous for this group or you need the view member email addresses permission to view the original message
to Caffe Users
Hello,
I would like to know how BN layer accumulates mean & variance statistics in multi-gpu mode, and if it is possible to use the same principle when batches are too small for low-budget scenarii like Fully-Convolutional Mode? I am assuming Gradient Accumulation wouldn't do the job, since it is not analytically equivalent.
Prasanna S
unread,
Feb 10, 2017, 2:22:36 PM2/10/17
Reply to author
Sign in to reply to author
Forward
Sign in to forward
Delete
You do not have permission to delete messages in this group
Copy link
Report message
Show original message
Either email addresses are anonymous for this group or you need the view member email addresses permission to view the original message
to Caffe Users
I don't think mean and variance are learnable parameters and thus update might not be called on them. I am also wondering about the same thing. If you find something, do let u