Size gets shrinked anyway at pooling stages(and this zero padding through all those layers gets from 100 pixels to something about 3-5 pixels), network produces nothing at this boundaries, added by padding, so I really can't see why add it at the first layer(network at next layers padded also but just with 1-2 pixels to compensate convolutional kernel size). Upsampling here comes though deconvolutional layer.
My suggestion now is following, network output from itermidiate layers gets concatenated at later stages with deconv layers output, so padding added to match the sizes. But not sure about that, since nowhere before deconvolutional layers blobs do not cropped to fixed size, size depends from, so there is few other ways: add padding when needed instead of wasting memory from the start, or more precise strides calculation to match sizes. But it's just suggestion, not sure whether it's true.
четверг, 1 сентября 2016 г., 17:25:47 UTC+3 пользователь xdtl написал: