Thanks! That's pretty interesting. I was resizing the labelmap instead, which is dumb since a lot of deconvolution layers are not trained
However, I just fear that there is a "theoretical" issue for Spatial Batch Normalization end-to-end, that is that a lot of correlated inputs lead to distorted means & variances unless the batch size is pretty big. Then do you think something like additional dropout or something breaking the correlation could help?
What I see so far is that training from scratch with Batch Normalization, even with small batch size, works well for densely labeled datasets like CityScapes or Camvid, does that make sense?
Anyway, thanks again for your generous support.
Etienne
PS : Saw your conference @GTC, took a snapshot with my shitty webcam ^^
![](https://lh3.googleusercontent.com/-f6g-e21CWWU/VxSnKESdacI/AAAAAAAAAzU/l5P4HWP0gncpfJ41ljY_TkjXQZDoxqaogCLcB/s200/evan%2Bshelhamer.png)