In this
Google's tuto, when they define the prior and posterior, they use a VariableLayer. From what I understood, this layer simply returns a trainable constant (the bias terms?) as output...
I think part of the objective pf the VariableLayer is to make the prior and posterior trainable.
1. How exactly does the VariableLayer helps us making them trainable? From the sequential way of building, it seems we're just passing some constants to the LambdaLayers...
2. Couldn't we just have used a usual Dense layer (by imposing a null kernel), instead of the VariableLayer?
3. What's the point in making the prior trainable?
Kind regards.