If I may ask for one more clarification, please.
I am using the old entropy_bottleneck layer, now as before, and I filter through it a 1-dimensional input (i.e. each batch runs an input of [N,128], where N is the minibatch size -- the number of elements of the dataset which is passed to the neural network in a single step).
If I have understood correctly, then, the command:
bits = tf.reduce_sum(tf.log(likelihoods), axis=1) / -np.log(2)
Returns a vector of dimensions Nx1, with the element [j,1] being the number of bits passed through the entropy bottleneck for the j-th element of the minibatch.
Have I understood this right?