base_lr
in the solver prototxt, but boost the blobs_lr
on the newly introduced layer. The idea is to have the rest of the model change very slowly with new data, but let the new layer learn fast. Additionally, we set stepsize
in the solver to a lower value than if we were training from scratch, since we’re virtually far along in training and therefore want the learning rate to go down faster. Note that we could also entirely prevent fine-tuning of all layers other than fc8_flickr
by setting their blobs_lr
to 0."--
You received this message because you are subscribed to the Google Groups "Caffe Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to caffe-users...@googlegroups.com.
To post to this group, send email to caffe...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/caffe-users/d0dde4c0-5403-40e6-bd1d-3725b3543b72%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
To view this discussion on the web visit https://groups.google.com/d/msgid/caffe-users/CAAV6PX%3DgCK3ACfsPFxwcAAaxhJEnBwS-33ZwTq0EJqnH2u8QLw%40mail.gmail.com.