- You have a well trained model of X different labels.
- You now want to add more Y labels, so you add your new labels data to your original dataset.
- Retrain the model with new 'fc8' layer (or any other last layer, model dependent) which now outputs X+Y values to the loss layer.
After following these steps, I find out that X labels accuracy outperforms the Y labels accuracy, which is reasonable of course, however I am wondering if there is a way to tell the back prop algorithm to finetune parts of the 'fc8' layer faster (much like using lt_mult param, but only for parts of that layer)