Training from snapshot not qualitatively the same (in accuracy) as continous training

31 views
Skip to first unread message

sarge

unread,
Jan 9, 2017, 2:07:24 PM1/9/17
to Caffe Users

I am seeking help concerning training Caffe AlexNet. A summary of the problem is that reported accuracy of continuous training of Caffe network (AlexNet on a randomly selected 200,000 ILSVRC 2012 images) differs from that of accuracy reported from resuming training from a snapshot of solver state.


The picture attached shows that reported accuracy differs based on number of iterations before snapshot and do not match training without snapshot. X-axis on the plot is the number of iterations (or number of images processed if you multiply each point by mini batch size of 100). The expectation is that all lines should match up irrespective of how many iterations before snapshot.


Please note that all experiments are setup the same way with same training/validation data and parameters (i.e., same seed, same mini batch size, fixed learning rate policy, zero momentum etc).


Thanks for your help.

caffe.png
Message has been deleted

Tian

unread,
Feb 25, 2019, 3:16:38 PM2/25/19
to Caffe Users
Weird. Can you share your solver file?
Reply all
Reply to author
Forward
0 new messages