This is the solver.protoxt I used for training.
# The train/test net protocol buffer definition
net: "train_val.prototxt"
# The base learning rate, momentum and the weight decay of the network.
base_lr: 0.0001
lr_policy: "fixed"
momentum: 0.95
weight_decay: 0.0005
gamma:0.99
stepsize: 40
# Display every 100 iterations
display: 100
# The maximum number of iterations
max_iter: 65000
# snapshot intermediate results
snapshot: 10000
snapshot_prefix: "./train/save_"
# solver mode: CPU or GPU
solver_mode: GPU
type: "AdaDelta"
delta: 1e-8
The issue that I am facing is that the lr_rate remains constant throughout all the iterations. Although I do realize that this can be because I have kept lr_policy as fixed, I cannot remove the lr_policy ( Isnt the point of Adadelta that the learning rate be calculated during the learning process ?? ). Also I have noticed that despite keeping type as "AdaDelta" caffe calls sgd_solver.cpp. please suggest modifications in my solver so as to use adadelta for training.