Diff in debug info log: Is these values normal or a Vanishing gradient ?

41 views
Skip to first unread message

p.Paul

unread,
May 26, 2017, 12:55:06 PM5/26/17
to Caffe Users
Hello, I am trying a regression problem. My network is not converging , I tried different learning rates, momentum, weight decay. Nothing worked for me. Initially the loss reduced and is now oscillating around 30
The following is from iteration 154000.That is why you do not see the initial loss reduction.  The following is the debug info ..Is this values normal, or is it a vanishing gradient problem?  I have no clue what is going wrong. I tried with different network architectures(AlexNet,VGG16). I would also try something else. But it would be great if you could suggest something about this.


Thank you very much in advance!


I0526
18:42:20.706812 16062 net.cpp:647]     [Backward] Layer conv3, param blob 1 diff: 0.0118202
I0526
18:42:20.706897 16062 net.cpp:636]     [Backward] Layer pool2, bottom blob norm2 diff: 0.000126111
I0526
18:42:20.707118 16062 net.cpp:636]     [Backward] Layer norm2, bottom blob conv2 diff: 0.000126107
I0526
18:42:20.707175 16062 net.cpp:636]     [Backward] Layer relu2, bottom blob conv2 diff: 9.69918e-05
I0526
18:42:20.707670 16062 net.cpp:636]     [Backward] Layer conv2, bottom blob pool1 diff: 0.000442692
I0526
18:42:20.707736 16062 net.cpp:647]     [Backward] Layer conv2, param blob 0 diff: 0.000754526
I0526
18:42:20.707794 16062 net.cpp:647]     [Backward] Layer conv2, param blob 1 diff: 0.0180862
I0526
18:42:20.707870 16062 net.cpp:636]     [Backward] Layer pool1, bottom blob norm1 diff: 8.73312e-05
I0526
18:42:20.707976 16062 net.cpp:636]     [Backward] Layer norm1, bottom blob pconv1 diff: 8.73285e-05
I0526
18:42:20.708031 16062 net.cpp:636]     [Backward] Layer relu1, bottom blob pconv1 diff: 7.55169e-06
I0526
18:42:20.708721 16062 net.cpp:647]     [Backward] Layer pconv1, param blob 0 diff: 0.000126193
I0526
18:42:20.708776 16062 net.cpp:647]     [Backward] Layer pconv1, param blob 1 diff: 0.00693271




loss4.png

p.Paul

unread,
May 26, 2017, 12:56:55 PM5/26/17
to Caffe Users
If it is a vanishing gradient problem What should I do, I have already ReLus in my net! Please help

p.Paul

unread,
May 28, 2017, 3:12:43 PM5/28/17
to Caffe Users
Hello, any idea?
Message has been deleted

p.Paul

unread,
Jun 2, 2017, 4:17:56 AM6/2/17
to Caffe Users
Closing issue, as this was very specific to the problem. Check your data, network architecture and there are many other variables like, optimizer, pretraining , solver parameters etc.

On Friday, May 26, 2017 at 6:55:06 PM UTC+2, p.Paul wrote:
Reply all
Reply to author
Forward
0 new messages