![](https://lh4.googleusercontent.com/-gKYXADZW_CY/VOyRh3vM7RI/AAAAAAAAAWc/7MvmNdPFKdc/s400/untitled.jpg)
![](https://lh5.googleusercontent.com/-dyuB2S-p-_k/VOyRTEf8geI/AAAAAAAAAWU/yBOGMxilz3I/s1600/matNet.png)
Hi Everyone,
I seem to be in a bit of a fix. I was trying to test caffe with HDF5 image classification and it was not converging. So, to eliminate variables and pin point potential issues, I intended to do a comparison with the matlab nn toolbox. To this end, I extracted an 1764 length HOG descriptor for each 64x64 sized input images. The total number of samples was 23667 images. Then I split it into train and validation ratio of 80:20.
Then I saved the images and labels into two HDF5 files for training and testing respectively.
Download data from onedrive. (~127 MB).
I used the same data to train the nn toolbox in matlab. The following is the image of the network. (1764 Input, 100 Tanh hidden, 8 linear output) And I am also attaching the confusion matrices. It can be seen that the net achieves around ~80% accuracy.
Now I ran the same experiment with caffe. I am attaching the solver prototext, the network file and also the stderr output. However after 70000 iterations the network only achieves 35% accuracy which is a far cry from the matlab toolbox accuracy.
What is going wrong?
I had to do this because the convnet was not converging on the original images. Now I see that even a simple MLP is also not converging. I have tried lowering the learning rate. I have tried both gaussian and xavier weight fillers, I also tried both TanH and ReLU activations. But it is not improving.
I have a paper deadline looming and I need to do this comparison fast. Any help in this regard would be appreciated.
Kind Regards,
Rick