Hi
I am trying to bulk predict images (a million ) using bvlc_alexnet model using a single GTX 980 card.
The python interface seem to be pretty slow even when i enable gpu_mode (1 image per sec).
So i switched to c++ version using MemoryData layer with a batch size of 256...
I am able to classify about 5 images per sec with this ...
My Question is:
1) What classification speed can i expect for 256x256 images in GTX 980 with a 256 batch size?
Caffe
paper claims ~2.5ms per image using k40... can i expect similar speed ?
2) Can i expect the similar speed in python interface also with Oversample = False?
./Zahoor