Varun Suresh @varun-suresh May 19 09:43
I'm curious if anyone here has experimented with VGG's caffe model. I'm not able to recreate their results on LFW. I used the pre-trained model available on the caffe model zoo.
elbamos @elbamos May 19 11:29
Everyone has played with it and the reproducibility of the results is pretty well established.
Varun Suresh @varun-suresh May 20 16:38
@elbamos - This is the procedure I followed -
1) Used dlib's face detection and landmark detection to align the faces. I used Openface's code to do this. I resize the images to be (224,224,3)
2) For the cropped images, I calculate the VGG Face descriptor. I load the image, subtract the mean and then re-shape them to be (1,3,224,224) and feed that as the input to the deep neural net.
3) I generate the labels.csv and reps.csv like in Openface and use the lfw.py script to evaluate it. I have modified the lfw.py to normalize the 4096-dimensional vectors before calculating the L2 distance.
Please let me know if you did something differently. Thanks!
elbamos @elbamos May 20 16:44
@varun-suresh I can't really evaluate your method. But I have certainly used VGG many times, and I'm positive that it works
well, let me say that differently: I'm positive the model detects useful features in many cases, and I'm confident that the original work is reproducible research.
Varun Suresh @varun-suresh May 20 17:07
I'm certain it is reproducible, just trying to figure out what I am missing. For a sanity check, I looked at this (https://github.com/AKSHAYUBHAT/TensorFace/blob/master/vggface/torch_verify.ipynb) and verified I was calculating the descriptor correctly.
Do you use the 4096-dimensional vector or the embedding? Thanks.