Thank you for making things more clear. I am trying to implement this algorithm in PyCaffe as well on ResNet-152 and there are issues.
I posted a separate question
here . This implementation is not working BTW.
Also i do not like how the author determines ReLU layers. "relu" in name is not very robust from my point of view. I tried the following
relu_layers = [(i, v) for i, v in enumerate(net.layers) if v.type.lower() == "relu"]
but i have issues with
net.blobs[layer_name] . Not all layers are present in
net.blobs.
I was also thinking about making custom ReLU layer with modified
backward() method ( a typical way to implement this algorithm in other frameworks like
theano ) but i need assistance on how to implement
setup() and
forward() as well since i cannot inherit from ReLU layer, only from generic Layer.