Hi Jonathan,
Thank you for your time and detailed reply.
What I understand from the link and google is:
SoftmaxWithLoss = -( L*log [(softmax(p)] + (1-L) log [(softmax(p)] )
SigmoidCrossEntropy = ( L*log [(Sign(p)] + (1-L) log [(Sign(p)] )
where, L is label and P logits from network
So basically they compute loss by same means except SoftmaxWithLoss applies softmax function on logits and SigmoidCrossEntropy applies signum function on logits before loss calculation.
Please correct me if my understanding is wrong.
Thank you.
Regards,
Anand