Hi,
I was reviewing the lecture notes, and I had a few questions regarding the risk decomposition of an arbitrary function, visible in the image below: (Page 11)
that is, why exactly can we decompose the risk of an arbitrary classifier into these two terms, and what do each of them represent? I understand the derivation steps afterwards, but I just can't understand why that would represent an arbitrary risk.
Theoretically, given that the loss function is the zero-one loss for classification it should be equivalent/equal to the formulation of risk: (Page 10)
But I just don't see how these two functions express the same thing.
I'd appreciate any kind of clarification.