If I am training a classifier on an imbalanced dataset, I can either use an infogain loss to weight the less frequent class more heavily, or I can oversample the less frequent class to make a balanced data set.
Are these equivalent, or are there theoretical or empirical advantages to either of these approaches?
-- Eli