Dear Andrew,
Yes, we're trying to better capture the clinical utility of algorithmic prescreening. Feedback on this year's scoring metric is more than welcome! We will revisit the scoring metric during next month's hiatus between the unofficial and official phases, so any feedback during the unofficial phase is especially helpful.
For the current scoring metric, the mean cost without algorithmic prescreening (i.e., the total cost c_0 divided by the number of patients) is 512 for the training set and a similar value for the validation set.
For comparison, the lowest possible mean cost with algorithmic prescreening (i.e., the total cost c_1 divided by the number of patients) is 311 for the training set and a similar value for the validation set.
These baseline scores are similar across the different partitions of the data because we partitioned the data so that the training, validation, and test sets have similar class prevalence rates.
As you noted, the current scoring metric does place a much higher cost on false negatives than on false positives, so a classifier that returns all positives (murmur present) does have a relatively low score. However, it is certainly possible for algorithms to achieve a lower/better cost.
Best,
Matt
(On behalf of the Challenge team.)
https://PhysioNetChallenges.org/https://PhysioNet.org/Please post questions and comments in the forum. However, if your question reveals information about your entry, then please email challenge at
physionet.org. We may post parts of our reply publicly if we feel that all Challengers should benefit from it. We will not answer emails about the Challenge to any other address. This email is maintained by a group. Please do not email us individually.