Is a fixed k really all that useful for multilabel prediction/testing?
I'm looking into using Fasttext to classify newspaper articles by topic (eg "education", "crime", "economy", "sport" etc).
I've got about 20 labels in total in my training set, and any given article might have anywhere from 0 to 10 relevant labels, say.
So:
fasttext predict mymodel.bin articles.txt 10
returns 10 labels for _every_ article - even if they don't really deserve that many labels (most articles should really have only 2 or 3 labels at most).
Using "predict-prob" instead gives me probabilities with which I can cull out the obviously-wrong ones, which is fine, albeit a little more cumbersome..
But it seems to me that instead of having a fixed k, it'd be more useful to have a probability threshold instead.
(And forgive me if I'm missing some basic bit of text-classification lore here - I'm new to this stuff :- )
Ben.