I am new to topic modeling and NLP in general. I have learned that perplexity is not correlated with human judgement but topic coherence metrics do. So I would like to measure coherence metrics on held-out data instead of perplexity. According to the article
Topic Coherence To Evaluate Topic Models, if I do not misunderstand, I will have to employ the extrinsic topic coherence metric on held-out data.
My first question is: Is c_v an extrinsic topic coherence metric?
My second question is: Why are some topic coherence measure results either nan or inf on held-out data?
For example:
CM = CoherenceModel(model=model, texts=test_data, dictionary=dictionary, coherence='c_uci')
CM.get_coherence_per_topic()
The result will be: [inf, inf, inf, inf, inf, inf, -13.283505155981141, inf, -13.999429177407952, inf]
Does this imply those topics with results inf do not present in the test data?
However, it looks like that in this graph, all topics present in the test data, except for the topic 10:

Thank you very much for your time and consideration.
Yours sincerely,
Vy Thuy Nguyen.