Hello,
Thanks for taking the time to read.
I'd just like some verification w.r.t. scoring for the WSJ recipe. AFAICT, it looks like Kaldi uses its own internal functions to compute WER in the WSJ corpus rather than using sclite. Also AFAICT, Kaldi uses a fixed insertion, deletion, and substitution penalty when aligning reference to hypothesis in compute-wer. My understanding of the NIST standard (such as sclite) is to use the weights 3, 3, and 4 when aligning. So it looks like Kaldi will generate different word error rates than sclite. Am I correct? Did I mess up somewhere?
Thanks again for your time,
Sean