Hi,
I have a task to recognize people counting in English with a closed vocabulary (numbers 0-29) in a supermarket. I collected around 30 hours of field data and chose to fine-tune a pre-trained model due to time and resource limitations.
The fine-tuning worked well (2% WER compared to 26% WER with the original model), but one issue persists.
During lattice generation in training, I get numerous WARNINGS like: "Did not reach requested beam in determinize-lattice: size exceeds maximum 50000000 bytes..." even after reducing the beam to 4.
How can a small vocabulary cause such lattice growth, especially when the model achieves 2% WER?
Any ideas on the issue?
--
Go to http://kaldi-asr.org/forums.html to find out how to join the kaldi-help group
---
You received this message because you are subscribed to the Google Groups "kaldi-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi-help+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kaldi-help/3e6d7a5a-425b-4955-9ee8-197fa9058484n%40googlegroups.com.
Thank you for your response, Dan.
The scenario you described is indeed reflected in my training set. That said, the steps/nnet3/chain/train.py command requires the --lat-dir parameter, so I’m wondering how the training can proceed without providing the lattices. Is it possible to run the training process without generating lattices?
One workaround I’ve considered is to allow partial results (by setting --allow-partial=true when calling nnet3-latgen-faster in align_lats.sh) to prevent the issue from escalating.
--
Go to http://kaldi-asr.org/forums.html to find out how to join the kaldi-help group
---
You received this message because you are subscribed to the Google Groups "kaldi-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi-help+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kaldi-help/e8383c7b-3eab-40b6-a9df-912a1a0b0f07n%40googlegroups.com.