I read that the combination of a generic and a domain specific LM improves the accuracy of a speech recognition model and therefore I wanted to create such a LM. However, I would like to give a higher weight on the domain specific LM.
I am using the mitlm tool to create language models and In the mitlm wiki branch I saw that there is a n-gram weighting technique which I am not sure if it does what I want. If this technique is for specifying the weight that you want to give on a language model, then I don't understand how it works.
If anybody is familiar with this tool and this technique, please let me know.
--
Go to http://kaldi-asr.org/forums.html find out how to join
---
You received this message because you are subscribed to the Google Groups "kaldi-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi-help+...@googlegroups.com.
To post to this group, send email to kaldi...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kaldi-help/6697d635-2028-483c-9b15-4c266594eabc%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
To view this discussion on the web visit https://groups.google.com/d/msgid/kaldi-help/CAFReZQbwfRGSRDpm%2BpabUDuh-8L98cUFVo0LkLsFTow1xJwHcw%40mail.gmail.com.
typically you find an interpolation weight that maximizes perplexity on a dev set.y.
On Tue, Jun 4, 2019 at 8:12 AM Itai Peer <ita...@gmail.com> wrote:
--
I don't familiar with this tools , but usually when a very easy to increase weight for portion of any training data for any purpose is to duplicate it , while leaving the other part as it is .As long of course that the training time is not an issue, and because LM training time is not so painful , you can easily do it .
בתאריך יום שלישי, 4 ביוני 2019 בשעה 13:30:24 UTC+3, מאת uRic Oresths:HelloI read that the combination of a generic and a domain specific LM improves the accuracy of a speech recognition model and therefore I wanted to create such a LM. However, I would like to give a higher weight on the domain specific LM.
I am using the mitlm tool to create language models and In the mitlm wiki branch I saw that there is a n-gram weighting technique which I am not sure if it does what I want. If this technique is for specifying the weight that you want to give on a language model, then I don't understand how it works.
If anybody is familiar with this tool and this technique, please let me know.
Go to http://kaldi-asr.org/forums.html find out how to join
---
You received this message because you are subscribed to the Google Groups "kaldi-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi...@googlegroups.com.
For example with the SRILM tool, there is the option to give weights to each language model that you want to merge (with interpolation).Isn't this a way to add bias somehow ?
I mean my goal was to not only merge two language models, but to give priority/higher weight to one of the language models ( the domain specific one )Your suggestion is to interpolate the domain specific text with a generic text and create a language model, and in addition to that to use half of the domain specific text as dev set ?