Hello,
Say I have two corpora, corpus A and B. I'm trying to perform data augmentation by adding transformed speakers from B to speakers in A. To do this I've trained both corpora separately to tri3b systems (with the same lang directory). Within a larger script I estimate fMLLR transforms for speakers from B to speaker dependent models of A speakers, using gmm-est-fmllr. However, because this is between two different models I first need to convert the alignments between the models using convert-ali. Finally, I combine the new transformed B speakers with the original A, align and train a neural net.
What I've found is that, if A and B are trained with the same parameters (numleaves and totgauss), this works kind-of successfully (-5% relative). However, I tried to optimise the parameters for A with a held-out dev set, and performing the above again I then get a huge amount of warnings when aligning the combined set (1300 vs 100), and the resulting WER after neural net training is doubled. My suspicion is that this is because I'm abusing convert-ali, and I'm not sure how it actually deals with the two models having different trees.
Am I using convert-ali wrongly? Would it be better to try to constrain the B models to the trees generated in A (or vice versa), is that possible?
Cheers,
Joachim