--
Go to http://kaldi-asr.org/forums.html find out how to join
---
You received this message because you are subscribed to the Google Groups "kaldi-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi-help+...@googlegroups.com.
To post to this group, send email to kaldi...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kaldi-help/67c3ec69-412e-488d-8391-3541f3920ffc%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
The reason we use MFCC is because they are more easily compressible, being decorrelated; we dump them to diskwith compression to 1 byte per coefficient. But we dump all the coefficients, so it's equivalent to filterbanks timesa full-rank matrix, no information is lost. For convolutional architectures, we convert them back into filterbanks inside thenetwork; see idct-layer.
Dan
On Sun, Jul 14, 2019 at 11:19 AM Titouan Parcollet <parcolle...@gmail.com> wrote:
Hi there!--I was wondering why is it still preferred to use mfcc_hires over standard fbanks for training nnet3 chain models? With neural networks, we don't need the features to be decorrelated. Does anyone has compared both features with a medium/big dataset?Thank you!
Go to http://kaldi-asr.org/forums.html find out how to join
---
You received this message because you are subscribed to the Google Groups "kaldi-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi...@googlegroups.com.
Hi,Dan.
Is idct-layer only compatible for convolutional layers or idct-layer can be used before any layers? e.g. tdnn, tdnnf and so on
The reason we use MFCC is because they are more easily compressible, being decorrelated; we dump them to diskwith compression to 1 byte per coefficient. But we dump all the coefficients, so it's equivalent to filterbanks timesa full-rank matrix, no information is lost. For convolutional architectures, we convert them back into filterbanks inside thenetwork; see idct-layer.
Dan
On Sun, Jul 14, 2019 at 11:19 AM Titouan Parcollet <parcolle...@gmail.com> wrote:
Hi there!--I was wondering why is it still preferred to use mfcc_hires over standard fbanks for training nnet3 chain models? With neural networks, we don't need the features to be decorrelated. Does anyone has compared both features with a medium/big dataset?Thank you!
Go to http://kaldi-asr.org/forums.html find out how to join
---
You received this message because you are subscribed to the Google Groups "kaldi-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi...@googlegroups.com.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi-help+...@googlegroups.com.
To post to this group, send email to kaldi...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kaldi-help/3bf5ec36-0bf9-4d63-adb6-8d4e2cea227e%40googlegroups.com.
the conv layers is just compute 3*3 on input seq*freq.tdnn is compute on all freq of concated frames like (-1,0,1).So tdnn computation is like dct or idct. right?
Another question. Is specaugment compatible for mfccs?
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi-help+...@googlegroups.com.
To post to this group, send email to kaldi...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kaldi-help/992b8f2b-2be3-40ae-8273-ffbd6dae2fa7%40googlegroups.com.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi-help+...@googlegroups.com.
To post to this group, send email to kaldi...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/kaldi-help/297fb03f-7eec-480b-bd89-6969f7287ffd%40googlegroups.com.