I can reasonably set those up for training, especially in the cloud as I would only need the VMs for a short time for training purposes. I just don't want to keep wasting time with what appears to not be working with my 4 CPU, 8 GB RAM VMWare VM that I'm running on premise. I'm wondering if a Tesla M2090 GPU (6 GB 1.8 GHz memory, 512 1.3 GHz cores) would be sufficient. If I run with GPUs, I don't think any hypervisors will pass it through so I'd need dedicated hardware so I'm also wondering if I'd be better off using a couple GPU-based hardware servers vs a larger number of VMs without GPUs (say, 3 hardware GPU-based servers vs 16 quad-core VMs in the public cloud).
I really appreciate your input!
Thanks,
Rhiannon
--
You received this message because you are subscribed to the Google Groups "kaldi-help" group.
To unsubscribe from this group and stop receiving emails from it, send an email to kaldi-help+...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
I see that tedlium has such a script:
local/online/run_nnet2_ms.sh
It looks like you have to run the run.sh at least until the tri3 stage before you can run that.
Dan
Well, the whole reason I was going to try nnet2 was because this server does not have any GPUs, and you'd mentioned that it would run a bit faster than nnet1 on a non-GPU system. I will eventually have a GPU server cluster on premise, but I'm not there yet. I have been planning on building a GPU VM in Amazon for training if I couldn't reasonably do it on prem. I've also been looking at finding pre-built language models that might work for now.
I'd gladly pay someone to help me get a working Kaldi installation going quickly, allowing me time to learn and optimize it, but haven't had luck in that avenue either. The project requires that the decoding be done on-prem and not outsourcing to a provider via an API, though that method might be necessary for initial testing until I can get something working.
I had asked previously but didn't get an answer - would a dual socket, quad core 3 GHz Xeon server (non VM) with 32 GB RAM and a lower-end / older GPU like a Tesla M2090 (6 GB, 512 cores) work reasonably well for training? I might go that route if it will and be done with it.