I'm not sure, but it's probably not faster than Sphinx. Sphinx was optimized for speed; kaldi is probably twice as accurate in terms of WER (or even more), but may not be as fast as Sphinx because we normally use different models, like neural nets, that are slower to evaluate, or at least bigger models. The online-nnet2 can be tuned to take around real-time if you have a good (fast) machine with one thread. The 'chain' models can be even faster (maybe twice faster than real time on a fast machine with one thread, if you tune the beam), but the online decoding is not ready yet (at least, not in the official Kaldi repository).
You could also look at the online GMM-based decoding (there is something in online2/, and there are scripts somewhere, look at egs/rm/s5/local/online/run_gmm.sh). That can be faster, depending on the beams, model sizes, etc.