Hello Amir,
the "Lee corpus" contains only 300 documents. It's quite possible the LDA results are not good, esp. when asking for 200 topics.
Also, you're doing only a single online training pass (try more passes, like 50), not removing common words (try dictionary.filter_extremes), use plain bag-of-words vectors (drop the log-entropy transformation), use fewer topics etc.
I tried LDA on Lee and got a correlation to human judgements of around ~0.4 (which is still way worse than LSI's 0.6).
You could also try LDA based on gibbs sampling (gensim's LdaModel is based on variational inference), which is a different method of training the model.
Incidentally, I just pushed a wrapper for Mallet's LDA to gensim, which is a great java implementation that uses gibbs sampling, hyperparameter optimizations etc. You can find it under `gensim.models.ldamallet`, using the develop branch of gensim from github.
Best,
Radim