Hello Seth,
what this means is that the "online LDA" algorithm by Hoffman et al.
(which is implemented in gensim), is "online" only in the sense of
"incremental". You can update the model with new documents,
incrementally. This will help improve/refine/converge the model.
But later batches have gradually less and less effect on the model, so
if you build a model over 100k docs, and later update it with another
100k docs, the impact of the latter will be much smaller. If documents
in both batches come from the same distribution (=no topic drift),
this is fine, and the order of the two batches doesn't matter much
(~not at all, in the limit of complete convergence). But if there is
topic drift, the order matters -- the resulting model will reflect
statistical patterns from the first batch more prominently than from
the second.
For more info, see the `kappa` and `tau` parameters from the article
"Online Learning for Latent Dirichlet Allocation".
HTH,
Radim
> is not truly online (the name of the [3]<
http://radimrehurek.com/gensim/wiki.html#id8> article