As described by the paper, PV-DBOW doesn't involve any NN-input-vectors per-word. There's one vector for the text, which alone is used to predict each individual word.
So while word-vectors still get randomly initialized (simply because of the way gensim code is shared with word2vec and other modes), they're not updated during training at all, and are still at their random initial values at the end of training.
Now, the DBOW training is very analogous to the Word2Vec "skip-gram" mode, but using vector(s) for the text-as-a-whole to predict target words, rather than just vector(s) for nearby-words... so it is very easy to combine with skip-gram word training. if you need the word-vectors too. You can turn this on with the `dbow_words=1` Doc2Vec initialization parameter.
Note it will slow training noticeably – roughly in proportion to your `window` value. (The `window` parameter only has any effect when doing context_word_vec(s) -> target_word training, not plain DBOW doc-vec -> target_word training.)
In comparison, Mikolov's demo '-sentence-vectors' patch to word2vec.c always includes word-training, because it uses an artificial pseudo-word at the start of each example, special-cased to participate in every 'window', as its sentence-vector-tag.
- Gordon