`hs=0, negative=0` would literally mean no output layer(s) to generate training errors for backpropagation, so the behavior is undefined. (If it doesn't error, any results are likely junk.)
(2)
Any of the modes which include random-choices during training/inference – that includes negative-sampling (`negative` > 0) or frequent-word downsampling (`sample` > 0) or the varying window-sizes in "DM" mode (`dm=1`) – shouldn't be giving deterministic results for repeated `infer_vector()` invocations.
If you seemed to get deterministic repeated results with `sample=1e05, negative=5`, are you sure you didn't do something else in that case make that so, like start fresh with an identically-loaded model for each attempt?
There's more discussion of what could achieve deterministic inference results in <
https://github.com/RaRe-Technologies/gensim/issues/447>. Using more iterations (the `steps` parameter of `infer_vector()`) should help improve the quality of inferred vectors, including making vectors from subsequent runs on the same text more similar to each other.
(3)
`sample` often helps to both speed training and improve vector-quality, for some downstream tasks (like the common analogies evaluation) – but there's no firm rules-of-thumb... it depends on your goals & data, so you need to explore different values.
(4)
Randomization of document order before each pass isn't strictly necessary; you may want to check if it really helps your vectors. (It may be worth one initial randomization, if there's a risk your original document order has grouped all occurrences of certain words/patterns-of-use/document-sizes together.)
- Gordon