As per Christoph's response, the 'GoogleNews' vectors contain plenty of n-grams. But note that:
* those were trained from Google's giant internal corpus of news articles as of ~2013 – if your domain isn't news, or involves terms that are novel or have shifted in sense since then, those vectors may be suboptimal
* while they used some variant of the statistical algorithm that's also implemented in gensim's `Phrases` class, run in multiple passes (as each pass only pairs previously-separate tokens), their exact parameters/tokenization/etc haven't been fully documented, as far as I can tell
If you had your own more-appropriate training texts, you could likely create more up-to-date vectors well-tuned for your domain.
You could use some statistical method, like that in the `Phrases` class - but note that the generated bigrams/etc may not be aesthetically-pleasing, or match your human-level understandings about what the true logical entities/phrases are. Its purely-statistical approach tends to miss things you'd like combined, and combine things you'd rather not - and tuning the thresholds only helps up to a point, where gaining some desired phrases costs others and vice-versa. The resulting processed text, with new bigram/etc tokens, can be very useful for classification/info-retrieval, but won't necessarily "look right" if presented to end-users.
If you have some other domain-specific way of identifying the key multi-word phrases, like a glossary or other heuristics, you could also use that to preprocess your text before Word2Vec training, to ensure your preferred multi-word phrases get word-vectors.
- Gordon