First, keep in mind that the loss-calculation of Gensim `Word2Vec` has a number of known problems which may limit its suitability for whatever-you're-hoping-it'll-provide, or require hackish workarounds. See open-issue <
https://github.com/RaRe-Technologies/gensim/issues/2617> for an overview, with links to related issues, of all that's missing.
However, there's a more foundational problem in your usage: `workers=-1` is not a supported value for Gensim. Its behavior will be undefined. (I suspect it results in no training happening, instantly. Do you have logging enabled at the INFO level to observe normal training steps happening & progressing, with sensible interim updates & internal counts reported, taking a reasonable amount of time?)
Separately, that's a pretty-small training set: even if training were happening, with a tiny 1360 unique-vocabulary (and even some of those words only having a meager `min_count=3` number of usage-examples), the final model might not be impressive in its generalizability to understanding the words' broad meanings. This algorithm really needs lots of realistic, varied data to show its most-useful benefits. (Though doing things like using a smaller `vector_size`, and more `epochs`, can manage *some* consistency/meaning out of smaller datasets, the 'ceiling' of how impressive the final model can get is still quite low with such parameter-tinkering. More varied & representative data always helps more.)
- Gordon