"Decently sized" is the key criteria here. With enough labelled data, you are better off not using pre-trained vectors, but rather training only on text from your problem domain. However, in practice, you are likely to find that "quantity has a quality all its own", and the pre-trained vectors give you an edge by incorporating information about the structure of language gleaned from the larger set. Of course, it is very easy to try training models with and without pre-trained vectors so as to empirically verify the claims of strange folk who talk to you on the internet!