model = word2vec.Word2Vec(corpus_file=input_file, sg=w2v_model == W2V_Model.SKIPGRAM, iter=100, min_count=1, size=300, workers=4, compute_loss=True, sample=0.5e4, negative=10, callbacks=[loss_callback])
def on_epoch_end(self, model): cumulative_loss = model.get_latest_training_loss() with open(self.loss_file_path, 'a') as f: f.write(f'{self.epoch}\t{cumulative_loss}\n')
This is part of the file tracking the losses:
32 126265624.033 128704760.034 131158848.035 133551176.036 134217728.037 134217728.038 134217728.039 134217728.040 134217728.041 134217728.0
.
.
.
And that's 134217728.0 till the end
2019-05-16 01:19:54,305 : INFO : collected 190104 word types from a corpus of 23928055 raw words and 1903865 sentences2019-05-16 01:19:54,305 : INFO : Loading a fresh vocabulary2019-05-16 01:19:54,747 : INFO : effective_min_count=1 retains 190104 unique words (100% of original 190104, drops 0)2019-05-16 01:19:54,748 : INFO : effective_min_count=1 leaves 23928055 word corpus (100% of original 23928055, drops 0)2019-05-16 01:19:55,435 : INFO : deleting the raw counts dictionary of 190104 items2019-05-16 01:19:55,438 : INFO : sample=5000 downsamples 50 most-common words2019-05-16 01:19:55,438 : INFO : downsampling leaves estimated 22549173 word corpus (94.2% of prior 23928055)