Hi all,
After training a fasttext model in genism (native fasttext and not the wrapper), I want to use the embeddings as a first layer in Keras for a deep Neural Network.
Basically:
model.wv.syn0 are the embeddings for vocabulary words.
model.wv.syn0_ngrams are the embeddings for the character n-grams.
It looks like e have 2 different dictionaries, 1 that maps the vocabulary words to integers, and 1 that maps the character n-grams to integers.
In Keras, the input layer should correspond to both the vocabulary words AND the character n-grams (subwords), right?
So is it my responsibility to create a merged dictionary that would uniquely map words and character n-grams to integers, and also combine the embeddings from model.wv.syn0 and model.wv.syn0_ngrams ?
Thanks!