I'm using the recently added model Universal Sentence Encoder available here
https://www.tensorflow.org/hub/modules/google/universal-sentence-encoder/1 for a Semantic Similarity Task on documents i.e. text with let's say 7-15 paragraphs (hence line separated by at least two end of line "\n\n") and 4-5 sentences for each paragraph (separated by end of line "\n"), like a poetry document.
My question is how to improve this result. The models on the Tensorflow hub can be trained, so I was wondering if using a model training like
embedded_text_feature_column = hub.text_embedding_column(
key="sentence", module_spec=hub_module, trainable=train_module)
But I'm not sure if this is the right methodology. Another option could (would) be using additional features other than text embedding features, like (maybe) the date /epoch of the text (let's say decades), or other info like a genre, but how to embed these features in a boosting style approach with the model hubs and the universal sentence encoder?