Thank you Gordon. Just so that I understand.
Suppose I have the following story:
"Sarah was eating a sandwich. Mary was drinking a beer. Michael was driving."
In order to build `sentences` from the story above, I can do either (just one document)
> sentences = [['Sarah', 'was', 'eating', 'a', 'sandwich', 'Mary', 'was', 'drinking', 'a', 'beer', 'Michael', 'was', 'driving']]
or (three different documents, the actual sentences)
> sentences = [['Sarah', 'was', 'eating', 'a', 'sandwich'], ['Mary', 'was', 'drinking', 'a', 'beer'], ['Michael', 'was', 'driving']]
and, as you explained, there isn't much difference between these two tokenizations, in terms of the resulting word embedding and its performance, right?