Just a general comment.
K-fold cross validation is great but comes at a cost since one must train-test k times. In traditional machine learning circles you will find cross-validation used almost everywhere and more often with small datasets.
However, in deep learning circles you will generally find it used much less since, usually, the gains to be had by training a single deep model for k times as long, outweighs the statistical security that k-fold cross validation provides. By "used much less" of course I mean never (you probably won't find a deep learning paper from Stanford or Berkeley that uses cross-validation). It is not uncommon for deep models to be trained for months (facenet = 5 mo+) making cross-validation a mute point.
Also note that deep learning systems are now being trained with datasets with 100s of billions of elements (google TensorFlow white paper released one week ago). The benefits of cross-validation diminish when dealing with such large datasets so this is another reason that it is not generally used in deep learning.