Hi all,
I am now considering scaling up the batch size to shorten the training time (by rewriting the script by DistributedDataParallel). I tried increasing the batch for the train_a script from 16 to 32, but this caused oscillation of the test loss (Average Corr. etc.) without score improvement.
Has anyone considered or tried increasing the batch size? It would be great if we could know recommendations for the way to tune the optimization parameters.
Best,
Yohsuke

--
You received this message because you are subscribed to the Google Groups "Orca Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to orca-users+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/orca-users/2b8e400b-b3fa-42b6-b29b-501083b6184an%40googlegroups.com.
