Hello,
To see if the OOM errors are caused by the Model itself, please test performing a batch prediction using the Google served AutoML model you trained, as listed here [1].
If that works, then the OOM is most likely due to your system's memory limit that is running the exported container [2], and you will need to increase your server hardware to resolve the issue.
Else, if the batch predictions to the Google served AutoML model fails [1] due to OOM, then it is recommended to open a Public Issue Tracker [3] to report the issue to the engineering team, as it may be a memory leak in Tensorflow (e.g. these reports [4]).