I have models exported to TFLite format. One is an fp16 model, and the other is an int8 model. How could I discriminate between these two models? I have searched on Google and issues in the Github TensorFlow repo, but there is no discussion about this topic.
I tried to read the scale of quantization of the input tensor in the two models. The scale of fp16 is zero, while the scale of int8 is 0.00392 approximately, as shown in the image below.
I would like to use the value of scale to discriminate, but I am afraid that other fp16 models may have a non-zero value of scale.
How could I determine whether a TFLite model is int8 quantized?