I am running the training example for Quantization aware training as is in Google Colab. You can find the tutorial here.
My exact code can be found here. I expected that all the layers in the output frozen model (.tflite) would be quantized (integer, int8 precisely)
However, I noticed a dequantize node present in the layers.
I am not sure what leads to this dequant node. This is not desirable and I would like to remove it, preferable with network surgery.