Hello,
I was trying a pre-quantized model in TFLite when I got an “NotImplementedError: Tensor type 9 is currently not supported” So I looked at the tflite.py ingestion code and found explicit checks that disable INT8 tensors. That was OK on old versions of TFLite, but according to the TFLite spec [https://www.tensorflow.org/lite/performance/quantization_spec] they have switched to symmetric quantization so values should be encoded as INT8.
I started adding code to support the INT8 data type and removing the assertions on the type checking for UINT8 only, but I wonder if there is a more disciplined way of doing this since I don’t want to break the ingestion process if there are assumptions on data types.
Also, is TVM going to support ingestion of pre-quantized models created with newer versions of TFLite? I know there is a quantizer in TVM, I have used it already, but there is still the case for ingesting pre-quantized models created in more recent versions of TFLite.
Thanks!