Hi all,
I notice that the latest TVM release https://github.com/dmlc/tvm/releases/tag/v0.5 suggests that backend support for int8 quantization has been added, along with a network quantizer for simple networks.
I can’t find any documentation on https://docs.tvm.ai/ related to this, or how to use this.
Where can I find clues on using TVM to run inference on int8 models?
Thanks,
Skanda