INT8 Quantization Not Performed in onnx2tf – Only Float32, Float16, and Dynamic Range Quantization Saved #740
Labels
Lack of reproduction methods
Lack of information or resources on how to reproduce
Issue Type
Feature Request
OS
Linux
onnx2tf version number
1.26.7
onnx version number
1.16.1
onnxruntime version number
1.18.1
onnxsim (onnx_simplifier) version number
0.4.33
tensorflow version number
2.17.0,
Download URL for ONNX
I am using my own SPNV2 ONNX
Parameter Replacement JSON
Description
I am trying to convert an ONNX model to TFLite using onnx2tf, but I noticed that INT8 quantization is not being performed correctly. Instead, the script only saves:
Float32 (.tflite)
Float16 (_float16.tflite)
Dynamic Range Quantization (_dynamic_range_quant.tflite)
However, I need Full Integer Quantization (INT8) because I plan to use the converted TFLite model for VectorBlox VNNX generation, which only supports fully integer models (both weights and activations must be INT8).
The text was updated successfully, but these errors were encountered: