You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
After calling quantsim.export(path, filename_prefix), I could not get int8 QNN ONNX model. My objective is to get an int8 ONNX model through aimet quant toolkit, which shows like the attached image below.
However, by calling quantsim.export(path, filename_prefix), I only can get pth files, encoding files and one fp32 ONNX model. Did I use the export functionality incorrectly? Or is any way to convert encoding files and the fp32 ONNX model to one int8 QNN model?
The text was updated successfully, but these errors were encountered:
@JiliangNi please use --keep_quant_nodes option with the qnn converters to see a QNN model with activation quant/dequant nodes. Without this option, quant nodes are stripped from the graph.
After calling quantsim.export(path, filename_prefix), I could not get int8 QNN ONNX model. My objective is to get an int8 ONNX model through aimet quant toolkit, which shows like the attached image below.
However, by calling quantsim.export(path, filename_prefix), I only can get pth files, encoding files and one fp32 ONNX model. Did I use the export functionality incorrectly? Or is any way to convert encoding files and the fp32 ONNX model to one int8 QNN model?
The text was updated successfully, but these errors were encountered: