You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Internally, I can see that it is calculated using offset and scale and changes to int8.
But I want to make sure that everyone only performs int operations inside, without int convertion process.
How do I convert the real type of weight to int in torch or onnx without Qualcomm SDK?
thanks,
The text was updated successfully, but these errors were encountered:
@IJS1016, the tools AIMET provide do quantization simulation, wherein we end up doing quantization and dequantization. So the resulting weight is kept in float. To see the actual quantization happening, we suggest users using Qualcomm SDK.
@quic-mangal I had the same question as @IJS1016, as I'm seeing similarly when reviewing the output model after adaround & AutoQuant. Given that I have an ultimate interest in the integer version of these weights, could you provide links/guidance on what you mean by using Qualcomm SDK to get these?
Hi, I have question for adaround quantization.
I applied adaround qauntization to my model.
the encodings file saids, weight's dtype is int.
like this.
"activation_encodings": { "150": [ { "bitwidth": 8, "dtype": "int", "is_symmetric": "False", "max": 3.8640635013580322, "min": 0.0, "offset": 0, "scale": 0.01515319012105465 } ], "154": [ { "bitwidth": 8, "dtype": "int", "is_symmetric": "False", "max": 3.8640635013580322, "min": 0.0, "offset": 0, "scale": 0.01515319012105465 } ],
but, the wieght's dtype is not int, float like this.
` [[ 0.0477, 0.0636, 0.3232],
[-0.2808, -0.3020, -0.3497],
[-0.1007, 0.2331, -0.4822]],
Internally, I can see that it is calculated using offset and scale and changes to int8.
But I want to make sure that everyone only performs int operations inside, without int convertion process.
How do I convert the real type of weight to int in torch or onnx without Qualcomm SDK?
thanks,
The text was updated successfully, but these errors were encountered: