Dequantize#
Dequantize operation converts a quantized (u8 or s8) tensor to a f32 tensor. It supports both per-tensor and per-channel asymmetric linear de-quantization. Rounding mode is library-implementation defined.
For per-tensor de-quantization:
For per-channel de-quantization, taking channel axis = 1 as an example:
where \(ic\) is the number of channels.
Operation Attributes#
|
Description |
Value Type |
|
|
---|---|---|---|---|
Specifies which de-quantization type is used |
string |
|
Optional |
|
Specifies dimension on which per-channel de-quantization is applied |
s64 |
A s64 value
in the
range of
[-r, r-1]
where r =
rank(src),
|
Optional |
|
Scalings applied on the src data |
f32 |
A f32 list
(only
contain one
element if
|
Required |
|
Offset values that maps to float zero |
s64 |
A s64 list
(only
contain one
element if
|
Required |
Execution Arguments#
The inputs and outputs must be provided according to the below index order when constructing an operation.
Inputs#
Index |
Argument Name |
Required or Optional |
---|---|---|
0 |
|
Required |
Outputs#
Index |
Argument Name |
Required or Optional |
---|---|---|
0 |
|
Required |
Supported Data Types#
Dequantize operation supports the following data type combinations.
Src |
Dst |
---|---|
s8, u8 |
f32 |
@note This operation is to support int8 quantization model.