Versioned name: Quantize-1

Category: lower_precision

Short description: Quantize converts a fp32 tensor to a quantized(int8 or uint8) tensor. It supports both per tensor and per channel asymmetric linear quantization. Output data type is specified in output tensor data_type. Nearest round is used in this OP. For per-tensor quantization:


For per-channel quantization, take channel axis = 2 as example:

\[q_{x_{...,i,...,...}}=round(x_{...,i,...,...}/scale_i+zp_i),i\in{[0, channelNum-1]}\]


  • qtype

    • Description: specifies which quantization type is used.

    • Range of values: “per_tensor” or “per_channel”

    • Type: string

    • Default value: “per_tensor”

    • Required: no

  • axis

    • Description: specifies dimension on which apply per-channel quantization. Only valid when qtype is “per_channel”.

    • Range of values: integers in [-d, d-1] where d = input_tensor.shape().size()

    • Type: int

    • Default value: 1

    • Required: no

  • scales

    • Description: apply in quantization formula.

    • Range of values: float values

    • Type: float[]

    • Default value: None

    • Required: yes

  • zps

    • Description: offset value that maps to float zero.

    • Range of values: integer values

    • Type: int[]

    • Default value: None

    • Required: yes


  • 1: input - fp32 tensor to be quantized. Required.


  • 1: output – quantized tensor.