Skip to main content
Version: 0.2

QATQuantizationSettings

CLASS - QATQuantizationSettings(

  • weights_num_bits: int =  8
  • activations_num_bits: int =  8
  • skip_tail_quantization: bool =  True
  • automatic_skip_quantization: bool =  True

)

Ancestors - (BaseQATQuantizationSettings, BaseQuantizationSettings)

Use this if you wish to do QAT Quantization.

Class Variables

  • weights_num_bits (int) - Number of Bits to use for the Weights if applicable.
  • activations_num_bits (int) - Number of Bits to use for the Activations if applicable.
  • skip_tail_quantization (bool) - Whether or not to automatically skip Quantization for the Tail of the Model. It is better to keep it 'True' if unsure.
  • automatic_skip_quantization (bool) - Whether or not to automatically skip Quantization for layers that are too sensitive. It is better to keep it 'True' if unsure. It does not override explicitly marked 'skipped_quantization = True' layers.