| advance: null |
| approach: post_training_static_quant |
| calib_iteration: 7 |
| framework: pytorch |
| op: |
| ? !!python/tuple |
| - quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.0.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.0.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.0.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.0.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.0.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.0.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.0.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.0.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.0.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.0.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.1.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.1.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.1.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.1.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.1.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.1.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.1.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.1.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.1.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.1.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.2.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.2.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.2.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.2.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.2.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.2.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.2.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.2.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.2.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.2.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.3.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.3.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.3.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.3.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.3.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.3.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.3.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.3.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.3.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.3.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.4.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.4.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.4.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.4.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.4.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.4.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.4.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.4.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.4.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.4.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.5.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.5.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.5.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.5.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.5.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.5.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.5.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.5.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.5.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.5.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.6.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.6.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.6.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.6.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.6.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.6.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.6.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.6.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.6.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.6.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.7.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.7.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.7.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.7.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.7.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.7.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.7.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.7.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.7.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.7.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.8.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.8.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.8.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.8.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.8.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.8.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.8.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.8.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.8.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.8.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.9.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.9.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.9.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.9.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.9.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.9.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.9.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.9.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.9.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.9.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.10.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.10.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.10.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.10.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.10.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.10.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.10.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.10.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.10.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.10.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.11.attention.self.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.11.attention.self.query |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.11.attention.self.key |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.11.attention.self.value |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.11.attention.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.11.attention.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.11.intermediate.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.11.intermediate.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.encoder.layer.11.output.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.encoder.layer.11.output.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - bert.pooler.quant |
| - QuantStub |
| : activation: |
| dtype: uint8 |
| scheme: asym |
| granularity: per_tensor |
| algorithm: minmax |
| ? !!python/tuple |
| - bert.pooler.dense |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
| ? !!python/tuple |
| - classifier |
| - Linear |
| : weight: |
| dtype: int8 |
| scheme: sym |
| granularity: per_channel |
| algorithm: minmax |
| bit: 7.0 |
| activation: |
| dtype: uint8 |
| scheme: sym |
| granularity: per_tensor |
| algorithm: kl |
|
|