[GSoC] Add block quantized models (#270)
Browse files* Gemm and MatMul block quantization support
* refactoring
* fix indentation
* node name independent
* Block quantization tool:
- constant weight category supported
- add data type saturation
- handled the case in which all the elements within a block are the same
benchmark script modified to support block quantized models
block quantized some models
* add missing block quantized models
* formatting
* add blocked models to eval script. Evaluation yunet
* Add sface and pphumanseg evaluation, block quantization tool fix, handpose blocked model fix, removed blocked CRNN EN,
* changed evaluation metric in block_quantize script and add verbose mode
* Add evaluation for PP-ResNet and Mobilenet
* changed file suffix and update readmes
* renamed int8bq
README.md
CHANGED
|
@@ -4,6 +4,9 @@ This model is contributed by Dong Xu (徐栋) from [watrix.ai](watrix.ai) (银
|
|
| 4 |
|
| 5 |
Please note that the model is trained with Chinese license plates, so the detection results of other license plates with this model may be limited.
|
| 6 |
|
|
|
|
|
|
|
|
|
|
| 7 |
## Demo
|
| 8 |
|
| 9 |
Run the following command to try the demo:
|
|
|
|
| 4 |
|
| 5 |
Please note that the model is trained with Chinese license plates, so the detection results of other license plates with this model may be limited.
|
| 6 |
|
| 7 |
+
**Note**:
|
| 8 |
+
- `license_plate_detection_lpd_yunet_2023mar_int8bq.onnx` represents the block-quantized version in int8 precision and is generated using [block_quantize.py](../../tools/quantize/block_quantize.py) with `block_size=64`.
|
| 9 |
+
|
| 10 |
## Demo
|
| 11 |
|
| 12 |
Run the following command to try the demo:
|