| license: apache-2.0 | |
| base_model: | |
| - intfloat/multilingual-e5-small | |
| the original model is OriginalModel.mlpackage with float32 | |
| # the quantization model info : | |
| ## quantization and input maxlength | |
| coreML: using linear quantize nbit=8 | |
| input max = 128 | |
| ## note | |
| i tried turn it into float 16, but it changed too much for its prediction. quantization using linear nbit=8, it works almost like the original. |