| library_name: colbert-onnx | |
| tags: | |
| - colbert | |
| - onnx | |
| - sentence-transformers | |
| - feature-extraction | |
| license: mit | |
| # model-test-onnx | |
| ONNX export of [unknown](https://huggingface.co/unknown) for fast CPU inference. | |
| ## Model Details | |
| - **Source Model**: [unknown](https://huggingface.co/unknown) | |
| - **Embedding Dimension**: unknown | |
| - **Format**: ONNX (FP32 + INT8) | |
| ## Files | |
| | File | Description | | |
| |------|-------------| | |
| | `model.onnx` | FP32 ONNX model | | |
| | `model_int8.onnx` | INT8 quantized model (faster) | | |
| | `tokenizer.json` | Tokenizer configuration | | |
| | `config_sentence_transformers.json` | Model configuration | | |
| ## Usage with colbert-onnx (Rust) | |
| ```rust | |
| use colbert_onnx::Colbert; | |
| let mut model = Colbert::from_pretrained("path/to/model")?; | |
| let embeddings = model.encode_documents(&["Hello world"])?; | |
| ``` | |
| ## Export Tool | |
| This model was exported using [pylate-onnx-export](https://github.com/lightonai/next-plaid/tree/main/onnx/python): | |
| ```bash | |
| pip install "pylate-onnx-export @ git+https://github.com/lightonai/next-plaid.git#subdirectory=onnx/python" | |
| pylate-onnx-export unknown --push-to-hub raphaelsty/model-test-onnx | |
| ``` | |