--- library_name: colbert-onnx tags: - colbert - onnx - sentence-transformers - feature-extraction license: mit --- # model-test-onnx ONNX export of [unknown](https://huggingface.co/unknown) for fast CPU inference. ## Model Details - **Source Model**: [unknown](https://huggingface.co/unknown) - **Embedding Dimension**: unknown - **Format**: ONNX (FP32 + INT8) ## Files | File | Description | |------|-------------| | `model.onnx` | FP32 ONNX model | | `model_int8.onnx` | INT8 quantized model (faster) | | `tokenizer.json` | Tokenizer configuration | | `config_sentence_transformers.json` | Model configuration | ## Usage with colbert-onnx (Rust) ```rust use colbert_onnx::Colbert; let mut model = Colbert::from_pretrained("path/to/model")?; let embeddings = model.encode_documents(&["Hello world"])?; ``` ## Export Tool This model was exported using [pylate-onnx-export](https://github.com/lightonai/next-plaid/tree/main/onnx/python): ```bash pip install "pylate-onnx-export @ git+https://github.com/lightonai/next-plaid.git#subdirectory=onnx/python" pylate-onnx-export unknown --push-to-hub raphaelsty/model-test-onnx ```