Update README.md
Browse files
README.md
CHANGED
|
@@ -119,7 +119,7 @@ nomic-embed-text-v2-moe is SoTA multilingual MoE text embedding model:
|
|
| 119 |
- **High Performance**: SoTA Multilingual performance compared to ~300M parameter models, competitive with models 2x in size
|
| 120 |
- **Multilinguality**: Supports ~100 languages and trained over 1.6B pairs
|
| 121 |
- **Flexible Embedding Dimension**: Trained with [Matryoshka Embeddings](https://arxiv.org/abs/2205.13147) with 3x reductions in storage cost with minimal performance degredations
|
| 122 |
-
- **Fully-
|
| 123 |
|
| 124 |
|
| 125 |
| Model | Params (M) | Emb Dim | BEIR | MIRACL | Pretrain Data | Finetune Data | Code |
|
|
@@ -151,6 +151,12 @@ nomic-embed-text-v2-moe is SoTA multilingual MoE text embedding model:
|
|
| 151 |
|
| 152 |
The model can be used through SentenceTransformers and Transformers.
|
| 153 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 154 |
**Important**: the text prompt *must* include a *task instruction prefix*, instructing the model which task is being performed.
|
| 155 |
|
| 156 |
For queries/questions, please use `search_query: ` and `search_document: ` for the corresponding document
|
|
|
|
| 119 |
- **High Performance**: SoTA Multilingual performance compared to ~300M parameter models, competitive with models 2x in size
|
| 120 |
- **Multilinguality**: Supports ~100 languages and trained over 1.6B pairs
|
| 121 |
- **Flexible Embedding Dimension**: Trained with [Matryoshka Embeddings](https://arxiv.org/abs/2205.13147) with 3x reductions in storage cost with minimal performance degredations
|
| 122 |
+
- **Fully Open-Source**: Model weights, [code](https://github.com/nomic-ai/contrastors), and training data (see code repo) released
|
| 123 |
|
| 124 |
|
| 125 |
| Model | Params (M) | Emb Dim | BEIR | MIRACL | Pretrain Data | Finetune Data | Code |
|
|
|
|
| 151 |
|
| 152 |
The model can be used through SentenceTransformers and Transformers.
|
| 153 |
|
| 154 |
+
For best performance on GPU, please install
|
| 155 |
+
|
| 156 |
+
```bash
|
| 157 |
+
pip install torch transformers einops git+https://github.com/nomic-ai/megablocks.git
|
| 158 |
+
```
|
| 159 |
+
|
| 160 |
**Important**: the text prompt *must* include a *task instruction prefix*, instructing the model which task is being performed.
|
| 161 |
|
| 162 |
For queries/questions, please use `search_query: ` and `search_document: ` for the corresponding document
|