Update README.md
Browse files
README.md
CHANGED
|
@@ -102,6 +102,19 @@ datasets:
|
|
| 102 |
|
| 103 |
F2LLM-v2 is a family of general-purpose, multilingual embedding models in 8 distinct sizes ranging from 80M to 14B. Trained on a curated composite of 60 million publicly available high-quality data, F2LLM-v2 supports more than 200 languages, with a particular emphasis on previously underserved mid- and low-resource languages.
|
| 104 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 105 |
## Usage
|
| 106 |
|
| 107 |
### With Sentence Transformers
|
|
@@ -168,4 +181,8 @@ similarity = query_embedding @ document_embeddings.T
|
|
| 168 |
print(similarity)
|
| 169 |
# tensor([[0.5703, 0.8281, 0.6953, 0.8203]], device='cuda:0',
|
| 170 |
# dtype=torch.bfloat16, grad_fn=<MmBackward0>)
|
| 171 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 102 |
|
| 103 |
F2LLM-v2 is a family of general-purpose, multilingual embedding models in 8 distinct sizes ranging from 80M to 14B. Trained on a curated composite of 60 million publicly available high-quality data, F2LLM-v2 supports more than 200 languages, with a particular emphasis on previously underserved mid- and low-resource languages.
|
| 104 |
|
| 105 |
+
F2LLM-v2 is fully open. We release base models in 5 sizes, instruct models in 8 sizes, the training data, the training code, and intermediate checkpoints. The three smallest instruct models are pruned and trained from the 0.6B base model.
|
| 106 |
+
|
| 107 |
+
| Model | Base | Instruct |
|
| 108 |
+
| ----- | ----------------------------------------------------------------------------------- | ------------------------------------------------------------------- |
|
| 109 |
+
| 80M | | [🤗F2LLM-v2-80M](https://huggingface.co/codefuse-ai/F2LLM-v2-80M) |
|
| 110 |
+
| 160M | | [🤗F2LLM-v2-160M](https://huggingface.co/codefuse-ai/F2LLM-v2-160M) |
|
| 111 |
+
| 330M | | [🤗F2LLM-v2-330M](https://huggingface.co/codefuse-ai/F2LLM-v2-330M) |
|
| 112 |
+
| 0.6B | [🤗F2LLM-v2-0.6B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-0.6B-Preview) | [🤗F2LLM-v2-0.6B](https://huggingface.co/codefuse-ai/F2LLM-v2-0.6B) |
|
| 113 |
+
| 1.7B | [🤗F2LLM-v2-1.7B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-1.7B-Preview) | [🤗F2LLM-v2-1.7B](https://huggingface.co/codefuse-ai/F2LLM-v2-1.7B) |
|
| 114 |
+
| 4B | [🤗F2LLM-v2-4B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-4B-Preview) | [🤗F2LLM-v2-4B](https://huggingface.co/codefuse-ai/F2LLM-v2-4B) |
|
| 115 |
+
| 8B | [🤗F2LLM-v2-8B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-8B-Preview) | [🤗F2LLM-v2-8B](https://huggingface.co/codefuse-ai/F2LLM-v2-8B) |
|
| 116 |
+
| 14B | [🤗F2LLM-v2-14B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-14B-Preview) | [🤗F2LLM-v2-14B](https://huggingface.co/codefuse-ai/F2LLM-v2-14B) |
|
| 117 |
+
|
| 118 |
## Usage
|
| 119 |
|
| 120 |
### With Sentence Transformers
|
|
|
|
| 181 |
print(similarity)
|
| 182 |
# tensor([[0.5703, 0.8281, 0.6953, 0.8203]], device='cuda:0',
|
| 183 |
# dtype=torch.bfloat16, grad_fn=<MmBackward0>)
|
| 184 |
+
```
|
| 185 |
+
|
| 186 |
+
## Intermediate Checkpoints
|
| 187 |
+
|
| 188 |
+
To facilitate future research, we release intermediate checkpoints in the `intermediate_checkpoints` branch.
|