Geralt-Targaryen commited on
Commit
b8eac16
·
verified ·
1 Parent(s): e04d1a0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -1
README.md CHANGED
@@ -102,6 +102,19 @@ datasets:
102
 
103
  F2LLM-v2 is a family of general-purpose, multilingual embedding models in 8 distinct sizes ranging from 80M to 14B. Trained on a curated composite of 60 million publicly available high-quality data, F2LLM-v2 supports more than 200 languages, with a particular emphasis on previously underserved mid- and low-resource languages.
104
 
 
 
 
 
 
 
 
 
 
 
 
 
 
105
  ## Usage
106
 
107
  ### With Sentence Transformers
@@ -168,4 +181,8 @@ similarity = query_embedding @ document_embeddings.T
168
  print(similarity)
169
  # tensor([[0.6328, 0.8555, 0.7148, 0.8398]], device='cuda:0',
170
  # dtype=torch.bfloat16, grad_fn=<MmBackward0>)
171
- ```
 
 
 
 
 
102
 
103
  F2LLM-v2 is a family of general-purpose, multilingual embedding models in 8 distinct sizes ranging from 80M to 14B. Trained on a curated composite of 60 million publicly available high-quality data, F2LLM-v2 supports more than 200 languages, with a particular emphasis on previously underserved mid- and low-resource languages.
104
 
105
+ F2LLM-v2 is fully open. We release base models in 5 sizes, instruct models in 8 sizes, the training data, the training code, and intermediate checkpoints. The three smallest instruct models are pruned and trained from the 0.6B base model.
106
+
107
+ | Model | Base | Instruct |
108
+ | ----- | ----------------------------------------------------------------------------------- | ------------------------------------------------------------------- |
109
+ | 80M | | [🤗F2LLM-v2-80M](https://huggingface.co/codefuse-ai/F2LLM-v2-80M) |
110
+ | 160M | | [🤗F2LLM-v2-160M](https://huggingface.co/codefuse-ai/F2LLM-v2-160M) |
111
+ | 330M | | [🤗F2LLM-v2-330M](https://huggingface.co/codefuse-ai/F2LLM-v2-330M) |
112
+ | 0.6B | [🤗F2LLM-v2-0.6B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-0.6B-Preview) | [🤗F2LLM-v2-0.6B](https://huggingface.co/codefuse-ai/F2LLM-v2-0.6B) |
113
+ | 1.7B | [🤗F2LLM-v2-1.7B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-1.7B-Preview) | [🤗F2LLM-v2-1.7B](https://huggingface.co/codefuse-ai/F2LLM-v2-1.7B) |
114
+ | 4B | [🤗F2LLM-v2-4B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-4B-Preview) | [🤗F2LLM-v2-4B](https://huggingface.co/codefuse-ai/F2LLM-v2-4B) |
115
+ | 8B | [🤗F2LLM-v2-8B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-8B-Preview) | [🤗F2LLM-v2-8B](https://huggingface.co/codefuse-ai/F2LLM-v2-8B) |
116
+ | 14B | [🤗F2LLM-v2-14B-Preview](https://huggingface.co/codefuse-ai/F2LLM-v2-14B-Preview) | [🤗F2LLM-v2-14B](https://huggingface.co/codefuse-ai/F2LLM-v2-14B) |
117
+
118
  ## Usage
119
 
120
  ### With Sentence Transformers
 
181
  print(similarity)
182
  # tensor([[0.6328, 0.8555, 0.7148, 0.8398]], device='cuda:0',
183
  # dtype=torch.bfloat16, grad_fn=<MmBackward0>)
184
+ ```
185
+
186
+ ## Intermediate Checkpoints
187
+
188
+ To facilitate future research, we release intermediate checkpoints in the `intermediate_checkpoints` branch.