Add new SentenceTransformer model
Browse files- README.md +160 -154
- model.safetensors +1 -1
README.md
CHANGED
|
@@ -7,87 +7,92 @@ tags:
|
|
| 7 |
- sentence-similarity
|
| 8 |
- feature-extraction
|
| 9 |
- generated_from_trainer
|
| 10 |
-
- dataset_size:
|
| 11 |
- loss:MatryoshkaLoss
|
| 12 |
- loss:MultipleNegativesRankingLoss
|
| 13 |
base_model: nomic-ai/modernbert-embed-base
|
| 14 |
widget:
|
| 15 |
-
- source_sentence:
|
|
|
|
| 16 |
sentences:
|
| 17 |
-
-
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
-
|
| 23 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
- However, deploying LLMs effectively in real-world applications often requires
|
| 25 |
LLM integration. This means embedding these models into systems, workflows, or
|
| 26 |
products where they can interact with other components like databases, APIs, user
|
| 27 |
interfaces, or even custom business logic
|
| 28 |
-
- source_sentence: What
|
|
|
|
| 29 |
sentences:
|
| 30 |
-
- .
|
| 31 |
-
|
| 32 |
-
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
answers based on internal documents and discussions.'
|
| 43 |
-
- source_sentence: Can the system retrieve documents even if the exact words weren't
|
| 44 |
-
used?
|
| 45 |
sentences:
|
| 46 |
-
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
where LLMs are paired with document databases (e.g., vector stores like Supabase,
|
| 54 |
-
Pinecone, or Weaviate) to answer questions with enterprise-specific knowledge.
|
| 55 |
-
- For instance, in a document management system, a user might type "policies about
|
| 56 |
-
sick leave", and the system—integrated with an LLM—could retrieve documents discussing
|
| 57 |
-
"medical leave", "employee absence", and "illness policies", even if those exact
|
| 58 |
-
words weren’t used.
|
| 59 |
-
- source_sentence: What are some techniques mentioned for mitigating challenges in
|
| 60 |
-
prompt engineering?
|
| 61 |
sentences:
|
| 62 |
-
-
|
| 63 |
-
|
| 64 |
-
- . These models are trained on massive text datasets and are capable of generating
|
| 65 |
-
coherent, context-aware language, answering questions, summarizing documents,
|
| 66 |
-
writing code, and more.
|
| 67 |
-
- 'Prompt Engineering: Designing effective prompts and interactions is a new and
|
| 68 |
-
still-evolving skill.
|
| 69 |
|
| 70 |
|
| 71 |
-
|
| 72 |
-
|
| 73 |
|
| 74 |
|
| 75 |
-
|
|
|
|
|
|
|
| 76 |
|
| 77 |
-
|
| 78 |
-
|
| 79 |
-
|
| 80 |
-
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
|
| 84 |
-
|
| 85 |
-
|
| 86 |
-
|
| 87 |
-
|
| 88 |
-
|
| 89 |
-
|
| 90 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 91 |
pipeline_tag: sentence-similarity
|
| 92 |
library_name: sentence-transformers
|
| 93 |
metrics:
|
|
@@ -117,10 +122,10 @@ model-index:
|
|
| 117 |
type: dim_768
|
| 118 |
metrics:
|
| 119 |
- type: cosine_accuracy@1
|
| 120 |
-
value: 0.
|
| 121 |
name: Cosine Accuracy@1
|
| 122 |
- type: cosine_accuracy@3
|
| 123 |
-
value: 0.
|
| 124 |
name: Cosine Accuracy@3
|
| 125 |
- type: cosine_accuracy@5
|
| 126 |
value: 0.8666666666666667
|
|
@@ -129,10 +134,10 @@ model-index:
|
|
| 129 |
value: 1.0
|
| 130 |
name: Cosine Accuracy@10
|
| 131 |
- type: cosine_precision@1
|
| 132 |
-
value: 0.
|
| 133 |
name: Cosine Precision@1
|
| 134 |
- type: cosine_precision@3
|
| 135 |
-
value: 0.
|
| 136 |
name: Cosine Precision@3
|
| 137 |
- type: cosine_precision@5
|
| 138 |
value: 0.17333333333333337
|
|
@@ -141,10 +146,10 @@ model-index:
|
|
| 141 |
value: 0.10000000000000003
|
| 142 |
name: Cosine Precision@10
|
| 143 |
- type: cosine_recall@1
|
| 144 |
-
value: 0.
|
| 145 |
name: Cosine Recall@1
|
| 146 |
- type: cosine_recall@3
|
| 147 |
-
value: 0.
|
| 148 |
name: Cosine Recall@3
|
| 149 |
- type: cosine_recall@5
|
| 150 |
value: 0.8666666666666667
|
|
@@ -153,13 +158,13 @@ model-index:
|
|
| 153 |
value: 1.0
|
| 154 |
name: Cosine Recall@10
|
| 155 |
- type: cosine_ndcg@10
|
| 156 |
-
value: 0.
|
| 157 |
name: Cosine Ndcg@10
|
| 158 |
- type: cosine_mrr@10
|
| 159 |
-
value: 0.
|
| 160 |
name: Cosine Mrr@10
|
| 161 |
- type: cosine_map@100
|
| 162 |
-
value: 0.
|
| 163 |
name: Cosine Map@100
|
| 164 |
- task:
|
| 165 |
type: information-retrieval
|
|
@@ -169,49 +174,49 @@ model-index:
|
|
| 169 |
type: dim_512
|
| 170 |
metrics:
|
| 171 |
- type: cosine_accuracy@1
|
| 172 |
-
value: 0.
|
| 173 |
name: Cosine Accuracy@1
|
| 174 |
- type: cosine_accuracy@3
|
| 175 |
value: 0.8
|
| 176 |
name: Cosine Accuracy@3
|
| 177 |
- type: cosine_accuracy@5
|
| 178 |
-
value: 0.
|
| 179 |
name: Cosine Accuracy@5
|
| 180 |
- type: cosine_accuracy@10
|
| 181 |
-
value:
|
| 182 |
name: Cosine Accuracy@10
|
| 183 |
- type: cosine_precision@1
|
| 184 |
-
value: 0.
|
| 185 |
name: Cosine Precision@1
|
| 186 |
- type: cosine_precision@3
|
| 187 |
-
value: 0.
|
| 188 |
name: Cosine Precision@3
|
| 189 |
- type: cosine_precision@5
|
| 190 |
-
value: 0.
|
| 191 |
name: Cosine Precision@5
|
| 192 |
- type: cosine_precision@10
|
| 193 |
-
value: 0.
|
| 194 |
name: Cosine Precision@10
|
| 195 |
- type: cosine_recall@1
|
| 196 |
-
value: 0.
|
| 197 |
name: Cosine Recall@1
|
| 198 |
- type: cosine_recall@3
|
| 199 |
value: 0.8
|
| 200 |
name: Cosine Recall@3
|
| 201 |
- type: cosine_recall@5
|
| 202 |
-
value: 0.
|
| 203 |
name: Cosine Recall@5
|
| 204 |
- type: cosine_recall@10
|
| 205 |
-
value:
|
| 206 |
name: Cosine Recall@10
|
| 207 |
- type: cosine_ndcg@10
|
| 208 |
-
value: 0.
|
| 209 |
name: Cosine Ndcg@10
|
| 210 |
- type: cosine_mrr@10
|
| 211 |
-
value: 0.
|
| 212 |
name: Cosine Mrr@10
|
| 213 |
- type: cosine_map@100
|
| 214 |
-
value: 0.
|
| 215 |
name: Cosine Map@100
|
| 216 |
- task:
|
| 217 |
type: information-retrieval
|
|
@@ -227,7 +232,7 @@ model-index:
|
|
| 227 |
value: 0.8
|
| 228 |
name: Cosine Accuracy@3
|
| 229 |
- type: cosine_accuracy@5
|
| 230 |
-
value: 0.
|
| 231 |
name: Cosine Accuracy@5
|
| 232 |
- type: cosine_accuracy@10
|
| 233 |
value: 1.0
|
|
@@ -236,10 +241,10 @@ model-index:
|
|
| 236 |
value: 0.6666666666666666
|
| 237 |
name: Cosine Precision@1
|
| 238 |
- type: cosine_precision@3
|
| 239 |
-
value: 0.
|
| 240 |
name: Cosine Precision@3
|
| 241 |
- type: cosine_precision@5
|
| 242 |
-
value: 0.
|
| 243 |
name: Cosine Precision@5
|
| 244 |
- type: cosine_precision@10
|
| 245 |
value: 0.10000000000000003
|
|
@@ -251,19 +256,19 @@ model-index:
|
|
| 251 |
value: 0.8
|
| 252 |
name: Cosine Recall@3
|
| 253 |
- type: cosine_recall@5
|
| 254 |
-
value: 0.
|
| 255 |
name: Cosine Recall@5
|
| 256 |
- type: cosine_recall@10
|
| 257 |
value: 1.0
|
| 258 |
name: Cosine Recall@10
|
| 259 |
- type: cosine_ndcg@10
|
| 260 |
-
value: 0.
|
| 261 |
name: Cosine Ndcg@10
|
| 262 |
- type: cosine_mrr@10
|
| 263 |
-
value: 0.
|
| 264 |
name: Cosine Mrr@10
|
| 265 |
- type: cosine_map@100
|
| 266 |
-
value: 0.
|
| 267 |
name: Cosine Map@100
|
| 268 |
- task:
|
| 269 |
type: information-retrieval
|
|
@@ -273,49 +278,49 @@ model-index:
|
|
| 273 |
type: dim_128
|
| 274 |
metrics:
|
| 275 |
- type: cosine_accuracy@1
|
| 276 |
-
value: 0.
|
| 277 |
name: Cosine Accuracy@1
|
| 278 |
- type: cosine_accuracy@3
|
| 279 |
-
value: 0.
|
| 280 |
name: Cosine Accuracy@3
|
| 281 |
- type: cosine_accuracy@5
|
| 282 |
-
value: 0.
|
| 283 |
name: Cosine Accuracy@5
|
| 284 |
- type: cosine_accuracy@10
|
| 285 |
-
value: 0.
|
| 286 |
name: Cosine Accuracy@10
|
| 287 |
- type: cosine_precision@1
|
| 288 |
-
value: 0.
|
| 289 |
name: Cosine Precision@1
|
| 290 |
- type: cosine_precision@3
|
| 291 |
-
value: 0.
|
| 292 |
name: Cosine Precision@3
|
| 293 |
- type: cosine_precision@5
|
| 294 |
-
value: 0.
|
| 295 |
name: Cosine Precision@5
|
| 296 |
- type: cosine_precision@10
|
| 297 |
-
value: 0.
|
| 298 |
name: Cosine Precision@10
|
| 299 |
- type: cosine_recall@1
|
| 300 |
-
value: 0.
|
| 301 |
name: Cosine Recall@1
|
| 302 |
- type: cosine_recall@3
|
| 303 |
-
value: 0.
|
| 304 |
name: Cosine Recall@3
|
| 305 |
- type: cosine_recall@5
|
| 306 |
-
value: 0.
|
| 307 |
name: Cosine Recall@5
|
| 308 |
- type: cosine_recall@10
|
| 309 |
-
value: 0.
|
| 310 |
name: Cosine Recall@10
|
| 311 |
- type: cosine_ndcg@10
|
| 312 |
-
value: 0.
|
| 313 |
name: Cosine Ndcg@10
|
| 314 |
- type: cosine_mrr@10
|
| 315 |
-
value: 0.
|
| 316 |
name: Cosine Mrr@10
|
| 317 |
- type: cosine_map@100
|
| 318 |
-
value: 0.
|
| 319 |
name: Cosine Map@100
|
| 320 |
- task:
|
| 321 |
type: information-retrieval
|
|
@@ -325,49 +330,49 @@ model-index:
|
|
| 325 |
type: dim_64
|
| 326 |
metrics:
|
| 327 |
- type: cosine_accuracy@1
|
| 328 |
-
value: 0.
|
| 329 |
name: Cosine Accuracy@1
|
| 330 |
- type: cosine_accuracy@3
|
| 331 |
-
value: 0.
|
| 332 |
name: Cosine Accuracy@3
|
| 333 |
- type: cosine_accuracy@5
|
| 334 |
value: 0.8
|
| 335 |
name: Cosine Accuracy@5
|
| 336 |
- type: cosine_accuracy@10
|
| 337 |
-
value: 0.
|
| 338 |
name: Cosine Accuracy@10
|
| 339 |
- type: cosine_precision@1
|
| 340 |
-
value: 0.
|
| 341 |
name: Cosine Precision@1
|
| 342 |
- type: cosine_precision@3
|
| 343 |
-
value: 0.
|
| 344 |
name: Cosine Precision@3
|
| 345 |
- type: cosine_precision@5
|
| 346 |
value: 0.16000000000000003
|
| 347 |
name: Cosine Precision@5
|
| 348 |
- type: cosine_precision@10
|
| 349 |
-
value: 0.
|
| 350 |
name: Cosine Precision@10
|
| 351 |
- type: cosine_recall@1
|
| 352 |
-
value: 0.
|
| 353 |
name: Cosine Recall@1
|
| 354 |
- type: cosine_recall@3
|
| 355 |
-
value: 0.
|
| 356 |
name: Cosine Recall@3
|
| 357 |
- type: cosine_recall@5
|
| 358 |
value: 0.8
|
| 359 |
name: Cosine Recall@5
|
| 360 |
- type: cosine_recall@10
|
| 361 |
-
value: 0.
|
| 362 |
name: Cosine Recall@10
|
| 363 |
- type: cosine_ndcg@10
|
| 364 |
-
value: 0.
|
| 365 |
name: Cosine Ndcg@10
|
| 366 |
- type: cosine_mrr@10
|
| 367 |
-
value: 0.
|
| 368 |
name: Cosine Mrr@10
|
| 369 |
- type: cosine_map@100
|
| 370 |
-
value: 0.
|
| 371 |
name: Cosine Map@100
|
| 372 |
---
|
| 373 |
|
|
@@ -421,9 +426,9 @@ from sentence_transformers import SentenceTransformer
|
|
| 421 |
model = SentenceTransformer("Nuf-hugginface/modernbert-embed-quickb")
|
| 422 |
# Run inference
|
| 423 |
sentences = [
|
| 424 |
-
'What
|
| 425 |
-
'.
|
| 426 |
-
'
|
| 427 |
]
|
| 428 |
embeddings = model.encode(sentences)
|
| 429 |
print(embeddings.shape)
|
|
@@ -468,23 +473,23 @@ You can finetune this model on your own dataset.
|
|
| 468 |
* Datasets: `dim_768`, `dim_512`, `dim_256`, `dim_128` and `dim_64`
|
| 469 |
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
| 470 |
|
| 471 |
-
| Metric | dim_768 | dim_512 | dim_256 | dim_128 | dim_64
|
| 472 |
-
|
| 473 |
-
| cosine_accuracy@1 | 0.
|
| 474 |
-
| cosine_accuracy@3 | 0.
|
| 475 |
-
| cosine_accuracy@5 | 0.8667 | 0.
|
| 476 |
-
| cosine_accuracy@10 | 1.0 |
|
| 477 |
-
| cosine_precision@1 | 0.
|
| 478 |
-
| cosine_precision@3 | 0.
|
| 479 |
-
| cosine_precision@5 | 0.1733 | 0.
|
| 480 |
-
| cosine_precision@10 | 0.1 | 0.
|
| 481 |
-
| cosine_recall@1 | 0.
|
| 482 |
-
| cosine_recall@3 | 0.
|
| 483 |
-
| cosine_recall@5 | 0.8667 | 0.
|
| 484 |
-
| cosine_recall@10 | 1.0 |
|
| 485 |
-
| **cosine_ndcg@10** | **0.
|
| 486 |
-
| cosine_mrr@10 | 0.
|
| 487 |
-
| cosine_map@100 | 0.
|
| 488 |
|
| 489 |
<!--
|
| 490 |
## Bias, Risks and Limitations
|
|
@@ -504,19 +509,19 @@ You can finetune this model on your own dataset.
|
|
| 504 |
|
| 505 |
#### Unnamed Dataset
|
| 506 |
|
| 507 |
-
* Size:
|
| 508 |
* Columns: <code>anchor</code> and <code>positive</code>
|
| 509 |
-
* Approximate statistics based on the first
|
| 510 |
| | anchor | positive |
|
| 511 |
|:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
|
| 512 |
| type | string | string |
|
| 513 |
-
| details | <ul><li>min: 8 tokens</li><li>mean: 13.
|
| 514 |
* Samples:
|
| 515 |
-
| anchor
|
| 516 |
-
|
| 517 |
-
| <code>What
|
| 518 |
-
| <code>
|
| 519 |
-
| <code>What
|
| 520 |
* Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters:
|
| 521 |
```json
|
| 522 |
{
|
|
@@ -675,12 +680,13 @@ You can finetune this model on your own dataset.
|
|
| 675 |
</details>
|
| 676 |
|
| 677 |
### Training Logs
|
| 678 |
-
| Epoch | Step
|
| 679 |
-
|
| 680 |
-
| 1.0 |
|
| 681 |
-
| 2.0
|
| 682 |
-
|
|
| 683 |
-
| 3.
|
|
|
|
| 684 |
|
| 685 |
* The bold row denotes the saved checkpoint.
|
| 686 |
|
|
|
|
| 7 |
- sentence-similarity
|
| 8 |
- feature-extraction
|
| 9 |
- generated_from_trainer
|
| 10 |
+
- dataset_size:127
|
| 11 |
- loss:MatryoshkaLoss
|
| 12 |
- loss:MultipleNegativesRankingLoss
|
| 13 |
base_model: nomic-ai/modernbert-embed-base
|
| 14 |
widget:
|
| 15 |
+
- source_sentence: What does 'multi-modal' refer to in the context of the services
|
| 16 |
+
mentioned?
|
| 17 |
sentences:
|
| 18 |
+
- '1. Chatbots and Virtual Assistants
|
| 19 |
+
|
| 20 |
+
One of the most visible LLM integrations is in chatbots. Tools like ChatGPT, Claude,
|
| 21 |
+
and Bard are themselves chatbot interfaces built on LLMs. Many businesses are
|
| 22 |
+
now integrating these models into their websites and customer support systems.'
|
| 23 |
+
- For example, e-commerce websites can deploy LLM-powered assistants to help customers
|
| 24 |
+
find products, track orders, or get personalized recommendations—much more effectively
|
| 25 |
+
than traditional rule-based bots.
|
| 26 |
+
- Some services, like ColBERT, Marqo, and ColQwen, specialize in integrating LLMs
|
| 27 |
+
into search pipelines for both text and multi-modal (text + image) content.
|
| 28 |
+
- source_sentence: What is one method mentioned for deploying LLMs?
|
| 29 |
+
sentences:
|
| 30 |
+
- However, deploying LLMs effectively in real-world applications often requires
|
| 31 |
+
LLM integration. This means embedding these models into systems, workflows, or
|
| 32 |
+
products where they can interact with other components like databases, APIs, user
|
| 33 |
+
interfaces, or even custom business logic
|
| 34 |
+
- Some services, like ColBERT, Marqo, and ColQwen, specialize in integrating LLMs
|
| 35 |
+
into search pipelines for both text and multi-modal (text + image) content.
|
| 36 |
- However, deploying LLMs effectively in real-world applications often requires
|
| 37 |
LLM integration. This means embedding these models into systems, workflows, or
|
| 38 |
products where they can interact with other components like databases, APIs, user
|
| 39 |
interfaces, or even custom business logic
|
| 40 |
+
- source_sentence: What will an LLM likely respond with when prompted about the capital
|
| 41 |
+
of France?
|
| 42 |
sentences:
|
| 43 |
+
- . For instance, a spam filter doesn’t just block emails with specific keywords—it
|
| 44 |
+
learns from thousands of examples what spam typically looks like.
|
| 45 |
+
- Over the past few years, the field of ML has advanced rapidly, especially in the
|
| 46 |
+
area of Natural Language Processing (NLP)—the ability of machines to understand
|
| 47 |
+
and generate human language. At the forefront of this progress are Large Language
|
| 48 |
+
Models (LLMs), such as OpenAI’s GPT (Generative Pre-trained Transformer), Google’s
|
| 49 |
+
PaLM, and Meta’s LLaMA
|
| 50 |
+
- For example, given a prompt like "The capital of France is", an LLM trained on
|
| 51 |
+
a wide range of texts will likely respond with "Paris". But beyond trivia, LLMs
|
| 52 |
+
can write essays, draft emails, simulate conversations, generate code snippets,
|
| 53 |
+
and much more.
|
| 54 |
+
- source_sentence: What might an LLM be connected to in a customer support chatbot?
|
|
|
|
|
|
|
|
|
|
| 55 |
sentences:
|
| 56 |
+
- . For instance, a spam filter doesn’t just block emails with specific keywords—it
|
| 57 |
+
learns from thousands of examples what spam typically looks like.
|
| 58 |
+
- . For example, integrating an LLM into a customer support chatbot might involve
|
| 59 |
+
connecting it to a company’s internal knowledge base, enabling it to answer customer
|
| 60 |
+
questions using accurate, up-to-date information.
|
| 61 |
+
- Large Language Models (LLMs) and Their Integrations
|
| 62 |
+
- source_sentence: What type of dialogues can LLMs simulate?
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 63 |
sentences:
|
| 64 |
+
- 'Hallucinations: LLMs can sometimes generate plausible-sounding but incorrect
|
| 65 |
+
or fictional information.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 66 |
|
| 67 |
|
| 68 |
+
Data Privacy: Sending sensitive data to third-party models raises privacy and
|
| 69 |
+
compliance concerns.
|
| 70 |
|
| 71 |
|
| 72 |
+
Cost and Latency: Running LLMs, especially large ones, can be computationally
|
| 73 |
+
expensive and slow.'
|
| 74 |
+
- '6. APIs and Developer Tools
|
| 75 |
|
| 76 |
+
Developers can integrate LLMs into their own apps using APIs provided by companies
|
| 77 |
+
like OpenAI, Anthropic, and Cohere. These APIs allow developers to send prompts
|
| 78 |
+
and receive intelligent outputs in return.
|
| 79 |
+
|
| 80 |
+
|
| 81 |
+
This enables custom applications like:
|
| 82 |
+
|
| 83 |
+
|
| 84 |
+
Smart assistants in mobile apps
|
| 85 |
+
|
| 86 |
+
|
| 87 |
+
AI-powered research tools
|
| 88 |
+
|
| 89 |
+
|
| 90 |
+
Voice interfaces'
|
| 91 |
+
- '5. Education and Learning Platforms
|
| 92 |
+
|
| 93 |
+
Educational tools like Khanmigo (from Khan Academy) and other tutoring platforms
|
| 94 |
+
are leveraging LLMs to provide real-time help to students. LLMs can break down
|
| 95 |
+
complex topics, provide feedback on writing, and simulate Socratic-style dialogues.'
|
| 96 |
pipeline_tag: sentence-similarity
|
| 97 |
library_name: sentence-transformers
|
| 98 |
metrics:
|
|
|
|
| 122 |
type: dim_768
|
| 123 |
metrics:
|
| 124 |
- type: cosine_accuracy@1
|
| 125 |
+
value: 0.6
|
| 126 |
name: Cosine Accuracy@1
|
| 127 |
- type: cosine_accuracy@3
|
| 128 |
+
value: 0.8666666666666667
|
| 129 |
name: Cosine Accuracy@3
|
| 130 |
- type: cosine_accuracy@5
|
| 131 |
value: 0.8666666666666667
|
|
|
|
| 134 |
value: 1.0
|
| 135 |
name: Cosine Accuracy@10
|
| 136 |
- type: cosine_precision@1
|
| 137 |
+
value: 0.6
|
| 138 |
name: Cosine Precision@1
|
| 139 |
- type: cosine_precision@3
|
| 140 |
+
value: 0.28888888888888886
|
| 141 |
name: Cosine Precision@3
|
| 142 |
- type: cosine_precision@5
|
| 143 |
value: 0.17333333333333337
|
|
|
|
| 146 |
value: 0.10000000000000003
|
| 147 |
name: Cosine Precision@10
|
| 148 |
- type: cosine_recall@1
|
| 149 |
+
value: 0.6
|
| 150 |
name: Cosine Recall@1
|
| 151 |
- type: cosine_recall@3
|
| 152 |
+
value: 0.8666666666666667
|
| 153 |
name: Cosine Recall@3
|
| 154 |
- type: cosine_recall@5
|
| 155 |
value: 0.8666666666666667
|
|
|
|
| 158 |
value: 1.0
|
| 159 |
name: Cosine Recall@10
|
| 160 |
- type: cosine_ndcg@10
|
| 161 |
+
value: 0.8025374182760189
|
| 162 |
name: Cosine Ndcg@10
|
| 163 |
- type: cosine_mrr@10
|
| 164 |
+
value: 0.74
|
| 165 |
name: Cosine Mrr@10
|
| 166 |
- type: cosine_map@100
|
| 167 |
+
value: 0.74
|
| 168 |
name: Cosine Map@100
|
| 169 |
- task:
|
| 170 |
type: information-retrieval
|
|
|
|
| 174 |
type: dim_512
|
| 175 |
metrics:
|
| 176 |
- type: cosine_accuracy@1
|
| 177 |
+
value: 0.6666666666666666
|
| 178 |
name: Cosine Accuracy@1
|
| 179 |
- type: cosine_accuracy@3
|
| 180 |
value: 0.8
|
| 181 |
name: Cosine Accuracy@3
|
| 182 |
- type: cosine_accuracy@5
|
| 183 |
+
value: 0.8
|
| 184 |
name: Cosine Accuracy@5
|
| 185 |
- type: cosine_accuracy@10
|
| 186 |
+
value: 0.9333333333333333
|
| 187 |
name: Cosine Accuracy@10
|
| 188 |
- type: cosine_precision@1
|
| 189 |
+
value: 0.6666666666666666
|
| 190 |
name: Cosine Precision@1
|
| 191 |
- type: cosine_precision@3
|
| 192 |
+
value: 0.2666666666666667
|
| 193 |
name: Cosine Precision@3
|
| 194 |
- type: cosine_precision@5
|
| 195 |
+
value: 0.16000000000000003
|
| 196 |
name: Cosine Precision@5
|
| 197 |
- type: cosine_precision@10
|
| 198 |
+
value: 0.09333333333333335
|
| 199 |
name: Cosine Precision@10
|
| 200 |
- type: cosine_recall@1
|
| 201 |
+
value: 0.6666666666666666
|
| 202 |
name: Cosine Recall@1
|
| 203 |
- type: cosine_recall@3
|
| 204 |
value: 0.8
|
| 205 |
name: Cosine Recall@3
|
| 206 |
- type: cosine_recall@5
|
| 207 |
+
value: 0.8
|
| 208 |
name: Cosine Recall@5
|
| 209 |
- type: cosine_recall@10
|
| 210 |
+
value: 0.9333333333333333
|
| 211 |
name: Cosine Recall@10
|
| 212 |
- type: cosine_ndcg@10
|
| 213 |
+
value: 0.7955687714024445
|
| 214 |
name: Cosine Ndcg@10
|
| 215 |
- type: cosine_mrr@10
|
| 216 |
+
value: 0.7527777777777779
|
| 217 |
name: Cosine Mrr@10
|
| 218 |
- type: cosine_map@100
|
| 219 |
+
value: 0.7583333333333333
|
| 220 |
name: Cosine Map@100
|
| 221 |
- task:
|
| 222 |
type: information-retrieval
|
|
|
|
| 232 |
value: 0.8
|
| 233 |
name: Cosine Accuracy@3
|
| 234 |
- type: cosine_accuracy@5
|
| 235 |
+
value: 0.8
|
| 236 |
name: Cosine Accuracy@5
|
| 237 |
- type: cosine_accuracy@10
|
| 238 |
value: 1.0
|
|
|
|
| 241 |
value: 0.6666666666666666
|
| 242 |
name: Cosine Precision@1
|
| 243 |
- type: cosine_precision@3
|
| 244 |
+
value: 0.2666666666666667
|
| 245 |
name: Cosine Precision@3
|
| 246 |
- type: cosine_precision@5
|
| 247 |
+
value: 0.16000000000000003
|
| 248 |
name: Cosine Precision@5
|
| 249 |
- type: cosine_precision@10
|
| 250 |
value: 0.10000000000000003
|
|
|
|
| 256 |
value: 0.8
|
| 257 |
name: Cosine Recall@3
|
| 258 |
- type: cosine_recall@5
|
| 259 |
+
value: 0.8
|
| 260 |
name: Cosine Recall@5
|
| 261 |
- type: cosine_recall@10
|
| 262 |
value: 1.0
|
| 263 |
name: Cosine Recall@10
|
| 264 |
- type: cosine_ndcg@10
|
| 265 |
+
value: 0.7985736897839496
|
| 266 |
name: Cosine Ndcg@10
|
| 267 |
- type: cosine_mrr@10
|
| 268 |
+
value: 0.7384126984126984
|
| 269 |
name: Cosine Mrr@10
|
| 270 |
- type: cosine_map@100
|
| 271 |
+
value: 0.7384126984126984
|
| 272 |
name: Cosine Map@100
|
| 273 |
- task:
|
| 274 |
type: information-retrieval
|
|
|
|
| 278 |
type: dim_128
|
| 279 |
metrics:
|
| 280 |
- type: cosine_accuracy@1
|
| 281 |
+
value: 0.6666666666666666
|
| 282 |
name: Cosine Accuracy@1
|
| 283 |
- type: cosine_accuracy@3
|
| 284 |
+
value: 0.8
|
| 285 |
name: Cosine Accuracy@3
|
| 286 |
- type: cosine_accuracy@5
|
| 287 |
+
value: 0.8
|
| 288 |
name: Cosine Accuracy@5
|
| 289 |
- type: cosine_accuracy@10
|
| 290 |
+
value: 0.8666666666666667
|
| 291 |
name: Cosine Accuracy@10
|
| 292 |
- type: cosine_precision@1
|
| 293 |
+
value: 0.6666666666666666
|
| 294 |
name: Cosine Precision@1
|
| 295 |
- type: cosine_precision@3
|
| 296 |
+
value: 0.2666666666666667
|
| 297 |
name: Cosine Precision@3
|
| 298 |
- type: cosine_precision@5
|
| 299 |
+
value: 0.16000000000000003
|
| 300 |
name: Cosine Precision@5
|
| 301 |
- type: cosine_precision@10
|
| 302 |
+
value: 0.08666666666666668
|
| 303 |
name: Cosine Precision@10
|
| 304 |
- type: cosine_recall@1
|
| 305 |
+
value: 0.6666666666666666
|
| 306 |
name: Cosine Recall@1
|
| 307 |
- type: cosine_recall@3
|
| 308 |
+
value: 0.8
|
| 309 |
name: Cosine Recall@3
|
| 310 |
- type: cosine_recall@5
|
| 311 |
+
value: 0.8
|
| 312 |
name: Cosine Recall@5
|
| 313 |
- type: cosine_recall@10
|
| 314 |
+
value: 0.8666666666666667
|
| 315 |
name: Cosine Recall@10
|
| 316 |
- type: cosine_ndcg@10
|
| 317 |
+
value: 0.7700616222307202
|
| 318 |
name: Cosine Ndcg@10
|
| 319 |
- type: cosine_mrr@10
|
| 320 |
+
value: 0.74
|
| 321 |
name: Cosine Mrr@10
|
| 322 |
- type: cosine_map@100
|
| 323 |
+
value: 0.7479365079365079
|
| 324 |
name: Cosine Map@100
|
| 325 |
- task:
|
| 326 |
type: information-retrieval
|
|
|
|
| 330 |
type: dim_64
|
| 331 |
metrics:
|
| 332 |
- type: cosine_accuracy@1
|
| 333 |
+
value: 0.6
|
| 334 |
name: Cosine Accuracy@1
|
| 335 |
- type: cosine_accuracy@3
|
| 336 |
+
value: 0.8
|
| 337 |
name: Cosine Accuracy@3
|
| 338 |
- type: cosine_accuracy@5
|
| 339 |
value: 0.8
|
| 340 |
name: Cosine Accuracy@5
|
| 341 |
- type: cosine_accuracy@10
|
| 342 |
+
value: 0.8
|
| 343 |
name: Cosine Accuracy@10
|
| 344 |
- type: cosine_precision@1
|
| 345 |
+
value: 0.6
|
| 346 |
name: Cosine Precision@1
|
| 347 |
- type: cosine_precision@3
|
| 348 |
+
value: 0.2666666666666667
|
| 349 |
name: Cosine Precision@3
|
| 350 |
- type: cosine_precision@5
|
| 351 |
value: 0.16000000000000003
|
| 352 |
name: Cosine Precision@5
|
| 353 |
- type: cosine_precision@10
|
| 354 |
+
value: 0.08000000000000002
|
| 355 |
name: Cosine Precision@10
|
| 356 |
- type: cosine_recall@1
|
| 357 |
+
value: 0.6
|
| 358 |
name: Cosine Recall@1
|
| 359 |
- type: cosine_recall@3
|
| 360 |
+
value: 0.8
|
| 361 |
name: Cosine Recall@3
|
| 362 |
- type: cosine_recall@5
|
| 363 |
value: 0.8
|
| 364 |
name: Cosine Recall@5
|
| 365 |
- type: cosine_recall@10
|
| 366 |
+
value: 0.8
|
| 367 |
name: Cosine Recall@10
|
| 368 |
- type: cosine_ndcg@10
|
| 369 |
+
value: 0.7174573004761944
|
| 370 |
name: Cosine Ndcg@10
|
| 371 |
- type: cosine_mrr@10
|
| 372 |
+
value: 0.6888888888888889
|
| 373 |
name: Cosine Mrr@10
|
| 374 |
- type: cosine_map@100
|
| 375 |
+
value: 0.7003968253968255
|
| 376 |
name: Cosine Map@100
|
| 377 |
---
|
| 378 |
|
|
|
|
| 426 |
model = SentenceTransformer("Nuf-hugginface/modernbert-embed-quickb")
|
| 427 |
# Run inference
|
| 428 |
sentences = [
|
| 429 |
+
'What type of dialogues can LLMs simulate?',
|
| 430 |
+
'5. Education and Learning Platforms\nEducational tools like Khanmigo (from Khan Academy) and other tutoring platforms are leveraging LLMs to provide real-time help to students. LLMs can break down complex topics, provide feedback on writing, and simulate Socratic-style dialogues.',
|
| 431 |
+
'6. APIs and Developer Tools\nDevelopers can integrate LLMs into their own apps using APIs provided by companies like OpenAI, Anthropic, and Cohere. These APIs allow developers to send prompts and receive intelligent outputs in return.\n\nThis enables custom applications like:\n\nSmart assistants in mobile apps\n\nAI-powered research tools\n\nVoice interfaces',
|
| 432 |
]
|
| 433 |
embeddings = model.encode(sentences)
|
| 434 |
print(embeddings.shape)
|
|
|
|
| 473 |
* Datasets: `dim_768`, `dim_512`, `dim_256`, `dim_128` and `dim_64`
|
| 474 |
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
| 475 |
|
| 476 |
+
| Metric | dim_768 | dim_512 | dim_256 | dim_128 | dim_64 |
|
| 477 |
+
|:--------------------|:-----------|:-----------|:-----------|:-----------|:-----------|
|
| 478 |
+
| cosine_accuracy@1 | 0.6 | 0.6667 | 0.6667 | 0.6667 | 0.6 |
|
| 479 |
+
| cosine_accuracy@3 | 0.8667 | 0.8 | 0.8 | 0.8 | 0.8 |
|
| 480 |
+
| cosine_accuracy@5 | 0.8667 | 0.8 | 0.8 | 0.8 | 0.8 |
|
| 481 |
+
| cosine_accuracy@10 | 1.0 | 0.9333 | 1.0 | 0.8667 | 0.8 |
|
| 482 |
+
| cosine_precision@1 | 0.6 | 0.6667 | 0.6667 | 0.6667 | 0.6 |
|
| 483 |
+
| cosine_precision@3 | 0.2889 | 0.2667 | 0.2667 | 0.2667 | 0.2667 |
|
| 484 |
+
| cosine_precision@5 | 0.1733 | 0.16 | 0.16 | 0.16 | 0.16 |
|
| 485 |
+
| cosine_precision@10 | 0.1 | 0.0933 | 0.1 | 0.0867 | 0.08 |
|
| 486 |
+
| cosine_recall@1 | 0.6 | 0.6667 | 0.6667 | 0.6667 | 0.6 |
|
| 487 |
+
| cosine_recall@3 | 0.8667 | 0.8 | 0.8 | 0.8 | 0.8 |
|
| 488 |
+
| cosine_recall@5 | 0.8667 | 0.8 | 0.8 | 0.8 | 0.8 |
|
| 489 |
+
| cosine_recall@10 | 1.0 | 0.9333 | 1.0 | 0.8667 | 0.8 |
|
| 490 |
+
| **cosine_ndcg@10** | **0.8025** | **0.7956** | **0.7986** | **0.7701** | **0.7175** |
|
| 491 |
+
| cosine_mrr@10 | 0.74 | 0.7528 | 0.7384 | 0.74 | 0.6889 |
|
| 492 |
+
| cosine_map@100 | 0.74 | 0.7583 | 0.7384 | 0.7479 | 0.7004 |
|
| 493 |
|
| 494 |
<!--
|
| 495 |
## Bias, Risks and Limitations
|
|
|
|
| 509 |
|
| 510 |
#### Unnamed Dataset
|
| 511 |
|
| 512 |
+
* Size: 127 training samples
|
| 513 |
* Columns: <code>anchor</code> and <code>positive</code>
|
| 514 |
+
* Approximate statistics based on the first 127 samples:
|
| 515 |
| | anchor | positive |
|
| 516 |
|:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
|
| 517 |
| type | string | string |
|
| 518 |
+
| details | <ul><li>min: 8 tokens</li><li>mean: 13.2 tokens</li><li>max: 20 tokens</li></ul> | <ul><li>min: 13 tokens</li><li>mean: 53.85 tokens</li><li>max: 86 tokens</li></ul> |
|
| 519 |
* Samples:
|
| 520 |
+
| anchor | positive |
|
| 521 |
+
|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
| 522 |
+
| <code>What documents could the system retrieve in relation to sick leave?</code> | <code>For instance, in a document management system, a user might type "policies about sick leave", and the system—integrated with an LLM—could retrieve documents discussing "medical leave", "employee absence", and "illness policies", even if those exact words weren’t used.</code> |
|
| 523 |
+
| <code>What is one of the most visible integrations of LLM technology?</code> | <code>1. Chatbots and Virtual Assistants<br>One of the most visible LLM integrations is in chatbots. Tools like ChatGPT, Claude, and Bard are themselves chatbot interfaces built on LLMs. Many businesses are now integrating these models into their websites and customer support systems.</code> |
|
| 524 |
+
| <code>What does AI stand for?</code> | <code>Introduction to AI, Machine Learning, LLMs, and Their Integration</code> |
|
| 525 |
* Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters:
|
| 526 |
```json
|
| 527 |
{
|
|
|
|
| 680 |
</details>
|
| 681 |
|
| 682 |
### Training Logs
|
| 683 |
+
| Epoch | Step | Training Loss | dim_768_cosine_ndcg@10 | dim_512_cosine_ndcg@10 | dim_256_cosine_ndcg@10 | dim_128_cosine_ndcg@10 | dim_64_cosine_ndcg@10 |
|
| 684 |
+
|:-------:|:-----:|:-------------:|:----------------------:|:----------------------:|:----------------------:|:----------------------:|:---------------------:|
|
| 685 |
+
| 1.0 | 4 | - | 0.7853 | 0.8214 | 0.7673 | 0.7586 | 0.6883 |
|
| 686 |
+
| **2.0** | **8** | **-** | **0.7764** | **0.7902** | **0.7686** | **0.7701** | **0.7321** |
|
| 687 |
+
| 2.5 | 10 | 13.8004 | - | - | - | - | - |
|
| 688 |
+
| 3.0 | 12 | - | 0.8028 | 0.7710 | 0.7932 | 0.7701 | 0.7175 |
|
| 689 |
+
| 4.0 | 16 | - | 0.8025 | 0.7956 | 0.7986 | 0.7701 | 0.7175 |
|
| 690 |
|
| 691 |
* The bold row denotes the saved checkpoint.
|
| 692 |
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 596070136
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7fa83288c96da5c91a8d0a7f680fa88e3010ec41496910019839fb29c898aa3
|
| 3 |
size 596070136
|