Nuf-hugginface commited on
Commit
2aec704
·
verified ·
1 Parent(s): 138adb2

Add new SentenceTransformer model

Browse files
Files changed (2) hide show
  1. README.md +160 -154
  2. model.safetensors +1 -1
README.md CHANGED
@@ -7,87 +7,92 @@ tags:
7
  - sentence-similarity
8
  - feature-extraction
9
  - generated_from_trainer
10
- - dataset_size:129
11
  - loss:MatryoshkaLoss
12
  - loss:MultipleNegativesRankingLoss
13
  base_model: nomic-ai/modernbert-embed-base
14
  widget:
15
- - source_sentence: In what contexts can LLMs be embedded according to the text?
 
16
  sentences:
17
- - Artificial Intelligence (AI) is the broad field of computer science that focuses
18
- on building systems capable of performing tasks that normally require human intelligence.
19
- These tasks include learning from experience, understanding language, recognizing
20
- patterns, and making decisions. AI powers everything from smart assistants like
21
- Siri to recommendation systems on Netflix and self-driving cars.
22
- - In software development, tools like GitHub Copilot integrate LLMs to assist programmers
23
- by generating code, commenting on functions, and detecting bugs.
 
 
 
 
 
 
 
 
 
 
 
24
  - However, deploying LLMs effectively in real-world applications often requires
25
  LLM integration. This means embedding these models into systems, workflows, or
26
  products where they can interact with other components like databases, APIs, user
27
  interfaces, or even custom business logic
28
- - source_sentence: What is one educational tool mentioned that uses LLMs?
 
29
  sentences:
30
- - . As organizations increasingly adopt these technologies, the ability to understand
31
- and apply LLMs will be a critical skill in the AI-powered future.
32
- - '5. Education and Learning Platforms
33
-
34
- Educational tools like Khanmigo (from Khan Academy) and other tutoring platforms
35
- are leveraging LLMs to provide real-time help to students. LLMs can break down
36
- complex topics, provide feedback on writing, and simulate Socratic-style dialogues.'
37
- - '7. Enterprise Integrations
38
-
39
- In enterprises, LLMs are being tied into internal systems like SharePoint, Slack,
40
- Jira, and Confluence to act as knowledge assistants. Employees can ask natural
41
- language questions like “What’s the latest update on Project Delta?” and get context-rich
42
- answers based on internal documents and discussions.'
43
- - source_sentence: Can the system retrieve documents even if the exact words weren't
44
- used?
45
  sentences:
46
- - '7. Enterprise Integrations
47
-
48
- In enterprises, LLMs are being tied into internal systems like SharePoint, Slack,
49
- Jira, and Confluence to act as knowledge assistants. Employees can ask natural
50
- language questions like “What’s the latest update on Project Delta?” and get context-rich
51
- answers based on internal documents and discussions.'
52
- - Companies are also experimenting with Retrieval-Augmented Generation (RAG)—a technique
53
- where LLMs are paired with document databases (e.g., vector stores like Supabase,
54
- Pinecone, or Weaviate) to answer questions with enterprise-specific knowledge.
55
- - For instance, in a document management system, a user might type "policies about
56
- sick leave", and the system—integrated with an LLM—could retrieve documents discussing
57
- "medical leave", "employee absence", and "illness policies", even if those exact
58
- words weren’t used.
59
- - source_sentence: What are some techniques mentioned for mitigating challenges in
60
- prompt engineering?
61
  sentences:
62
- - . These include text generation, summarization, translation, question answering,
63
- code generation, and more.
64
- - . These models are trained on massive text datasets and are capable of generating
65
- coherent, context-aware language, answering questions, summarizing documents,
66
- writing code, and more.
67
- - 'Prompt Engineering: Designing effective prompts and interactions is a new and
68
- still-evolving skill.
69
 
70
 
71
- Mitigating these challenges often involves techniques like prompt tuning, fine-tuning,
72
- hybrid search, caching, and using smaller models for certain tasks.
73
 
74
 
75
- The Future of LLM Integrations
 
 
76
 
77
- As LLMs evolve, we’ll see deeper and more seamless integration into everyday tools.
78
- The future points to:'
79
- - source_sentence: What are these models trained on?
80
- sentences:
81
- - Ultimately, the integration of LLMs across platforms, tools, and workflows is
82
- transforming how we interact with information and machines—making software more
83
- conversational, intelligent, and context-aware.
84
- - . These models are trained on massive text datasets and are capable of generating
85
- coherent, context-aware language, answering questions, summarizing documents,
86
- writing code, and more.
87
- - LLMs work by learning statistical relationships between words and phrases, allowing
88
- them to predict and generate language that feels natural. The power of these models
89
- lies not only in their size but also in the diversity of tasks they can perform
90
- with little to no task-specific training
 
 
 
 
 
 
91
  pipeline_tag: sentence-similarity
92
  library_name: sentence-transformers
93
  metrics:
@@ -117,10 +122,10 @@ model-index:
117
  type: dim_768
118
  metrics:
119
  - type: cosine_accuracy@1
120
- value: 0.7333333333333333
121
  name: Cosine Accuracy@1
122
  - type: cosine_accuracy@3
123
- value: 0.8
124
  name: Cosine Accuracy@3
125
  - type: cosine_accuracy@5
126
  value: 0.8666666666666667
@@ -129,10 +134,10 @@ model-index:
129
  value: 1.0
130
  name: Cosine Accuracy@10
131
  - type: cosine_precision@1
132
- value: 0.7333333333333333
133
  name: Cosine Precision@1
134
  - type: cosine_precision@3
135
- value: 0.26666666666666666
136
  name: Cosine Precision@3
137
  - type: cosine_precision@5
138
  value: 0.17333333333333337
@@ -141,10 +146,10 @@ model-index:
141
  value: 0.10000000000000003
142
  name: Cosine Precision@10
143
  - type: cosine_recall@1
144
- value: 0.7333333333333333
145
  name: Cosine Recall@1
146
  - type: cosine_recall@3
147
- value: 0.8
148
  name: Cosine Recall@3
149
  - type: cosine_recall@5
150
  value: 0.8666666666666667
@@ -153,13 +158,13 @@ model-index:
153
  value: 1.0
154
  name: Cosine Recall@10
155
  - type: cosine_ndcg@10
156
- value: 0.8434763926535543
157
  name: Cosine Ndcg@10
158
  - type: cosine_mrr@10
159
- value: 0.7969312169312168
160
  name: Cosine Mrr@10
161
  - type: cosine_map@100
162
- value: 0.7969312169312168
163
  name: Cosine Map@100
164
  - task:
165
  type: information-retrieval
@@ -169,49 +174,49 @@ model-index:
169
  type: dim_512
170
  metrics:
171
  - type: cosine_accuracy@1
172
- value: 0.7333333333333333
173
  name: Cosine Accuracy@1
174
  - type: cosine_accuracy@3
175
  value: 0.8
176
  name: Cosine Accuracy@3
177
  - type: cosine_accuracy@5
178
- value: 0.8666666666666667
179
  name: Cosine Accuracy@5
180
  - type: cosine_accuracy@10
181
- value: 1.0
182
  name: Cosine Accuracy@10
183
  - type: cosine_precision@1
184
- value: 0.7333333333333333
185
  name: Cosine Precision@1
186
  - type: cosine_precision@3
187
- value: 0.26666666666666666
188
  name: Cosine Precision@3
189
  - type: cosine_precision@5
190
- value: 0.17333333333333337
191
  name: Cosine Precision@5
192
  - type: cosine_precision@10
193
- value: 0.10000000000000003
194
  name: Cosine Precision@10
195
  - type: cosine_recall@1
196
- value: 0.7333333333333333
197
  name: Cosine Recall@1
198
  - type: cosine_recall@3
199
  value: 0.8
200
  name: Cosine Recall@3
201
  - type: cosine_recall@5
202
- value: 0.8666666666666667
203
  name: Cosine Recall@5
204
  - type: cosine_recall@10
205
- value: 1.0
206
  name: Cosine Recall@10
207
  - type: cosine_ndcg@10
208
- value: 0.8422851622170473
209
  name: Cosine Ndcg@10
210
  - type: cosine_mrr@10
211
- value: 0.7957407407407406
212
  name: Cosine Mrr@10
213
  - type: cosine_map@100
214
- value: 0.7957407407407406
215
  name: Cosine Map@100
216
  - task:
217
  type: information-retrieval
@@ -227,7 +232,7 @@ model-index:
227
  value: 0.8
228
  name: Cosine Accuracy@3
229
  - type: cosine_accuracy@5
230
- value: 0.8666666666666667
231
  name: Cosine Accuracy@5
232
  - type: cosine_accuracy@10
233
  value: 1.0
@@ -236,10 +241,10 @@ model-index:
236
  value: 0.6666666666666666
237
  name: Cosine Precision@1
238
  - type: cosine_precision@3
239
- value: 0.26666666666666666
240
  name: Cosine Precision@3
241
  - type: cosine_precision@5
242
- value: 0.17333333333333337
243
  name: Cosine Precision@5
244
  - type: cosine_precision@10
245
  value: 0.10000000000000003
@@ -251,19 +256,19 @@ model-index:
251
  value: 0.8
252
  name: Cosine Recall@3
253
  - type: cosine_recall@5
254
- value: 0.8666666666666667
255
  name: Cosine Recall@5
256
  - type: cosine_recall@10
257
  value: 1.0
258
  name: Cosine Recall@10
259
  - type: cosine_ndcg@10
260
- value: 0.810143059320221
261
  name: Cosine Ndcg@10
262
  - type: cosine_mrr@10
263
- value: 0.7524867724867724
264
  name: Cosine Mrr@10
265
  - type: cosine_map@100
266
- value: 0.7524867724867724
267
  name: Cosine Map@100
268
  - task:
269
  type: information-retrieval
@@ -273,49 +278,49 @@ model-index:
273
  type: dim_128
274
  metrics:
275
  - type: cosine_accuracy@1
276
- value: 0.5333333333333333
277
  name: Cosine Accuracy@1
278
  - type: cosine_accuracy@3
279
- value: 0.7333333333333333
280
  name: Cosine Accuracy@3
281
  - type: cosine_accuracy@5
282
- value: 0.8666666666666667
283
  name: Cosine Accuracy@5
284
  - type: cosine_accuracy@10
285
- value: 0.9333333333333333
286
  name: Cosine Accuracy@10
287
  - type: cosine_precision@1
288
- value: 0.5333333333333333
289
  name: Cosine Precision@1
290
  - type: cosine_precision@3
291
- value: 0.2444444444444445
292
  name: Cosine Precision@3
293
  - type: cosine_precision@5
294
- value: 0.17333333333333337
295
  name: Cosine Precision@5
296
  - type: cosine_precision@10
297
- value: 0.09333333333333335
298
  name: Cosine Precision@10
299
  - type: cosine_recall@1
300
- value: 0.5333333333333333
301
  name: Cosine Recall@1
302
  - type: cosine_recall@3
303
- value: 0.7333333333333333
304
  name: Cosine Recall@3
305
  - type: cosine_recall@5
306
- value: 0.8666666666666667
307
  name: Cosine Recall@5
308
  - type: cosine_recall@10
309
- value: 0.9333333333333333
310
  name: Cosine Recall@10
311
  - type: cosine_ndcg@10
312
- value: 0.7245635799179159
313
  name: Cosine Ndcg@10
314
  - type: cosine_mrr@10
315
- value: 0.6588888888888889
316
  name: Cosine Mrr@10
317
  - type: cosine_map@100
318
- value: 0.6630555555555555
319
  name: Cosine Map@100
320
  - task:
321
  type: information-retrieval
@@ -325,49 +330,49 @@ model-index:
325
  type: dim_64
326
  metrics:
327
  - type: cosine_accuracy@1
328
- value: 0.4666666666666667
329
  name: Cosine Accuracy@1
330
  - type: cosine_accuracy@3
331
- value: 0.6
332
  name: Cosine Accuracy@3
333
  - type: cosine_accuracy@5
334
  value: 0.8
335
  name: Cosine Accuracy@5
336
  - type: cosine_accuracy@10
337
- value: 0.8666666666666667
338
  name: Cosine Accuracy@10
339
  - type: cosine_precision@1
340
- value: 0.4666666666666667
341
  name: Cosine Precision@1
342
  - type: cosine_precision@3
343
- value: 0.2
344
  name: Cosine Precision@3
345
  - type: cosine_precision@5
346
  value: 0.16000000000000003
347
  name: Cosine Precision@5
348
  - type: cosine_precision@10
349
- value: 0.08666666666666668
350
  name: Cosine Precision@10
351
  - type: cosine_recall@1
352
- value: 0.4666666666666667
353
  name: Cosine Recall@1
354
  - type: cosine_recall@3
355
- value: 0.6
356
  name: Cosine Recall@3
357
  - type: cosine_recall@5
358
  value: 0.8
359
  name: Cosine Recall@5
360
  - type: cosine_recall@10
361
- value: 0.8666666666666667
362
  name: Cosine Recall@10
363
  - type: cosine_ndcg@10
364
- value: 0.6490228576040539
365
  name: Cosine Ndcg@10
366
  - type: cosine_mrr@10
367
- value: 0.58
368
  name: Cosine Mrr@10
369
  - type: cosine_map@100
370
- value: 0.5892352092352092
371
  name: Cosine Map@100
372
  ---
373
 
@@ -421,9 +426,9 @@ from sentence_transformers import SentenceTransformer
421
  model = SentenceTransformer("Nuf-hugginface/modernbert-embed-quickb")
422
  # Run inference
423
  sentences = [
424
- 'What are these models trained on?',
425
- '. These models are trained on massive text datasets and are capable of generating coherent, context-aware language, answering questions, summarizing documents, writing code, and more.',
426
- 'Ultimately, the integration of LLMs across platforms, tools, and workflows is transforming how we interact with information and machines—making software more conversational, intelligent, and context-aware.',
427
  ]
428
  embeddings = model.encode(sentences)
429
  print(embeddings.shape)
@@ -468,23 +473,23 @@ You can finetune this model on your own dataset.
468
  * Datasets: `dim_768`, `dim_512`, `dim_256`, `dim_128` and `dim_64`
469
  * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
470
 
471
- | Metric | dim_768 | dim_512 | dim_256 | dim_128 | dim_64 |
472
- |:--------------------|:-----------|:-----------|:-----------|:-----------|:----------|
473
- | cosine_accuracy@1 | 0.7333 | 0.7333 | 0.6667 | 0.5333 | 0.4667 |
474
- | cosine_accuracy@3 | 0.8 | 0.8 | 0.8 | 0.7333 | 0.6 |
475
- | cosine_accuracy@5 | 0.8667 | 0.8667 | 0.8667 | 0.8667 | 0.8 |
476
- | cosine_accuracy@10 | 1.0 | 1.0 | 1.0 | 0.9333 | 0.8667 |
477
- | cosine_precision@1 | 0.7333 | 0.7333 | 0.6667 | 0.5333 | 0.4667 |
478
- | cosine_precision@3 | 0.2667 | 0.2667 | 0.2667 | 0.2444 | 0.2 |
479
- | cosine_precision@5 | 0.1733 | 0.1733 | 0.1733 | 0.1733 | 0.16 |
480
- | cosine_precision@10 | 0.1 | 0.1 | 0.1 | 0.0933 | 0.0867 |
481
- | cosine_recall@1 | 0.7333 | 0.7333 | 0.6667 | 0.5333 | 0.4667 |
482
- | cosine_recall@3 | 0.8 | 0.8 | 0.8 | 0.7333 | 0.6 |
483
- | cosine_recall@5 | 0.8667 | 0.8667 | 0.8667 | 0.8667 | 0.8 |
484
- | cosine_recall@10 | 1.0 | 1.0 | 1.0 | 0.9333 | 0.8667 |
485
- | **cosine_ndcg@10** | **0.8435** | **0.8423** | **0.8101** | **0.7246** | **0.649** |
486
- | cosine_mrr@10 | 0.7969 | 0.7957 | 0.7525 | 0.6589 | 0.58 |
487
- | cosine_map@100 | 0.7969 | 0.7957 | 0.7525 | 0.6631 | 0.5892 |
488
 
489
  <!--
490
  ## Bias, Risks and Limitations
@@ -504,19 +509,19 @@ You can finetune this model on your own dataset.
504
 
505
  #### Unnamed Dataset
506
 
507
- * Size: 129 training samples
508
  * Columns: <code>anchor</code> and <code>positive</code>
509
- * Approximate statistics based on the first 129 samples:
510
  | | anchor | positive |
511
  |:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
512
  | type | string | string |
513
- | details | <ul><li>min: 8 tokens</li><li>mean: 13.8 tokens</li><li>max: 25 tokens</li></ul> | <ul><li>min: 13 tokens</li><li>mean: 53.68 tokens</li><li>max: 86 tokens</li></ul> |
514
  * Samples:
515
- | anchor | positive |
516
- |:-----------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
517
- | <code>What is the primary ability discussed in the text?</code> | <code>. This generalization ability makes them incredibly useful across industries—from customer service and education to software development and healthcare.</code> |
518
- | <code>How many tasks are listed in the text?</code> | <code>. These include text generation, summarization, translation, question answering, code generation, and more.</code> |
519
- | <code>What are examples of chatbot tools mentioned in the text?</code> | <code>1. Chatbots and Virtual Assistants<br>One of the most visible LLM integrations is in chatbots. Tools like ChatGPT, Claude, and Bard are themselves chatbot interfaces built on LLMs. Many businesses are now integrating these models into their websites and customer support systems.</code> |
520
  * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters:
521
  ```json
522
  {
@@ -675,12 +680,13 @@ You can finetune this model on your own dataset.
675
  </details>
676
 
677
  ### Training Logs
678
- | Epoch | Step | Training Loss | dim_768_cosine_ndcg@10 | dim_512_cosine_ndcg@10 | dim_256_cosine_ndcg@10 | dim_128_cosine_ndcg@10 | dim_64_cosine_ndcg@10 |
679
- |:-------:|:------:|:-------------:|:----------------------:|:----------------------:|:----------------------:|:----------------------:|:---------------------:|
680
- | 1.0 | 5 | - | 0.8450 | 0.8003 | 0.8117 | 0.7009 | 0.6370 |
681
- | 2.0 | 10 | 12.0802 | 0.8427 | 0.8222 | 0.8055 | 0.6979 | 0.6608 |
682
- | **3.0** | **15** | **-** | **0.8435** | **0.8423** | **0.8101** | **0.7246** | **0.649** |
683
- | 3.2424 | 16 | - | 0.8435 | 0.8423 | 0.8101 | 0.7246 | 0.6490 |
 
684
 
685
  * The bold row denotes the saved checkpoint.
686
 
 
7
  - sentence-similarity
8
  - feature-extraction
9
  - generated_from_trainer
10
+ - dataset_size:127
11
  - loss:MatryoshkaLoss
12
  - loss:MultipleNegativesRankingLoss
13
  base_model: nomic-ai/modernbert-embed-base
14
  widget:
15
+ - source_sentence: What does 'multi-modal' refer to in the context of the services
16
+ mentioned?
17
  sentences:
18
+ - '1. Chatbots and Virtual Assistants
19
+
20
+ One of the most visible LLM integrations is in chatbots. Tools like ChatGPT, Claude,
21
+ and Bard are themselves chatbot interfaces built on LLMs. Many businesses are
22
+ now integrating these models into their websites and customer support systems.'
23
+ - For example, e-commerce websites can deploy LLM-powered assistants to help customers
24
+ find products, track orders, or get personalized recommendations—much more effectively
25
+ than traditional rule-based bots.
26
+ - Some services, like ColBERT, Marqo, and ColQwen, specialize in integrating LLMs
27
+ into search pipelines for both text and multi-modal (text + image) content.
28
+ - source_sentence: What is one method mentioned for deploying LLMs?
29
+ sentences:
30
+ - However, deploying LLMs effectively in real-world applications often requires
31
+ LLM integration. This means embedding these models into systems, workflows, or
32
+ products where they can interact with other components like databases, APIs, user
33
+ interfaces, or even custom business logic
34
+ - Some services, like ColBERT, Marqo, and ColQwen, specialize in integrating LLMs
35
+ into search pipelines for both text and multi-modal (text + image) content.
36
  - However, deploying LLMs effectively in real-world applications often requires
37
  LLM integration. This means embedding these models into systems, workflows, or
38
  products where they can interact with other components like databases, APIs, user
39
  interfaces, or even custom business logic
40
+ - source_sentence: What will an LLM likely respond with when prompted about the capital
41
+ of France?
42
  sentences:
43
+ - . For instance, a spam filter doesn’t just block emails with specific keywords—it
44
+ learns from thousands of examples what spam typically looks like.
45
+ - Over the past few years, the field of ML has advanced rapidly, especially in the
46
+ area of Natural Language Processing (NLP)—the ability of machines to understand
47
+ and generate human language. At the forefront of this progress are Large Language
48
+ Models (LLMs), such as OpenAI’s GPT (Generative Pre-trained Transformer), Google’s
49
+ PaLM, and Meta’s LLaMA
50
+ - For example, given a prompt like "The capital of France is", an LLM trained on
51
+ a wide range of texts will likely respond with "Paris". But beyond trivia, LLMs
52
+ can write essays, draft emails, simulate conversations, generate code snippets,
53
+ and much more.
54
+ - source_sentence: What might an LLM be connected to in a customer support chatbot?
 
 
 
55
  sentences:
56
+ - . For instance, a spam filter doesn’t just block emails with specific keywords—it
57
+ learns from thousands of examples what spam typically looks like.
58
+ - . For example, integrating an LLM into a customer support chatbot might involve
59
+ connecting it to a company’s internal knowledge base, enabling it to answer customer
60
+ questions using accurate, up-to-date information.
61
+ - Large Language Models (LLMs) and Their Integrations
62
+ - source_sentence: What type of dialogues can LLMs simulate?
 
 
 
 
 
 
 
 
63
  sentences:
64
+ - 'Hallucinations: LLMs can sometimes generate plausible-sounding but incorrect
65
+ or fictional information.
 
 
 
 
 
66
 
67
 
68
+ Data Privacy: Sending sensitive data to third-party models raises privacy and
69
+ compliance concerns.
70
 
71
 
72
+ Cost and Latency: Running LLMs, especially large ones, can be computationally
73
+ expensive and slow.'
74
+ - '6. APIs and Developer Tools
75
 
76
+ Developers can integrate LLMs into their own apps using APIs provided by companies
77
+ like OpenAI, Anthropic, and Cohere. These APIs allow developers to send prompts
78
+ and receive intelligent outputs in return.
79
+
80
+
81
+ This enables custom applications like:
82
+
83
+
84
+ Smart assistants in mobile apps
85
+
86
+
87
+ AI-powered research tools
88
+
89
+
90
+ Voice interfaces'
91
+ - '5. Education and Learning Platforms
92
+
93
+ Educational tools like Khanmigo (from Khan Academy) and other tutoring platforms
94
+ are leveraging LLMs to provide real-time help to students. LLMs can break down
95
+ complex topics, provide feedback on writing, and simulate Socratic-style dialogues.'
96
  pipeline_tag: sentence-similarity
97
  library_name: sentence-transformers
98
  metrics:
 
122
  type: dim_768
123
  metrics:
124
  - type: cosine_accuracy@1
125
+ value: 0.6
126
  name: Cosine Accuracy@1
127
  - type: cosine_accuracy@3
128
+ value: 0.8666666666666667
129
  name: Cosine Accuracy@3
130
  - type: cosine_accuracy@5
131
  value: 0.8666666666666667
 
134
  value: 1.0
135
  name: Cosine Accuracy@10
136
  - type: cosine_precision@1
137
+ value: 0.6
138
  name: Cosine Precision@1
139
  - type: cosine_precision@3
140
+ value: 0.28888888888888886
141
  name: Cosine Precision@3
142
  - type: cosine_precision@5
143
  value: 0.17333333333333337
 
146
  value: 0.10000000000000003
147
  name: Cosine Precision@10
148
  - type: cosine_recall@1
149
+ value: 0.6
150
  name: Cosine Recall@1
151
  - type: cosine_recall@3
152
+ value: 0.8666666666666667
153
  name: Cosine Recall@3
154
  - type: cosine_recall@5
155
  value: 0.8666666666666667
 
158
  value: 1.0
159
  name: Cosine Recall@10
160
  - type: cosine_ndcg@10
161
+ value: 0.8025374182760189
162
  name: Cosine Ndcg@10
163
  - type: cosine_mrr@10
164
+ value: 0.74
165
  name: Cosine Mrr@10
166
  - type: cosine_map@100
167
+ value: 0.74
168
  name: Cosine Map@100
169
  - task:
170
  type: information-retrieval
 
174
  type: dim_512
175
  metrics:
176
  - type: cosine_accuracy@1
177
+ value: 0.6666666666666666
178
  name: Cosine Accuracy@1
179
  - type: cosine_accuracy@3
180
  value: 0.8
181
  name: Cosine Accuracy@3
182
  - type: cosine_accuracy@5
183
+ value: 0.8
184
  name: Cosine Accuracy@5
185
  - type: cosine_accuracy@10
186
+ value: 0.9333333333333333
187
  name: Cosine Accuracy@10
188
  - type: cosine_precision@1
189
+ value: 0.6666666666666666
190
  name: Cosine Precision@1
191
  - type: cosine_precision@3
192
+ value: 0.2666666666666667
193
  name: Cosine Precision@3
194
  - type: cosine_precision@5
195
+ value: 0.16000000000000003
196
  name: Cosine Precision@5
197
  - type: cosine_precision@10
198
+ value: 0.09333333333333335
199
  name: Cosine Precision@10
200
  - type: cosine_recall@1
201
+ value: 0.6666666666666666
202
  name: Cosine Recall@1
203
  - type: cosine_recall@3
204
  value: 0.8
205
  name: Cosine Recall@3
206
  - type: cosine_recall@5
207
+ value: 0.8
208
  name: Cosine Recall@5
209
  - type: cosine_recall@10
210
+ value: 0.9333333333333333
211
  name: Cosine Recall@10
212
  - type: cosine_ndcg@10
213
+ value: 0.7955687714024445
214
  name: Cosine Ndcg@10
215
  - type: cosine_mrr@10
216
+ value: 0.7527777777777779
217
  name: Cosine Mrr@10
218
  - type: cosine_map@100
219
+ value: 0.7583333333333333
220
  name: Cosine Map@100
221
  - task:
222
  type: information-retrieval
 
232
  value: 0.8
233
  name: Cosine Accuracy@3
234
  - type: cosine_accuracy@5
235
+ value: 0.8
236
  name: Cosine Accuracy@5
237
  - type: cosine_accuracy@10
238
  value: 1.0
 
241
  value: 0.6666666666666666
242
  name: Cosine Precision@1
243
  - type: cosine_precision@3
244
+ value: 0.2666666666666667
245
  name: Cosine Precision@3
246
  - type: cosine_precision@5
247
+ value: 0.16000000000000003
248
  name: Cosine Precision@5
249
  - type: cosine_precision@10
250
  value: 0.10000000000000003
 
256
  value: 0.8
257
  name: Cosine Recall@3
258
  - type: cosine_recall@5
259
+ value: 0.8
260
  name: Cosine Recall@5
261
  - type: cosine_recall@10
262
  value: 1.0
263
  name: Cosine Recall@10
264
  - type: cosine_ndcg@10
265
+ value: 0.7985736897839496
266
  name: Cosine Ndcg@10
267
  - type: cosine_mrr@10
268
+ value: 0.7384126984126984
269
  name: Cosine Mrr@10
270
  - type: cosine_map@100
271
+ value: 0.7384126984126984
272
  name: Cosine Map@100
273
  - task:
274
  type: information-retrieval
 
278
  type: dim_128
279
  metrics:
280
  - type: cosine_accuracy@1
281
+ value: 0.6666666666666666
282
  name: Cosine Accuracy@1
283
  - type: cosine_accuracy@3
284
+ value: 0.8
285
  name: Cosine Accuracy@3
286
  - type: cosine_accuracy@5
287
+ value: 0.8
288
  name: Cosine Accuracy@5
289
  - type: cosine_accuracy@10
290
+ value: 0.8666666666666667
291
  name: Cosine Accuracy@10
292
  - type: cosine_precision@1
293
+ value: 0.6666666666666666
294
  name: Cosine Precision@1
295
  - type: cosine_precision@3
296
+ value: 0.2666666666666667
297
  name: Cosine Precision@3
298
  - type: cosine_precision@5
299
+ value: 0.16000000000000003
300
  name: Cosine Precision@5
301
  - type: cosine_precision@10
302
+ value: 0.08666666666666668
303
  name: Cosine Precision@10
304
  - type: cosine_recall@1
305
+ value: 0.6666666666666666
306
  name: Cosine Recall@1
307
  - type: cosine_recall@3
308
+ value: 0.8
309
  name: Cosine Recall@3
310
  - type: cosine_recall@5
311
+ value: 0.8
312
  name: Cosine Recall@5
313
  - type: cosine_recall@10
314
+ value: 0.8666666666666667
315
  name: Cosine Recall@10
316
  - type: cosine_ndcg@10
317
+ value: 0.7700616222307202
318
  name: Cosine Ndcg@10
319
  - type: cosine_mrr@10
320
+ value: 0.74
321
  name: Cosine Mrr@10
322
  - type: cosine_map@100
323
+ value: 0.7479365079365079
324
  name: Cosine Map@100
325
  - task:
326
  type: information-retrieval
 
330
  type: dim_64
331
  metrics:
332
  - type: cosine_accuracy@1
333
+ value: 0.6
334
  name: Cosine Accuracy@1
335
  - type: cosine_accuracy@3
336
+ value: 0.8
337
  name: Cosine Accuracy@3
338
  - type: cosine_accuracy@5
339
  value: 0.8
340
  name: Cosine Accuracy@5
341
  - type: cosine_accuracy@10
342
+ value: 0.8
343
  name: Cosine Accuracy@10
344
  - type: cosine_precision@1
345
+ value: 0.6
346
  name: Cosine Precision@1
347
  - type: cosine_precision@3
348
+ value: 0.2666666666666667
349
  name: Cosine Precision@3
350
  - type: cosine_precision@5
351
  value: 0.16000000000000003
352
  name: Cosine Precision@5
353
  - type: cosine_precision@10
354
+ value: 0.08000000000000002
355
  name: Cosine Precision@10
356
  - type: cosine_recall@1
357
+ value: 0.6
358
  name: Cosine Recall@1
359
  - type: cosine_recall@3
360
+ value: 0.8
361
  name: Cosine Recall@3
362
  - type: cosine_recall@5
363
  value: 0.8
364
  name: Cosine Recall@5
365
  - type: cosine_recall@10
366
+ value: 0.8
367
  name: Cosine Recall@10
368
  - type: cosine_ndcg@10
369
+ value: 0.7174573004761944
370
  name: Cosine Ndcg@10
371
  - type: cosine_mrr@10
372
+ value: 0.6888888888888889
373
  name: Cosine Mrr@10
374
  - type: cosine_map@100
375
+ value: 0.7003968253968255
376
  name: Cosine Map@100
377
  ---
378
 
 
426
  model = SentenceTransformer("Nuf-hugginface/modernbert-embed-quickb")
427
  # Run inference
428
  sentences = [
429
+ 'What type of dialogues can LLMs simulate?',
430
+ '5. Education and Learning Platforms\nEducational tools like Khanmigo (from Khan Academy) and other tutoring platforms are leveraging LLMs to provide real-time help to students. LLMs can break down complex topics, provide feedback on writing, and simulate Socratic-style dialogues.',
431
+ '6. APIs and Developer Tools\nDevelopers can integrate LLMs into their own apps using APIs provided by companies like OpenAI, Anthropic, and Cohere. These APIs allow developers to send prompts and receive intelligent outputs in return.\n\nThis enables custom applications like:\n\nSmart assistants in mobile apps\n\nAI-powered research tools\n\nVoice interfaces',
432
  ]
433
  embeddings = model.encode(sentences)
434
  print(embeddings.shape)
 
473
  * Datasets: `dim_768`, `dim_512`, `dim_256`, `dim_128` and `dim_64`
474
  * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
475
 
476
+ | Metric | dim_768 | dim_512 | dim_256 | dim_128 | dim_64 |
477
+ |:--------------------|:-----------|:-----------|:-----------|:-----------|:-----------|
478
+ | cosine_accuracy@1 | 0.6 | 0.6667 | 0.6667 | 0.6667 | 0.6 |
479
+ | cosine_accuracy@3 | 0.8667 | 0.8 | 0.8 | 0.8 | 0.8 |
480
+ | cosine_accuracy@5 | 0.8667 | 0.8 | 0.8 | 0.8 | 0.8 |
481
+ | cosine_accuracy@10 | 1.0 | 0.9333 | 1.0 | 0.8667 | 0.8 |
482
+ | cosine_precision@1 | 0.6 | 0.6667 | 0.6667 | 0.6667 | 0.6 |
483
+ | cosine_precision@3 | 0.2889 | 0.2667 | 0.2667 | 0.2667 | 0.2667 |
484
+ | cosine_precision@5 | 0.1733 | 0.16 | 0.16 | 0.16 | 0.16 |
485
+ | cosine_precision@10 | 0.1 | 0.0933 | 0.1 | 0.0867 | 0.08 |
486
+ | cosine_recall@1 | 0.6 | 0.6667 | 0.6667 | 0.6667 | 0.6 |
487
+ | cosine_recall@3 | 0.8667 | 0.8 | 0.8 | 0.8 | 0.8 |
488
+ | cosine_recall@5 | 0.8667 | 0.8 | 0.8 | 0.8 | 0.8 |
489
+ | cosine_recall@10 | 1.0 | 0.9333 | 1.0 | 0.8667 | 0.8 |
490
+ | **cosine_ndcg@10** | **0.8025** | **0.7956** | **0.7986** | **0.7701** | **0.7175** |
491
+ | cosine_mrr@10 | 0.74 | 0.7528 | 0.7384 | 0.74 | 0.6889 |
492
+ | cosine_map@100 | 0.74 | 0.7583 | 0.7384 | 0.7479 | 0.7004 |
493
 
494
  <!--
495
  ## Bias, Risks and Limitations
 
509
 
510
  #### Unnamed Dataset
511
 
512
+ * Size: 127 training samples
513
  * Columns: <code>anchor</code> and <code>positive</code>
514
+ * Approximate statistics based on the first 127 samples:
515
  | | anchor | positive |
516
  |:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
517
  | type | string | string |
518
+ | details | <ul><li>min: 8 tokens</li><li>mean: 13.2 tokens</li><li>max: 20 tokens</li></ul> | <ul><li>min: 13 tokens</li><li>mean: 53.85 tokens</li><li>max: 86 tokens</li></ul> |
519
  * Samples:
520
+ | anchor | positive |
521
+ |:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
522
+ | <code>What documents could the system retrieve in relation to sick leave?</code> | <code>For instance, in a document management system, a user might type "policies about sick leave", and the system—integrated with an LLM—could retrieve documents discussing "medical leave", "employee absence", and "illness policies", even if those exact words weren’t used.</code> |
523
+ | <code>What is one of the most visible integrations of LLM technology?</code> | <code>1. Chatbots and Virtual Assistants<br>One of the most visible LLM integrations is in chatbots. Tools like ChatGPT, Claude, and Bard are themselves chatbot interfaces built on LLMs. Many businesses are now integrating these models into their websites and customer support systems.</code> |
524
+ | <code>What does AI stand for?</code> | <code>Introduction to AI, Machine Learning, LLMs, and Their Integration</code> |
525
  * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters:
526
  ```json
527
  {
 
680
  </details>
681
 
682
  ### Training Logs
683
+ | Epoch | Step | Training Loss | dim_768_cosine_ndcg@10 | dim_512_cosine_ndcg@10 | dim_256_cosine_ndcg@10 | dim_128_cosine_ndcg@10 | dim_64_cosine_ndcg@10 |
684
+ |:-------:|:-----:|:-------------:|:----------------------:|:----------------------:|:----------------------:|:----------------------:|:---------------------:|
685
+ | 1.0 | 4 | - | 0.7853 | 0.8214 | 0.7673 | 0.7586 | 0.6883 |
686
+ | **2.0** | **8** | **-** | **0.7764** | **0.7902** | **0.7686** | **0.7701** | **0.7321** |
687
+ | 2.5 | 10 | 13.8004 | - | - | - | - | - |
688
+ | 3.0 | 12 | - | 0.8028 | 0.7710 | 0.7932 | 0.7701 | 0.7175 |
689
+ | 4.0 | 16 | - | 0.8025 | 0.7956 | 0.7986 | 0.7701 | 0.7175 |
690
 
691
  * The bold row denotes the saved checkpoint.
692
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:940b2d4ff23ed21583158f999073c662a4aef925a21b9ce34e0d4737565b9db3
3
  size 596070136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7fa83288c96da5c91a8d0a7f680fa88e3010ec41496910019839fb29c898aa3
3
  size 596070136