Nuf-hugginface commited on
Commit
e279263
·
verified ·
1 Parent(s): fae28ea

Add new SentenceTransformer model

Browse files
Files changed (2) hide show
  1. README.md +115 -113
  2. model.safetensors +1 -1
README.md CHANGED
@@ -7,63 +7,64 @@ tags:
7
  - sentence-similarity
8
  - feature-extraction
9
  - generated_from_trainer
10
- - dataset_size:44
11
  - loss:MatryoshkaLoss
12
  - loss:MultipleNegativesRankingLoss
13
  base_model: nomic-ai/modernbert-embed-base
14
  widget:
15
- - source_sentence: Which technologies are mentioned in the title of the text?
16
  sentences:
17
- - Introduction to AI, Machine Learning, LLMs, and Their Integration
18
- - Introduction to AI, Machine Learning, LLMs, and Their Integration
19
  - Furthermore, advanced integrations might include fine-tuning the LLM on domain-specific
20
  data, or pairing it with retrieval-augmented generation (RAG) pipelines. In RAG
21
  systems, the model first retrieves relevant documents from a database (like a
22
  knowledge base), then generates a response using that context—significantly improving
23
  the relevance and accuracy of the answers.
24
- - source_sentence: What is the ability of machines to understand and generate human
25
- language called?
26
- sentences:
 
27
  - . For example, integrating an LLM into a customer support chatbot might involve
28
  connecting it to a company’s internal knowledge base, enabling it to answer customer
29
  questions using accurate, up-to-date information.
30
- - Over the past few years, the field of ML has advanced rapidly, especially in the
31
- area of Natural Language Processing (NLP)—the ability of machines to understand
32
- and generate human language. At the forefront of this progress are Large Language
33
- Models (LLMs), such as OpenAI’s GPT (Generative Pre-trained Transformer), Google’s
34
- PaLM, and Meta’s LLaMA
35
- - A major subset of AI is Machine Learning (ML), which involves algorithms that
36
- learn from data rather than being explicitly programmed. Instead of writing detailed
37
- instructions for every task, ML models find patterns in large datasets and use
38
- these patterns to make predictions or decisions
39
- - source_sentence: What is the purpose of embedding LLMs into systems?
40
  sentences:
 
 
 
 
 
41
  - However, deploying LLMs effectively in real-world applications often requires
42
  LLM integration. This means embedding these models into systems, workflows, or
43
  products where they can interact with other components like databases, APIs, user
44
  interfaces, or even custom business logic
45
- - Over the past few years, the field of ML has advanced rapidly, especially in the
46
- area of Natural Language Processing (NLP)—the ability of machines to understand
47
- and generate human language. At the forefront of this progress are Large Language
48
- Models (LLMs), such as OpenAI’s GPT (Generative Pre-trained Transformer), Google’s
49
- PaLM, and Meta’s LLaMA
 
 
 
 
 
 
 
 
50
  - Introduction to AI, Machine Learning, LLMs, and Their Integration
51
- - source_sentence: What do ML algorithms learn from?
 
52
  sentences:
53
- - A major subset of AI is Machine Learning (ML), which involves algorithms that
54
- learn from data rather than being explicitly programmed. Instead of writing detailed
55
- instructions for every task, ML models find patterns in large datasets and use
56
- these patterns to make predictions or decisions
57
  - LLMs work by learning statistical relationships between words and phrases, allowing
58
  them to predict and generate language that feels natural. The power of these models
59
  lies not only in their size but also in the diversity of tasks they can perform
60
  with little to no task-specific training
61
- - Over the past few years, the field of ML has advanced rapidly, especially in the
62
- area of Natural Language Processing (NLP)—the ability of machines to understand
63
- and generate human language. At the forefront of this progress are Large Language
64
- Models (LLMs), such as OpenAI’s GPT (Generative Pre-trained Transformer), Google’s
65
- PaLM, and Meta’s LLaMA
66
- - source_sentence: What is required for effectively deploying LLMs in real-world applications?
 
67
  sentences:
68
  - . For instance, a spam filter doesn’t just block emails with specific keywords—it
69
  learns from thousands of examples what spam typically looks like.
@@ -72,10 +73,11 @@ widget:
72
  systems, the model first retrieves relevant documents from a database (like a
73
  knowledge base), then generates a response using that context—significantly improving
74
  the relevance and accuracy of the answers.
75
- - However, deploying LLMs effectively in real-world applications often requires
76
- LLM integration. This means embedding these models into systems, workflows, or
77
- products where they can interact with other components like databases, APIs, user
78
- interfaces, or even custom business logic
 
79
  pipeline_tag: sentence-similarity
80
  library_name: sentence-transformers
81
  metrics:
@@ -105,10 +107,10 @@ model-index:
105
  type: dim_768
106
  metrics:
107
  - type: cosine_accuracy@1
108
- value: 1.0
109
  name: Cosine Accuracy@1
110
  - type: cosine_accuracy@3
111
- value: 1.0
112
  name: Cosine Accuracy@3
113
  - type: cosine_accuracy@5
114
  value: 1.0
@@ -117,22 +119,22 @@ model-index:
117
  value: 1.0
118
  name: Cosine Accuracy@10
119
  - type: cosine_precision@1
120
- value: 1.0
121
  name: Cosine Precision@1
122
  - type: cosine_precision@3
123
- value: 0.3333333333333333
124
  name: Cosine Precision@3
125
  - type: cosine_precision@5
126
- value: 0.2
127
  name: Cosine Precision@5
128
  - type: cosine_precision@10
129
- value: 0.1
130
  name: Cosine Precision@10
131
  - type: cosine_recall@1
132
- value: 1.0
133
  name: Cosine Recall@1
134
  - type: cosine_recall@3
135
- value: 1.0
136
  name: Cosine Recall@3
137
  - type: cosine_recall@5
138
  value: 1.0
@@ -141,13 +143,13 @@ model-index:
141
  value: 1.0
142
  name: Cosine Recall@10
143
  - type: cosine_ndcg@10
144
- value: 1.0
145
  name: Cosine Ndcg@10
146
  - type: cosine_mrr@10
147
- value: 1.0
148
  name: Cosine Mrr@10
149
  - type: cosine_map@100
150
- value: 1.0
151
  name: Cosine Map@100
152
  - task:
153
  type: information-retrieval
@@ -157,7 +159,7 @@ model-index:
157
  type: dim_512
158
  metrics:
159
  - type: cosine_accuracy@1
160
- value: 1.0
161
  name: Cosine Accuracy@1
162
  - type: cosine_accuracy@3
163
  value: 1.0
@@ -169,19 +171,19 @@ model-index:
169
  value: 1.0
170
  name: Cosine Accuracy@10
171
  - type: cosine_precision@1
172
- value: 1.0
173
  name: Cosine Precision@1
174
  - type: cosine_precision@3
175
  value: 0.3333333333333333
176
  name: Cosine Precision@3
177
  - type: cosine_precision@5
178
- value: 0.2
179
  name: Cosine Precision@5
180
  - type: cosine_precision@10
181
- value: 0.1
182
  name: Cosine Precision@10
183
  - type: cosine_recall@1
184
- value: 1.0
185
  name: Cosine Recall@1
186
  - type: cosine_recall@3
187
  value: 1.0
@@ -193,13 +195,13 @@ model-index:
193
  value: 1.0
194
  name: Cosine Recall@10
195
  - type: cosine_ndcg@10
196
- value: 1.0
197
  name: Cosine Ndcg@10
198
  - type: cosine_mrr@10
199
- value: 1.0
200
  name: Cosine Mrr@10
201
  - type: cosine_map@100
202
- value: 1.0
203
  name: Cosine Map@100
204
  - task:
205
  type: information-retrieval
@@ -209,7 +211,7 @@ model-index:
209
  type: dim_256
210
  metrics:
211
  - type: cosine_accuracy@1
212
- value: 1.0
213
  name: Cosine Accuracy@1
214
  - type: cosine_accuracy@3
215
  value: 1.0
@@ -221,19 +223,19 @@ model-index:
221
  value: 1.0
222
  name: Cosine Accuracy@10
223
  - type: cosine_precision@1
224
- value: 1.0
225
  name: Cosine Precision@1
226
  - type: cosine_precision@3
227
  value: 0.3333333333333333
228
  name: Cosine Precision@3
229
  - type: cosine_precision@5
230
- value: 0.2
231
  name: Cosine Precision@5
232
  - type: cosine_precision@10
233
- value: 0.1
234
  name: Cosine Precision@10
235
  - type: cosine_recall@1
236
- value: 1.0
237
  name: Cosine Recall@1
238
  - type: cosine_recall@3
239
  value: 1.0
@@ -245,13 +247,13 @@ model-index:
245
  value: 1.0
246
  name: Cosine Recall@10
247
  - type: cosine_ndcg@10
248
- value: 1.0
249
  name: Cosine Ndcg@10
250
  - type: cosine_mrr@10
251
- value: 1.0
252
  name: Cosine Mrr@10
253
  - type: cosine_map@100
254
- value: 1.0
255
  name: Cosine Map@100
256
  - task:
257
  type: information-retrieval
@@ -261,10 +263,10 @@ model-index:
261
  type: dim_128
262
  metrics:
263
  - type: cosine_accuracy@1
264
- value: 0.8
265
  name: Cosine Accuracy@1
266
  - type: cosine_accuracy@3
267
- value: 1.0
268
  name: Cosine Accuracy@3
269
  - type: cosine_accuracy@5
270
  value: 1.0
@@ -273,22 +275,22 @@ model-index:
273
  value: 1.0
274
  name: Cosine Accuracy@10
275
  - type: cosine_precision@1
276
- value: 0.8
277
  name: Cosine Precision@1
278
  - type: cosine_precision@3
279
- value: 0.3333333333333333
280
  name: Cosine Precision@3
281
  - type: cosine_precision@5
282
- value: 0.2
283
  name: Cosine Precision@5
284
  - type: cosine_precision@10
285
- value: 0.1
286
  name: Cosine Precision@10
287
  - type: cosine_recall@1
288
- value: 0.8
289
  name: Cosine Recall@1
290
  - type: cosine_recall@3
291
- value: 1.0
292
  name: Cosine Recall@3
293
  - type: cosine_recall@5
294
  value: 1.0
@@ -297,13 +299,13 @@ model-index:
297
  value: 1.0
298
  name: Cosine Recall@10
299
  - type: cosine_ndcg@10
300
- value: 0.9261859507142916
301
  name: Cosine Ndcg@10
302
  - type: cosine_mrr@10
303
- value: 0.9
304
  name: Cosine Mrr@10
305
  - type: cosine_map@100
306
- value: 0.9
307
  name: Cosine Map@100
308
  - task:
309
  type: information-retrieval
@@ -313,7 +315,7 @@ model-index:
313
  type: dim_64
314
  metrics:
315
  - type: cosine_accuracy@1
316
- value: 0.8
317
  name: Cosine Accuracy@1
318
  - type: cosine_accuracy@3
319
  value: 1.0
@@ -325,19 +327,19 @@ model-index:
325
  value: 1.0
326
  name: Cosine Accuracy@10
327
  - type: cosine_precision@1
328
- value: 0.8
329
  name: Cosine Precision@1
330
  - type: cosine_precision@3
331
  value: 0.3333333333333333
332
  name: Cosine Precision@3
333
  - type: cosine_precision@5
334
- value: 0.2
335
  name: Cosine Precision@5
336
  - type: cosine_precision@10
337
- value: 0.1
338
  name: Cosine Precision@10
339
  - type: cosine_recall@1
340
- value: 0.8
341
  name: Cosine Recall@1
342
  - type: cosine_recall@3
343
  value: 1.0
@@ -349,13 +351,13 @@ model-index:
349
  value: 1.0
350
  name: Cosine Recall@10
351
  - type: cosine_ndcg@10
352
- value: 0.9261859507142916
353
  name: Cosine Ndcg@10
354
  - type: cosine_mrr@10
355
- value: 0.9
356
  name: Cosine Mrr@10
357
  - type: cosine_map@100
358
- value: 0.9
359
  name: Cosine Map@100
360
  ---
361
 
@@ -409,9 +411,9 @@ from sentence_transformers import SentenceTransformer
409
  model = SentenceTransformer("Nuf-hugginface/modernbert-embed-quickb")
410
  # Run inference
411
  sentences = [
412
- 'What is required for effectively deploying LLMs in real-world applications?',
413
- 'However, deploying LLMs effectively in real-world applications often requires LLM integration. This means embedding these models into systems, workflows, or products where they can interact with other components like databases, APIs, user interfaces, or even custom business logic',
414
- '. For instance, a spam filter doesn’t just block emails with specific keywords—it learns from thousands of examples what spam typically looks like.',
415
  ]
416
  embeddings = model.encode(sentences)
417
  print(embeddings.shape)
@@ -456,23 +458,23 @@ You can finetune this model on your own dataset.
456
  * Datasets: `dim_768`, `dim_512`, `dim_256`, `dim_128` and `dim_64`
457
  * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
458
 
459
- | Metric | dim_768 | dim_512 | dim_256 | dim_128 | dim_64 |
460
- |:--------------------|:--------|:--------|:--------|:-----------|:-----------|
461
- | cosine_accuracy@1 | 1.0 | 1.0 | 1.0 | 0.8 | 0.8 |
462
- | cosine_accuracy@3 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
463
- | cosine_accuracy@5 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
464
- | cosine_accuracy@10 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
465
- | cosine_precision@1 | 1.0 | 1.0 | 1.0 | 0.8 | 0.8 |
466
- | cosine_precision@3 | 0.3333 | 0.3333 | 0.3333 | 0.3333 | 0.3333 |
467
- | cosine_precision@5 | 0.2 | 0.2 | 0.2 | 0.2 | 0.2 |
468
- | cosine_precision@10 | 0.1 | 0.1 | 0.1 | 0.1 | 0.1 |
469
- | cosine_recall@1 | 1.0 | 1.0 | 1.0 | 0.8 | 0.8 |
470
- | cosine_recall@3 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
471
- | cosine_recall@5 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
472
- | cosine_recall@10 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
473
- | **cosine_ndcg@10** | **1.0** | **1.0** | **1.0** | **0.9262** | **0.9262** |
474
- | cosine_mrr@10 | 1.0 | 1.0 | 1.0 | 0.9 | 0.9 |
475
- | cosine_map@100 | 1.0 | 1.0 | 1.0 | 0.9 | 0.9 |
476
 
477
  <!--
478
  ## Bias, Risks and Limitations
@@ -492,19 +494,19 @@ You can finetune this model on your own dataset.
492
 
493
  #### Unnamed Dataset
494
 
495
- * Size: 44 training samples
496
  * Columns: <code>anchor</code> and <code>positive</code>
497
- * Approximate statistics based on the first 44 samples:
498
  | | anchor | positive |
499
  |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
500
  | type | string | string |
501
- | details | <ul><li>min: 8 tokens</li><li>mean: 12.68 tokens</li><li>max: 16 tokens</li></ul> | <ul><li>min: 15 tokens</li><li>mean: 47.25 tokens</li><li>max: 83 tokens</li></ul> |
502
  * Samples:
503
- | anchor | positive |
504
- |:-----------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
505
- | <code>What role does generalization ability play in customer service?</code> | <code>. This generalization ability makes them incredibly useful across industries—from customer service and education to software development and healthcare.</code> |
506
- | <code>Do LLMs require task-specific training to perform tasks?</code> | <code>LLMs work by learning statistical relationships between words and phrases, allowing them to predict and generate language that feels natural. The power of these models lies not only in their size but also in the diversity of tasks they can perform with little to no task-specific training</code> |
507
- | <code>What does a spam filter learn from?</code> | <code>. For instance, a spam filter doesn’t just block emails with specific keywords—it learns from thousands of examples what spam typically looks like.</code> |
508
  * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters:
509
  ```json
510
  {
@@ -665,8 +667,8 @@ You can finetune this model on your own dataset.
665
  ### Training Logs
666
  | Epoch | Step | dim_768_cosine_ndcg@10 | dim_512_cosine_ndcg@10 | dim_256_cosine_ndcg@10 | dim_128_cosine_ndcg@10 | dim_64_cosine_ndcg@10 |
667
  |:-------:|:-----:|:----------------------:|:----------------------:|:----------------------:|:----------------------:|:---------------------:|
668
- | **1.0** | **2** | **1.0** | **1.0** | **1.0** | **0.9262** | **0.9** |
669
- | 2.0 | 4 | 1.0 | 1.0 | 1.0 | 0.9262 | 0.9262 |
670
 
671
  * The bold row denotes the saved checkpoint.
672
 
 
7
  - sentence-similarity
8
  - feature-extraction
9
  - generated_from_trainer
10
+ - dataset_size:46
11
  - loss:MatryoshkaLoss
12
  - loss:MultipleNegativesRankingLoss
13
  base_model: nomic-ai/modernbert-embed-base
14
  widget:
15
+ - source_sentence: What two factors contribute to the power of LLMs?
16
  sentences:
 
 
17
  - Furthermore, advanced integrations might include fine-tuning the LLM on domain-specific
18
  data, or pairing it with retrieval-augmented generation (RAG) pipelines. In RAG
19
  systems, the model first retrieves relevant documents from a database (like a
20
  knowledge base), then generates a response using that context—significantly improving
21
  the relevance and accuracy of the answers.
22
+ - LLMs work by learning statistical relationships between words and phrases, allowing
23
+ them to predict and generate language that feels natural. The power of these models
24
+ lies not only in their size but also in the diversity of tasks they can perform
25
+ with little to no task-specific training
26
  - . For example, integrating an LLM into a customer support chatbot might involve
27
  connecting it to a company’s internal knowledge base, enabling it to answer customer
28
  questions using accurate, up-to-date information.
29
+ - source_sentence: What is one method mentioned for fine-tuning the LLM?
 
 
 
 
 
 
 
 
 
30
  sentences:
31
+ - Furthermore, advanced integrations might include fine-tuning the LLM on domain-specific
32
+ data, or pairing it with retrieval-augmented generation (RAG) pipelines. In RAG
33
+ systems, the model first retrieves relevant documents from a database (like a
34
+ knowledge base), then generates a response using that context—significantly improving
35
+ the relevance and accuracy of the answers.
36
  - However, deploying LLMs effectively in real-world applications often requires
37
  LLM integration. This means embedding these models into systems, workflows, or
38
  products where they can interact with other components like databases, APIs, user
39
  interfaces, or even custom business logic
40
+ - . As organizations increasingly adopt these technologies, the ability to understand
41
+ and apply LLMs will be a critical skill in the AI-powered future.
42
+ - source_sentence: What are some tasks that AI is capable of performing?
43
+ sentences:
44
+ - Artificial Intelligence (AI) is the broad field of computer science that focuses
45
+ on building systems capable of performing tasks that normally require human intelligence.
46
+ These tasks include learning from experience, understanding language, recognizing
47
+ patterns, and making decisions. AI powers everything from smart assistants like
48
+ Siri to recommendation systems on Netflix and self-driving cars.
49
+ - In summary, AI and ML form the foundation for intelligent automation, while LLMs
50
+ represent a breakthrough in language understanding and generation. Integrating
51
+ these models into real-world systems unlocks practical value, turning raw intelligence
52
+ into tangible solutions
53
  - Introduction to AI, Machine Learning, LLMs, and Their Integration
54
+ - source_sentence: What is the abbreviation for Large Language Models as mentioned
55
+ in the text?
56
  sentences:
 
 
 
 
57
  - LLMs work by learning statistical relationships between words and phrases, allowing
58
  them to predict and generate language that feels natural. The power of these models
59
  lies not only in their size but also in the diversity of tasks they can perform
60
  with little to no task-specific training
61
+ - LLMs work by learning statistical relationships between words and phrases, allowing
62
+ them to predict and generate language that feels natural. The power of these models
63
+ lies not only in their size but also in the diversity of tasks they can perform
64
+ with little to no task-specific training
65
+ - . As organizations increasingly adopt these technologies, the ability to understand
66
+ and apply LLMs will be a critical skill in the AI-powered future.
67
+ - source_sentence: What does the use of RAG systems improve according to the text?
68
  sentences:
69
  - . For instance, a spam filter doesn’t just block emails with specific keywords—it
70
  learns from thousands of examples what spam typically looks like.
 
73
  systems, the model first retrieves relevant documents from a database (like a
74
  knowledge base), then generates a response using that context—significantly improving
75
  the relevance and accuracy of the answers.
76
+ - Furthermore, advanced integrations might include fine-tuning the LLM on domain-specific
77
+ data, or pairing it with retrieval-augmented generation (RAG) pipelines. In RAG
78
+ systems, the model first retrieves relevant documents from a database (like a
79
+ knowledge base), then generates a response using that context—significantly improving
80
+ the relevance and accuracy of the answers.
81
  pipeline_tag: sentence-similarity
82
  library_name: sentence-transformers
83
  metrics:
 
107
  type: dim_768
108
  metrics:
109
  - type: cosine_accuracy@1
110
+ value: 0.6666666666666666
111
  name: Cosine Accuracy@1
112
  - type: cosine_accuracy@3
113
+ value: 0.8333333333333334
114
  name: Cosine Accuracy@3
115
  - type: cosine_accuracy@5
116
  value: 1.0
 
119
  value: 1.0
120
  name: Cosine Accuracy@10
121
  - type: cosine_precision@1
122
+ value: 0.6666666666666666
123
  name: Cosine Precision@1
124
  - type: cosine_precision@3
125
+ value: 0.27777777777777773
126
  name: Cosine Precision@3
127
  - type: cosine_precision@5
128
+ value: 0.19999999999999998
129
  name: Cosine Precision@5
130
  - type: cosine_precision@10
131
+ value: 0.09999999999999999
132
  name: Cosine Precision@10
133
  - type: cosine_recall@1
134
+ value: 0.6666666666666666
135
  name: Cosine Recall@1
136
  - type: cosine_recall@3
137
+ value: 0.8333333333333334
138
  name: Cosine Recall@3
139
  - type: cosine_recall@5
140
  value: 1.0
 
143
  value: 1.0
144
  name: Cosine Recall@10
145
  - type: cosine_ndcg@10
146
+ value: 0.8436010519408085
147
  name: Cosine Ndcg@10
148
  - type: cosine_mrr@10
149
+ value: 0.7916666666666666
150
  name: Cosine Mrr@10
151
  - type: cosine_map@100
152
+ value: 0.7916666666666666
153
  name: Cosine Map@100
154
  - task:
155
  type: information-retrieval
 
159
  type: dim_512
160
  metrics:
161
  - type: cosine_accuracy@1
162
+ value: 0.6666666666666666
163
  name: Cosine Accuracy@1
164
  - type: cosine_accuracy@3
165
  value: 1.0
 
171
  value: 1.0
172
  name: Cosine Accuracy@10
173
  - type: cosine_precision@1
174
+ value: 0.6666666666666666
175
  name: Cosine Precision@1
176
  - type: cosine_precision@3
177
  value: 0.3333333333333333
178
  name: Cosine Precision@3
179
  - type: cosine_precision@5
180
+ value: 0.19999999999999998
181
  name: Cosine Precision@5
182
  - type: cosine_precision@10
183
+ value: 0.09999999999999999
184
  name: Cosine Precision@10
185
  - type: cosine_recall@1
186
+ value: 0.6666666666666666
187
  name: Cosine Recall@1
188
  - type: cosine_recall@3
189
  value: 1.0
 
195
  value: 1.0
196
  name: Cosine Recall@10
197
  - type: cosine_ndcg@10
198
+ value: 0.8769765845238192
199
  name: Cosine Ndcg@10
200
  - type: cosine_mrr@10
201
+ value: 0.8333333333333334
202
  name: Cosine Mrr@10
203
  - type: cosine_map@100
204
+ value: 0.8333333333333334
205
  name: Cosine Map@100
206
  - task:
207
  type: information-retrieval
 
211
  type: dim_256
212
  metrics:
213
  - type: cosine_accuracy@1
214
+ value: 0.6666666666666666
215
  name: Cosine Accuracy@1
216
  - type: cosine_accuracy@3
217
  value: 1.0
 
223
  value: 1.0
224
  name: Cosine Accuracy@10
225
  - type: cosine_precision@1
226
+ value: 0.6666666666666666
227
  name: Cosine Precision@1
228
  - type: cosine_precision@3
229
  value: 0.3333333333333333
230
  name: Cosine Precision@3
231
  - type: cosine_precision@5
232
+ value: 0.19999999999999998
233
  name: Cosine Precision@5
234
  - type: cosine_precision@10
235
+ value: 0.09999999999999999
236
  name: Cosine Precision@10
237
  - type: cosine_recall@1
238
+ value: 0.6666666666666666
239
  name: Cosine Recall@1
240
  - type: cosine_recall@3
241
  value: 1.0
 
247
  value: 1.0
248
  name: Cosine Recall@10
249
  - type: cosine_ndcg@10
250
+ value: 0.8551549589285763
251
  name: Cosine Ndcg@10
252
  - type: cosine_mrr@10
253
+ value: 0.8055555555555557
254
  name: Cosine Mrr@10
255
  - type: cosine_map@100
256
+ value: 0.8055555555555557
257
  name: Cosine Map@100
258
  - task:
259
  type: information-retrieval
 
263
  type: dim_128
264
  metrics:
265
  - type: cosine_accuracy@1
266
+ value: 0.6666666666666666
267
  name: Cosine Accuracy@1
268
  - type: cosine_accuracy@3
269
+ value: 0.8333333333333334
270
  name: Cosine Accuracy@3
271
  - type: cosine_accuracy@5
272
  value: 1.0
 
275
  value: 1.0
276
  name: Cosine Accuracy@10
277
  - type: cosine_precision@1
278
+ value: 0.6666666666666666
279
  name: Cosine Precision@1
280
  - type: cosine_precision@3
281
+ value: 0.27777777777777773
282
  name: Cosine Precision@3
283
  - type: cosine_precision@5
284
+ value: 0.19999999999999998
285
  name: Cosine Precision@5
286
  - type: cosine_precision@10
287
+ value: 0.09999999999999999
288
  name: Cosine Precision@10
289
  - type: cosine_recall@1
290
+ value: 0.6666666666666666
291
  name: Cosine Recall@1
292
  - type: cosine_recall@3
293
+ value: 0.8333333333333334
294
  name: Cosine Recall@3
295
  - type: cosine_recall@5
296
  value: 1.0
 
299
  value: 1.0
300
  name: Cosine Recall@10
301
  - type: cosine_ndcg@10
302
+ value: 0.8217794263455654
303
  name: Cosine Ndcg@10
304
  - type: cosine_mrr@10
305
+ value: 0.763888888888889
306
  name: Cosine Mrr@10
307
  - type: cosine_map@100
308
+ value: 0.763888888888889
309
  name: Cosine Map@100
310
  - task:
311
  type: information-retrieval
 
315
  type: dim_64
316
  metrics:
317
  - type: cosine_accuracy@1
318
+ value: 0.8333333333333334
319
  name: Cosine Accuracy@1
320
  - type: cosine_accuracy@3
321
  value: 1.0
 
327
  value: 1.0
328
  name: Cosine Accuracy@10
329
  - type: cosine_precision@1
330
+ value: 0.8333333333333334
331
  name: Cosine Precision@1
332
  - type: cosine_precision@3
333
  value: 0.3333333333333333
334
  name: Cosine Precision@3
335
  - type: cosine_precision@5
336
+ value: 0.19999999999999998
337
  name: Cosine Precision@5
338
  - type: cosine_precision@10
339
+ value: 0.09999999999999999
340
  name: Cosine Precision@10
341
  - type: cosine_recall@1
342
+ value: 0.8333333333333334
343
  name: Cosine Recall@1
344
  - type: cosine_recall@3
345
  value: 1.0
 
351
  value: 1.0
352
  name: Cosine Recall@10
353
  - type: cosine_ndcg@10
354
+ value: 0.9166666666666666
355
  name: Cosine Ndcg@10
356
  - type: cosine_mrr@10
357
+ value: 0.888888888888889
358
  name: Cosine Mrr@10
359
  - type: cosine_map@100
360
+ value: 0.888888888888889
361
  name: Cosine Map@100
362
  ---
363
 
 
411
  model = SentenceTransformer("Nuf-hugginface/modernbert-embed-quickb")
412
  # Run inference
413
  sentences = [
414
+ 'What does the use of RAG systems improve according to the text?',
415
+ 'Furthermore, advanced integrations might include fine-tuning the LLM on domain-specific data, or pairing it with retrieval-augmented generation (RAG) pipelines. In RAG systems, the model first retrieves relevant documents from a database (like a knowledge base), then generates a response using that context—significantly improving the relevance and accuracy of the answers.',
416
+ 'Furthermore, advanced integrations might include fine-tuning the LLM on domain-specific data, or pairing it with retrieval-augmented generation (RAG) pipelines. In RAG systems, the model first retrieves relevant documents from a database (like a knowledge base), then generates a response using that context—significantly improving the relevance and accuracy of the answers.',
417
  ]
418
  embeddings = model.encode(sentences)
419
  print(embeddings.shape)
 
458
  * Datasets: `dim_768`, `dim_512`, `dim_256`, `dim_128` and `dim_64`
459
  * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
460
 
461
+ | Metric | dim_768 | dim_512 | dim_256 | dim_128 | dim_64 |
462
+ |:--------------------|:-----------|:----------|:-----------|:-----------|:-----------|
463
+ | cosine_accuracy@1 | 0.6667 | 0.6667 | 0.6667 | 0.6667 | 0.8333 |
464
+ | cosine_accuracy@3 | 0.8333 | 1.0 | 1.0 | 0.8333 | 1.0 |
465
+ | cosine_accuracy@5 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
466
+ | cosine_accuracy@10 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
467
+ | cosine_precision@1 | 0.6667 | 0.6667 | 0.6667 | 0.6667 | 0.8333 |
468
+ | cosine_precision@3 | 0.2778 | 0.3333 | 0.3333 | 0.2778 | 0.3333 |
469
+ | cosine_precision@5 | 0.2 | 0.2 | 0.2 | 0.2 | 0.2 |
470
+ | cosine_precision@10 | 0.1 | 0.1 | 0.1 | 0.1 | 0.1 |
471
+ | cosine_recall@1 | 0.6667 | 0.6667 | 0.6667 | 0.6667 | 0.8333 |
472
+ | cosine_recall@3 | 0.8333 | 1.0 | 1.0 | 0.8333 | 1.0 |
473
+ | cosine_recall@5 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
474
+ | cosine_recall@10 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
475
+ | **cosine_ndcg@10** | **0.8436** | **0.877** | **0.8552** | **0.8218** | **0.9167** |
476
+ | cosine_mrr@10 | 0.7917 | 0.8333 | 0.8056 | 0.7639 | 0.8889 |
477
+ | cosine_map@100 | 0.7917 | 0.8333 | 0.8056 | 0.7639 | 0.8889 |
478
 
479
  <!--
480
  ## Bias, Risks and Limitations
 
494
 
495
  #### Unnamed Dataset
496
 
497
+ * Size: 46 training samples
498
  * Columns: <code>anchor</code> and <code>positive</code>
499
+ * Approximate statistics based on the first 46 samples:
500
  | | anchor | positive |
501
  |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
502
  | type | string | string |
503
+ | details | <ul><li>min: 9 tokens</li><li>mean: 13.28 tokens</li><li>max: 18 tokens</li></ul> | <ul><li>min: 15 tokens</li><li>mean: 47.54 tokens</li><li>max: 83 tokens</li></ul> |
504
  * Samples:
505
+ | anchor | positive |
506
+ |:------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
507
+ | <code>What does RAG stand for in the context of the text?</code> | <code>Furthermore, advanced integrations might include fine-tuning the LLM on domain-specific data, or pairing it with retrieval-augmented generation (RAG) pipelines. In RAG systems, the model first retrieves relevant documents from a database (like a knowledge base), then generates a response using that context—significantly improving the relevance and accuracy of the answers.</code> |
508
+ | <code>What type of information can the LLM use to answer customer questions?</code> | <code>. For example, integrating an LLM into a customer support chatbot might involve connecting it to a company’s internal knowledge base, enabling it to answer customer questions using accurate, up-to-date information.</code> |
509
+ | <code>What do AI and ML form the foundation for?</code> | <code>In summary, AI and ML form the foundation for intelligent automation, while LLMs represent a breakthrough in language understanding and generation. Integrating these models into real-world systems unlocks practical value, turning raw intelligence into tangible solutions</code> |
510
  * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters:
511
  ```json
512
  {
 
667
  ### Training Logs
668
  | Epoch | Step | dim_768_cosine_ndcg@10 | dim_512_cosine_ndcg@10 | dim_256_cosine_ndcg@10 | dim_128_cosine_ndcg@10 | dim_64_cosine_ndcg@10 |
669
  |:-------:|:-----:|:----------------------:|:----------------------:|:----------------------:|:----------------------:|:---------------------:|
670
+ | 1.0 | 2 | 0.8244 | 0.8770 | 0.8244 | 0.8029 | 0.8552 |
671
+ | **2.0** | **4** | **0.8436** | **0.877** | **0.8552** | **0.8218** | **0.9167** |
672
 
673
  * The bold row denotes the saved checkpoint.
674
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0e68ee945d1bbc67d5cc8f65e86d365fe1ff4f668bc04a3c9b4f7049249f7a0
3
  size 596070136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32dfd9d19e38eafcef8a97038c57c0f09c730e4f42354f9efc84c9b49b1aab11
3
  size 596070136