radoslavralev commited on
Commit
b2f3977
·
verified ·
1 Parent(s): 96c60a1

Add new SentenceTransformer model

Browse files
1_Pooling/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "word_embedding_dimension": 384,
3
- "pooling_mode_cls_token": false,
4
- "pooling_mode_mean_tokens": true,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
 
1
  {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
README.md CHANGED
@@ -7,86 +7,112 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:111470
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: thenlper/gte-small
11
  widget:
12
- - source_sentence: why are some rocks radioactive
13
  sentences:
14
- - Radioactive accessory minerals such as zircon may contribute to the radioactivity
15
- of a mineral which is otherwise non-radioactive by calculation. Many granites
16
- or other igneous rocks contain some radioactivity because of minor, but highly
17
- radioactive, accessory minerals.re = mineral density (S Atomic number / Molecular
18
- Weight) where re is the electron density in grams/cc.efinition. Radioactivity
19
- in minerals are caused by the inclusion of naturally-occurring radioactive elements
20
- in the mineral's composition. The degree of radioactivity is dependent on the
21
- concentration and isotope present in the mineral.
22
- - Taking B-complex vitamins, which include vitamin B12, can cause urine to have
23
- a bright yellow or even orange color, but check with your doctor to be sure that's
24
- what is going on in your case. B vitamins are water-soluble vitamins, which means
25
- that what your body doesn't use is excreted in your urine. Riboflavin (vitamin
26
- B2) is especially likely to cause this color change in urine. Several medications
27
- can also turn urine a bright yellow or orange color. Changes in urine color may
28
- also signal certain health problems.
29
- - Radioactive material is just another name for a group of unstable atoms that emit
30
- ionizing radiation. These groups of unstable atoms emit radiation because they
31
- try to become stable. Radioactive materials emit radiation in a process called
32
- radioactive decay.
33
- - source_sentence: How was your experience of Lucid dreaming at home?
34
  sentences:
35
- - How was your experience of Lucid dreaming at home?
36
- - How was your experience of Lucid dreaming outside the home?
37
- - "Bournemouth /Ë\x88bÉ\x94É\x99rnmÉ\x99θ/ is a large coastal resort town on the\
38
- \ south coast of England directly to the east of the Jurassic Coast, a 96-mile\
39
- \ (155 km) World Heritage Site. According to the 2011 census, the town has a population\
40
- \ of 183,491 making it the largest settlement in Dorset.he Bournemouth Eye is\
41
- \ a helium-filled balloon attached to a steel cable in the town's lower gardens.\
42
- \ The spherical balloon is 69 m (226 ft) in circumference and carries an enclosed,\
43
- \ steel gondola. Rising to a height of 150 m (492 ft), it provides a panoramic\
44
- \ view of the surrounding area for up to 28 passengers."
45
- - source_sentence: what is iraq's dominant religion
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
  sentences:
47
- - 'If you are working, consider taking maternity leave as early as you can. This
48
- makes sense anyway because carrying twins is hard work, and most twins arrive
49
- earlier than single babies (NCCWCH 2011: 128) . More than half of twins arrive
50
- early, before 37 weeks (NCCWCH 2011: 120, Tamba 2012) .Talk to your midwife or
51
- doctor if you are feeling down about your pregnancy (NICE 2011) .f you are working,
52
- consider taking maternity leave as early as you can. This makes sense anyway because
53
- carrying twins is hard work, and most twins arrive earlier than single babies
54
- (NCCWCH 2011: 128) . More than half of twins arrive early, before 37 weeks (NCCWCH
55
- 2011: 120, Tamba 2012) .'
56
- - "Introduction. Although Iranâ\x80\x99s state religion is Shiite Islam and the\
57
- \ majority of its population is ethnically Persian, millions of minorities from\
58
- \ various ethnic, religious, and linguistic backgrounds also reside in Iran. Among\
59
- \ these groups are ethnic Kurds, Baluchis, and Azeris.lthough Iranâ\x80\x99s state\
60
- \ religion is Shiite Islam and the majority of its population is ethnically Persian,\
61
- \ millions of minorities from various ethnic, religious, and linguistic backgrounds\
62
- \ also reside in Iran."
63
- - In today's Republic of Iraq, where Islam is the state religion and claims the
64
- beliefs of 95 percent of the population, the majority of Iraqis identify with
65
- Arab culture. The second-largest cultural group is the Kurds, who are in the highlands
66
- and mountain valleys of the north in a politically autonomous settlement.
67
- - source_sentence: how many years of education are needed to become a pediatric nurse
68
  sentences:
69
- - In terms of educational background, pediatric nurse requirements include either
70
- an Associate's or a Bachelor's degree in Nursing. An Associate's degree (ADN)
71
- typically takes two years to complete, while a Bachelor's degree (BSN) typically
72
- takes four years. ADN programs are usually offered by community colleges.
73
- - "Photo of Oxford Suites Sonoma County - Rohnert Park - Rohnert Park, CA, United\
74
- \ States Photo of Oxford Suites Sonoma County - Rohnert Park - Rohnert Park, CA,\
75
- \ United States Living area with king bed by Monique' M. â\x80\x9CAnd there's\
76
- \ a complimentary reception with 2 drinks, soup and salad bar nightly.â\x80\x9D\
77
- \ in 2 reviews"
78
- - 'From there, additional training specific to the care of children is required.
79
- Pediatric nurses can become certified in the field and may choose to further specialize
80
- in a particular area. Program Levels: Associate''s degree, bachelor''s degree.'
81
- - source_sentence: Schliemann recognized five shafts and cleared them like the graves
82
- mentioned by Pausanias .
 
 
 
 
 
 
 
 
 
 
83
  sentences:
84
- - IBM banned the usage of the POWER5+ in its System p5 510Q, 520Q, 550Q and 560Q
85
- servers.
86
- - Schliemann cleared five shafts and recognized them as the graves mentioned by
87
- Pausania .
88
- - Schliemann recognized five shafts and cleared them like the graves mentioned by
89
- Pausanias .
 
 
 
 
90
  pipeline_tag: sentence-similarity
91
  library_name: sentence-transformers
92
  metrics:
@@ -106,7 +132,7 @@ metrics:
106
  - cosine_mrr@10
107
  - cosine_map@100
108
  model-index:
109
- - name: SentenceTransformer based on thenlper/gte-small
110
  results:
111
  - task:
112
  type: information-retrieval
@@ -119,46 +145,46 @@ model-index:
119
  value: 0.24
120
  name: Cosine Accuracy@1
121
  - type: cosine_accuracy@3
122
- value: 0.46
123
  name: Cosine Accuracy@3
124
  - type: cosine_accuracy@5
125
- value: 0.54
126
  name: Cosine Accuracy@5
127
  - type: cosine_accuracy@10
128
- value: 0.66
129
  name: Cosine Accuracy@10
130
  - type: cosine_precision@1
131
  value: 0.24
132
  name: Cosine Precision@1
133
  - type: cosine_precision@3
134
- value: 0.15333333333333332
135
  name: Cosine Precision@3
136
  - type: cosine_precision@5
137
- value: 0.10800000000000001
138
  name: Cosine Precision@5
139
  - type: cosine_precision@10
140
- value: 0.066
141
  name: Cosine Precision@10
142
  - type: cosine_recall@1
143
  value: 0.24
144
  name: Cosine Recall@1
145
  - type: cosine_recall@3
146
- value: 0.46
147
  name: Cosine Recall@3
148
  - type: cosine_recall@5
149
- value: 0.54
150
  name: Cosine Recall@5
151
  - type: cosine_recall@10
152
- value: 0.66
153
  name: Cosine Recall@10
154
  - type: cosine_ndcg@10
155
- value: 0.44291538669701197
156
  name: Cosine Ndcg@10
157
  - type: cosine_mrr@10
158
- value: 0.3744682539682539
159
  name: Cosine Mrr@10
160
  - type: cosine_map@100
161
- value: 0.3864402026320918
162
  name: Cosine Map@100
163
  - task:
164
  type: information-retrieval
@@ -168,49 +194,49 @@ model-index:
168
  type: NanoNQ
169
  metrics:
170
  - type: cosine_accuracy@1
171
- value: 0.24
172
  name: Cosine Accuracy@1
173
  - type: cosine_accuracy@3
174
- value: 0.34
175
  name: Cosine Accuracy@3
176
  - type: cosine_accuracy@5
177
- value: 0.38
178
  name: Cosine Accuracy@5
179
  - type: cosine_accuracy@10
180
- value: 0.5
181
  name: Cosine Accuracy@10
182
  - type: cosine_precision@1
183
- value: 0.24
184
  name: Cosine Precision@1
185
  - type: cosine_precision@3
186
- value: 0.11333333333333333
187
  name: Cosine Precision@3
188
  - type: cosine_precision@5
189
- value: 0.08
190
  name: Cosine Precision@5
191
  - type: cosine_precision@10
192
- value: 0.052000000000000005
193
  name: Cosine Precision@10
194
  - type: cosine_recall@1
195
- value: 0.22
196
  name: Cosine Recall@1
197
  - type: cosine_recall@3
198
- value: 0.32
199
  name: Cosine Recall@3
200
  - type: cosine_recall@5
201
- value: 0.37
202
  name: Cosine Recall@5
203
  - type: cosine_recall@10
204
- value: 0.48
205
  name: Cosine Recall@10
206
  - type: cosine_ndcg@10
207
- value: 0.3425602412795631
208
  name: Cosine Ndcg@10
209
  - type: cosine_mrr@10
210
- value: 0.31126984126984125
211
  name: Cosine Mrr@10
212
  - type: cosine_map@100
213
- value: 0.30988313931877554
214
  name: Cosine Map@100
215
  - task:
216
  type: nano-beir
@@ -220,63 +246,63 @@ model-index:
220
  type: NanoBEIR_mean
221
  metrics:
222
  - type: cosine_accuracy@1
223
- value: 0.24
224
  name: Cosine Accuracy@1
225
  - type: cosine_accuracy@3
226
- value: 0.4
227
  name: Cosine Accuracy@3
228
  - type: cosine_accuracy@5
229
- value: 0.46
230
  name: Cosine Accuracy@5
231
  - type: cosine_accuracy@10
232
- value: 0.5800000000000001
233
  name: Cosine Accuracy@10
234
  - type: cosine_precision@1
235
- value: 0.24
236
  name: Cosine Precision@1
237
  - type: cosine_precision@3
238
- value: 0.13333333333333333
239
  name: Cosine Precision@3
240
  - type: cosine_precision@5
241
- value: 0.094
242
  name: Cosine Precision@5
243
  - type: cosine_precision@10
244
- value: 0.059000000000000004
245
  name: Cosine Precision@10
246
  - type: cosine_recall@1
247
- value: 0.22999999999999998
248
  name: Cosine Recall@1
249
  - type: cosine_recall@3
250
- value: 0.39
251
  name: Cosine Recall@3
252
  - type: cosine_recall@5
253
- value: 0.455
254
  name: Cosine Recall@5
255
  - type: cosine_recall@10
256
- value: 0.5700000000000001
257
  name: Cosine Recall@10
258
  - type: cosine_ndcg@10
259
- value: 0.39273781398828755
260
  name: Cosine Ndcg@10
261
  - type: cosine_mrr@10
262
- value: 0.3428690476190476
263
  name: Cosine Mrr@10
264
  - type: cosine_map@100
265
- value: 0.34816167097543366
266
  name: Cosine Map@100
267
  ---
268
 
269
- # SentenceTransformer based on thenlper/gte-small
270
 
271
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [thenlper/gte-small](https://huggingface.co/thenlper/gte-small). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
272
 
273
  ## Model Details
274
 
275
  ### Model Description
276
  - **Model Type:** Sentence Transformer
277
- - **Base model:** [thenlper/gte-small](https://huggingface.co/thenlper/gte-small) <!-- at revision 17e1f347d17fe144873b1201da91788898c639cd -->
278
  - **Maximum Sequence Length:** 128 tokens
279
- - **Output Dimensionality:** 384 dimensions
280
  - **Similarity Function:** Cosine Similarity
281
  <!-- - **Training Dataset:** Unknown -->
282
  <!-- - **Language:** Unknown -->
@@ -292,9 +318,8 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [t
292
 
293
  ```
294
  SentenceTransformer(
295
- (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'BertModel'})
296
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
297
- (2): Normalize()
298
  )
299
  ```
300
 
@@ -316,20 +341,20 @@ from sentence_transformers import SentenceTransformer
316
  model = SentenceTransformer("redis/model-b-structured")
317
  # Run inference
318
  sentences = [
319
- 'Schliemann recognized five shafts and cleared them like the graves mentioned by Pausanias .',
320
- 'Schliemann recognized five shafts and cleared them like the graves mentioned by Pausanias .',
321
- 'Schliemann cleared five shafts and recognized them as the graves mentioned by Pausania .',
322
  ]
323
  embeddings = model.encode(sentences)
324
  print(embeddings.shape)
325
- # [3, 384]
326
 
327
  # Get the similarity scores for the embeddings
328
  similarities = model.similarity(embeddings, embeddings)
329
  print(similarities)
330
- # tensor([[1.0000, 1.0000, 0.9879],
331
- # [1.0000, 1.0000, 0.9879],
332
- # [0.9879, 0.9879, 1.0000]])
333
  ```
334
 
335
  <!--
@@ -367,21 +392,21 @@ You can finetune this model on your own dataset.
367
 
368
  | Metric | NanoMSMARCO | NanoNQ |
369
  |:--------------------|:------------|:-----------|
370
- | cosine_accuracy@1 | 0.24 | 0.24 |
371
- | cosine_accuracy@3 | 0.46 | 0.34 |
372
- | cosine_accuracy@5 | 0.54 | 0.38 |
373
- | cosine_accuracy@10 | 0.66 | 0.5 |
374
- | cosine_precision@1 | 0.24 | 0.24 |
375
- | cosine_precision@3 | 0.1533 | 0.1133 |
376
- | cosine_precision@5 | 0.108 | 0.08 |
377
- | cosine_precision@10 | 0.066 | 0.052 |
378
- | cosine_recall@1 | 0.24 | 0.22 |
379
- | cosine_recall@3 | 0.46 | 0.32 |
380
- | cosine_recall@5 | 0.54 | 0.37 |
381
- | cosine_recall@10 | 0.66 | 0.48 |
382
- | **cosine_ndcg@10** | **0.4429** | **0.3426** |
383
- | cosine_mrr@10 | 0.3745 | 0.3113 |
384
- | cosine_map@100 | 0.3864 | 0.3099 |
385
 
386
  #### Nano BEIR
387
 
@@ -399,21 +424,21 @@ You can finetune this model on your own dataset.
399
 
400
  | Metric | Value |
401
  |:--------------------|:-----------|
402
- | cosine_accuracy@1 | 0.24 |
403
- | cosine_accuracy@3 | 0.4 |
404
- | cosine_accuracy@5 | 0.46 |
405
- | cosine_accuracy@10 | 0.58 |
406
- | cosine_precision@1 | 0.24 |
407
- | cosine_precision@3 | 0.1333 |
408
- | cosine_precision@5 | 0.094 |
409
- | cosine_precision@10 | 0.059 |
410
- | cosine_recall@1 | 0.23 |
411
- | cosine_recall@3 | 0.39 |
412
- | cosine_recall@5 | 0.455 |
413
- | cosine_recall@10 | 0.57 |
414
- | **cosine_ndcg@10** | **0.3927** |
415
- | cosine_mrr@10 | 0.3429 |
416
- | cosine_map@100 | 0.3482 |
417
 
418
  <!--
419
  ## Bias, Risks and Limitations
@@ -439,13 +464,13 @@ You can finetune this model on your own dataset.
439
  | | anchor | positive | negative |
440
  |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
441
  | type | string | string | string |
442
- | details | <ul><li>min: 4 tokens</li><li>mean: 10.95 tokens</li><li>max: 60 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 67.57 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 66.64 tokens</li><li>max: 128 tokens</li></ul> |
443
  * Samples:
444
- | anchor | positive | negative |
445
- |:----------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
446
- | <code>how far is sandos caracol eco resort from cancun airport</code> | <code>The Sandos Caracol Eco Resort is 2 miles from the Church of Guadalupe and a 45-minute drive from Cancun Cancún. Airport The Gran Coral Golf Riviera maya is located within the same estate as The. Sandos we speak your! Language Hotel: rooms, 680 Hotel: Chain Sandos & Hotels. resorts</code> | <code>Featuring a spa, 8 restaurants and 2 outdoor pools, Sandos Caracol Eco Resort is set on Playa del Carmen Beach, overlooking Cozumel Island. Its rooms have balconies overlooking the Caribbean Sea. Sandos Caracol Eco Resort is in beautiful gardens and features bright accommodations.</code> |
447
- | <code>can eggs expire</code> | <code>Here is a link from Georgia Eggs Commission about eggs and expiration dates. The following is from Swedish Medical Center Eggs: If you ve purchased a carton of eggs before the date expires, you should be able to use them safely for three to five weeks after expiration.ere is a link from Georgia Eggs Commission about eggs and expiration dates. The following is from Swedish Medical Center Eggs: If you ve purchased a carton of eggs before the date expires, you should be able to use them safely for three to five weeks after expiration.</code> | <code>The answer to this question may surprise you: while uncooked eggs typically last four to five weeks when properly refrigerated, hard-boiled eggs will only last about a week. This is because egg shells, which are highly porous, are sprayed before sale with a thin coating of mineral oil that seals the egg.</code> |
448
- | <code>how old are first graders?</code> | <code>First Grade Worksheets Online. 6 and 7 year old kids get their first taste of real schooling in first grade. Help children learn the basics in math, reading, language and science with our printable first grade worksheets. Spelling Worksheets for 1st Grade.</code> | <code>Average BMI percentile-for-age values were 59.5 (28.8) for first-graders, 59.5 (30.5) for third-graders, and 62.4 (31.7) for fifth-graders. The number of participants classified as obese was 144 (25.6% of first-graders, 28.5% of third-graders, and 34.5% of fifth-graders). The percentage of students who reported a reasonable height or weight ranged from 20% (first grade, height) to 92% (fifth grade, weight) (Table). In general, self-report ability was better in older children and when self-reporting weight.</code> |
449
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
450
  ```json
451
  {
@@ -465,13 +490,13 @@ You can finetune this model on your own dataset.
465
  | | anchor | positive | negative |
466
  |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
467
  | type | string | string | string |
468
- | details | <ul><li>min: 4 tokens</li><li>mean: 11.11 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 67.99 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 66.08 tokens</li><li>max: 128 tokens</li></ul> |
469
  * Samples:
470
- | anchor | positive | negative |
471
- |:----------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
472
- | <code>In 1883 , the first schools were built in the vicinity for 400 white and 60 black students .</code> | <code>In 1883 , the first schools were built in the vicinity for 400 white and 60 black students .</code> | <code>In 1883 , the first schools in the area were built for 400 black students and 60 white students .</code> |
473
- | <code>what is the origin of the name haja</code> | <code>Haja is a Muslim baby Girl name, it is an Arabic originated name. Haja name meaning is In the heart condition through and the lucky number associated with Haja is 5. Find all the relevant details about the Haja Meaning, Origin, Lucky Number and Religion from this page. Average rating of Haja is 1 stars, based on 0 reviews.</code> | <code>Synonomis with the exclamation commonly used in urban circles Holla. Haba is derived from the term, Holla Bitches, which became Haba Litches, which eventually evolved to Habalicious, and finally became just Haba. When seeing a fine female passing by, Russell exclaimed, Haba.</code> |
474
- | <code>what causes itch rash</code> | <code>A rash is a noticeable change in the texture or color of the skin. The skin may become itchy, bumpy, chapped, scaly, or otherwise irritated. Rashes are caused by a wide range of conditions, including allergies, medication, cosmetics, and various diseases. The rash is often reddish and itchy, with a scaly texture. 2 bug bites: tick bites are of particular concern, as they can transmit disease. 3 psoriasis: a scaly, itchy, red rash that forms along the scalp and joints. 4 dandruff: an itchy, flaky rash on the scalp.</code> | <code>Causes of Similar Symptoms to Behind knee rash. Research the causes of these symptoms that are similar to, or related to, the symptom Behind knee rash: 1 Behind knee itch (14 causes). 2 Knee rash (18 causes).3 Knee pain (122 causes). 4 Knee tingling (6 causes). 5 Knee symptoms (149 causes). 6 Skin itch (1068 causes). 7 Skin rash (461 causes). 8 Insect bite.auses of Similar Symptoms to Behind knee rash. Research the causes of these symptoms that are similar to, or related to, the symptom Behind knee rash: 1 Behind knee itch (14 causes). 2 Knee rash (18 causes).</code> |
475
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
476
  ```json
477
  {
@@ -487,9 +512,9 @@ You can finetune this model on your own dataset.
487
  - `eval_strategy`: steps
488
  - `per_device_train_batch_size`: 128
489
  - `per_device_eval_batch_size`: 128
490
- - `learning_rate`: 8e-05
491
- - `weight_decay`: 0.005
492
- - `max_steps`: 3375
493
  - `warmup_ratio`: 0.1
494
  - `fp16`: True
495
  - `dataloader_drop_last`: True
@@ -516,14 +541,14 @@ You can finetune this model on your own dataset.
516
  - `gradient_accumulation_steps`: 1
517
  - `eval_accumulation_steps`: None
518
  - `torch_empty_cache_steps`: None
519
- - `learning_rate`: 8e-05
520
- - `weight_decay`: 0.005
521
  - `adam_beta1`: 0.9
522
  - `adam_beta2`: 0.999
523
  - `adam_epsilon`: 1e-08
524
  - `max_grad_norm`: 1.0
525
  - `num_train_epochs`: 3.0
526
- - `max_steps`: 3375
527
  - `lr_scheduler_type`: linear
528
  - `lr_scheduler_kwargs`: {}
529
  - `warmup_ratio`: 0.1
@@ -630,20 +655,9 @@ You can finetune this model on your own dataset.
630
  ### Training Logs
631
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
632
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
633
- | 0 | 0 | - | 4.9037 | 0.6259 | 0.6583 | 0.6421 |
634
- | 0.2874 | 250 | 3.7176 | 3.0286 | 0.5347 | 0.4519 | 0.4933 |
635
- | 0.5747 | 500 | 3.111 | 2.9929 | 0.4499 | 0.4041 | 0.4270 |
636
- | 0.8621 | 750 | 3.0734 | 2.9741 | 0.4816 | 0.3752 | 0.4284 |
637
- | 1.1494 | 1000 | 3.0287 | 2.9680 | 0.4802 | 0.3422 | 0.4112 |
638
- | 1.4368 | 1250 | 3.0024 | 2.9618 | 0.4850 | 0.3506 | 0.4178 |
639
- | 1.7241 | 1500 | 2.9962 | 2.9568 | 0.4677 | 0.3843 | 0.4260 |
640
- | 2.0115 | 1750 | 2.9903 | 2.9532 | 0.4694 | 0.3430 | 0.4062 |
641
- | 2.2989 | 2000 | 2.9473 | 2.9527 | 0.4446 | 0.3497 | 0.3972 |
642
- | 2.5862 | 2250 | 2.9458 | 2.9519 | 0.4300 | 0.3340 | 0.3820 |
643
- | 2.8736 | 2500 | 2.9392 | 2.9500 | 0.4570 | 0.3466 | 0.4018 |
644
- | 3.1609 | 2750 | 2.9292 | 2.9500 | 0.4529 | 0.3474 | 0.4001 |
645
- | 3.4483 | 3000 | 2.9165 | 2.9502 | 0.4507 | 0.3365 | 0.3936 |
646
- | 3.7356 | 3250 | 2.9151 | 2.9491 | 0.4429 | 0.3426 | 0.3927 |
647
 
648
 
649
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:111470
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: Alibaba-NLP/gte-modernbert-base
11
  widget:
12
+ - source_sentence: when was the first elephant brought to america
13
  sentences:
14
+ - Old Bet The first elephant brought to the United States was in 1796, aboard the
15
+ America which set sail from Calcutta for New York on December 3, 1795.[4] However,
16
+ it is not certain that this was Old Bet.[2] The first references to Old Bet start
17
+ in 1804 in Boston as part of a menagerie.[1] In 1808, while residing in Somers,
18
+ New York, Hachaliah Bailey purchased the menagerie elephant for $1,000 and named
19
+ it "Old Bet".[5][6]
20
+ - Cronus Rhea secretly gave birth to Zeus in Crete, and handed Cronus a stone wrapped
21
+ in swaddling clothes, also known as the Omphalos Stone, which he promptly swallowed,
22
+ thinking that it was his son.
23
+ - Renal artery One or two accessory renal arteries are frequently found, especially
24
+ on the left side since they usually arise from the aorta, and may come off above
25
+ (more common) or below the main artery. Instead of entering the kidney at the
26
+ hilus, they usually pierce the upper or lower part of the organ.
27
+ - source_sentence: who won the india's next superstar grand finale
 
 
 
 
 
 
28
  sentences:
29
+ - India's Next Superstars India's Next Superstars is a talent-search Indian reality
30
+ TV show, which premiered on Star Plus and is streamed on Hotstar.[1] Karan Johar
31
+ and Rohit Shetty are the judges for the show. [2] Aman Gandotra and Natasha Bharadwaj
32
+ were declared winners of 2018 season. Shruti Sharma won a 'Special Mention' award.
33
+ Runners up in the male category were Aashish Mehrotra and Harshvardhan Deo and
34
+ in the female category were Naina Singh and Shruti Sharma. [3]
35
+ - India national cricket team India was invited to The Imperial Cricket Council
36
+ in 1926, and made their debut as a Test playing nation in England in 1932, led
37
+ by CK Nayudu, who was considered as the best Indian batsman at the time.[14] The
38
+ one-off Test match between the two sides was played at Lord's in London. The team
39
+ was not strong in their batting at this point and went on to lose by 158 runs.[15]
40
+ In 1933, the first Test series in India was played between India and England with
41
+ matches in Bombay, Calcutta (now Kolkata) and Madras (now Chennai). England won
42
+ the series 2–0.[16] The Indian team continued to improve throughout the 1930s
43
+ and '40s but did not achieve an international victory during this period. In the
44
+ early 1940s, India didn't play any Test cricket due to the Second World War. The
45
+ team's first series as an independent country was in late 1947 against Sir Donald
46
+ Bradman's Invincibles (a name given to the Australia national cricket team of
47
+ that time). It was also the first Test series India played which was not against
48
+ England. Australia won the five-match series 4–0, with Bradman tormenting the
49
+ Indian bowling in his final Australian summer.[17] India subsequently played their
50
+ first Test series at home not against England against the West Indies in 1948.
51
+ West Indies won the 5-Test series 1–0.[18]
52
+ - Hindi Medium (film) Raj Batra (Irrfan Khan) is a rich businessman from Delhi staying
53
+ with his wife Mita (Saba Qamar). They studied in a Hindi Medium school but wants
54
+ their 5 year old daughter, Pia (Dishita Sehgal), to be admitted to one of the
55
+ top schools in Delhi. The top school, 'Delhi Grammar School', has a condition
56
+ that they will admit students who reside within 3km radius, so the family moves
57
+ to Vasant Vihar.
58
+ - source_sentence: i am human and nothing of that which is human is alien to me meaning
59
  sentences:
60
+ - America's Got Talent Introduced in season nine, the "Golden Buzzer" is located
61
+ on the center of the judges' desk and may be used once per season by each judge.
62
+ In season 9, a judge could press the golden buzzer to save an act from elimination,
63
+ regardless of the number of X's earned from the other judges. Starting in season
64
+ 10 and onward, any act that receives a golden buzzer advances directly to the
65
+ live show; and in season 11, the hosts also were given the power to use the golden
66
+ buzzer. The golden buzzer is also used in the Judge Cuts format.
67
+ - You'll Never Walk Alone "You'll Never Walk Alone" is a show tune from the 1945
68
+ Rodgers and Hammerstein musical Carousel. In the second act of the musical, Nettie
69
+ Fowler, the cousin of the female protagonist Julie Jordan, sings "You'll Never
70
+ Walk Alone" to comfort and encourage Julie when her husband, Billy Bigelow, the
71
+ male lead, commits suicide after a failed robbery attempt. It is reprised in the
72
+ final scene to encourage a graduation class of which Louise (Billy and Julie's
73
+ daughter) is a member. The now invisible Billy, who has been granted the chance
74
+ to return to Earth for one day in order to redeem himself, watches the ceremony
75
+ and is able to silently motivate the unhappy Louise to join in the song.
76
+ - 'Terence One famous quotation by Terence reads: "Homo sum, humani nihil a me alienum
77
+ puto", or "I am human, and I think that nothing of that which is human is alien
78
+ to me." This appeared in his play Heauton Timorumenos.'
79
+ - source_sentence: what do glial cells do in the brain
 
80
  sentences:
81
+ - 'Neuroglia Neuroglia, also called glial cells or simply glia, are non-neuronal
82
+ cells in the central nervous system (brain and spinal cord) and the peripheral
83
+ nervous system. They maintain homeostasis, form myelin, and provide support and
84
+ protection for neurons.[1] In the central nervous system, glial cells include
85
+ oligodendrocytes, astrocytes, ependymal cells and microglia, and in the peripheral
86
+ nervous system glial cells include Schwann cells and satellite cells. They have
87
+ four main functions: (1) To surround neurons and hold them in place (2) To supply
88
+ nutrients and oxygen to neurons (3) To insulate one neuron from another (4) To
89
+ destroy pathogens and remove dead neurons. They also play a role in neurotransmission
90
+ and synaptic connections,[2] and in physiological processes like breathing,[3][4]
91
+ .'
92
+ - The Mother (How I Met Your Mother) Tracy McConnell, better known as "The Mother",
93
+ is the title character from the CBS television sitcom How I Met Your Mother. The
94
+ show, narrated by Future Ted, tells the story of how Ted Mosby met The Mother.
95
+ Tracy McConnell appears in 8 episodes from "Lucky Penny" to "The Time Travelers"
96
+ as an unseen character; she was first seen fully in "Something New" and was promoted
97
+ to a main character in season 9. The Mother is played by Cristin Milioti.
98
+ - Marsupial Marsupials are any members of the mammalian infraclass Marsupialia.
99
+ All extant marsupials are endemic to Australasia and the Americas. A distinctive
100
+ characteristic common to these species is that most of the young are carried in
101
+ a pouch. Well-known marsupials include kangaroos, wallabies, koalas, possums,
102
+ opossums, wombats, and Tasmanian devils. Some lesser-known marsupials are the
103
+ potoroo and the quokka.
104
+ - source_sentence: 'It was Easipower that said :'
105
  sentences:
106
+ - United States presidential election However, federal law does specify that all
107
+ electors must be selected on the same day, which is "the first Tuesday after the
108
+ first Monday in November," i.e., a Tuesday no earlier than November 2 and no later
109
+ than November 8.[17] Today, the states and the District of Columbia each conduct
110
+ their own popular elections on Election Day to help determine their respective
111
+ slate of electors. Thus, the presidential election is really an amalgamation of
112
+ separate and simultaneous state elections instead of a single national election
113
+ run by the federal government.
114
+ - It is said that Easipower was ,
115
+ - 'It was Easipower that said :'
116
  pipeline_tag: sentence-similarity
117
  library_name: sentence-transformers
118
  metrics:
 
132
  - cosine_mrr@10
133
  - cosine_map@100
134
  model-index:
135
+ - name: SentenceTransformer based on Alibaba-NLP/gte-modernbert-base
136
  results:
137
  - task:
138
  type: information-retrieval
 
145
  value: 0.24
146
  name: Cosine Accuracy@1
147
  - type: cosine_accuracy@3
148
+ value: 0.52
149
  name: Cosine Accuracy@3
150
  - type: cosine_accuracy@5
151
+ value: 0.56
152
  name: Cosine Accuracy@5
153
  - type: cosine_accuracy@10
154
+ value: 0.64
155
  name: Cosine Accuracy@10
156
  - type: cosine_precision@1
157
  value: 0.24
158
  name: Cosine Precision@1
159
  - type: cosine_precision@3
160
+ value: 0.1733333333333333
161
  name: Cosine Precision@3
162
  - type: cosine_precision@5
163
+ value: 0.11200000000000002
164
  name: Cosine Precision@5
165
  - type: cosine_precision@10
166
+ value: 0.06400000000000002
167
  name: Cosine Precision@10
168
  - type: cosine_recall@1
169
  value: 0.24
170
  name: Cosine Recall@1
171
  - type: cosine_recall@3
172
+ value: 0.52
173
  name: Cosine Recall@3
174
  - type: cosine_recall@5
175
+ value: 0.56
176
  name: Cosine Recall@5
177
  - type: cosine_recall@10
178
+ value: 0.64
179
  name: Cosine Recall@10
180
  - type: cosine_ndcg@10
181
+ value: 0.44801117912488453
182
  name: Cosine Ndcg@10
183
  - type: cosine_mrr@10
184
+ value: 0.3859444444444445
185
  name: Cosine Mrr@10
186
  - type: cosine_map@100
187
+ value: 0.39907679444975275
188
  name: Cosine Map@100
189
  - task:
190
  type: information-retrieval
 
194
  type: NanoNQ
195
  metrics:
196
  - type: cosine_accuracy@1
197
+ value: 0.3
198
  name: Cosine Accuracy@1
199
  - type: cosine_accuracy@3
200
+ value: 0.56
201
  name: Cosine Accuracy@3
202
  - type: cosine_accuracy@5
203
+ value: 0.66
204
  name: Cosine Accuracy@5
205
  - type: cosine_accuracy@10
206
+ value: 0.78
207
  name: Cosine Accuracy@10
208
  - type: cosine_precision@1
209
+ value: 0.3
210
  name: Cosine Precision@1
211
  - type: cosine_precision@3
212
+ value: 0.18666666666666668
213
  name: Cosine Precision@3
214
  - type: cosine_precision@5
215
+ value: 0.132
216
  name: Cosine Precision@5
217
  - type: cosine_precision@10
218
+ value: 0.08199999999999999
219
  name: Cosine Precision@10
220
  - type: cosine_recall@1
221
+ value: 0.3
222
  name: Cosine Recall@1
223
  - type: cosine_recall@3
224
+ value: 0.56
225
  name: Cosine Recall@3
226
  - type: cosine_recall@5
227
+ value: 0.64
228
  name: Cosine Recall@5
229
  - type: cosine_recall@10
230
+ value: 0.76
231
  name: Cosine Recall@10
232
  - type: cosine_ndcg@10
233
+ value: 0.521342140364588
234
  name: Cosine Ndcg@10
235
  - type: cosine_mrr@10
236
+ value: 0.44460317460317456
237
  name: Cosine Mrr@10
238
  - type: cosine_map@100
239
+ value: 0.4511292484432019
240
  name: Cosine Map@100
241
  - task:
242
  type: nano-beir
 
246
  type: NanoBEIR_mean
247
  metrics:
248
  - type: cosine_accuracy@1
249
+ value: 0.27
250
  name: Cosine Accuracy@1
251
  - type: cosine_accuracy@3
252
+ value: 0.54
253
  name: Cosine Accuracy@3
254
  - type: cosine_accuracy@5
255
+ value: 0.6100000000000001
256
  name: Cosine Accuracy@5
257
  - type: cosine_accuracy@10
258
+ value: 0.71
259
  name: Cosine Accuracy@10
260
  - type: cosine_precision@1
261
+ value: 0.27
262
  name: Cosine Precision@1
263
  - type: cosine_precision@3
264
+ value: 0.18
265
  name: Cosine Precision@3
266
  - type: cosine_precision@5
267
+ value: 0.12200000000000001
268
  name: Cosine Precision@5
269
  - type: cosine_precision@10
270
+ value: 0.07300000000000001
271
  name: Cosine Precision@10
272
  - type: cosine_recall@1
273
+ value: 0.27
274
  name: Cosine Recall@1
275
  - type: cosine_recall@3
276
+ value: 0.54
277
  name: Cosine Recall@3
278
  - type: cosine_recall@5
279
+ value: 0.6000000000000001
280
  name: Cosine Recall@5
281
  - type: cosine_recall@10
282
+ value: 0.7
283
  name: Cosine Recall@10
284
  - type: cosine_ndcg@10
285
+ value: 0.4846766597447363
286
  name: Cosine Ndcg@10
287
  - type: cosine_mrr@10
288
+ value: 0.41527380952380955
289
  name: Cosine Mrr@10
290
  - type: cosine_map@100
291
+ value: 0.4251030214464773
292
  name: Cosine Map@100
293
  ---
294
 
295
+ # SentenceTransformer based on Alibaba-NLP/gte-modernbert-base
296
 
297
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
298
 
299
  ## Model Details
300
 
301
  ### Model Description
302
  - **Model Type:** Sentence Transformer
303
+ - **Base model:** [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) <!-- at revision e7f32e3c00f91d699e8c43b53106206bcc72bb22 -->
304
  - **Maximum Sequence Length:** 128 tokens
305
+ - **Output Dimensionality:** 768 dimensions
306
  - **Similarity Function:** Cosine Similarity
307
  <!-- - **Training Dataset:** Unknown -->
308
  <!-- - **Language:** Unknown -->
 
318
 
319
  ```
320
  SentenceTransformer(
321
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'ModernBertModel'})
322
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
 
323
  )
324
  ```
325
 
 
341
  model = SentenceTransformer("redis/model-b-structured")
342
  # Run inference
343
  sentences = [
344
+ 'It was Easipower that said :',
345
+ 'It was Easipower that said :',
346
+ 'It is said that Easipower was ,',
347
  ]
348
  embeddings = model.encode(sentences)
349
  print(embeddings.shape)
350
+ # [3, 768]
351
 
352
  # Get the similarity scores for the embeddings
353
  similarities = model.similarity(embeddings, embeddings)
354
  print(similarities)
355
+ # tensor([[1.0001, 1.0001, 0.1242],
356
+ # [1.0001, 1.0001, 0.1242],
357
+ # [0.1242, 0.1242, 1.0001]])
358
  ```
359
 
360
  <!--
 
392
 
393
  | Metric | NanoMSMARCO | NanoNQ |
394
  |:--------------------|:------------|:-----------|
395
+ | cosine_accuracy@1 | 0.24 | 0.3 |
396
+ | cosine_accuracy@3 | 0.52 | 0.56 |
397
+ | cosine_accuracy@5 | 0.56 | 0.66 |
398
+ | cosine_accuracy@10 | 0.64 | 0.78 |
399
+ | cosine_precision@1 | 0.24 | 0.3 |
400
+ | cosine_precision@3 | 0.1733 | 0.1867 |
401
+ | cosine_precision@5 | 0.112 | 0.132 |
402
+ | cosine_precision@10 | 0.064 | 0.082 |
403
+ | cosine_recall@1 | 0.24 | 0.3 |
404
+ | cosine_recall@3 | 0.52 | 0.56 |
405
+ | cosine_recall@5 | 0.56 | 0.64 |
406
+ | cosine_recall@10 | 0.64 | 0.76 |
407
+ | **cosine_ndcg@10** | **0.448** | **0.5213** |
408
+ | cosine_mrr@10 | 0.3859 | 0.4446 |
409
+ | cosine_map@100 | 0.3991 | 0.4511 |
410
 
411
  #### Nano BEIR
412
 
 
424
 
425
  | Metric | Value |
426
  |:--------------------|:-----------|
427
+ | cosine_accuracy@1 | 0.27 |
428
+ | cosine_accuracy@3 | 0.54 |
429
+ | cosine_accuracy@5 | 0.61 |
430
+ | cosine_accuracy@10 | 0.71 |
431
+ | cosine_precision@1 | 0.27 |
432
+ | cosine_precision@3 | 0.18 |
433
+ | cosine_precision@5 | 0.122 |
434
+ | cosine_precision@10 | 0.073 |
435
+ | cosine_recall@1 | 0.27 |
436
+ | cosine_recall@3 | 0.54 |
437
+ | cosine_recall@5 | 0.6 |
438
+ | cosine_recall@10 | 0.7 |
439
+ | **cosine_ndcg@10** | **0.4847** |
440
+ | cosine_mrr@10 | 0.4153 |
441
+ | cosine_map@100 | 0.4251 |
442
 
443
  <!--
444
  ## Bias, Risks and Limitations
 
464
  | | anchor | positive | negative |
465
  |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
466
  | type | string | string | string |
467
+ | details | <ul><li>min: 6 tokens</li><li>mean: 13.83 tokens</li><li>max: 45 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 91.42 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 90.36 tokens</li><li>max: 128 tokens</li></ul> |
468
  * Samples:
469
+ | anchor | positive | negative |
470
+ |:--------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
471
+ | <code>which state is home to the arizona ice tea beverage company</code> | <code>Arizona Beverage Company Arizona Beverages USA (stylized as AriZona) is an American producer of many flavors of iced tea, juice cocktails and energy drinks based in Woodbury, New York.[2] Arizona's first product was made available in 1992.</code> | <code>Arya Vaishya Arya Vaishya (Arya Vysya) is an Indian caste. Orthodox Arya Vaishyas follow rituals prescribed in the Vasavi Puranam, a religious text written in the late Middle Ages. Their kuladevata is Vasavi. The community were formerly known as Komati Chettiars in Tamil Nadu but now prefer to be referred to as Arya Vaishya.[1]</code> |
472
+ | <code>when were afro-american and africana studies programs founded in colleges and universities</code> | <code>African-American studies Programs and departments of African-American studies were first created in the 1960s and 1970s as a result of inter-ethnic student and faculty activism at many universities, sparked by a five-month strike for black studies at San Francisco State. In February 1968, San Francisco State hired sociologist Nathan Hare to coordinate the first black studies program and write a proposal for the first Department of Black Studies; the department was created in September 1968 and gained official status at the end of the five-months strike in the spring of 1969. The creation of programs and departments in Black studies was a common demand of protests and sit-ins by minority students and their allies, who felt that their cultures and interests were underserved by the traditional academic structures.</code> | <code>Maze Runner: The Death Cure Maze Runner: The Death Cure was originally set to be released on February 17, 2017, in the United States by 20th Century Fox, but the studio rescheduled the film's release for January 26, 2018 in theatres and IMAX, allowing time for O'Brien to recover from injuries he sustained during filming. The film received mixed reviews from critics and grossed over $284 million worldwide.</code> |
473
+ | <code>who recorded the song total eclipse of the heart</code> | <code>Bonnie Tyler Bonnie Tyler (born Gaynor Hopkins; 8 June 1951) is a Welsh singer, known for her distinctive husky voice. Tyler came to prominence with the release of her 1977 album The World Starts Tonight and its singles "Lost in France" and "More Than a Lover". Her 1978 single "It's a Heartache" reached number four on the UK Singles Chart, and number three on the US Billboard Hot 100.</code> | <code>Manny Pacquiao vs. Juan Manuel Márquez IV Marquez defeated Pacquiao by knockout with one second remaining in the sixth round. It was named Fight of the Year and Knockout of the Year by Ring Magazine, with round five garnering Round of the Year honors.[2]</code> |
474
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
475
  ```json
476
  {
 
490
  | | anchor | positive | negative |
491
  |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
492
  | type | string | string | string |
493
+ | details | <ul><li>min: 6 tokens</li><li>mean: 13.69 tokens</li><li>max: 45 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 90.17 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 89.67 tokens</li><li>max: 128 tokens</li></ul> |
494
  * Samples:
495
+ | anchor | positive | negative |
496
+ |:----------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
497
+ | <code>In early July , Steve Whitley , the criminal father of Harper Whitley and Garrett Whitley , and brother of Benny Cameron .</code> | <code>In early July , Steve Whitley , the criminal father of Harper Whitley and Garrett Whitley , and brother of Benny Cameron .</code> | <code>In early July , Garrett Whitley , who is the criminal father of Harper Whitley and Steve Whitley , and the brother of Benny Cameron , appeared .</code> |
498
+ | <code>when will the next season of house of cards be released</code> | <code>House of Cards (season 6) The sixth and final season of the American political drama web television series House of Cards was confirmed by Netflix on December 4, 2017, and is scheduled to be released on November 2, 2018. Unlike previous seasons that consisted of thirteen episodes each, the sixth season will consist of only eight. The season will not include former lead actor Kevin Spacey, who was fired from the show due to sexual misconduct allegations.</code> | <code>Wild 'n Out For the first four seasons, the show filmed from Los Angeles/Hollywood and aired on MTV. The first run episodes were suspended as Mr. Renaissance Entertainment became Ncredible Entertainment in 2012. Upon being revived in 2012, the show was produced in New York City and aired on MTV2 during Seasons 5–7, it also returned to that location for Season 9. In 2016, the show returned to airing new episodes on MTV and also for the first time since Season 4, production is in Los Angeles.</code> |
499
+ | <code>who played the father on father knows best</code> | <code>Father Knows Best The series began August 25, 1949, on NBC Radio. Set in the Midwest, it starred Robert Young as the General Insurance agent Jim Anderson. His wife Margaret was first portrayed by June Whitley and later by Jean Vander Pyl. The Anderson children were Betty (Rhoda Williams), Bud (Ted Donaldson), and Kathy (Norma Jean Nilsson). Others in the cast were Eleanor Audley, Herb Vigran and Sam Edwards. Sponsored through most of its run by General Foods, the series was heard Thursday evenings on NBC until March 25, 1954.</code> | <code>List of To Kill a Mockingbird characters Maycomb children believe he is a horrible person, due to the rumors spread about him and a trial he underwent as a teenager. It is implied during the story that Boo is a very lonely man who attempts to reach out to Jem and Scout for love and friendship, such as leaving them small gifts and figures in a tree knothole. Scout finally meets him at the very end of the book, when he saves the children's lives from Bob Ewell. Scout describes him as being sickly white, with a thin mouth, thin and feathery hair, and grey eyes, almost as if he were blind. During the same night, when Boo whispers to Scout to walk him back to the Radley house, Scout takes a moment to picture what it would be like to be Boo Radley. While standing on his porch, she realizes his "exile" inside his house is really not that lonely.</code> |
500
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
501
  ```json
502
  {
 
512
  - `eval_strategy`: steps
513
  - `per_device_train_batch_size`: 128
514
  - `per_device_eval_batch_size`: 128
515
+ - `learning_rate`: 4e-05
516
+ - `weight_decay`: 0.01
517
+ - `max_steps`: 703
518
  - `warmup_ratio`: 0.1
519
  - `fp16`: True
520
  - `dataloader_drop_last`: True
 
541
  - `gradient_accumulation_steps`: 1
542
  - `eval_accumulation_steps`: None
543
  - `torch_empty_cache_steps`: None
544
+ - `learning_rate`: 4e-05
545
+ - `weight_decay`: 0.01
546
  - `adam_beta1`: 0.9
547
  - `adam_beta2`: 0.999
548
  - `adam_epsilon`: 1e-08
549
  - `max_grad_norm`: 1.0
550
  - `num_train_epochs`: 3.0
551
+ - `max_steps`: 703
552
  - `lr_scheduler_type`: linear
553
  - `lr_scheduler_kwargs`: {}
554
  - `warmup_ratio`: 0.1
 
655
  ### Training Logs
656
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
657
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
658
+ | 0 | 0 | - | 4.3452 | 0.6530 | 0.6552 | 0.6541 |
659
+ | 0.2874 | 250 | 3.1166 | 2.9191 | 0.4629 | 0.5508 | 0.5069 |
660
+ | 0.5747 | 500 | 2.9043 | 2.8945 | 0.4480 | 0.5213 | 0.4847 |
 
 
 
 
 
 
 
 
 
 
 
661
 
662
 
663
  ### Framework Versions
config_sentence_transformers.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "model_type": "SentenceTransformer",
3
  "__version__": {
4
  "sentence_transformers": "5.2.0",
5
  "transformers": "4.57.3",
@@ -10,5 +9,6 @@
10
  "document": ""
11
  },
12
  "default_prompt_name": null,
13
- "similarity_fn_name": "cosine"
 
14
  }
 
1
  {
 
2
  "__version__": {
3
  "sentence_transformers": "5.2.0",
4
  "transformers": "4.57.3",
 
9
  "document": ""
10
  },
11
  "default_prompt_name": null,
12
+ "similarity_fn_name": "cosine",
13
+ "model_type": "SentenceTransformer"
14
  }
modules.json CHANGED
@@ -10,11 +10,5 @@
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
13
- },
14
- {
15
- "idx": 2,
16
- "name": "2",
17
- "path": "2_Normalize",
18
- "type": "sentence_transformers.models.Normalize"
19
  }
20
  ]
 
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
 
 
 
 
 
 
13
  }
14
  ]