File size: 54,627 Bytes
8616223
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
---
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- generated_from_trainer
- dataset_size:589508
- loss:CachedInfonce
base_model: jinaai/jina-embeddings-v3
widget:
- source_sentence: What are some examples of postgraduate fellowships in the United
    States and Canada
  sentences:
  - "Fellowships as a training program\nFellowships may involve a short placement\
    \ for capacity building, e.g., to get more experience in government, such as the\
    \ American Association for the Advancement of Science's fellowships and the American\
    \ Academy of Arts and Sciences Fellowship programs. Some institutions offer fellowships\
    \ as a professional training program as well as a financial grant, such as the\
    \ Balsillie School of International Affairs, where tuition and other fees are\
    \ paid by the fellowship. Fellowships as a special membership grade\n\nFellows\
    \ are often the highest grade of membership of many professional associations\
    \ or learned societies, for example, the Chartered Institute of Arbitrators, the\
    \ Chartered Governance Institute or Royal College of Surgeons. Lower grades are\
    \ referred to as members (who typically share voting rights with the fellows),\
    \ or associates (who may or may not, depending on whether \"associate\" status\
    \ is a form of full membership). Additional grades of membership exist in, for\
    \ example, the IEEE and the ACM. Fellowships of this type can be awarded as a\
    \ title of honor in their own right, e.g. the Fellowship of the Royal Society\
    \ (FRS). Exclusive learned societies such as the Royal Society have Fellow as\
    \ the only grade of membership. Appointment as an honorary fellow in a learned\
    \ or professional society can be either to honour exceptional achievement or service\
    \ within the professional domain of the awarding body or to honour contributions\
    \ related to the domain from someone who is professionally outside it. Membership\
    \ of the awarding body may or may not be a requirement. How a fellowship is awarded\
    \ varies for each society, but may typically involve some or all of these:\n A\
    \ qualifying period in a lower grade\n Passing a series of examinations\n Nomination\
    \ by two existing fellows who know the applicant professionally\n Evidence of\
    \ continued formal training post-qualification\n Evidence of substantial achievement\
    \ in the subject area\n Submission of a thesis or portfolio of works which will\
    \ be examined\n Election by a vote of the fellowship\n\nIn ancient universities\n\
    \nAt the ancient universities of the University of Oxford, the University of Cambridge,\
    \ and Trinity College, Dublin, members of the teaching staff typically have two\
    \ affiliations: one as a reader, lecturer, or other academic rank within a department\
    \ of the university, as at other universities, and a second affiliation as a fellow\
    \ of one of the colleges of the university. The fellows, sometimes referred to\
    \ as university dons, form the governing body of the college. They may elect a\
    \ council to handle day-to-day management."
  - If you are an enrolled domestic or international student studying a full degree
    program, you may be eligible to study overseas! We have over 70 partner institutions
    worldwide and the opportunities are endless. Visit our USC International and Study
    Overseas blog to learn more about the amazing experiences our students are having
    abroad.
  - 'The title (senior) fellow can also be bestowed to an academic member of staff
    upon retirement who continues to be affiliated to a university in the United Kingdom.
    The term teaching fellow or teaching assistant is used, in the United States and
    United Kingdom, in secondary school, high school and middle school setting for
    students or adults that assist a teacher with one or more classes. Medical fellowships


    In US medical institutions, a fellow refers to someone who has completed residency
    training (e.g. in internal medicine, pediatrics, general surgery, etc.) and is
    currently in a 1 to 3 year subspecialty training program (e.g. cardiology, pediatric
    nephrology, transplant surgery, etc.). Research fellowships


    As an academic position


    The title of research fellow may be used to denote an academic position at a university
    or a similar institution; it is roughly equivalent to the title of lecturer in
    the Commonwealth teaching career pathway. As a financial grant

    Research fellow may also refer to the recipient of academic financial grant or
    scholarship. For example, in Germany, institutions such as the Alexander von Humboldt
    Foundation offer research fellowship for postdoctoral research and refer to the
    holder as research fellows, while the award holder may formally hold a specific
    academic title at their home institution (e.g., Privatdozent). These are often
    shortened to the name of the programme or organization, e.g. Dorothy Hodgkin Fellow
    rather than Dorothy Hodgkin Research Fellow, except where this might cause confusion
    with another fellowship, (e.g. Royal Society University Research Fellowship.)'
  - "In the context of graduate school in the United States and Canada, a fellow is\
    \ a recipient of a postgraduate fellowship. Examples include the NSF Graduate\
    \ Research Fellowship, the DoD National Defense Science and Engineering Graduate\
    \ Fellowship, the DOE Computational Science Graduate Fellowship, the Guggenheim\
    \ Fellowship, the Rosenthal Fellowship, the Frank Knox Memorial Fellowship, the\
    \ Woodrow Wilson Teaching Fellowship and the Presidential Management Fellowship.\
    \ It is granted to prospective or current students, on the basis of their academic\
    \ or research achievements. In the UK, research fellowships are awarded to support\
    \ postdoctoral researchers such as those funded by the Wellcome Trust and the\
    \ Biotechnology and Biological Sciences Research Council (BBSRC). At ETH Zurich,\
    \ postdoctoral fellowships support incoming researchers. The MacArthur Fellows\
    \ Program (aka \"genius grant\") as prestigious research fellowship awarded in\
    \ the United States. Fellowships as a training program\nFellowships may involve\
    \ a short placement for capacity building, e.g., to get more experience in government,\
    \ such as the American Association for the Advancement of Science's fellowships\
    \ and the American Academy of Arts and Sciences Fellowship programs. Some institutions\
    \ offer fellowships as a professional training program as well as a financial\
    \ grant, such as the Balsillie School of International Affairs, where tuition\
    \ and other fees are paid by the fellowship. Fellowships as a special membership\
    \ grade\n\nFellows are often the highest grade of membership of many professional\
    \ associations or learned societies, for example, the Chartered Institute of Arbitrators,\
    \ the Chartered Governance Institute or Royal College of Surgeons. Lower grades\
    \ are referred to as members (who typically share voting rights with the fellows),\
    \ or associates (who may or may not, depending on whether \"associate\" status\
    \ is a form of full membership). Additional grades of membership exist in, for\
    \ example, the IEEE and the ACM. Fellowships of this type can be awarded as a\
    \ title of honor in their own right, e.g. the Fellowship of the Royal Society\
    \ (FRS). Exclusive learned societies such as the Royal Society have Fellow as\
    \ the only grade of membership. Appointment as an honorary fellow in a learned\
    \ or professional society can be either to honour exceptional achievement or service\
    \ within the professional domain of the awarding body or to honour contributions\
    \ related to the domain from someone who is professionally outside it. Membership\
    \ of the awarding body may or may not be a requirement. How a fellowship is awarded\
    \ varies for each society, but may typically involve some or all of these:\n A\
    \ qualifying period in a lower grade\n Passing a series of examinations\n Nomination\
    \ by two existing fellows who know the applicant professionally\n Evidence of\
    \ continued formal training post-qualification\n Evidence of substantial achievement\
    \ in the subject area\n Submission of a thesis or portfolio of works which will\
    \ be examined\n Election by a vote of the fellowship\n\nIn ancient universities\n\
    \nAt the ancient universities of the University of Oxford, the University of Cambridge,\
    \ and Trinity College, Dublin, members of the teaching staff typically have two\
    \ affiliations: one as a reader, lecturer, or other academic rank within a department\
    \ of the university, as at other universities, and a second affiliation as a fellow\
    \ of one of the colleges of the university. The fellows, sometimes referred to\
    \ as university dons, form the governing body of the college. They may elect a\
    \ council to handle day-to-day management. All fellows are entitled to certain\
    \ privileges within their colleges, which may include dining at High Table (free\
    \ of charge) and possibly the right to a room in college (free of charge). At\
    \ Cambridge, retired academics may remain fellows. At Oxford, however, a Governing\
    \ Body fellow would normally be elected a fellow emeritus and would leave the\
    \ Governing Body upon his or her retirement. Distinguished old members of the\
    \ college, or its benefactors and friends, might also be elected 'Honorary Fellow',\
    \ normally for life; but beyond limited dining rights this is merely an honour.\
    \ Most Oxford colleges have 'Fellows by Special Election' or 'Supernumerary Fellows',\
    \ who may be members of the teaching staff, but not necessarily members of the\
    \ Governing Body. Some senior administrators of a college such as bursars are\
    \ made fellows, and thereby become members of the governing body, because of their\
    \ importance to the running of a college."
- source_sentence: What kind of plants or decorations are described as popular, fresh,
    and plentiful in the garden at this time of year
  sentences:
  - Enjoy the beautiful scent of gardenia, rosemary, and lavender from your garden.
    Hurry this will not last.
  - Things have been given the opportunity to grow whichever they want but not out
    of neglect per se. Somehow it adds whimsy and mystery to the courtyard.
  - I thought it might be fun to show how this garden goes though the season. The
    perennials will be the fastest to clean up.. clear out the pathways, and bed them
    down well.
  - I'm not surprised that they are so popular, they are fresh, green, with the jolly
    berries AND plentiful in the garden, this time of year, and also, so very decorative.
    I am trying to use as little light and dof as possible, a challenge that I love.
- source_sentence: When was the Santa Venera church in Avola constructed
  sentences:
  - The dome collapsed in the earthquake of 1848, and was not reconstructed until
    1962 by the engineer Pietro Lojacono. The decorated three story facade, flanked
    by volutes and obelisks, houses a statue of Saint Venera, patron of Avola, above
    the central portal.
  - 'Santa Venera is a Baroque style church located on Piazza Teatro in the town of
    Avola, province of Siracusa, region of Sicily, Italy. History and description

    Construction of a church at the site took place from 1713-1715 using designs attributed
    to Michelangelo Alessi.'
  - 'The Saint Bavo Church (Dutch: Sint-Bavokerk, Sint-Baafskerk) is a Dutch Reformed
    church building in Aardenburg, Netherlands. The church was founded in 959 by monks
    of the Saint Bavo''s Abbey in Ghent. Due to a rise in population this small church
    was replaced by a Romanesque church which burned down in 1202. In 1220 the current
    tower, nave and transept were built.'
  - The decorated three story facade, flanked by volutes and obelisks, houses a statue
    of Saint Venera, patron of Avola, above the central portal. The interior has three
    naves.
- source_sentence: What is the last dream the speaker mentions
  sentences:
  - Have you ever felt like the dreams you had have never become reality? Have you
    ever felt like you need someone to spark the flame for you
  - I'm very new to this, so I'm not sure what I'm doing with the technical side of
    things. Please bear with me if I've got anything wrong. "Night Thoughts And Dreams"
    is the first thing I've written in about two years. I used to write all the time,
    but then I just stopped, however "Sherlock" and Benedict Cumberbatch have inspired
    me to have another go.
  - I had a fantastic phone conversation with my brother today. I also had a nightmare
    where a man pulled off his skin like a shirt.
  - They don't bury me without my uniform." "My last dream is to be in Cooperstown-to
    be with those guys.
- source_sentence: What is the description of the Myrmecoleon and what are its two
    interpretations
  sentences:
  - 'The stone lies at the bottom of the sea and comes to life early in the morning.
    When it rises from its resting-place to the surface of the sea, it opens its mouth
    and takes in some heavenly dew, and the rays of the sun shine around it; thus
    there grows within the stone a most precious, shining pearl indeed, conceived
    from the heavenly dew and given lustre by the rays of the sun." Interpretations


    There are two interpretations of what a Myrmecoleon is. In one version, the antlion
    is so called because it is the "lion of ants", a large ant or small animal that
    hides in the dust and kills ants. In the other version, it is a beast that is
    the result of a mating between a lion and an ant. It has the face of a lion and
    the body of an ant, with each part having its appropriate nature. Because the
    lion part will only eat meat and the ant part can only digest grain, the ant-lion
    starves.'
  - It is found in Medieval bestiaries such as the Hortus Sanitatis of Jacob Meydenbach.
    It is also referenced in some sources as a Formicaleon (Antlion), Formicaleun
    or Mirmicioleon.
  - Microdiprion is a genus of sawflies belonging to the family Diprionidae.
  - Macrodon is a genus of marine ray-finned fishes belonging to the family Sciaenidae,
    the drums and croakers.
pipeline_tag: sentence-similarity
library_name: sentence-transformers
---

# SentenceTransformer based on jinaai/jina-embeddings-v3

This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [jinaai/jina-embeddings-v3](https://huggingface.co/jinaai/jina-embeddings-v3) on the hard_negative_merged dataset. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.

## Model Details

### Model Description
- **Model Type:** Sentence Transformer
- **Base model:** [jinaai/jina-embeddings-v3](https://huggingface.co/jinaai/jina-embeddings-v3) <!-- at revision f1944de8402dcd5f2b03f822a4bc22a7f2de2eb9 -->
- **Maximum Sequence Length:** 2048 tokens
- **Output Dimensionality:** 1024 dimensions
- **Similarity Function:** Cosine Similarity
- **Training Dataset:**
    - hard_negative_merged
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->

### Model Sources

- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)

### Full Model Architecture

```
SentenceTransformer(
  (transformer): Transformer(
    (auto_model): XLMRobertaLoRA(
      (roberta): XLMRobertaModel(
        (embeddings): XLMRobertaEmbeddings(
          (word_embeddings): ParametrizedEmbedding(
            250002, 1024, padding_idx=1
            (parametrizations): ModuleDict(
              (weight): ParametrizationList(
                (0): LoRAParametrization()
              )
            )
          )
          (token_type_embeddings): ParametrizedEmbedding(
            1, 1024
            (parametrizations): ModuleDict(
              (weight): ParametrizationList(
                (0): LoRAParametrization()
              )
            )
          )
        )
        (emb_drop): Dropout(p=0.1, inplace=False)
        (emb_ln): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
        (encoder): XLMRobertaEncoder(
          (layers): ModuleList(
            (0-23): 24 x Block(
              (mixer): MHA(
                (rotary_emb): RotaryEmbedding()
                (Wqkv): ParametrizedLinearResidual(
                  in_features=1024, out_features=3072, bias=True
                  (parametrizations): ModuleDict(
                    (weight): ParametrizationList(
                      (0): LoRAParametrization()
                    )
                  )
                )
                (inner_attn): FlashSelfAttention(
                  (drop): Dropout(p=0.1, inplace=False)
                )
                (inner_cross_attn): FlashCrossAttention(
                  (drop): Dropout(p=0.1, inplace=False)
                )
                (out_proj): ParametrizedLinear(
                  in_features=1024, out_features=1024, bias=True
                  (parametrizations): ModuleDict(
                    (weight): ParametrizationList(
                      (0): LoRAParametrization()
                    )
                  )
                )
              )
              (dropout1): Dropout(p=0.1, inplace=False)
              (drop_path1): StochasticDepth(p=0.0, mode=row)
              (norm1): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
              (mlp): Mlp(
                (fc1): ParametrizedLinear(
                  in_features=1024, out_features=4096, bias=True
                  (parametrizations): ModuleDict(
                    (weight): ParametrizationList(
                      (0): LoRAParametrization()
                    )
                  )
                )
                (fc2): ParametrizedLinear(
                  in_features=4096, out_features=1024, bias=True
                  (parametrizations): ModuleDict(
                    (weight): ParametrizationList(
                      (0): LoRAParametrization()
                    )
                  )
                )
              )
              (dropout2): Dropout(p=0.1, inplace=False)
              (drop_path2): StochasticDepth(p=0.0, mode=row)
              (norm2): LayerNorm((1024,), eps=1e-05, elementwise_affine=True)
            )
          )
        )
        (pooler): XLMRobertaPooler(
          (dense): ParametrizedLinear(
            in_features=1024, out_features=1024, bias=True
            (parametrizations): ModuleDict(
              (weight): ParametrizationList(
                (0): LoRAParametrization()
              )
            )
          )
          (activation): Tanh()
        )
      )
    )
  )
  (pooler): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
  (normalizer): Normalize()
)
```

## Usage

### Direct Usage (Sentence Transformers)

First install the Sentence Transformers library:

```bash
pip install -U sentence-transformers
```

Then you can load this model and run inference.
```python
from sentence_transformers import SentenceTransformer

# Download from the 🤗 Hub
model = SentenceTransformer("Jrinky/jina_final_temp")
# Run inference
sentences = [
    'What is the description of the Myrmecoleon and what are its two interpretations',
    'The stone lies at the bottom of the sea and comes to life early in the morning. When it rises from its resting-place to the surface of the sea, it opens its mouth and takes in some heavenly dew, and the rays of the sun shine around it; thus there grows within the stone a most precious, shining pearl indeed, conceived from the heavenly dew and given lustre by the rays of the sun." Interpretations\n\nThere are two interpretations of what a Myrmecoleon is. In one version, the antlion is so called because it is the "lion of ants", a large ant or small animal that hides in the dust and kills ants. In the other version, it is a beast that is the result of a mating between a lion and an ant. It has the face of a lion and the body of an ant, with each part having its appropriate nature. Because the lion part will only eat meat and the ant part can only digest grain, the ant-lion starves.',
    'It is found in Medieval bestiaries such as the Hortus Sanitatis of Jacob Meydenbach. It is also referenced in some sources as a Formicaleon (Antlion), Formicaleun or Mirmicioleon.',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 1024]

# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
```

<!--
### Direct Usage (Transformers)

<details><summary>Click to see the direct usage in Transformers</summary>

</details>
-->

<!--
### Downstream Usage (Sentence Transformers)

You can finetune this model on your own dataset.

<details><summary>Click to expand</summary>

</details>
-->

<!--
### Out-of-Scope Use

*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->

<!--
## Bias, Risks and Limitations

*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->

<!--
### Recommendations

*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->

## Training Details

### Training Dataset

#### hard_negative_merged

* Dataset: hard_negative_merged
* Size: 589,508 training samples
* Columns: <code>anchor</code>, <code>positive</code>, <code>negative_1</code>, <code>negative_2</code>, and <code>negative_3</code>
* Approximate statistics based on the first 1000 samples:
  |         | anchor                                                                            | positive                                                                             | negative_1                                                                           | negative_2                                                                           | negative_3                                                                           |
  |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
  | type    | string                                                                            | string                                                                               | string                                                                               | string                                                                               | string                                                                               |
  | details | <ul><li>min: 6 tokens</li><li>mean: 17.37 tokens</li><li>max: 37 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 122.81 tokens</li><li>max: 2048 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 128.36 tokens</li><li>max: 2048 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 110.47 tokens</li><li>max: 1920 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 103.93 tokens</li><li>max: 2048 tokens</li></ul> |
* Samples:
  | anchor                                                                                    | positive                                                                                                                                                                                                                                                                                                                                                                                             | negative_1                                                                                                                                                                                                                                                      | negative_2                                                                                                                                                                                                                                                                                                                                                                            | negative_3                                                                                                                                                                                                                                                                                                                        |
  |:------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
  | <code>What does the plot of the story revolve around</code>                               | <code>Respawn points are created when the player accumulates enough blood collected from slain enemies or in-level blood pickups, and idles a certain distance away from immediate level hazards. Plot<br>The plot follows the events of an unnamed young girl's arrival at the Lafcadio Academy for Troubled Young Ladies.</code>                                                                   | <code>An really interesting idea behind the story and one that had me unable to put it down some nights! View all my reviews</code>                                                                                                                             | <code>And everything has such meaning and depth behind it. Nothing is just said casually, and it is all so thoughfully laced with emotion and words to draw you in to the story itself.</code>                                                                                                                                                                                        | <code>It has a terribly implication that this flashback may be lasting more than a chapter. It's not as if we aren't learning anything of importance. I'm just not curious where this is going. I'm wondering when it'll finally be over. Not something you want from your audience as a story teller. In no simple terms.</code> |
  | <code>What type of warranty is offered with the Zhumell Signature 10x42 binoculars</code> | <code>The Signature is also backed by Zhumell's full, 25-year, no-fault warranty, ensuring a lifetime of worry-free viewing. The Zhumell Signature 10x42 binoculars will give you plenty of power - whenever you need it, for as long as you need it!</code>                                                                                                                                         | <code>This item is backed by a Limited Lifetime Warranty. In the event this item should fail due to manufacturing defects during intended use, we will exchange the part free of charge (excludes shipping charges) for the original purchaser.</code>          | <code>if you have different ideas or better suggestion ,be free to leave message . Warranty and terms:<br>-Warranty year is 1 year under normal use,the warranty period is a year from the date of original purchase.</code>                                                                                                                                                          | <code>We have more than 55 years of experience designing, manufacturing and refining custom optical lenses for use in a range of industries. Our production staff follows strict ISO 9001 standards and uses state-of-the-art metrology equipment to test finished lenses for quality and performance.</code>                     |
  | <code>When did he announce his retirement from all professional rugby</code>              | <code>He was named in the Pro12 Dream Teams at the end of the 2014/15 and 2016/17 seasons. In April 2021 he announced his retirement from all professional rugby. International career<br><br>Qualifying to play internationally for Scotland through his Glasgow-born mother, on 24 October 2012 he was named in the full Scottish national team for the 2012 end-of-year rugby union tests.</code> | <code>After retiring from full-time professional football, he worked as a production controller before becoming a sales administrator for International Computers Limited. He lived in Southampton for the rest of his life and died on 28 January 2014.</code> | <code>On December 15 2018, it was announced that he had left WWE voluntarily. Professional boxing record<br>{| class="wikitable" style="text-align:center;"<br>| style="text-align:center;" colspan="8" | 6 Wins (3 knockouts, 3 decisions),  0 Losses, 0 Draws<br>|-  style="text-align:center; background:#e3e3e3;"<br>|  style="border-style:none none solid solid;" | Res.</code> | <code>Since retiring from football he has worked as a journalist for the Professional Footballers' Association. References<br><br>English men's footballers<br>Bristol City F.C. players<br>Kidderminster Harriers F.C. players<br>Yeovil Town F.C.</code>                                                                        |
* Loss: <code>cachedselfloss2.CachedInfonce</code> with these parameters:
  ```json
  {
      "scale": 20.0,
      "similarity_fct": "cos_sim"
  }
  ```

### Evaluation Dataset

#### hard_negative_merged

* Dataset: hard_negative_merged
* Size: 589,508 evaluation samples
* Columns: <code>anchor</code>, <code>positive</code>, <code>negative_1</code>, <code>negative_2</code>, and <code>negative_3</code>
* Approximate statistics based on the first 1000 samples:
  |         | anchor                                                                            | positive                                                                             | negative_1                                                                           | negative_2                                                                           | negative_3                                                                           |
  |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
  | type    | string                                                                            | string                                                                               | string                                                                               | string                                                                               | string                                                                               |
  | details | <ul><li>min: 4 tokens</li><li>mean: 17.27 tokens</li><li>max: 39 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 120.45 tokens</li><li>max: 2031 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 123.54 tokens</li><li>max: 2018 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 114.85 tokens</li><li>max: 1860 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 115.74 tokens</li><li>max: 1605 tokens</li></ul> |
* Samples:
  | anchor                                                                                                                | positive                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           | negative_1                                                                                                                                                                                                                                                                                                    | negative_2                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                     | negative_3                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          |
  |:----------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
  | <code>What could the term 'Golia' refer to</code>                                                                     | <code>Golia may refer to:<br><br>Golia (surname)<br>Golia, Ganjam<br>Golia Monastery<br>1226 Golia</code>                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                          | <code>Gouka may refer to:<br><br> 9708 Gouka, a main-belt asteroid after the Dutch astronomer Adriaan Gouka<br> Eric Gouka (born 1970), Dutch cricketer<br> Gouka, Benin, a town and arrondissement</code>                                                                                                    | <code>Gottschelia is a genus of liverworts belonging to the family Cephaloziellaceae.</code>                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   | <code>Agila may refer to:<br><br>Agila I (died 554), Visigothic king<br>Agila II (died 714), Visigothic king<br>Agila 2, the first Filipino satellite<br>Agila (album), a 1996 album by Spanish rock band Extremoduro<br>Agila (film), a 1980 Philippine film directed by Eddie Romero<br>Agila (TV series), a 1987 Philippine teledrama series<br>Agila Town, Benue State, Nigeria<br>Opel Agila or Vauxhall Agila, a city car<br><br>See also<br>Agila division, the 10th Infantry Division of the Philippine Army<br>Aguila (disambiguation)</code>                                                                                                                                                                                                                                                              |
  | <code>What is the timeframe in which Itera plans to potentially make an agreement with a financial institution</code> | <code>As Itera's President Igor Makarov reported at today's meeting of the Russian Gas Society in Moscow, the gas company could make an agreement with a financial institution, which would make the most profitable and optimum offer, in the next two to three months. According to him, they are currently holding negotiations with several financial enterprises, which specialize in introducing companies to the financial market.</code>                                                                                                                                                                                                                   | <code>The process from receipt of the funding proposal to completion of due diligence is incredibly quick, with a goal of 30 days. After initial evaluation of their proposals, a selected number of start-ups, usually 6 to 8, are asked to make preliminary presentations to the steering committee.</code> | <code>Coinexchange, Cryptopia, YoBit, HitBtc, Binance, Bittrex<br>Q1 2018 : Partners announced (Debit card & Merchants) We are currently in negotiation with major payment providers to offer you a worldwide usable card. Q1/2 2018 : ETHX Beta Wallet release (Android, Windows, iOS) and debit cart pre-order<br>Q3 201 : More partnerships Wider range of companies accepting ETHX. First targets are the biggest e-commerce websites. We will release a beta application to collect user reviews and answer to the community. The app is expected to come out in Q1 2018 on Android and later on iOS. We are very sensitive about our community welfare, so we try to do our best to keep our members informed about the latest news. The app will also help us to inform and get suggestions. Ethereum X is community driven. If you are also a cryptography and distributed ledger tech-nology enthusiast and want to support the project, please feel free to contact us. Additional developers as well as community managers for our social...</code> | <code>The project will be floated in the market for solicitation of expression of interest from the potential investors in June 2017. The land slots will be awarded to the successful bidders based on evaluation by the end of August, 2017. The Monitoring and Evaluation (M&E) of forest sites, awarded to successful bidders, will be done in collaboration with the Forestry, Wildlife & Fisheries Department, Government of the Punjab, as per the provisions of PPP Act, 2014, and The Punjab Forest (Amendment) Act, 2016. Revenue sharing will be done in this initiative. The Company in order to effectively reach out to the business community is organizing seminars in collaboration with various Chambers of Commerce & Industry to sensitize business groups to invest in the opportunity.</code> |
  | <code>What role does File History play in the issue being discussed</code>                                            | <code>What has File History got to do with the problem<br>I don't know but maybe someone at DC does<br>I post the question..... get lots of ideas and methods to remove the naughty files, but I still don't know why deleting file history worked unless the file history is tacked onto the file somehow<br>Since then I've been checking more of the "includes folders" for more over-long files and trying to figure what to do with them. The files are easy to find once you start paying attention<br>Open a folder and if it contains extra long files a scroll bar appears at the bottom of the page<br>Found some more files and started playing.</code> | <code>Newspapers feature stories about lost computers and memory sticks but a more common and longstanding problem is about staff accessing records that they have no right to see. It has always been possible for staff to look at paper records, and in most cases, there is no track of record.</code>    | <code>In data vault it is referred to as the record source. Background <br>The need to identify systems of record can become acute in organizations where management information systems have been built by taking output data from multiple source systems, re-processing this data, and then re-presenting the result for a new business use.</code>                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                         | <code>The idea of preservation, in the sense of both immortalization and protection is addressed. How do we decide what to remember from history, and what do we leave out</code>                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   |
* Loss: <code>cachedselfloss2.CachedInfonce</code> with these parameters:
  ```json
  {
      "scale": 20.0,
      "similarity_fct": "cos_sim"
  }
  ```

### Training Hyperparameters
#### Non-Default Hyperparameters

- `eval_strategy`: steps
- `per_device_train_batch_size`: 500
- `per_device_eval_batch_size`: 500
- `learning_rate`: 2e-05
- `num_train_epochs`: 10
- `warmup_ratio`: 0.1
- `bf16`: True
- `batch_sampler`: no_duplicates

#### All Hyperparameters
<details><summary>Click to expand</summary>

- `overwrite_output_dir`: False
- `do_predict`: False
- `eval_strategy`: steps
- `prediction_loss_only`: True
- `per_device_train_batch_size`: 500
- `per_device_eval_batch_size`: 500
- `per_gpu_train_batch_size`: None
- `per_gpu_eval_batch_size`: None
- `gradient_accumulation_steps`: 1
- `eval_accumulation_steps`: None
- `torch_empty_cache_steps`: None
- `learning_rate`: 2e-05
- `weight_decay`: 0.0
- `adam_beta1`: 0.9
- `adam_beta2`: 0.999
- `adam_epsilon`: 1e-08
- `max_grad_norm`: 1.0
- `num_train_epochs`: 10
- `max_steps`: -1
- `lr_scheduler_type`: linear
- `lr_scheduler_kwargs`: {}
- `warmup_ratio`: 0.1
- `warmup_steps`: 0
- `log_level`: passive
- `log_level_replica`: warning
- `log_on_each_node`: True
- `logging_nan_inf_filter`: True
- `save_safetensors`: True
- `save_on_each_node`: False
- `save_only_model`: False
- `restore_callback_states_from_checkpoint`: False
- `no_cuda`: False
- `use_cpu`: False
- `use_mps_device`: False
- `seed`: 42
- `data_seed`: None
- `jit_mode_eval`: False
- `use_ipex`: False
- `bf16`: True
- `fp16`: False
- `fp16_opt_level`: O1
- `half_precision_backend`: auto
- `bf16_full_eval`: False
- `fp16_full_eval`: False
- `tf32`: None
- `local_rank`: 0
- `ddp_backend`: None
- `tpu_num_cores`: None
- `tpu_metrics_debug`: False
- `debug`: []
- `dataloader_drop_last`: True
- `dataloader_num_workers`: 0
- `dataloader_prefetch_factor`: None
- `past_index`: -1
- `disable_tqdm`: False
- `remove_unused_columns`: True
- `label_names`: None
- `load_best_model_at_end`: False
- `ignore_data_skip`: False
- `fsdp`: []
- `fsdp_min_num_params`: 0
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
- `tp_size`: 0
- `fsdp_transformer_layer_cls_to_wrap`: None
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
- `deepspeed`: None
- `label_smoothing_factor`: 0.0
- `optim`: adamw_torch
- `optim_args`: None
- `adafactor`: False
- `group_by_length`: False
- `length_column_name`: length
- `ddp_find_unused_parameters`: None
- `ddp_bucket_cap_mb`: None
- `ddp_broadcast_buffers`: False
- `dataloader_pin_memory`: True
- `dataloader_persistent_workers`: False
- `skip_memory_metrics`: True
- `use_legacy_prediction_loop`: False
- `push_to_hub`: False
- `resume_from_checkpoint`: None
- `hub_model_id`: None
- `hub_strategy`: every_save
- `hub_private_repo`: None
- `hub_always_push`: False
- `gradient_checkpointing`: False
- `gradient_checkpointing_kwargs`: None
- `include_inputs_for_metrics`: False
- `include_for_metrics`: []
- `eval_do_concat_batches`: True
- `fp16_backend`: auto
- `push_to_hub_model_id`: None
- `push_to_hub_organization`: None
- `mp_parameters`: 
- `auto_find_batch_size`: False
- `full_determinism`: False
- `torchdynamo`: None
- `ray_scope`: last
- `ddp_timeout`: 1800
- `torch_compile`: False
- `torch_compile_backend`: None
- `torch_compile_mode`: None
- `dispatch_batches`: None
- `split_batches`: None
- `include_tokens_per_second`: False
- `include_num_input_tokens_seen`: False
- `neftune_noise_alpha`: None
- `optim_target_modules`: None
- `batch_eval_metrics`: False
- `eval_on_start`: False
- `use_liger_kernel`: False
- `eval_use_gather_object`: False
- `average_tokens_across_devices`: False
- `prompts`: None
- `batch_sampler`: no_duplicates
- `multi_dataset_batch_sampler`: proportional

</details>

### Training Logs
| Epoch  | Step | Training Loss | Validation Loss |
|:------:|:----:|:-------------:|:---------------:|
| 0.1786 | 40   | 8.7768        | 8.5959          |
| 0.3571 | 80   | 8.8187        | 8.5129          |
| 0.5357 | 120  | 8.6175        | 8.2742          |
| 0.7143 | 160  | 8.0868        | 7.8954          |
| 0.8929 | 200  | 7.5681        | 7.3531          |
| 1.0714 | 240  | 7.0288        | 6.5431          |
| 1.25   | 280  | 6.2266        | 5.8462          |
| 1.4286 | 320  | 5.4682        | 5.2924          |
| 1.6071 | 360  | 5.0398        | 4.8148          |
| 1.7857 | 400  | 4.5158        | 4.4110          |
| 1.9643 | 440  | 4.184         | 4.0419          |
| 2.1429 | 480  | 3.7868        | 3.7165          |
| 2.3214 | 520  | 3.6258        | 3.4216          |
| 2.5    | 560  | 3.2262        | 3.1530          |
| 2.6786 | 600  | 3.0175        | 2.9128          |
| 2.8571 | 640  | 2.75          | 2.6999          |
| 3.0357 | 680  | 2.4915        | 2.5085          |


### Framework Versions
- Python: 3.10.14
- Sentence Transformers: 3.4.1
- Transformers: 4.50.0
- PyTorch: 2.3.1+cu121
- Accelerate: 1.5.2
- Datasets: 3.4.1
- Tokenizers: 0.21.1

## Citation

### BibTeX

#### Sentence Transformers
```bibtex
@inproceedings{reimers-2019-sentence-bert,
    title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
    author = "Reimers, Nils and Gurevych, Iryna",
    booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
    month = "11",
    year = "2019",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/1908.10084",
}
```

#### CachedInfonce
```bibtex
@misc{gao2021scaling,
    title={Scaling Deep Contrastive Learning Batch Size under Memory Limited Setup},
    author={Luyu Gao and Yunyi Zhang and Jiawei Han and Jamie Callan},
    year={2021},
    eprint={2101.06983},
    archivePrefix={arXiv},
    primaryClass={cs.LG}
}
```

<!--
## Glossary

*Clearly define terms in order to be accessible across audiences.*
-->

<!--
## Model Card Authors

*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->

<!--
## Model Card Contact

*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
-->