Culture-and-Morality-Lab commited on
Commit
cee3a90
·
verified ·
1 Parent(s): 2ef3dc7

Upload 10 files

Browse files
README.md ADDED
@@ -0,0 +1,626 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - dense
7
+ - generated_from_trainer
8
+ - dataset_size:11180
9
+ - loss:CosineSimilarityLoss
10
+ widget:
11
+ - source_sentence: '" My cousin said he share hoes with his brothers. He said sharing
12
+ is caring and he love his brothers 😂"'
13
+ sentences:
14
+ - Clerical Error Led to Costa Rica's First Legal Gay Marriage
15
+ - 'UPDATE: The leasing office is going to tape notes on ALL tenants doors in our
16
+ building as to not single anyone out. However, we were not the first to complain.
17
+ Our maintenance guy also caught them smoking in the breezeway of our building
18
+ and told them to cut it out. Hopefully the note helps, if not, I don''t care enough
19
+ to make it a bigger issue and I''m only here for the next 8 months so smoke em
20
+ if you got em I guess.
21
+
22
+
23
+ Alright bare/bear with me here...
24
+
25
+
26
+ My husband and I live in a pretty decent apartment complex, we''re on the top
27
+ floor, and we are cool with most of our neighbors. We''re 90% sure about which
28
+ neighbor is smoking the devils lettuce, as it only really started when our next-door
29
+ neighbors new roommate moved in.
30
+
31
+
32
+ I have nothing against smoking, in any capacity, but it''s literally all our apartment
33
+ has smelled like the last few months. It''s starting to permeate our clothes and
34
+ furniture it''s so bad. My husband and I both work for the government (and are
35
+ drug tested) so this is not ideal. At first (before we realized what was actually
36
+ going on) we thought a skunk had sprayed near our apartment and was just coming
37
+ in through a window. Well, it''s winter now, no windows are open, and every day
38
+ we wake up to and come home to SKUNK. We have told the apartment complex about
39
+ this, and nothing has changed.
40
+
41
+
42
+ WIBTA if I left a note on this neighbors door? Something to the effect of "Hey
43
+ fellow apartment dwellers, pot is fun. However, consider blowing the smoke out
44
+ into a paper towel roll that has a dryer sheet at the end. Or vape it. Or eat
45
+ it. My entire apartment smells like skunky pot and I''m over it. Love, your neighbors"
46
+
47
+
48
+ Thoughts?
49
+
50
+
51
+ Edit: When we messaged the complex about it, we didn''t single out them or their
52
+ apartment, just said "someone in our building." They actually have since responded
53
+ saying that they''d "send a message to our building about it," so we''ll see what
54
+ happens! I''ll hold off on the note/talking to them in person for now, but if
55
+ it continues I''m definitely considering buying two smoke buddies and wrapping
56
+ them up nicely in Christmas paper with a little (not aggressive) card.
57
+
58
+
59
+ Edit 2: Enough people are asking so I wanted to clarify, **I do not live in a
60
+ state where it is legal.**'
61
+ - 'And just as Kamala Harris is dedicated to building an America for all peoples,
62
+ so too am I.
63
+
64
+
65
+ That is why the idea that freedom of religion somehow comes primary is so offensive
66
+ to the very idea of an America for all.
67
+
68
+
69
+ Some religions will cite their bigotry as having a basis in scripture. That will
70
+ apply to bigotry hitting the LDS community as well.
71
+
72
+
73
+ According to Christian doctrine, [the LDS church is fundamentally heretical](
74
+
75
+
76
+ Without discrimination protections, a catholic, muslim, jewish or christian business
77
+ can discriminate against the members of the LDS church without recourse.
78
+
79
+
80
+ That is why the protection of anti-discrimination measures must inevitably step
81
+ on the toes of religious bigotry. Religion is cited to justify bigotry. Just as
82
+ it was by the LDS church.'
83
+ - source_sentence: Thomson Buys Test Provider Capstar TORONTO (Reuters) - Electronic
84
+ publisher Thomson Corp. TOC.TO said on Monday it will buy test provider Capstar
85
+ from Educational Testing Service, the latest in a string of acquisitions designed
86
+ to flesh out its product offerings.
87
+ sentences:
88
+ - 'This is one of the most brilliant movies that I have seen in recent times. Goes
89
+ way above even any international movie of any repute. I am really surprised why
90
+ this has not received the recognition it deserved. Sonali Kulkarni winning the
91
+ National Award is perhaps the only consoling fact. Renuka Daftardar simply amazes
92
+ as she speaks volumes through her eyes. There are a few scenes that stand out:
93
+ When Gauri comes back from the city on Krishna''s wedding, she and Krishna meet
94
+ for the first time in many years. Krishna notices a change in Gauri, but not a
95
+ single line of dialogue is said. The entire gamut of emotions is conveyed through
96
+ subtle mannerisms and the eyes. There''s another towards the end when Krishna
97
+ pleads to Abhay Kulkarni to marry Gauri instead. If you are not moved by that
98
+ scene, you don''t have a heart.Watch this movie for sheer movie-making brilliance
99
+ and acting capabilities.'
100
+ - I put all my genius into my life I put only my talent into my works.
101
+ - '" we dont trust these niggas all these bitches "'
102
+ - source_sentence: 'Project 2025 happens. We already know what will happen. The only
103
+ thing we don''t know his how quick Republican plan to burn things down.
104
+
105
+
106
+ I think they will move quick. The pointy hoods are off. Everything is being said
107
+ out loud. They have their own website for fucks sake.
108
+
109
+
110
+ "We''re going to make the country so successful again, I''m not going to have
111
+ time for retribution, and remember this: our ultimate retribution is success."
112
+ - Trump'
113
+ sentences:
114
+ - 'I have a stepdaughter who i have never gotten along with . There is only a three
115
+ year age difference, so I''m not entirely blaming her, but i think my husband
116
+ did a shitty job raising her. She is just not nice, and not just to me, but she
117
+ is rude to wait staff, entitled, something of a mean girl. She has always made
118
+ snide comments about our age difference and me being a gold digger because I quit
119
+ my job. I told her once as an April Fool''s joke that her dad was leaving me everything
120
+ in the will, and she didn''t even respond to me, she just tattled to daddy.
121
+
122
+
123
+ She graduated in May with a PhD, and before she realized that she wasn''t going
124
+ to have a graduation because of what''s going on right now, she made a big deal
125
+ of announcing to everyone that I wasn''t invited. My husband told me to just deal
126
+ with it, because it is her graduation.
127
+
128
+
129
+ She came over the other night with her boyfriend of three years (she''s 26 and
130
+ he is 38, important later) and announced that they are getting married. Honestly
131
+ I was kind of pissed off. My husband is wealthy, but her boyfriend is legit rich,
132
+ so she is doing exactly what I did, marrying a man with a large age gap, and marrying
133
+ a man with a lot of money. I''m not saying that she doesn''t love him, but I think
134
+ she is a huge hypocrite to not acknowledge that just maybe i love her dad.
135
+
136
+
137
+ I told her that was kind of hypocritical and ironic and she just rolled her eyes,
138
+ but after the fact my husband got mad and said I was trying to steal her spotlight
139
+ and make her engagement about me, but he is always super defensive about her.'
140
+ - Paul Hamm Wins All-Around Event After falling on the landing of his vault, Paul
141
+ Hamm put on two of the most spectacular routines of his career to win Olympic
142
+ gold, the first ever by a U.S. man.
143
+ - "This morning I (31m) woke up and it was cold as hell in my apartment. I didn't\
144
+ \ realize the weather was getting a little cooler today and didn't adjust the\
145
+ \ thermostat accordingly, I just left it on full blast all night since its been\
146
+ \ hot lately. I woke up in the morning and had to take a shower before work and\
147
+ \ the whole apartment was freezing. \n\nI wasn't really in the mood to be cold\
148
+ \ as hell and shivering when I got out of the shower and had to do my hair and\
149
+ \ all that, so after I cleaned myself off, I just stood towards the back of the\
150
+ \ shower and turned the water to as hot as it goes and let it run for about ten\
151
+ \ or fifteen minutes. After a couple minutes I opened the bathroom door to start\
152
+ \ letting the heat and steam out so it wouldn't be so cold. I got out of the shower\
153
+ \ and it was steamy as hell and so nice and comfortable. The bathroom mirror was\
154
+ \ all cloudy from the steam.\n\nWhile I was doing my hair my fiance comes in and\
155
+ \ asks why the apartment is so foggy, so I told her what I did. I figured it was\
156
+ \ no big deal but she made it seem like I was a huge dick for it, saying it's\
157
+ \ stupid and childish and a waste of water, but I really feel like she was overreacting\
158
+ \ and just looking for an argument or something. What do you guys think? AITA?"
159
+ - source_sentence: Jake decides that the car would just need to be cleaned again in
160
+ the future so it would be a waste of time.
161
+ sentences:
162
+ - 'You guys think if I got a big face tattoo that said #metoo it would stop the
163
+ comments drunk men make at me while I do my job????'
164
+ - "please stop your oppressive editing\n TAB+TAB\nHi. I see that you're not as stupid\
165
+ \ as the other sysop, trying to block ip addresses and stuff. But all my warnings\
166
+ \ apply to you too. And if you're still wondering what this is about or who the\
167
+ \ fuck I am, you know what to do. Just do a whois on my ip117.201.39.75"
168
+ - Mindy cuts off ties with John and finds other friends who won't abuse her.
169
+ - source_sentence: Besides that which the men brought him that were over the tributes,
170
+ and the merchants, and they that sold by retail, and all the kings of Arabia,
171
+ and the governors of the country.
172
+ sentences:
173
+ - I don't look for much to come out of government ownership as long as we have Democrats
174
+ and Republicans.
175
+ - If this needs a federal mandate and 100% global consensus, than leaders like Macron
176
+ should let us renegotiate. As it stands right now, this agreement is 100% toothless.
177
+ There are no penalties for not following through with it.
178
+ - 'Barclays to Buy CIBC Credit Card Unit for \$293 Mln (Update1) Aug. 18 (Bloomberg)
179
+ -- Barclays Plc, the UK #39;s third- biggest bank, agreed to buy Juniper Financial
180
+ Corp. from Canadian Imperial Bank of Commerce for \$293 million in cash to help
181
+ expand its credit card business in North America.'
182
+ pipeline_tag: sentence-similarity
183
+ library_name: sentence-transformers
184
+ metrics:
185
+ - pearson_cosine
186
+ - spearman_cosine
187
+ model-index:
188
+ - name: SentenceTransformer
189
+ results:
190
+ - task:
191
+ type: semantic-similarity
192
+ name: Semantic Similarity
193
+ dataset:
194
+ name: similarity
195
+ type: similarity
196
+ metrics:
197
+ - type: pearson_cosine
198
+ value: 0.3879415946781869
199
+ name: Pearson Cosine
200
+ - type: spearman_cosine
201
+ value: 0.40484905925030346
202
+ name: Spearman Cosine
203
+ ---
204
+
205
+ # SentenceTransformer
206
+
207
+ This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
208
+
209
+ ## Model Details
210
+
211
+ ### Model Description
212
+ - **Model Type:** Sentence Transformer
213
+ <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
214
+ - **Maximum Sequence Length:** 512 tokens
215
+ - **Output Dimensionality:** 1024 dimensions
216
+ - **Similarity Function:** Cosine Similarity
217
+ <!-- - **Training Dataset:** Unknown -->
218
+ <!-- - **Language:** Unknown -->
219
+ <!-- - **License:** Unknown -->
220
+
221
+ ### Model Sources
222
+
223
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
224
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
225
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
226
+
227
+ ### Full Model Architecture
228
+
229
+ ```
230
+ SentenceTransformer(
231
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'BertModel'})
232
+ (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
233
+ (2): Normalize()
234
+ )
235
+ ```
236
+
237
+ ## Usage
238
+
239
+ ### Direct Usage (Sentence Transformers)
240
+
241
+ First install the Sentence Transformers library:
242
+
243
+ ```bash
244
+ pip install -U sentence-transformers
245
+ ```
246
+
247
+ Then you can load this model and run inference.
248
+ ```python
249
+ from sentence_transformers import SentenceTransformer
250
+
251
+ # Download from the 🤗 Hub
252
+ model = SentenceTransformer("sentence_transformers_model_id")
253
+ # Run inference
254
+ sentences = [
255
+ 'Besides that which the men brought him that were over the tributes, and the merchants, and they that sold by retail, and all the kings of Arabia, and the governors of the country.',
256
+ 'If this needs a federal mandate and 100% global consensus, than leaders like Macron should let us renegotiate. As it stands right now, this agreement is 100% toothless. There are no penalties for not following through with it.',
257
+ "I don't look for much to come out of government ownership as long as we have Democrats and Republicans.",
258
+ ]
259
+ embeddings = model.encode(sentences)
260
+ print(embeddings.shape)
261
+ # [3, 1024]
262
+
263
+ # Get the similarity scores for the embeddings
264
+ similarities = model.similarity(embeddings, embeddings)
265
+ print(similarities)
266
+ # tensor([[1.0000, 0.5648, 0.5502],
267
+ # [0.5648, 1.0000, 0.7965],
268
+ # [0.5502, 0.7965, 1.0000]])
269
+ ```
270
+
271
+ <!--
272
+ ### Direct Usage (Transformers)
273
+
274
+ <details><summary>Click to see the direct usage in Transformers</summary>
275
+
276
+ </details>
277
+ -->
278
+
279
+ <!--
280
+ ### Downstream Usage (Sentence Transformers)
281
+
282
+ You can finetune this model on your own dataset.
283
+
284
+ <details><summary>Click to expand</summary>
285
+
286
+ </details>
287
+ -->
288
+
289
+ <!--
290
+ ### Out-of-Scope Use
291
+
292
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
293
+ -->
294
+
295
+ ## Evaluation
296
+
297
+ ### Metrics
298
+
299
+ #### Semantic Similarity
300
+
301
+ * Dataset: `similarity`
302
+ * Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
303
+
304
+ | Metric | Value |
305
+ |:--------------------|:-----------|
306
+ | pearson_cosine | 0.3879 |
307
+ | **spearman_cosine** | **0.4048** |
308
+
309
+ <!--
310
+ ## Bias, Risks and Limitations
311
+
312
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
313
+ -->
314
+
315
+ <!--
316
+ ### Recommendations
317
+
318
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
319
+ -->
320
+
321
+ ## Training Details
322
+
323
+ ### Training Dataset
324
+
325
+ #### Unnamed Dataset
326
+
327
+ * Size: 11,180 training samples
328
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
329
+ * Approximate statistics based on the first 1000 samples:
330
+ | | sentence_0 | sentence_1 | label |
331
+ |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:---------------------------------------------------------------|
332
+ | type | string | string | float |
333
+ | details | <ul><li>min: 6 tokens</li><li>mean: 104.26 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 118.5 tokens</li><li>max: 512 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.53</li><li>max: 1.0</li></ul> |
334
+ * Samples:
335
+ | sentence_0 | sentence_1 | label |
336
+ |:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------|
337
+ | <code>He worked at Rothschild as an investment banker. Great. Am I supposed to be alarmed that France elected a technocrat who has worked in the private banking sector? <br><br>I also don't give a shit about what macron does in his personal life. Clearly the French people don't either.</code> | <code>Chad runs over the raccoon since it's been bothering him anyway.</code> | <code>0.3535533905932737</code> |
338
+ | <code>Amazing effects for a movie of this time. A primer of the uselessness of war and how war becomes a nurturer of itself.A wonderful thing about this movie is it is now public domain and available at archive.org. No charge, no sign up necessary. Watch it in one sitting and you will be propelled.I plan to share this flick with as many people as possible as I had never heard of it before and I am a hard core sci fi fan.I would like to see how others react to this movie.Watch it.Rate it.Tell us what you think.</code> | <code>First off, I must say that I made the mistake of watching the Election films out of sequence. I say unfortunately, because after seeing Election 2 first, Election seems a bit of a disappointment. Both films are gangster epics that are similar in form. And while Election is an enjoyable piece of cinema... it's just not nearly as good as it's sequel.In the first Election installment, we are shown the two competitors for Chairman; Big D and Lok. After a few scenes of discussion amongst the "Uncle's" as to who should have the Chairman title, they (almost unanimously) decide That Lok (Simon Yam) will helm the Triads. Suffice to say this doesn't go over very well with competitor Big D (Tony Leung Ka Fai) and in a bid to influence the takeover, Big D kidnaps two of the uncles in order to sway the election board to his side. This has disastrous results and heads the triads into an all out war. Lok is determined to become Chairman but won't become official until he can recover the "Dragon Head ...</code> | <code>0.7071067811865475</code> |
339
+ | <code>MY SINCERE APOLOGIES 2U WHO I'VE OFFENDED WITH ALLEGATIONS OF COMPLACENT COWARDS & ASSHOLES FOR CLIMATE CHANGE INDIFFERENCE!</code> | <code>yeah man fucking disgusting. as if we didn't waste enough time at work</code> | <code>1.0</code> |
340
+ * Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
341
+ ```json
342
+ {
343
+ "loss_fct": "torch.nn.modules.loss.MSELoss"
344
+ }
345
+ ```
346
+
347
+ ### Training Hyperparameters
348
+ #### Non-Default Hyperparameters
349
+
350
+ - `eval_strategy`: steps
351
+ - `per_device_train_batch_size`: 32
352
+ - `per_device_eval_batch_size`: 32
353
+ - `fp16`: True
354
+ - `multi_dataset_batch_sampler`: round_robin
355
+
356
+ #### All Hyperparameters
357
+ <details><summary>Click to expand</summary>
358
+
359
+ - `overwrite_output_dir`: False
360
+ - `do_predict`: False
361
+ - `eval_strategy`: steps
362
+ - `prediction_loss_only`: True
363
+ - `per_device_train_batch_size`: 32
364
+ - `per_device_eval_batch_size`: 32
365
+ - `per_gpu_train_batch_size`: None
366
+ - `per_gpu_eval_batch_size`: None
367
+ - `gradient_accumulation_steps`: 1
368
+ - `eval_accumulation_steps`: None
369
+ - `torch_empty_cache_steps`: None
370
+ - `learning_rate`: 5e-05
371
+ - `weight_decay`: 0.0
372
+ - `adam_beta1`: 0.9
373
+ - `adam_beta2`: 0.999
374
+ - `adam_epsilon`: 1e-08
375
+ - `max_grad_norm`: 1
376
+ - `num_train_epochs`: 3
377
+ - `max_steps`: -1
378
+ - `lr_scheduler_type`: linear
379
+ - `lr_scheduler_kwargs`: {}
380
+ - `warmup_ratio`: 0.0
381
+ - `warmup_steps`: 0
382
+ - `log_level`: passive
383
+ - `log_level_replica`: warning
384
+ - `log_on_each_node`: True
385
+ - `logging_nan_inf_filter`: True
386
+ - `save_safetensors`: True
387
+ - `save_on_each_node`: False
388
+ - `save_only_model`: False
389
+ - `restore_callback_states_from_checkpoint`: False
390
+ - `no_cuda`: False
391
+ - `use_cpu`: False
392
+ - `use_mps_device`: False
393
+ - `seed`: 42
394
+ - `data_seed`: None
395
+ - `jit_mode_eval`: False
396
+ - `use_ipex`: False
397
+ - `bf16`: False
398
+ - `fp16`: True
399
+ - `fp16_opt_level`: O1
400
+ - `half_precision_backend`: auto
401
+ - `bf16_full_eval`: False
402
+ - `fp16_full_eval`: False
403
+ - `tf32`: None
404
+ - `local_rank`: 0
405
+ - `ddp_backend`: None
406
+ - `tpu_num_cores`: None
407
+ - `tpu_metrics_debug`: False
408
+ - `debug`: []
409
+ - `dataloader_drop_last`: False
410
+ - `dataloader_num_workers`: 0
411
+ - `dataloader_prefetch_factor`: None
412
+ - `past_index`: -1
413
+ - `disable_tqdm`: False
414
+ - `remove_unused_columns`: True
415
+ - `label_names`: None
416
+ - `load_best_model_at_end`: False
417
+ - `ignore_data_skip`: False
418
+ - `fsdp`: []
419
+ - `fsdp_min_num_params`: 0
420
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
421
+ - `fsdp_transformer_layer_cls_to_wrap`: None
422
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
423
+ - `deepspeed`: None
424
+ - `label_smoothing_factor`: 0.0
425
+ - `optim`: adamw_torch
426
+ - `optim_args`: None
427
+ - `adafactor`: False
428
+ - `group_by_length`: False
429
+ - `length_column_name`: length
430
+ - `ddp_find_unused_parameters`: None
431
+ - `ddp_bucket_cap_mb`: None
432
+ - `ddp_broadcast_buffers`: False
433
+ - `dataloader_pin_memory`: True
434
+ - `dataloader_persistent_workers`: False
435
+ - `skip_memory_metrics`: True
436
+ - `use_legacy_prediction_loop`: False
437
+ - `push_to_hub`: False
438
+ - `resume_from_checkpoint`: None
439
+ - `hub_model_id`: None
440
+ - `hub_strategy`: every_save
441
+ - `hub_private_repo`: None
442
+ - `hub_always_push`: False
443
+ - `gradient_checkpointing`: False
444
+ - `gradient_checkpointing_kwargs`: None
445
+ - `include_inputs_for_metrics`: False
446
+ - `include_for_metrics`: []
447
+ - `eval_do_concat_batches`: True
448
+ - `fp16_backend`: auto
449
+ - `push_to_hub_model_id`: None
450
+ - `push_to_hub_organization`: None
451
+ - `mp_parameters`:
452
+ - `auto_find_batch_size`: False
453
+ - `full_determinism`: False
454
+ - `torchdynamo`: None
455
+ - `ray_scope`: last
456
+ - `ddp_timeout`: 1800
457
+ - `torch_compile`: False
458
+ - `torch_compile_backend`: None
459
+ - `torch_compile_mode`: None
460
+ - `dispatch_batches`: None
461
+ - `split_batches`: None
462
+ - `include_tokens_per_second`: False
463
+ - `include_num_input_tokens_seen`: False
464
+ - `neftune_noise_alpha`: None
465
+ - `optim_target_modules`: None
466
+ - `batch_eval_metrics`: False
467
+ - `eval_on_start`: False
468
+ - `use_liger_kernel`: False
469
+ - `eval_use_gather_object`: False
470
+ - `average_tokens_across_devices`: False
471
+ - `prompts`: None
472
+ - `batch_sampler`: batch_sampler
473
+ - `multi_dataset_batch_sampler`: round_robin
474
+ - `router_mapping`: {}
475
+ - `learning_rate_mapping`: {}
476
+
477
+ </details>
478
+
479
+ ### Training Logs
480
+ | Epoch | Step | Training Loss | similarity_spearman_cosine |
481
+ |:------:|:----:|:-------------:|:--------------------------:|
482
+ | 0.0286 | 10 | - | 0.2006 |
483
+ | 0.0571 | 20 | - | 0.2012 |
484
+ | 0.0857 | 30 | - | 0.2023 |
485
+ | 0.1143 | 40 | - | 0.2036 |
486
+ | 0.1429 | 50 | - | 0.2054 |
487
+ | 0.1714 | 60 | - | 0.2081 |
488
+ | 0.2 | 70 | - | 0.2098 |
489
+ | 0.2286 | 80 | - | 0.2115 |
490
+ | 0.2571 | 90 | - | 0.2128 |
491
+ | 0.2857 | 100 | - | 0.2149 |
492
+ | 0.3143 | 110 | - | 0.2177 |
493
+ | 0.3429 | 120 | - | 0.2207 |
494
+ | 0.3714 | 130 | - | 0.2243 |
495
+ | 0.4 | 140 | - | 0.2278 |
496
+ | 0.4286 | 150 | - | 0.2310 |
497
+ | 0.4571 | 160 | - | 0.2332 |
498
+ | 0.4857 | 170 | - | 0.2350 |
499
+ | 0.5143 | 180 | - | 0.2361 |
500
+ | 0.5429 | 190 | - | 0.2360 |
501
+ | 0.5714 | 200 | - | 0.2369 |
502
+ | 0.6 | 210 | - | 0.2423 |
503
+ | 0.6286 | 220 | - | 0.2533 |
504
+ | 0.6571 | 230 | - | 0.2691 |
505
+ | 0.6857 | 240 | - | 0.2808 |
506
+ | 0.7143 | 250 | - | 0.2889 |
507
+ | 0.7429 | 260 | - | 0.2960 |
508
+ | 0.7714 | 270 | - | 0.2939 |
509
+ | 0.8 | 280 | - | 0.3007 |
510
+ | 0.8286 | 290 | - | 0.3010 |
511
+ | 0.8571 | 300 | - | 0.3016 |
512
+ | 0.8857 | 310 | - | 0.3035 |
513
+ | 0.9143 | 320 | - | 0.3078 |
514
+ | 0.9429 | 330 | - | 0.3138 |
515
+ | 0.9714 | 340 | - | 0.3206 |
516
+ | 1.0 | 350 | - | 0.3234 |
517
+ | 1.0286 | 360 | - | 0.3299 |
518
+ | 1.0571 | 370 | - | 0.3367 |
519
+ | 1.0857 | 380 | - | 0.3267 |
520
+ | 1.1143 | 390 | - | 0.3307 |
521
+ | 1.1429 | 400 | - | 0.3359 |
522
+ | 1.1714 | 410 | - | 0.3417 |
523
+ | 1.2 | 420 | - | 0.3504 |
524
+ | 1.2286 | 430 | - | 0.3324 |
525
+ | 1.2571 | 440 | - | 0.3365 |
526
+ | 1.2857 | 450 | - | 0.3580 |
527
+ | 1.3143 | 460 | - | 0.3622 |
528
+ | 1.3429 | 470 | - | 0.3073 |
529
+ | 1.3714 | 480 | - | 0.3596 |
530
+ | 1.4 | 490 | - | 0.3473 |
531
+ | 1.4286 | 500 | 0.1278 | 0.3573 |
532
+ | 1.4571 | 510 | - | 0.3539 |
533
+ | 1.4857 | 520 | - | 0.3355 |
534
+ | 1.5143 | 530 | - | 0.3299 |
535
+ | 1.5429 | 540 | - | 0.3559 |
536
+ | 1.5714 | 550 | - | 0.3285 |
537
+ | 1.6 | 560 | - | 0.3435 |
538
+ | 1.6286 | 570 | - | 0.3654 |
539
+ | 1.6571 | 580 | - | 0.3824 |
540
+ | 1.6857 | 590 | - | 0.3426 |
541
+ | 1.7143 | 600 | - | 0.3413 |
542
+ | 1.7429 | 610 | - | 0.3395 |
543
+ | 1.7714 | 620 | - | 0.3492 |
544
+ | 1.8 | 630 | - | 0.3664 |
545
+ | 1.8286 | 640 | - | 0.3634 |
546
+ | 1.8571 | 650 | - | 0.3392 |
547
+ | 1.8857 | 660 | - | 0.3686 |
548
+ | 1.9143 | 670 | - | 0.3722 |
549
+ | 1.9429 | 680 | - | 0.3557 |
550
+ | 1.9714 | 690 | - | 0.3896 |
551
+ | 2.0 | 700 | - | 0.3908 |
552
+ | 2.0286 | 710 | - | 0.3859 |
553
+ | 2.0571 | 720 | - | 0.3536 |
554
+ | 2.0857 | 730 | - | 0.3606 |
555
+ | 2.1143 | 740 | - | 0.3638 |
556
+ | 2.1429 | 750 | - | 0.3713 |
557
+ | 2.1714 | 760 | - | 0.3704 |
558
+ | 2.2 | 770 | - | 0.3441 |
559
+ | 2.2286 | 780 | - | 0.3435 |
560
+ | 2.2571 | 790 | - | 0.3668 |
561
+ | 2.2857 | 800 | - | 0.3735 |
562
+ | 2.3143 | 810 | - | 0.3373 |
563
+ | 2.3429 | 820 | - | 0.3474 |
564
+ | 2.3714 | 830 | - | 0.3560 |
565
+ | 2.4 | 840 | - | 0.3028 |
566
+ | 2.4286 | 850 | - | 0.3485 |
567
+ | 2.4571 | 860 | - | 0.3604 |
568
+ | 2.4857 | 870 | - | 0.3769 |
569
+ | 2.5143 | 880 | - | 0.3600 |
570
+ | 2.5429 | 890 | - | 0.3916 |
571
+ | 2.5714 | 900 | - | 0.3957 |
572
+ | 2.6 | 910 | - | 0.3797 |
573
+ | 2.6286 | 920 | - | 0.3875 |
574
+ | 2.6571 | 930 | - | 0.3978 |
575
+ | 2.6857 | 940 | - | 0.3951 |
576
+ | 2.7143 | 950 | - | 0.3831 |
577
+ | 2.7429 | 960 | - | 0.3912 |
578
+ | 2.7714 | 970 | - | 0.3800 |
579
+ | 2.8 | 980 | - | 0.3955 |
580
+ | 2.8286 | 990 | - | 0.3976 |
581
+ | 2.8571 | 1000 | 0.1036 | 0.4048 |
582
+
583
+
584
+ ### Framework Versions
585
+ - Python: 3.11.9
586
+ - Sentence Transformers: 5.1.0
587
+ - Transformers: 4.49.0
588
+ - PyTorch: 2.8.0+cu128
589
+ - Accelerate: 1.10.0
590
+ - Datasets: 2.14.4
591
+ - Tokenizers: 0.21.0
592
+
593
+ ## Citation
594
+
595
+ ### BibTeX
596
+
597
+ #### Sentence Transformers
598
+ ```bibtex
599
+ @inproceedings{reimers-2019-sentence-bert,
600
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
601
+ author = "Reimers, Nils and Gurevych, Iryna",
602
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
603
+ month = "11",
604
+ year = "2019",
605
+ publisher = "Association for Computational Linguistics",
606
+ url = "https://arxiv.org/abs/1908.10084",
607
+ }
608
+ ```
609
+
610
+ <!--
611
+ ## Glossary
612
+
613
+ *Clearly define terms in order to be accessible across audiences.*
614
+ -->
615
+
616
+ <!--
617
+ ## Model Card Authors
618
+
619
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
620
+ -->
621
+
622
+ <!--
623
+ ## Model Card Contact
624
+
625
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
626
+ -->
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./models/pretrained/gte-large",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 1024,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 4096,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.49.0",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 30522
26
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "SentenceTransformer",
3
+ "__version__": {
4
+ "sentence_transformers": "5.1.0",
5
+ "transformers": "4.49.0",
6
+ "pytorch": "2.8.0+cu128"
7
+ },
8
+ "prompts": {
9
+ "query": "",
10
+ "document": ""
11
+ },
12
+ "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
14
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a846270ef79ec7517531a94c9d816903d7b59602b410352185a48014ca8d0128
3
+ size 1340612432
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "max_length": 128,
50
+ "model_max_length": 512,
51
+ "pad_to_multiple_of": null,
52
+ "pad_token": "[PAD]",
53
+ "pad_token_type_id": 0,
54
+ "padding_side": "right",
55
+ "sep_token": "[SEP]",
56
+ "stride": 0,
57
+ "strip_accents": null,
58
+ "tokenize_chinese_chars": true,
59
+ "tokenizer_class": "BertTokenizer",
60
+ "truncation_side": "right",
61
+ "truncation_strategy": "longest_first",
62
+ "unk_token": "[UNK]"
63
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff