radoslavralev commited on
Commit
61bbeb2
·
verified ·
1 Parent(s): 2039136

Training in progress, step 5000

Browse files
1_Pooling/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "word_embedding_dimension": 384,
3
- "pooling_mode_cls_token": false,
4
- "pooling_mode_mean_tokens": true,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
 
1
  {
2
+ "word_embedding_dimension": 512,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
Information-Retrieval_evaluation_BeIR-touche2020-subset-test_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.7551020408163265,0.8979591836734694,0.9387755102040817,1.0,0.7551020408163265,0.01681883497544576,0.6802721088435373,0.045273391320759554,0.6489795918367346,0.07198855467813525,0.563265306122449,0.12417651370073833,0.8328474246841594,0.6023501966395867,0.2434385293084787
Information-Retrieval_evaluation_NanoArguAna_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.18,0.5,0.66,0.74,0.18,0.18,0.16666666666666663,0.5,0.13200000000000003,0.66,0.07400000000000001,0.74,0.3599682539682539,0.45218312003145433,0.3658170202780539
Information-Retrieval_evaluation_NanoClimateFEVER_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.1,0.32,0.44,0.6,0.1,0.04333333333333333,0.11333333333333333,0.154,0.092,0.214,0.066,0.2723333333333333,0.23579365079365078,0.18832347198247595,0.13278630044723194
Information-Retrieval_evaluation_NanoDBPedia_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.58,0.82,0.86,0.88,0.58,0.05774032197349101,0.5266666666666666,0.13525154291594316,0.444,0.17429307751101658,0.36199999999999993,0.24557535689700863,0.7,0.46037876007701023,0.32047705412103555
Information-Retrieval_evaluation_NanoFEVER_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.56,0.8,0.88,0.88,0.56,0.5266666666666666,0.28,0.7533333333333333,0.184,0.8333333333333333,0.092,0.8333333333333333,0.6789999999999999,0.7016221865098926,0.6483203115492273
Information-Retrieval_evaluation_NanoFiQA2018_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.24,0.38,0.42,0.52,0.24,0.13307936507936508,0.15333333333333332,0.21543650793650795,0.11600000000000002,0.2601031746031746,0.07600000000000001,0.3241587301587302,0.3178809523809524,0.2635447070179544,0.22615568818713977
Information-Retrieval_evaluation_NanoHotpotQA_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.5,0.58,0.62,0.68,0.5,0.25,0.2333333333333333,0.35,0.16399999999999998,0.41,0.094,0.47,0.5529999999999999,0.43627646197603637,0.37828612061221206
Information-Retrieval_evaluation_NanoMSMARCO_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.26,0.52,0.6,0.62,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.062,0.62,0.40519047619047627,0.45904886208148177,0.4260102142025637
Information-Retrieval_evaluation_NanoNFCorpus_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.32,0.5,0.52,0.64,0.32,0.01260685895239504,0.31333333333333335,0.03463231741620742,0.27599999999999997,0.06850002262294216,0.24399999999999994,0.10851262864104039,0.4212380952380952,0.2703690747449406,0.10029196368651581
Information-Retrieval_evaluation_NanoNQ_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.32,0.5,0.6,0.62,0.32,0.3,0.1733333333333333,0.47,0.128,0.58,0.066,0.6,0.4272222222222222,0.4619884812398348,0.42411983365963474
Information-Retrieval_evaluation_NanoQuoraRetrieval_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.94,1.0,1.0,1.0,0.94,0.8273333333333334,0.4133333333333333,0.9653333333333333,0.25999999999999995,0.9793333333333334,0.13799999999999998,0.9966666666666666,0.9666666666666667,0.9697624312418531,0.9551233766233765
Information-Retrieval_evaluation_NanoSCIDOCS_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.42,0.6,0.72,0.82,0.42,0.08866666666666667,0.33333333333333326,0.20866666666666664,0.272,0.2806666666666667,0.17999999999999997,0.3696666666666666,0.540047619047619,0.36082794471047336,0.2862806075456748
Information-Retrieval_evaluation_NanoSciFact_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.38,0.46,0.5,0.58,0.38,0.345,0.15999999999999998,0.43,0.10800000000000003,0.475,0.068,0.58,0.44026984126984126,0.46384622999765257,0.43257979600699104
Information-Retrieval_evaluation_NanoTouche2020_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.46938775510204084,0.8367346938775511,0.9387755102040817,1.0,0.46938775510204084,0.032657982947973084,0.44897959183673464,0.09621881460341672,0.42040816326530606,0.1425551052100505,0.3346938775510204,0.22061476067159091,0.6573129251700679,0.3807140713282222,0.2698119698398041
NanoBEIR_evaluation_mean_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accuracy@10,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-Precision@10,cosine-Recall@10,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
+ -1,-1,0.4053375196232339,0.601287284144427,0.6737519623233909,0.7369230769230769,0.4053375196232339,0.23516034838101724,0.26838304552590264,0.37175942432349296,0.20895447409733128,0.43675267025234743,0.1428226059654631,0.4908354981821823,0.5156608233036803,0.45145275407225244,0.3820046351353431
README.md CHANGED
@@ -5,231 +5,51 @@ tags:
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
- - dataset_size:713743
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
- - source_sentence: 'Abraham Lincoln: Why is the Gettysburg Address so memorable?'
13
  sentences:
14
- - 'Abraham Lincoln: Why is the Gettysburg Address so memorable?'
15
- - What does the Gettysburg Address really mean?
16
- - What is eatalo.com?
17
- - source_sentence: Has the influence of Ancient Carthage in science, math, and society
18
- been underestimated?
19
  sentences:
20
- - How does one earn money online without an investment from home?
21
- - Has the influence of Ancient Carthage in science, math, and society been underestimated?
22
- - Has the influence of the Ancient Etruscans in science and math been underestimated?
23
- - source_sentence: Is there any app that shares charging to others like share it how
24
- we transfer files?
25
  sentences:
26
- - How do you think of Chinese claims that the present Private Arbitration is illegal,
27
- its verdict violates the UNCLOS and is illegal?
28
- - Is there any app that shares charging to others like share it how we transfer
29
- files?
30
- - Are there any platforms that provides end-to-end encryption for file transfer/
31
- sharing?
32
- - source_sentence: Why AAP’s MLA Dinesh Mohaniya has been arrested?
33
  sentences:
34
- - What are your views on the latest sex scandal by AAP MLA Sandeep Kumar?
35
- - What is a dc current? What are some examples?
36
- - Why AAP’s MLA Dinesh Mohaniya has been arrested?
37
- - source_sentence: What is the difference between economic growth and economic development?
38
  sentences:
39
- - How cold can the Gobi Desert get, and how do its average temperatures compare
40
- to the ones in the Simpson Desert?
41
- - the difference between economic growth and economic development is What?
42
- - What is the difference between economic growth and economic development?
43
  pipeline_tag: sentence-similarity
44
  library_name: sentence-transformers
45
- metrics:
46
- - cosine_accuracy@1
47
- - cosine_accuracy@3
48
- - cosine_accuracy@5
49
- - cosine_accuracy@10
50
- - cosine_precision@1
51
- - cosine_precision@3
52
- - cosine_precision@5
53
- - cosine_precision@10
54
- - cosine_recall@1
55
- - cosine_recall@3
56
- - cosine_recall@5
57
- - cosine_recall@10
58
- - cosine_ndcg@10
59
- - cosine_mrr@10
60
- - cosine_map@100
61
- model-index:
62
- - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
63
- results:
64
- - task:
65
- type: information-retrieval
66
- name: Information Retrieval
67
- dataset:
68
- name: NanoMSMARCO
69
- type: NanoMSMARCO
70
- metrics:
71
- - type: cosine_accuracy@1
72
- value: 0.26
73
- name: Cosine Accuracy@1
74
- - type: cosine_accuracy@3
75
- value: 0.52
76
- name: Cosine Accuracy@3
77
- - type: cosine_accuracy@5
78
- value: 0.6
79
- name: Cosine Accuracy@5
80
- - type: cosine_accuracy@10
81
- value: 0.62
82
- name: Cosine Accuracy@10
83
- - type: cosine_precision@1
84
- value: 0.26
85
- name: Cosine Precision@1
86
- - type: cosine_precision@3
87
- value: 0.1733333333333333
88
- name: Cosine Precision@3
89
- - type: cosine_precision@5
90
- value: 0.12
91
- name: Cosine Precision@5
92
- - type: cosine_precision@10
93
- value: 0.062
94
- name: Cosine Precision@10
95
- - type: cosine_recall@1
96
- value: 0.26
97
- name: Cosine Recall@1
98
- - type: cosine_recall@3
99
- value: 0.52
100
- name: Cosine Recall@3
101
- - type: cosine_recall@5
102
- value: 0.6
103
- name: Cosine Recall@5
104
- - type: cosine_recall@10
105
- value: 0.62
106
- name: Cosine Recall@10
107
- - type: cosine_ndcg@10
108
- value: 0.45904886208148177
109
- name: Cosine Ndcg@10
110
- - type: cosine_mrr@10
111
- value: 0.40519047619047627
112
- name: Cosine Mrr@10
113
- - type: cosine_map@100
114
- value: 0.4260102142025637
115
- name: Cosine Map@100
116
- - task:
117
- type: information-retrieval
118
- name: Information Retrieval
119
- dataset:
120
- name: NanoNQ
121
- type: NanoNQ
122
- metrics:
123
- - type: cosine_accuracy@1
124
- value: 0.32
125
- name: Cosine Accuracy@1
126
- - type: cosine_accuracy@3
127
- value: 0.5
128
- name: Cosine Accuracy@3
129
- - type: cosine_accuracy@5
130
- value: 0.6
131
- name: Cosine Accuracy@5
132
- - type: cosine_accuracy@10
133
- value: 0.62
134
- name: Cosine Accuracy@10
135
- - type: cosine_precision@1
136
- value: 0.32
137
- name: Cosine Precision@1
138
- - type: cosine_precision@3
139
- value: 0.1733333333333333
140
- name: Cosine Precision@3
141
- - type: cosine_precision@5
142
- value: 0.128
143
- name: Cosine Precision@5
144
- - type: cosine_precision@10
145
- value: 0.066
146
- name: Cosine Precision@10
147
- - type: cosine_recall@1
148
- value: 0.3
149
- name: Cosine Recall@1
150
- - type: cosine_recall@3
151
- value: 0.47
152
- name: Cosine Recall@3
153
- - type: cosine_recall@5
154
- value: 0.58
155
- name: Cosine Recall@5
156
- - type: cosine_recall@10
157
- value: 0.6
158
- name: Cosine Recall@10
159
- - type: cosine_ndcg@10
160
- value: 0.4619884812398348
161
- name: Cosine Ndcg@10
162
- - type: cosine_mrr@10
163
- value: 0.4272222222222222
164
- name: Cosine Mrr@10
165
- - type: cosine_map@100
166
- value: 0.42411471333193373
167
- name: Cosine Map@100
168
- - task:
169
- type: nano-beir
170
- name: Nano BEIR
171
- dataset:
172
- name: NanoBEIR mean
173
- type: NanoBEIR_mean
174
- metrics:
175
- - type: cosine_accuracy@1
176
- value: 0.29000000000000004
177
- name: Cosine Accuracy@1
178
- - type: cosine_accuracy@3
179
- value: 0.51
180
- name: Cosine Accuracy@3
181
- - type: cosine_accuracy@5
182
- value: 0.6
183
- name: Cosine Accuracy@5
184
- - type: cosine_accuracy@10
185
- value: 0.62
186
- name: Cosine Accuracy@10
187
- - type: cosine_precision@1
188
- value: 0.29000000000000004
189
- name: Cosine Precision@1
190
- - type: cosine_precision@3
191
- value: 0.1733333333333333
192
- name: Cosine Precision@3
193
- - type: cosine_precision@5
194
- value: 0.124
195
- name: Cosine Precision@5
196
- - type: cosine_precision@10
197
- value: 0.064
198
- name: Cosine Precision@10
199
- - type: cosine_recall@1
200
- value: 0.28
201
- name: Cosine Recall@1
202
- - type: cosine_recall@3
203
- value: 0.495
204
- name: Cosine Recall@3
205
- - type: cosine_recall@5
206
- value: 0.59
207
- name: Cosine Recall@5
208
- - type: cosine_recall@10
209
- value: 0.61
210
- name: Cosine Recall@10
211
- - type: cosine_ndcg@10
212
- value: 0.4605186716606583
213
- name: Cosine Ndcg@10
214
- - type: cosine_mrr@10
215
- value: 0.41620634920634925
216
- name: Cosine Mrr@10
217
- - type: cosine_map@100
218
- value: 0.4250624637672487
219
- name: Cosine Map@100
220
  ---
221
 
222
- # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
223
 
224
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
225
 
226
  ## Model Details
227
 
228
  ### Model Description
229
  - **Model Type:** Sentence Transformer
230
- - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
231
  - **Maximum Sequence Length:** 128 tokens
232
- - **Output Dimensionality:** 384 dimensions
233
  - **Similarity Function:** Cosine Similarity
234
  <!-- - **Training Dataset:** Unknown -->
235
  <!-- - **Language:** Unknown -->
@@ -246,8 +66,7 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [s
246
  ```
247
  SentenceTransformer(
248
  (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'BertModel'})
249
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
250
- (2): Normalize()
251
  )
252
  ```
253
 
@@ -266,23 +85,23 @@ Then you can load this model and run inference.
266
  from sentence_transformers import SentenceTransformer
267
 
268
  # Download from the 🤗 Hub
269
- model = SentenceTransformer("redis/model-b-structured")
270
  # Run inference
271
  sentences = [
272
- 'What is the difference between economic growth and economic development?',
273
- 'What is the difference between economic growth and economic development?',
274
- 'the difference between economic growth and economic development is What?',
275
  ]
276
  embeddings = model.encode(sentences)
277
  print(embeddings.shape)
278
- # [3, 384]
279
 
280
  # Get the similarity scores for the embeddings
281
  similarities = model.similarity(embeddings, embeddings)
282
  print(similarities)
283
- # tensor([[ 1.0000, 1.0000, -0.0482],
284
- # [ 1.0000, 1.0000, -0.0482],
285
- # [-0.0482, -0.0482, 1.0000]])
286
  ```
287
 
288
  <!--
@@ -309,65 +128,6 @@ You can finetune this model on your own dataset.
309
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
310
  -->
311
 
312
- ## Evaluation
313
-
314
- ### Metrics
315
-
316
- #### Information Retrieval
317
-
318
- * Datasets: `NanoMSMARCO` and `NanoNQ`
319
- * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
320
-
321
- | Metric | NanoMSMARCO | NanoNQ |
322
- |:--------------------|:------------|:----------|
323
- | cosine_accuracy@1 | 0.26 | 0.32 |
324
- | cosine_accuracy@3 | 0.52 | 0.5 |
325
- | cosine_accuracy@5 | 0.6 | 0.6 |
326
- | cosine_accuracy@10 | 0.62 | 0.62 |
327
- | cosine_precision@1 | 0.26 | 0.32 |
328
- | cosine_precision@3 | 0.1733 | 0.1733 |
329
- | cosine_precision@5 | 0.12 | 0.128 |
330
- | cosine_precision@10 | 0.062 | 0.066 |
331
- | cosine_recall@1 | 0.26 | 0.3 |
332
- | cosine_recall@3 | 0.52 | 0.47 |
333
- | cosine_recall@5 | 0.6 | 0.58 |
334
- | cosine_recall@10 | 0.62 | 0.6 |
335
- | **cosine_ndcg@10** | **0.459** | **0.462** |
336
- | cosine_mrr@10 | 0.4052 | 0.4272 |
337
- | cosine_map@100 | 0.426 | 0.4241 |
338
-
339
- #### Nano BEIR
340
-
341
- * Dataset: `NanoBEIR_mean`
342
- * Evaluated with [<code>NanoBEIREvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.NanoBEIREvaluator) with these parameters:
343
- ```json
344
- {
345
- "dataset_names": [
346
- "msmarco",
347
- "nq"
348
- ],
349
- "dataset_id": "lightonai/NanoBEIR-en"
350
- }
351
- ```
352
-
353
- | Metric | Value |
354
- |:--------------------|:-----------|
355
- | cosine_accuracy@1 | 0.29 |
356
- | cosine_accuracy@3 | 0.51 |
357
- | cosine_accuracy@5 | 0.6 |
358
- | cosine_accuracy@10 | 0.62 |
359
- | cosine_precision@1 | 0.29 |
360
- | cosine_precision@3 | 0.1733 |
361
- | cosine_precision@5 | 0.124 |
362
- | cosine_precision@10 | 0.064 |
363
- | cosine_recall@1 | 0.28 |
364
- | cosine_recall@3 | 0.495 |
365
- | cosine_recall@5 | 0.59 |
366
- | cosine_recall@10 | 0.61 |
367
- | **cosine_ndcg@10** | **0.4605** |
368
- | cosine_mrr@10 | 0.4162 |
369
- | cosine_map@100 | 0.4251 |
370
-
371
  <!--
372
  ## Bias, Risks and Limitations
373
 
@@ -386,49 +146,23 @@ You can finetune this model on your own dataset.
386
 
387
  #### Unnamed Dataset
388
 
389
- * Size: 713,743 training samples
390
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
391
- * Approximate statistics based on the first 1000 samples:
392
- | | anchor | positive | negative |
393
- |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
394
- | type | string | string | string |
395
- | details | <ul><li>min: 6 tokens</li><li>mean: 16.07 tokens</li><li>max: 53 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.03 tokens</li><li>max: 53 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.81 tokens</li><li>max: 58 tokens</li></ul> |
396
- * Samples:
397
- | anchor | positive | negative |
398
- |:-------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------|
399
- | <code>Which one is better Linux OS? Ubuntu or Mint?</code> | <code>Why do you use Linux Mint?</code> | <code>Which one is not better Linux OS ? Ubuntu or Mint ?</code> |
400
- | <code>What is flow?</code> | <code>What is flow?</code> | <code>What are flow lines?</code> |
401
- | <code>How is Trump planning to get Mexico to pay for his supposed wall?</code> | <code>How is it possible for Donald Trump to force Mexico to pay for the wall?</code> | <code>Why do we connect the positive terminal before the negative terminal to ground in a vehicle battery?</code> |
402
- * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
403
- ```json
404
- {
405
- "scale": 7.0,
406
- "similarity_fct": "cos_sim",
407
- "gather_across_devices": false
408
- }
409
- ```
410
-
411
- ### Evaluation Dataset
412
-
413
- #### Unnamed Dataset
414
-
415
- * Size: 40,000 evaluation samples
416
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
417
  * Approximate statistics based on the first 1000 samples:
418
- | | anchor | positive | negative |
419
  |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
420
  | type | string | string | string |
421
- | details | <ul><li>min: 6 tokens</li><li>mean: 15.52 tokens</li><li>max: 74 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.51 tokens</li><li>max: 74 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.79 tokens</li><li>max: 69 tokens</li></ul> |
422
  * Samples:
423
- | anchor | positive | negative |
424
- |:-------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------|
425
- | <code>Why are all my questions on Quora marked needing improvement?</code> | <code>Why are all my questions immediately being marked as needing improvement?</code> | <code>For a post-graduate student in IIT, is it allowed to take an external scholarship as a top-up to his/her MHRD assistantship?</code> |
426
- | <code>Can blue butter fly needle with vaccum tube be reused? Is it HIV risk? . Heard the needle is too small to be reused . Had blood draw at clinic?</code> | <code>Can blue butter fly needle with vaccum tube be reused? Is it HIV risk? . Heard the needle is too small to be reused . Had blood draw at clinic?</code> | <code>Can blue butter fly needle with vaccum tube be reused not ? Is it HIV risk ? . Heard the needle is too small to be reused . Had blood draw at clinic ?</code> |
427
- | <code>Why do people still believe the world is flat?</code> | <code>Why are there still people who believe the world is flat?</code> | <code>I'm not able to buy Udemy course .it is not accepting mine and my friends debit card.my card can be used for Flipkart .how to purchase now?</code> |
428
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
429
  ```json
430
  {
431
- "scale": 7.0,
432
  "similarity_fct": "cos_sim",
433
  "gather_across_devices": false
434
  }
@@ -437,49 +171,36 @@ You can finetune this model on your own dataset.
437
  ### Training Hyperparameters
438
  #### Non-Default Hyperparameters
439
 
440
- - `eval_strategy`: steps
441
- - `per_device_train_batch_size`: 128
442
- - `per_device_eval_batch_size`: 128
443
- - `learning_rate`: 2e-05
444
- - `weight_decay`: 0.0001
445
- - `max_steps`: 5000
446
- - `warmup_ratio`: 0.1
447
  - `fp16`: True
448
- - `dataloader_drop_last`: True
449
- - `dataloader_num_workers`: 1
450
- - `dataloader_prefetch_factor`: 1
451
- - `load_best_model_at_end`: True
452
- - `optim`: adamw_torch
453
- - `ddp_find_unused_parameters`: False
454
- - `push_to_hub`: True
455
- - `hub_model_id`: redis/model-b-structured
456
- - `eval_on_start`: True
457
 
458
  #### All Hyperparameters
459
  <details><summary>Click to expand</summary>
460
 
461
  - `overwrite_output_dir`: False
462
  - `do_predict`: False
463
- - `eval_strategy`: steps
464
  - `prediction_loss_only`: True
465
- - `per_device_train_batch_size`: 128
466
- - `per_device_eval_batch_size`: 128
467
  - `per_gpu_train_batch_size`: None
468
  - `per_gpu_eval_batch_size`: None
469
  - `gradient_accumulation_steps`: 1
470
  - `eval_accumulation_steps`: None
471
  - `torch_empty_cache_steps`: None
472
- - `learning_rate`: 2e-05
473
- - `weight_decay`: 0.0001
474
  - `adam_beta1`: 0.9
475
  - `adam_beta2`: 0.999
476
  - `adam_epsilon`: 1e-08
477
- - `max_grad_norm`: 1.0
478
- - `num_train_epochs`: 3.0
479
- - `max_steps`: 5000
480
  - `lr_scheduler_type`: linear
481
  - `lr_scheduler_kwargs`: {}
482
- - `warmup_ratio`: 0.1
483
  - `warmup_steps`: 0
484
  - `log_level`: passive
485
  - `log_level_replica`: warning
@@ -507,14 +228,14 @@ You can finetune this model on your own dataset.
507
  - `tpu_num_cores`: None
508
  - `tpu_metrics_debug`: False
509
  - `debug`: []
510
- - `dataloader_drop_last`: True
511
- - `dataloader_num_workers`: 1
512
- - `dataloader_prefetch_factor`: 1
513
  - `past_index`: -1
514
  - `disable_tqdm`: False
515
  - `remove_unused_columns`: True
516
  - `label_names`: None
517
- - `load_best_model_at_end`: True
518
  - `ignore_data_skip`: False
519
  - `fsdp`: []
520
  - `fsdp_min_num_params`: 0
@@ -524,23 +245,23 @@ You can finetune this model on your own dataset.
524
  - `parallelism_config`: None
525
  - `deepspeed`: None
526
  - `label_smoothing_factor`: 0.0
527
- - `optim`: adamw_torch
528
  - `optim_args`: None
529
  - `adafactor`: False
530
  - `group_by_length`: False
531
  - `length_column_name`: length
532
  - `project`: huggingface
533
  - `trackio_space_id`: trackio
534
- - `ddp_find_unused_parameters`: False
535
  - `ddp_bucket_cap_mb`: None
536
  - `ddp_broadcast_buffers`: False
537
  - `dataloader_pin_memory`: True
538
  - `dataloader_persistent_workers`: False
539
  - `skip_memory_metrics`: True
540
  - `use_legacy_prediction_loop`: False
541
- - `push_to_hub`: True
542
  - `resume_from_checkpoint`: None
543
- - `hub_model_id`: redis/model-b-structured
544
  - `hub_strategy`: every_save
545
  - `hub_private_repo`: None
546
  - `hub_always_push`: False
@@ -567,43 +288,31 @@ You can finetune this model on your own dataset.
567
  - `neftune_noise_alpha`: None
568
  - `optim_target_modules`: None
569
  - `batch_eval_metrics`: False
570
- - `eval_on_start`: True
571
  - `use_liger_kernel`: False
572
  - `liger_kernel_config`: None
573
  - `eval_use_gather_object`: False
574
  - `average_tokens_across_devices`: True
575
  - `prompts`: None
576
  - `batch_sampler`: batch_sampler
577
- - `multi_dataset_batch_sampler`: proportional
578
  - `router_mapping`: {}
579
  - `learning_rate_mapping`: {}
580
 
581
  </details>
582
 
583
  ### Training Logs
584
- | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
585
- |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
586
- | 0 | 0 | - | 0.7908 | 0.5540 | 0.5931 | 0.5735 |
587
- | 0.0448 | 250 | 0.7632 | 0.4756 | 0.5373 | 0.5302 | 0.5337 |
588
- | 0.0897 | 500 | 0.5825 | 0.4308 | 0.5277 | 0.4949 | 0.5113 |
589
- | 0.1345 | 750 | 0.5438 | 0.4161 | 0.5180 | 0.5039 | 0.5110 |
590
- | 0.1793 | 1000 | 0.5277 | 0.4070 | 0.5008 | 0.4875 | 0.4942 |
591
- | 0.2242 | 1250 | 0.516 | 0.4012 | 0.4983 | 0.4779 | 0.4881 |
592
- | 0.2690 | 1500 | 0.5049 | 0.3962 | 0.4923 | 0.4777 | 0.4850 |
593
- | 0.3138 | 1750 | 0.4966 | 0.3931 | 0.4789 | 0.4769 | 0.4779 |
594
- | 0.3587 | 2000 | 0.493 | 0.3894 | 0.4792 | 0.4616 | 0.4704 |
595
- | 0.4035 | 2250 | 0.4852 | 0.3866 | 0.4828 | 0.4749 | 0.4788 |
596
- | 0.4484 | 2500 | 0.4815 | 0.3841 | 0.4589 | 0.4559 | 0.4574 |
597
- | 0.4932 | 2750 | 0.4761 | 0.3820 | 0.4647 | 0.4539 | 0.4593 |
598
- | 0.5380 | 3000 | 0.4747 | 0.3796 | 0.4588 | 0.4493 | 0.4540 |
599
- | 0.5829 | 3250 | 0.4722 | 0.3786 | 0.4588 | 0.4458 | 0.4523 |
600
- | 0.6277 | 3500 | 0.4725 | 0.3774 | 0.4587 | 0.4537 | 0.4562 |
601
- | 0.6725 | 3750 | 0.4692 | 0.3766 | 0.4561 | 0.4621 | 0.4591 |
602
- | 0.7174 | 4000 | 0.4664 | 0.3763 | 0.4584 | 0.4395 | 0.4489 |
603
- | 0.7622 | 4250 | 0.4659 | 0.3747 | 0.4645 | 0.4586 | 0.4616 |
604
- | 0.8070 | 4500 | 0.464 | 0.3742 | 0.4619 | 0.4479 | 0.4549 |
605
- | 0.8519 | 4750 | 0.4662 | 0.3739 | 0.4590 | 0.4498 | 0.4544 |
606
- | 0.8967 | 5000 | 0.4662 | 0.3739 | 0.4590 | 0.4620 | 0.4605 |
607
 
608
 
609
  ### Framework Versions
@@ -612,7 +321,7 @@ You can finetune this model on your own dataset.
612
  - Transformers: 4.57.3
613
  - PyTorch: 2.9.1+cu128
614
  - Accelerate: 1.12.0
615
- - Datasets: 2.21.0
616
  - Tokenizers: 0.22.1
617
 
618
  ## Citation
 
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
+ - dataset_size:100000
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: prajjwal1/bert-small
11
  widget:
12
+ - source_sentence: How do I calculate IQ?
13
  sentences:
14
+ - What is the easiest way to know my IQ?
15
+ - How do I calculate not IQ ?
16
+ - What are some creative and innovative business ideas with less investment in India?
17
+ - source_sentence: How can I learn martial arts in my home?
 
18
  sentences:
19
+ - How can I learn martial arts by myself?
20
+ - What are the advantages and disadvantages of investing in gold?
21
+ - Can people see that I have looked at their pictures on instagram if I am not following
22
+ them?
23
+ - source_sentence: When Enterprise picks you up do you have to take them back?
24
  sentences:
25
+ - Are there any software Training institute in Tuticorin?
26
+ - When Enterprise picks you up do you have to take them back?
27
+ - When Enterprise picks you up do them have to take youback?
28
+ - source_sentence: What are some non-capital goods?
 
 
 
29
  sentences:
30
+ - What are capital goods?
31
+ - How is the value of [math]\pi[/math] calculated?
32
+ - What are some non-capital goods?
33
+ - source_sentence: What is the QuickBooks technical support phone number in New York?
34
  sentences:
35
+ - What caused the Great Depression?
36
+ - Can I apply for PR in Canada?
37
+ - Which is the best QuickBooks Hosting Support Number in New York?
 
38
  pipeline_tag: sentence-similarity
39
  library_name: sentence-transformers
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ---
41
 
42
+ # SentenceTransformer based on prajjwal1/bert-small
43
 
44
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [prajjwal1/bert-small](https://huggingface.co/prajjwal1/bert-small). It maps sentences & paragraphs to a 512-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
45
 
46
  ## Model Details
47
 
48
  ### Model Description
49
  - **Model Type:** Sentence Transformer
50
+ - **Base model:** [prajjwal1/bert-small](https://huggingface.co/prajjwal1/bert-small) <!-- at revision 0ec5f86f27c1a77d704439db5e01c307ea11b9d4 -->
51
  - **Maximum Sequence Length:** 128 tokens
52
+ - **Output Dimensionality:** 512 dimensions
53
  - **Similarity Function:** Cosine Similarity
54
  <!-- - **Training Dataset:** Unknown -->
55
  <!-- - **Language:** Unknown -->
 
66
  ```
67
  SentenceTransformer(
68
  (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'BertModel'})
69
+ (1): Pooling({'word_embedding_dimension': 512, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
 
70
  )
71
  ```
72
 
 
85
  from sentence_transformers import SentenceTransformer
86
 
87
  # Download from the 🤗 Hub
88
+ model = SentenceTransformer("sentence_transformers_model_id")
89
  # Run inference
90
  sentences = [
91
+ 'What is the QuickBooks technical support phone number in New York?',
92
+ 'Which is the best QuickBooks Hosting Support Number in New York?',
93
+ 'Can I apply for PR in Canada?',
94
  ]
95
  embeddings = model.encode(sentences)
96
  print(embeddings.shape)
97
+ # [3, 512]
98
 
99
  # Get the similarity scores for the embeddings
100
  similarities = model.similarity(embeddings, embeddings)
101
  print(similarities)
102
+ # tensor([[1.0000, 0.8563, 0.0594],
103
+ # [0.8563, 1.0000, 0.1245],
104
+ # [0.0594, 0.1245, 1.0000]])
105
  ```
106
 
107
  <!--
 
128
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
129
  -->
130
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  <!--
132
  ## Bias, Risks and Limitations
133
 
 
146
 
147
  #### Unnamed Dataset
148
 
149
+ * Size: 100,000 training samples
150
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
151
  * Approximate statistics based on the first 1000 samples:
152
+ | | sentence_0 | sentence_1 | sentence_2 |
153
  |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
154
  | type | string | string | string |
155
+ | details | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.68 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 16.37 tokens</li><li>max: 67 tokens</li></ul> |
156
  * Samples:
157
+ | sentence_0 | sentence_1 | sentence_2 |
158
+ |:-----------------------------------------------------------------|:-----------------------------------------------------------------|:----------------------------------------------------------------------------------|
159
+ | <code>Is masturbating bad for boys?</code> | <code>Is masturbating bad for boys?</code> | <code>How harmful or unhealthy is masturbation?</code> |
160
+ | <code>Does a train engine move in reverse?</code> | <code>Does a train engine move in reverse?</code> | <code>Time moves forward, not in reverse. Doesn't that make time a vector?</code> |
161
+ | <code>What is the most badass thing anyone has ever done?</code> | <code>What is the most badass thing anyone has ever done?</code> | <code>anyone is the most badass thing Whathas ever done?</code> |
162
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
163
  ```json
164
  {
165
+ "scale": 20.0,
166
  "similarity_fct": "cos_sim",
167
  "gather_across_devices": false
168
  }
 
171
  ### Training Hyperparameters
172
  #### Non-Default Hyperparameters
173
 
174
+ - `per_device_train_batch_size`: 64
175
+ - `per_device_eval_batch_size`: 64
 
 
 
 
 
176
  - `fp16`: True
177
+ - `multi_dataset_batch_sampler`: round_robin
 
 
 
 
 
 
 
 
178
 
179
  #### All Hyperparameters
180
  <details><summary>Click to expand</summary>
181
 
182
  - `overwrite_output_dir`: False
183
  - `do_predict`: False
184
+ - `eval_strategy`: no
185
  - `prediction_loss_only`: True
186
+ - `per_device_train_batch_size`: 64
187
+ - `per_device_eval_batch_size`: 64
188
  - `per_gpu_train_batch_size`: None
189
  - `per_gpu_eval_batch_size`: None
190
  - `gradient_accumulation_steps`: 1
191
  - `eval_accumulation_steps`: None
192
  - `torch_empty_cache_steps`: None
193
+ - `learning_rate`: 5e-05
194
+ - `weight_decay`: 0.0
195
  - `adam_beta1`: 0.9
196
  - `adam_beta2`: 0.999
197
  - `adam_epsilon`: 1e-08
198
+ - `max_grad_norm`: 1
199
+ - `num_train_epochs`: 3
200
+ - `max_steps`: -1
201
  - `lr_scheduler_type`: linear
202
  - `lr_scheduler_kwargs`: {}
203
+ - `warmup_ratio`: 0.0
204
  - `warmup_steps`: 0
205
  - `log_level`: passive
206
  - `log_level_replica`: warning
 
228
  - `tpu_num_cores`: None
229
  - `tpu_metrics_debug`: False
230
  - `debug`: []
231
+ - `dataloader_drop_last`: False
232
+ - `dataloader_num_workers`: 0
233
+ - `dataloader_prefetch_factor`: None
234
  - `past_index`: -1
235
  - `disable_tqdm`: False
236
  - `remove_unused_columns`: True
237
  - `label_names`: None
238
+ - `load_best_model_at_end`: False
239
  - `ignore_data_skip`: False
240
  - `fsdp`: []
241
  - `fsdp_min_num_params`: 0
 
245
  - `parallelism_config`: None
246
  - `deepspeed`: None
247
  - `label_smoothing_factor`: 0.0
248
+ - `optim`: adamw_torch_fused
249
  - `optim_args`: None
250
  - `adafactor`: False
251
  - `group_by_length`: False
252
  - `length_column_name`: length
253
  - `project`: huggingface
254
  - `trackio_space_id`: trackio
255
+ - `ddp_find_unused_parameters`: None
256
  - `ddp_bucket_cap_mb`: None
257
  - `ddp_broadcast_buffers`: False
258
  - `dataloader_pin_memory`: True
259
  - `dataloader_persistent_workers`: False
260
  - `skip_memory_metrics`: True
261
  - `use_legacy_prediction_loop`: False
262
+ - `push_to_hub`: False
263
  - `resume_from_checkpoint`: None
264
+ - `hub_model_id`: None
265
  - `hub_strategy`: every_save
266
  - `hub_private_repo`: None
267
  - `hub_always_push`: False
 
288
  - `neftune_noise_alpha`: None
289
  - `optim_target_modules`: None
290
  - `batch_eval_metrics`: False
291
+ - `eval_on_start`: False
292
  - `use_liger_kernel`: False
293
  - `liger_kernel_config`: None
294
  - `eval_use_gather_object`: False
295
  - `average_tokens_across_devices`: True
296
  - `prompts`: None
297
  - `batch_sampler`: batch_sampler
298
+ - `multi_dataset_batch_sampler`: round_robin
299
  - `router_mapping`: {}
300
  - `learning_rate_mapping`: {}
301
 
302
  </details>
303
 
304
  ### Training Logs
305
+ | Epoch | Step | Training Loss |
306
+ |:------:|:----:|:-------------:|
307
+ | 0.3199 | 500 | 0.4294 |
308
+ | 0.6398 | 1000 | 0.1268 |
309
+ | 0.9597 | 1500 | 0.1 |
310
+ | 1.2796 | 2000 | 0.0792 |
311
+ | 1.5995 | 2500 | 0.0706 |
312
+ | 1.9194 | 3000 | 0.0687 |
313
+ | 2.2393 | 3500 | 0.0584 |
314
+ | 2.5592 | 4000 | 0.057 |
315
+ | 2.8791 | 4500 | 0.0581 |
 
 
 
 
 
 
 
 
 
 
 
 
316
 
317
 
318
  ### Framework Versions
 
321
  - Transformers: 4.57.3
322
  - PyTorch: 2.9.1+cu128
323
  - Accelerate: 1.12.0
324
+ - Datasets: 4.4.2
325
  - Tokenizers: 0.22.1
326
 
327
  ## Citation
config.json CHANGED
@@ -15,7 +15,7 @@
15
  "max_position_embeddings": 512,
16
  "model_type": "bert",
17
  "num_attention_heads": 12,
18
- "num_hidden_layers": 6,
19
  "pad_token_id": 0,
20
  "position_embedding_type": "absolute",
21
  "transformers_version": "4.57.3",
 
15
  "max_position_embeddings": 512,
16
  "model_type": "bert",
17
  "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
  "pad_token_id": 0,
20
  "position_embedding_type": "absolute",
21
  "transformers_version": "4.57.3",
config_sentence_transformers.json CHANGED
@@ -1,10 +1,10 @@
1
  {
 
2
  "__version__": {
3
  "sentence_transformers": "5.2.0",
4
  "transformers": "4.57.3",
5
  "pytorch": "2.9.1+cu128"
6
  },
7
- "model_type": "SentenceTransformer",
8
  "prompts": {
9
  "query": "",
10
  "document": ""
 
1
  {
2
+ "model_type": "SentenceTransformer",
3
  "__version__": {
4
  "sentence_transformers": "5.2.0",
5
  "transformers": "4.57.3",
6
  "pytorch": "2.9.1+cu128"
7
  },
 
8
  "prompts": {
9
  "query": "",
10
  "document": ""
eval/Information-Retrieval_evaluation_NanoMSMARCO_results.csv CHANGED
@@ -20,3 +20,24 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accurac
20
  0.8070301291248206,4500,0.26,0.52,0.6,0.64,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.064,0.64,0.4035,0.4618541944054583,0.4223204303148985
21
  0.8518651362984218,4750,0.26,0.52,0.6,0.62,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.062,0.62,0.40519047619047627,0.45904886208148177,0.42619789929009394
22
  0.896700143472023,5000,0.26,0.52,0.6,0.62,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.062,0.62,0.40519047619047627,0.45904886208148177,0.4260102142025637
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  0.8070301291248206,4500,0.26,0.52,0.6,0.64,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.064,0.64,0.4035,0.4618541944054583,0.4223204303148985
21
  0.8518651362984218,4750,0.26,0.52,0.6,0.62,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.062,0.62,0.40519047619047627,0.45904886208148177,0.42619789929009394
22
  0.896700143472023,5000,0.26,0.52,0.6,0.62,0.26,0.26,0.1733333333333333,0.52,0.12,0.6,0.062,0.62,0.40519047619047627,0.45904886208148177,0.4260102142025637
23
+ 0,0,0.38,0.62,0.68,0.8,0.38,0.38,0.20666666666666667,0.62,0.136,0.68,0.08,0.8,0.5215238095238095,0.5887227948760404,0.5325402082997991
24
+ 0.04483500717360115,250,0.36,0.58,0.68,0.74,0.36,0.36,0.19333333333333333,0.58,0.136,0.68,0.07400000000000001,0.74,0.4905555555555555,0.5511895065270163,0.506031756095913
25
+ 0.0896700143472023,500,0.36,0.56,0.64,0.76,0.36,0.36,0.18666666666666668,0.56,0.128,0.64,0.07600000000000001,0.76,0.4851904761904761,0.5508845690199249,0.49791509453033606
26
+ 0.13450502152080343,750,0.32,0.58,0.64,0.74,0.32,0.32,0.19333333333333333,0.58,0.128,0.64,0.07400000000000001,0.74,0.4641904761904761,0.531009021314192,0.47885677958247
27
+ 0.1793400286944046,1000,0.34,0.58,0.64,0.74,0.34,0.34,0.19333333333333333,0.58,0.128,0.64,0.07400000000000001,0.74,0.47469047619047616,0.5389184272521197,0.4900991890622418
28
+ 0.22417503586800575,1250,0.36,0.6,0.62,0.76,0.36,0.36,0.19999999999999996,0.6,0.124,0.62,0.07600000000000001,0.76,0.4822142857142857,0.5481489205832354,0.49567847940262133
29
+ 0.26901004304160686,1500,0.32,0.52,0.62,0.8,0.32,0.32,0.1733333333333333,0.52,0.124,0.62,0.08,0.8,0.45526984126984116,0.5360854329130652,0.4661566612305466
30
+ 0.31384505021520803,1750,0.32,0.54,0.64,0.76,0.32,0.32,0.18,0.54,0.128,0.64,0.07600000000000001,0.76,0.45913492063492056,0.5310625456594542,0.47177852901012957
31
+ 0.3586800573888092,2000,0.32,0.54,0.6,0.78,0.32,0.32,0.18,0.54,0.12000000000000002,0.6,0.078,0.78,0.451579365079365,0.5289782036371049,0.4624455348572996
32
+ 0.4035150645624103,2250,0.32,0.54,0.6,0.76,0.32,0.32,0.17999999999999997,0.54,0.12000000000000002,0.6,0.07600000000000001,0.76,0.45301587301587287,0.5259546976245565,0.46591934426473436
33
+ 0.4483500717360115,2500,0.3,0.54,0.62,0.74,0.3,0.3,0.17999999999999997,0.54,0.124,0.62,0.07400000000000001,0.74,0.44035714285714284,0.5121499347260792,0.4559007233900851
34
+ 0.4931850789096126,2750,0.32,0.54,0.62,0.78,0.32,0.32,0.17999999999999997,0.54,0.124,0.62,0.078,0.78,0.4564920634920634,0.5332727558279641,0.4681851283092703
35
+ 0.5380200860832137,3000,0.28,0.54,0.58,0.78,0.28,0.28,0.17999999999999997,0.54,0.11600000000000002,0.58,0.078,0.78,0.4334682539682539,0.5155839048190068,0.44574141585928134
36
+ 0.582855093256815,3250,0.3,0.54,0.6,0.76,0.3,0.3,0.17999999999999997,0.54,0.12000000000000002,0.6,0.07600000000000001,0.76,0.43893650793650785,0.5152963962963514,0.4530906280704235
37
+ 0.6276901004304161,3500,0.28,0.54,0.62,0.78,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.078,0.78,0.43210317460317454,0.5148372013060255,0.4444357120357121
38
+ 0.6725251076040172,3750,0.28,0.52,0.62,0.76,0.28,0.28,0.1733333333333333,0.52,0.124,0.62,0.07600000000000001,0.76,0.4268571428571428,0.5061450417456256,0.4408749948308773
39
+ 0.7173601147776184,4000,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43823809523809515,0.5239447884675306,0.44793010449227444
40
+ 0.7621951219512195,4250,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43435714285714283,0.5208366715209551,0.4441001581538667
41
+ 0.8070301291248206,4500,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43813492063492065,0.5239338733662281,0.44783147531841366
42
+ 0.8518651362984218,4750,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43884920634920627,0.5246486116281323,0.4486126806792123
43
+ 0.896700143472023,5000,0.28,0.54,0.62,0.8,0.28,0.28,0.17999999999999997,0.54,0.124,0.62,0.08,0.8,0.43837301587301575,0.5241911345526384,0.4480711307258977
eval/Information-Retrieval_evaluation_NanoNQ_results.csv CHANGED
@@ -20,3 +20,24 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accurac
20
  0.8070301291248206,4500,0.3,0.5,0.58,0.62,0.3,0.28,0.1733333333333333,0.47,0.124,0.56,0.066,0.6,0.40888888888888886,0.44788049874909824,0.40536711434283007
21
  0.8518651362984218,4750,0.3,0.5,0.6,0.62,0.3,0.28,0.1733333333333333,0.47,0.128,0.58,0.066,0.6,0.4098333333333333,0.4497951196284565,0.40828431891707495
22
  0.896700143472023,5000,0.32,0.5,0.6,0.62,0.32,0.3,0.1733333333333333,0.47,0.128,0.58,0.066,0.6,0.4272222222222222,0.4619884812398348,0.42411471333193373
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  0.8070301291248206,4500,0.3,0.5,0.58,0.62,0.3,0.28,0.1733333333333333,0.47,0.124,0.56,0.066,0.6,0.40888888888888886,0.44788049874909824,0.40536711434283007
21
  0.8518651362984218,4750,0.3,0.5,0.6,0.62,0.3,0.28,0.1733333333333333,0.47,0.128,0.58,0.066,0.6,0.4098333333333333,0.4497951196284565,0.40828431891707495
22
  0.896700143472023,5000,0.32,0.5,0.6,0.62,0.32,0.3,0.1733333333333333,0.47,0.128,0.58,0.066,0.6,0.4272222222222222,0.4619884812398348,0.42411471333193373
23
+ 0,0,0.44,0.62,0.7,0.72,0.44,0.42,0.20666666666666664,0.59,0.15200000000000002,0.69,0.07800000000000001,0.71,0.5425,0.5785694545006416,0.5396412232598975
24
+ 0.04483500717360115,250,0.38,0.58,0.6,0.72,0.38,0.36,0.20666666666666664,0.56,0.128,0.58,0.07800000000000001,0.71,0.4943809523809524,0.5408265737093888,0.49188573556883314
25
+ 0.0896700143472023,500,0.38,0.56,0.6,0.68,0.38,0.37,0.19333333333333333,0.53,0.128,0.58,0.07200000000000001,0.66,0.4859444444444444,0.5218051504362269,0.4845106506090576
26
+ 0.13450502152080343,750,0.38,0.58,0.6,0.66,0.38,0.37,0.2,0.55,0.128,0.58,0.07,0.64,0.48938095238095236,0.5203266781182219,0.48645510419847765
27
+ 0.1793400286944046,1000,0.4,0.58,0.6,0.62,0.4,0.38,0.20666666666666664,0.56,0.128,0.58,0.066,0.6,0.494,0.5128357141707354,0.49102779405243174
28
+ 0.22417503586800575,1250,0.36,0.56,0.6,0.6,0.36,0.35,0.2,0.54,0.128,0.58,0.064,0.58,0.46333333333333326,0.4880570938247723,0.46640313682745294
29
+ 0.26901004304160686,1500,0.4,0.56,0.6,0.62,0.4,0.38,0.2,0.54,0.128,0.58,0.066,0.6,0.4855555555555556,0.5059849863772411,0.482704723788614
30
+ 0.31384505021520803,1750,0.42,0.56,0.6,0.62,0.42,0.4,0.2,0.54,0.128,0.58,0.066,0.6,0.4945555555555556,0.512489916289035,0.49025505979636536
31
+ 0.3586800573888092,2000,0.42,0.58,0.6,0.62,0.42,0.4,0.20666666666666664,0.56,0.128,0.58,0.066,0.6,0.5008333333333334,0.5176601528382082,0.4966428998524106
32
+ 0.4035150645624103,2250,0.4,0.58,0.6,0.62,0.4,0.39,0.2,0.55,0.128,0.58,0.066,0.6,0.49083333333333334,0.5117467474553212,0.48883870214938285
33
+ 0.4483500717360115,2500,0.4,0.58,0.6,0.62,0.4,0.39,0.2,0.55,0.128,0.58,0.066,0.6,0.4883333333333333,0.5104804067864038,0.4870282637039752
34
+ 0.4931850789096126,2750,0.38,0.58,0.6,0.62,0.38,0.37,0.20666666666666664,0.56,0.128,0.58,0.066,0.6,0.47700000000000004,0.5026062641182331,0.4774604900053835
35
+ 0.5380200860832137,3000,0.38,0.54,0.58,0.62,0.38,0.37,0.18666666666666665,0.51,0.124,0.56,0.066,0.6,0.4686666666666667,0.49487523447422943,0.46727110910305214
36
+ 0.582855093256815,3250,0.38,0.58,0.6,0.6,0.38,0.37,0.2,0.55,0.128,0.58,0.064,0.58,0.47166666666666673,0.4933562630442434,0.47192459079370797
37
+ 0.6276901004304161,3500,0.4,0.56,0.6,0.62,0.4,0.39,0.19333333333333333,0.53,0.128,0.58,0.066,0.6,0.4891666666666667,0.510897686812855,0.48816169105329865
38
+ 0.6725251076040172,3750,0.4,0.56,0.6,0.6,0.4,0.38,0.19333333333333333,0.53,0.128,0.58,0.064,0.58,0.4746666666666667,0.4921241368113464,0.470711152240727
39
+ 0.7173601147776184,4000,0.36,0.54,0.6,0.62,0.36,0.35,0.18666666666666665,0.51,0.128,0.58,0.066,0.6,0.4598333333333333,0.48863474295754583,0.4590118602451427
40
+ 0.7621951219512195,4250,0.36,0.54,0.6,0.62,0.36,0.35,0.18666666666666665,0.51,0.128,0.58,0.066,0.6,0.4603333333333333,0.488983216964966,0.45999092949283493
41
+ 0.8070301291248206,4500,0.36,0.54,0.6,0.62,0.36,0.35,0.18666666666666665,0.51,0.128,0.58,0.066,0.6,0.457,0.48636462189353685,0.4563546379037804
42
+ 0.8518651362984218,4750,0.36,0.56,0.6,0.62,0.36,0.35,0.19333333333333333,0.53,0.128,0.58,0.066,0.6,0.4591666666666667,0.4882790917414258,0.4588111953162766
43
+ 0.896700143472023,5000,0.36,0.56,0.6,0.62,0.36,0.35,0.19333333333333333,0.53,0.128,0.58,0.066,0.6,0.4625,0.490897686812855,0.46206363135240186
eval/NanoBEIR_evaluation_mean_results.csv CHANGED
@@ -20,3 +20,24 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Accurac
20
  0.8070301291248206,4500,0.28,0.51,0.59,0.63,0.28,0.27,0.1733333333333333,0.495,0.122,0.5800000000000001,0.065,0.62,0.4061944444444444,0.45486734657727823,0.41384377232886427
21
  0.8518651362984218,4750,0.28,0.51,0.6,0.62,0.28,0.27,0.1733333333333333,0.495,0.124,0.59,0.064,0.61,0.4075119047619048,0.45442199085496915,0.4172411091035845
22
  0.896700143472023,5000,0.29000000000000004,0.51,0.6,0.62,0.29000000000000004,0.28,0.1733333333333333,0.495,0.124,0.59,0.064,0.61,0.41620634920634925,0.4605186716606583,0.4250624637672487
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  0.8070301291248206,4500,0.28,0.51,0.59,0.63,0.28,0.27,0.1733333333333333,0.495,0.122,0.5800000000000001,0.065,0.62,0.4061944444444444,0.45486734657727823,0.41384377232886427
21
  0.8518651362984218,4750,0.28,0.51,0.6,0.62,0.28,0.27,0.1733333333333333,0.495,0.124,0.59,0.064,0.61,0.4075119047619048,0.45442199085496915,0.4172411091035845
22
  0.896700143472023,5000,0.29000000000000004,0.51,0.6,0.62,0.29000000000000004,0.28,0.1733333333333333,0.495,0.124,0.59,0.064,0.61,0.41620634920634925,0.4605186716606583,0.4250624637672487
23
+ 0,0,0.41000000000000003,0.62,0.69,0.76,0.41000000000000003,0.4,0.20666666666666667,0.605,0.14400000000000002,0.685,0.07900000000000001,0.755,0.5320119047619047,0.583646124688341,0.5360907157798482
24
+ 0.04483500717360115,250,0.37,0.58,0.64,0.73,0.37,0.36,0.19999999999999998,0.5700000000000001,0.132,0.63,0.07600000000000001,0.725,0.492468253968254,0.5460080401182026,0.4989587458323731
25
+ 0.0896700143472023,500,0.37,0.56,0.62,0.72,0.37,0.365,0.19,0.545,0.128,0.61,0.07400000000000001,0.71,0.48556746031746023,0.5363448597280759,0.4912128725696968
26
+ 0.13450502152080343,750,0.35,0.58,0.62,0.7,0.35,0.345,0.19666666666666666,0.565,0.128,0.61,0.07200000000000001,0.69,0.47678571428571426,0.525667849716207,0.48265594189047384
27
+ 0.1793400286944046,1000,0.37,0.58,0.62,0.6799999999999999,0.37,0.36,0.19999999999999998,0.5700000000000001,0.128,0.61,0.07,0.6699999999999999,0.4843452380952381,0.5258770707114275,0.4905634915573368
28
+ 0.22417503586800575,1250,0.36,0.5800000000000001,0.61,0.6799999999999999,0.36,0.355,0.19999999999999998,0.5700000000000001,0.126,0.6,0.07,0.6699999999999999,0.4727738095238095,0.5181030072040038,0.48104080811503713
29
+ 0.26901004304160686,1500,0.36,0.54,0.61,0.71,0.36,0.35,0.18666666666666665,0.53,0.126,0.6,0.07300000000000001,0.7,0.4704126984126984,0.5210352096451532,0.4744306925095803
30
+ 0.31384505021520803,1750,0.37,0.55,0.62,0.69,0.37,0.36,0.19,0.54,0.128,0.61,0.07100000000000001,0.6799999999999999,0.47684523809523804,0.5217762309742446,0.48101679440324746
31
+ 0.3586800573888092,2000,0.37,0.56,0.6,0.7,0.37,0.36,0.1933333333333333,0.55,0.12400000000000001,0.59,0.07200000000000001,0.69,0.4762063492063492,0.5233191782376565,0.4795442173548551
32
+ 0.4035150645624103,2250,0.36,0.56,0.6,0.69,0.36,0.355,0.19,0.545,0.12400000000000001,0.59,0.07100000000000001,0.6799999999999999,0.4719246031746031,0.5188507225399388,0.4773790232070586
33
+ 0.4483500717360115,2500,0.35,0.56,0.61,0.6799999999999999,0.35,0.345,0.19,0.545,0.126,0.6,0.07,0.6699999999999999,0.4643452380952381,0.5113151707562416,0.47146449354703013
34
+ 0.4931850789096126,2750,0.35,0.56,0.61,0.7,0.35,0.345,0.1933333333333333,0.55,0.126,0.6,0.07200000000000001,0.69,0.4667460317460317,0.5179395099730986,0.4728228091573269
35
+ 0.5380200860832137,3000,0.33,0.54,0.58,0.7,0.33,0.325,0.1833333333333333,0.525,0.12000000000000001,0.5700000000000001,0.07200000000000001,0.69,0.45106746031746026,0.5052295696466181,0.45650626248116677
36
+ 0.582855093256815,3250,0.33999999999999997,0.56,0.6,0.6799999999999999,0.33999999999999997,0.33499999999999996,0.19,0.545,0.12400000000000001,0.59,0.07,0.6699999999999999,0.4553015873015873,0.5043263296702973,0.4625076094320657
37
+ 0.6276901004304161,3500,0.34,0.55,0.61,0.7,0.34,0.335,0.18666666666666665,0.535,0.126,0.6,0.07200000000000001,0.69,0.4606349206349206,0.5128674440594403,0.46629870154450537
38
+ 0.6725251076040172,3750,0.34,0.54,0.61,0.6799999999999999,0.34,0.33,0.18333333333333332,0.525,0.126,0.6,0.07,0.6699999999999999,0.45076190476190475,0.499134589278486,0.4557930735358021
39
+ 0.7173601147776184,4000,0.32,0.54,0.61,0.71,0.32,0.315,0.1833333333333333,0.525,0.126,0.6,0.07300000000000001,0.7,0.4490357142857142,0.5062897657125383,0.45347098236870853
40
+ 0.7621951219512195,4250,0.32,0.54,0.61,0.71,0.32,0.315,0.1833333333333333,0.525,0.126,0.6,0.07300000000000001,0.7,0.4473452380952381,0.5049099442429605,0.4520455438233508
41
+ 0.8070301291248206,4500,0.32,0.54,0.61,0.71,0.32,0.315,0.1833333333333333,0.525,0.126,0.6,0.07300000000000001,0.7,0.4475674603174603,0.5051492476298824,0.452093056611097
42
+ 0.8518651362984218,4750,0.32,0.55,0.61,0.71,0.32,0.315,0.18666666666666665,0.535,0.126,0.6,0.07300000000000001,0.7,0.44900793650793647,0.506463851684779,0.45371193799774445
43
+ 0.896700143472023,5000,0.32,0.55,0.61,0.71,0.32,0.315,0.18666666666666665,0.535,0.126,0.6,0.07300000000000001,0.7,0.4504365079365079,0.5075444106827467,0.4550673810391498
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2666bef76509b97bc72f0c314f0f5b45d6eb54098fdecdfc61400524864eb46
3
- size 90864192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db59c43bf96bbee3c0ba2dacf6445ee23cd7102f00450058446032338eff15a7
3
+ size 133462128
modules.json CHANGED
@@ -10,11 +10,5 @@
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
13
- },
14
- {
15
- "idx": 2,
16
- "name": "2",
17
- "path": "2_Normalize",
18
- "type": "sentence_transformers.models.Normalize"
19
  }
20
  ]
 
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
 
 
 
 
 
 
13
  }
14
  ]
tokenizer_config.json CHANGED
@@ -48,7 +48,7 @@
48
  "extra_special_tokens": {},
49
  "mask_token": "[MASK]",
50
  "max_length": 128,
51
- "model_max_length": 256,
52
  "never_split": null,
53
  "pad_to_multiple_of": null,
54
  "pad_token": "[PAD]",
 
48
  "extra_special_tokens": {},
49
  "mask_token": "[MASK]",
50
  "max_length": 128,
51
+ "model_max_length": 128,
52
  "never_split": null,
53
  "pad_to_multiple_of": null,
54
  "pad_token": "[PAD]",