JohanHeinsen commited on
Commit
b125252
·
verified ·
1 Parent(s): 39fc4e1

Push model using huggingface_hub.

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,418 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - setfit
4
+ - sentence-transformers
5
+ - text-classification
6
+ - generated_from_setfit_trainer
7
+ widget:
8
+ - text: Studiosus N. Krog, Skoleholder i Ribe, haver Leilighed til at tage et par
9
+ unge Mennesker i Kost og Logemente, som skulde behøve Underviisning ved ham, mod
10
+ billig Betaling, da de med Omgangen og Læremaaden skal blive tiente og for nøiede.
11
+ - text: Min Karl Mads Hansen, demitteret fra Landmilicesessionen for Frederichsorg
12
+ Amt den 17 Marti d. A., af 25 lægd No. 77, Hagerup Sogn, 36 Aar, er mod min Vidende
13
+ og Villie undvigt sin Tieneste, var iklædt hvid Vadmels Kiole, rund Hat og Skoe
14
+ paa Fødderne. Da jeg haver grundet Aarsag at forfølge ham, saa advares alle og
15
+ enhver ikke at modtage, huse eller hæle ham, mindre betroe ham noget enten i mit
16
+ eller andres Navn, men tvertimod enten at angive hans Opholdssted, eller mod en
17
+ passende Belønning foranstalte ham tilbagebragt i hans ulovlig forladte Tieneste,
18
+ paa det de ved Deeltagelse ikke skal paadrage sig den Straf, som Lov og Anordninger
19
+ i slig Tilfælde bestemmer. J. F. Menz, Bager, boende ved Amagerbroe.
20
+ - text: En ganske nye Vand-Filtrum af Holms Fabrik i Kjøbenhavn, destillerende 50
21
+ Potter Vand om Dagen er tilkjøbs i Stokkemarke Præstegaard.
22
+ - text: 'Rusland. Den russiske Regjerings heftige Forbittrelse mod Engelland - hedder
23
+ det i en London, ner Avis: bryder nu les i dens Journaler. Den Moskauer Avis paastaaer
24
+ f. Ex., at den næste Fred slutning mellem Rusland og Storbrittanien torde blive
25
+ undertegnet i Calcutta. Denne Trudsel, tilføier den engelske Avis: er intet Pralerie,
26
+ men et Project, hvis Udførelse i flere Aar har beskjæftiget det Petersborgske
27
+ Kabinet. Under det Paaskud at knytte Handelsforbindelser med Lande i det indre
28
+ Asien, have Russerne udvidet deres militaire Rekognosceringer indtil Grændserne
29
+ af det engelske Jndien, og paa en Maade iforveien gjort Udkastet til en Militairvei
30
+ der hen. Flene Eventyrere have, skjulte under allehaandForklædninger, i dette
31
+ Øiemed vovet sig lige til Punab, paa Hindostans Grændse. Vi ville ikke give os
32
+ af med at spaae, men det troe vi, at de næste 10 Aar ville hidføre store og uventede
33
+ Begivenheder i Asien. - Paa Londons Børs var det en heel almindelig Efterretning,
34
+ at den russiske Regjering skal have opfordret endeel tydske Stater til at forhøie
35
+ Jndførselstolden paa engelske Vare saaledes, at den kommer meget nær et Jnførselsforbud.
36
+ Den brittiske Handelstand er naturligviis bleven meget forbittret herover. Ligeledes
37
+ har den Efterretning gjort Opsigt i Wien, at 6 wallakiske Regimenter ere paa Keiser
38
+ Nicolais Befaling blevne indlemmede i den russiske Armee. J St. Petersborg hed
39
+ det, at Don Carlos af Spaniens ældste Søn skal forloves med en nordisk Prindsesse.'
40
+ - text: Hos Undertegnede erholdes i Commission heftede Exemplarer a 2 Rbdlr. i Sedler
41
+ af den nys i Kiøbenhavn udkomne Jule= og Nytaarsgave Lyra ved Jørgen Henrich,
42
+ Berner Rottbøll Sadolin, Boghandler og Bogbinder.
43
+ metrics:
44
+ - accuracy
45
+ - f1
46
+ - precision
47
+ - recall
48
+ pipeline_tag: text-classification
49
+ library_name: setfit
50
+ inference: true
51
+ base_model: JohanHeinsen/Old_News_Segmentation_SBERT_V0.1
52
+ model-index:
53
+ - name: SetFit with JohanHeinsen/Old_News_Segmentation_SBERT_V0.1
54
+ results:
55
+ - task:
56
+ type: text-classification
57
+ name: Text Classification
58
+ dataset:
59
+ name: Unknown
60
+ type: unknown
61
+ split: test
62
+ metrics:
63
+ - type: accuracy
64
+ value: 0.998960498960499
65
+ name: Accuracy
66
+ - type: f1
67
+ value: 0.9915966386554622
68
+ name: F1
69
+ - type: precision
70
+ value: 0.9833333333333333
71
+ name: Precision
72
+ - type: recall
73
+ value: 1.0
74
+ name: Recall
75
+ ---
76
+
77
+ # SetFit with JohanHeinsen/Old_News_Segmentation_SBERT_V0.1
78
+
79
+ This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [JohanHeinsen/Old_News_Segmentation_SBERT_V0.1](https://huggingface.co/JohanHeinsen/Old_News_Segmentation_SBERT_V0.1) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
80
+
81
+ The model has been trained using an efficient few-shot learning technique that involves:
82
+
83
+ 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
84
+ 2. Training a classification head with features from the fine-tuned Sentence Transformer.
85
+
86
+ ## Model Details
87
+
88
+ ### Model Description
89
+ - **Model Type:** SetFit
90
+ - **Sentence Transformer body:** [JohanHeinsen/Old_News_Segmentation_SBERT_V0.1](https://huggingface.co/JohanHeinsen/Old_News_Segmentation_SBERT_V0.1)
91
+ - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
92
+ - **Maximum Sequence Length:** 512 tokens
93
+ - **Number of Classes:** 2 classes
94
+ <!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
95
+ <!-- - **Language:** Unknown -->
96
+ <!-- - **License:** Unknown -->
97
+
98
+ ### Model Sources
99
+
100
+ - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
101
+ - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
102
+ - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
103
+
104
+ ### Model Labels
105
+ | Label | Examples |
106
+ |:------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
107
+ | 0 | <ul><li>'En meget brav gammel adelig Dame i Augsburg, har legeret 600,000 Gylden til et Pigeinstitues Oprettelse.'</li><li>'Efter indkommen Anmeldelse fra vedkommende Strandtoldbetjent er der løst af Havet inddrevet: paa Østeragger Strand 1 Oxhoved Viin mkt. J Feene paa Tolbøl Strand et Ditto Dito med samme Mærke, paa Hvidberg v. A. Strand 1 Ditto Dito mkt. DL. R. paa Ørum Strand 1 Ditto Dito mkt. 1 Pupaa Steenberg Strand 1 Ditto Dito mkt. NeEieren eller Eierne til fornævnte Oxhoveder Vine indkaldes herved sub poena præclusi et perpetui silentii med Aar og Dags Varsel at indfinde sig ved Amtet for at legitimere Eiendomsretten, hvorefter det indkommende Auctionsbeløb, med Fradrag af alle lovlige Udgifter, skal vorde Vedkommende udbetalt. Thisted Amthuus, den 24de August 1833. Faye.'</li><li>'Ved Tallotteriets 1212te Trækning i Altona den 12te April udkom følgende Nummere:'</li></ul> |
108
+ | 1 | <ul><li>'En Pige 15 Aar gammel, liden af Vext, navnlig Anne Marie, er den 25 May 1761. fra sine Forældre undvigt, og da hende en Arv er tilfalden, saa ombedes hun, eller hvo hende skulde forekomme, at formode hende at indfinde sig hos mig, boende i Nyeboder i Kiøbenhavn paa Elsdyrs-Længden i No. 18, som er hendes Fader, Christen Matros ved 4de Divisions 8de Compagnie.'</li><li>'At fra Kronborg Fæstnings Arbeide den 2 Oct. Sidst er undvigt uærlige Slave Hans Hansen, fød i Roeskilde, 42 Aar gl., liden af Vext, maadelig af Lemmer, blaae af Øine og bruun af Haar, det bekiendtgiøres herved til alle og enhvers Efterretning ligesom man og tillige vil have enhver anmodet at anholde denne for den offentlige Sikkerhed farlige Person, hvor som helst han skulde antræffes, og derefter henbringe ham til nærmeste Arresthuus til Bevaring, hvorfra han, naar saadant Commandant-skabet paa Kronborg tilmeldes, strax skal vorde afhentet, og de paa hans Anholdelse, Arrest og Forplegning anvendte Bekostninger, samt de sædvanlige Opbringerpenge bliver betalt, og tiener tillige til Underretning, at fornævnte Slave ved sin Undvigelse ei havde andet end bare Skiorte paa Kroppen, men Slave Buxer, Strømper og Skoe paa Benene, og en rund Hat paa Hovedet, og har desuden et stort Ar paa det ene Been fra en langvarig Beenskade.'</li><li>'Af Kongens Regiment har Mousqueteer Carl Sverling absenteret sig, samme var klæd i en graa Frakke, rød Manchesters Vest og Buxer, koparret af Ansigt, 23 Aar gl. 65, Tom. Høy; den som tager ham op, levere ham til Casernene imod Douceur efter Forordningen.'</li></ul> |
109
+
110
+ ## Evaluation
111
+
112
+ ### Metrics
113
+ | Label | Accuracy | F1 | Precision | Recall |
114
+ |:--------|:---------|:-------|:----------|:-------|
115
+ | **all** | 0.9990 | 0.9916 | 0.9833 | 1.0 |
116
+
117
+ ## Uses
118
+
119
+ ### Direct Use for Inference
120
+
121
+ First install the SetFit library:
122
+
123
+ ```bash
124
+ pip install setfit
125
+ ```
126
+
127
+ Then you can load this model and run inference.
128
+
129
+ ```python
130
+ from setfit import SetFitModel
131
+
132
+ # Download from the 🤗 Hub
133
+ model = SetFitModel.from_pretrained("setfit_model_id")
134
+ # Run inference
135
+ preds = model("En ganske nye Vand-Filtrum af Holms Fabrik i Kjøbenhavn, destillerende 50 Potter Vand om Dagen er tilkjøbs i Stokkemarke Præstegaard.")
136
+ ```
137
+
138
+ <!--
139
+ ### Downstream Use
140
+
141
+ *List how someone could finetune this model on their own dataset.*
142
+ -->
143
+
144
+ <!--
145
+ ### Out-of-Scope Use
146
+
147
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
148
+ -->
149
+
150
+ <!--
151
+ ## Bias, Risks and Limitations
152
+
153
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
154
+ -->
155
+
156
+ <!--
157
+ ### Recommendations
158
+
159
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
160
+ -->
161
+
162
+ ## Training Details
163
+
164
+ ### Training Set Metrics
165
+ | Training set | Min | Median | Max |
166
+ |:-------------|:----|:--------|:-----|
167
+ | Word count | 5 | 88.9318 | 1999 |
168
+
169
+ | Label | Training Sample Count |
170
+ |:------|:----------------------|
171
+ | 0 | 2093 |
172
+ | 1 | 149 |
173
+
174
+ ### Training Hyperparameters
175
+ - batch_size: (12, 12)
176
+ - num_epochs: (2, 2)
177
+ - max_steps: -1
178
+ - sampling_strategy: oversampling
179
+ - num_iterations: 12
180
+ - body_learning_rate: (2e-05, 2e-05)
181
+ - head_learning_rate: 2e-05
182
+ - loss: CosineSimilarityLoss
183
+ - distance_metric: cosine_distance
184
+ - margin: 0.25
185
+ - end_to_end: False
186
+ - use_amp: False
187
+ - warmup_proportion: 0.1
188
+ - l2_weight: 0.01
189
+ - seed: 42
190
+ - eval_max_steps: -1
191
+ - load_best_model_at_end: False
192
+
193
+ ### Training Results
194
+ | Epoch | Step | Training Loss | Validation Loss |
195
+ |:------:|:----:|:-------------:|:---------------:|
196
+ | 0.0002 | 1 | 0.5665 | - |
197
+ | 0.0112 | 50 | 0.4302 | - |
198
+ | 0.0223 | 100 | 0.3677 | - |
199
+ | 0.0335 | 150 | 0.1981 | - |
200
+ | 0.0446 | 200 | 0.0642 | - |
201
+ | 0.0558 | 250 | 0.0272 | - |
202
+ | 0.0669 | 300 | 0.0083 | - |
203
+ | 0.0781 | 350 | 0.0114 | - |
204
+ | 0.0892 | 400 | 0.0038 | - |
205
+ | 0.1004 | 450 | 0.0036 | - |
206
+ | 0.1115 | 500 | 0.0023 | - |
207
+ | 0.1227 | 550 | 0.005 | - |
208
+ | 0.1338 | 600 | 0.0031 | - |
209
+ | 0.1450 | 650 | 0.0011 | - |
210
+ | 0.1561 | 700 | 0.0038 | - |
211
+ | 0.1673 | 750 | 0.0001 | - |
212
+ | 0.1784 | 800 | 0.0005 | - |
213
+ | 0.1896 | 850 | 0.0019 | - |
214
+ | 0.2007 | 900 | 0.0016 | - |
215
+ | 0.2119 | 950 | 0.0001 | - |
216
+ | 0.2230 | 1000 | 0.0014 | - |
217
+ | 0.2342 | 1050 | 0.0022 | - |
218
+ | 0.2453 | 1100 | 0.0021 | - |
219
+ | 0.2565 | 1150 | 0.0018 | - |
220
+ | 0.2676 | 1200 | 0.0002 | - |
221
+ | 0.2788 | 1250 | 0.0 | - |
222
+ | 0.2899 | 1300 | 0.0019 | - |
223
+ | 0.3011 | 1350 | 0.0 | - |
224
+ | 0.3122 | 1400 | 0.0 | - |
225
+ | 0.3234 | 1450 | 0.0036 | - |
226
+ | 0.3345 | 1500 | 0.0 | - |
227
+ | 0.3457 | 1550 | 0.0 | - |
228
+ | 0.3568 | 1600 | 0.0 | - |
229
+ | 0.3680 | 1650 | 0.0 | - |
230
+ | 0.3791 | 1700 | 0.0 | - |
231
+ | 0.3903 | 1750 | 0.0018 | - |
232
+ | 0.4014 | 1800 | 0.0001 | - |
233
+ | 0.4126 | 1850 | 0.0017 | - |
234
+ | 0.4237 | 1900 | 0.0 | - |
235
+ | 0.4349 | 1950 | 0.0 | - |
236
+ | 0.4460 | 2000 | 0.0 | - |
237
+ | 0.4572 | 2050 | 0.0035 | - |
238
+ | 0.4683 | 2100 | 0.0034 | - |
239
+ | 0.4795 | 2150 | 0.0036 | - |
240
+ | 0.4906 | 2200 | 0.0017 | - |
241
+ | 0.5018 | 2250 | 0.0056 | - |
242
+ | 0.5129 | 2300 | 0.0006 | - |
243
+ | 0.5241 | 2350 | 0.0 | - |
244
+ | 0.5352 | 2400 | 0.0 | - |
245
+ | 0.5464 | 2450 | 0.0 | - |
246
+ | 0.5575 | 2500 | 0.0016 | - |
247
+ | 0.5687 | 2550 | 0.0014 | - |
248
+ | 0.5798 | 2600 | 0.0 | - |
249
+ | 0.5910 | 2650 | 0.0012 | - |
250
+ | 0.6021 | 2700 | 0.0001 | - |
251
+ | 0.6133 | 2750 | 0.0 | - |
252
+ | 0.6244 | 2800 | 0.0 | - |
253
+ | 0.6356 | 2850 | 0.0 | - |
254
+ | 0.6467 | 2900 | 0.0 | - |
255
+ | 0.6579 | 2950 | 0.0 | - |
256
+ | 0.6690 | 3000 | 0.0016 | - |
257
+ | 0.6802 | 3050 | 0.0 | - |
258
+ | 0.6913 | 3100 | 0.0 | - |
259
+ | 0.7025 | 3150 | 0.0 | - |
260
+ | 0.7136 | 3200 | 0.0017 | - |
261
+ | 0.7248 | 3250 | 0.0012 | - |
262
+ | 0.7360 | 3300 | 0.0002 | - |
263
+ | 0.7471 | 3350 | 0.0 | - |
264
+ | 0.7583 | 3400 | 0.0 | - |
265
+ | 0.7694 | 3450 | 0.0 | - |
266
+ | 0.7806 | 3500 | 0.0 | - |
267
+ | 0.7917 | 3550 | 0.0 | - |
268
+ | 0.8029 | 3600 | 0.0 | - |
269
+ | 0.8140 | 3650 | 0.0 | - |
270
+ | 0.8252 | 3700 | 0.0 | - |
271
+ | 0.8363 | 3750 | 0.0 | - |
272
+ | 0.8475 | 3800 | 0.0 | - |
273
+ | 0.8586 | 3850 | 0.0 | - |
274
+ | 0.8698 | 3900 | 0.0 | - |
275
+ | 0.8809 | 3950 | 0.0 | - |
276
+ | 0.8921 | 4000 | 0.0 | - |
277
+ | 0.9032 | 4050 | 0.0 | - |
278
+ | 0.9144 | 4100 | 0.0 | - |
279
+ | 0.9255 | 4150 | 0.0 | - |
280
+ | 0.9367 | 4200 | 0.0 | - |
281
+ | 0.9478 | 4250 | 0.0 | - |
282
+ | 0.9590 | 4300 | 0.0 | - |
283
+ | 0.9701 | 4350 | 0.0 | - |
284
+ | 0.9813 | 4400 | 0.0 | - |
285
+ | 0.9924 | 4450 | 0.0 | - |
286
+ | 1.0036 | 4500 | 0.0 | - |
287
+ | 1.0147 | 4550 | 0.0 | - |
288
+ | 1.0259 | 4600 | 0.0 | - |
289
+ | 1.0370 | 4650 | 0.0 | - |
290
+ | 1.0482 | 4700 | 0.0 | - |
291
+ | 1.0593 | 4750 | 0.0 | - |
292
+ | 1.0705 | 4800 | 0.0 | - |
293
+ | 1.0816 | 4850 | 0.0 | - |
294
+ | 1.0928 | 4900 | 0.0 | - |
295
+ | 1.1039 | 4950 | 0.0 | - |
296
+ | 1.1151 | 5000 | 0.0 | - |
297
+ | 1.1262 | 5050 | 0.0 | - |
298
+ | 1.1374 | 5100 | 0.0 | - |
299
+ | 1.1485 | 5150 | 0.0 | - |
300
+ | 1.1597 | 5200 | 0.0 | - |
301
+ | 1.1708 | 5250 | 0.0 | - |
302
+ | 1.1820 | 5300 | 0.0 | - |
303
+ | 1.1931 | 5350 | 0.0 | - |
304
+ | 1.2043 | 5400 | 0.0 | - |
305
+ | 1.2154 | 5450 | 0.0 | - |
306
+ | 1.2266 | 5500 | 0.0 | - |
307
+ | 1.2377 | 5550 | 0.0 | - |
308
+ | 1.2489 | 5600 | 0.0 | - |
309
+ | 1.2600 | 5650 | 0.0 | - |
310
+ | 1.2712 | 5700 | 0.0 | - |
311
+ | 1.2823 | 5750 | 0.0 | - |
312
+ | 1.2935 | 5800 | 0.0 | - |
313
+ | 1.3046 | 5850 | 0.0 | - |
314
+ | 1.3158 | 5900 | 0.0 | - |
315
+ | 1.3269 | 5950 | 0.0 | - |
316
+ | 1.3381 | 6000 | 0.0 | - |
317
+ | 1.3492 | 6050 | 0.0 | - |
318
+ | 1.3604 | 6100 | 0.0 | - |
319
+ | 1.3715 | 6150 | 0.0 | - |
320
+ | 1.3827 | 6200 | 0.0 | - |
321
+ | 1.3938 | 6250 | 0.0 | - |
322
+ | 1.4050 | 6300 | 0.0 | - |
323
+ | 1.4161 | 6350 | 0.0 | - |
324
+ | 1.4273 | 6400 | 0.0 | - |
325
+ | 1.4384 | 6450 | 0.0 | - |
326
+ | 1.4496 | 6500 | 0.0 | - |
327
+ | 1.4607 | 6550 | 0.0 | - |
328
+ | 1.4719 | 6600 | 0.0 | - |
329
+ | 1.4831 | 6650 | 0.0 | - |
330
+ | 1.4942 | 6700 | 0.0 | - |
331
+ | 1.5054 | 6750 | 0.0 | - |
332
+ | 1.5165 | 6800 | 0.0 | - |
333
+ | 1.5277 | 6850 | 0.0 | - |
334
+ | 1.5388 | 6900 | 0.0 | - |
335
+ | 1.5500 | 6950 | 0.0 | - |
336
+ | 1.5611 | 7000 | 0.0 | - |
337
+ | 1.5723 | 7050 | 0.0 | - |
338
+ | 1.5834 | 7100 | 0.0 | - |
339
+ | 1.5946 | 7150 | 0.0 | - |
340
+ | 1.6057 | 7200 | 0.0 | - |
341
+ | 1.6169 | 7250 | 0.0 | - |
342
+ | 1.6280 | 7300 | 0.0 | - |
343
+ | 1.6392 | 7350 | 0.0 | - |
344
+ | 1.6503 | 7400 | 0.0 | - |
345
+ | 1.6615 | 7450 | 0.0 | - |
346
+ | 1.6726 | 7500 | 0.0 | - |
347
+ | 1.6838 | 7550 | 0.0 | - |
348
+ | 1.6949 | 7600 | 0.0 | - |
349
+ | 1.7061 | 7650 | 0.0 | - |
350
+ | 1.7172 | 7700 | 0.0 | - |
351
+ | 1.7284 | 7750 | 0.0 | - |
352
+ | 1.7395 | 7800 | 0.0 | - |
353
+ | 1.7507 | 7850 | 0.0 | - |
354
+ | 1.7618 | 7900 | 0.0 | - |
355
+ | 1.7730 | 7950 | 0.0 | - |
356
+ | 1.7841 | 8000 | 0.0 | - |
357
+ | 1.7953 | 8050 | 0.0 | - |
358
+ | 1.8064 | 8100 | 0.0 | - |
359
+ | 1.8176 | 8150 | 0.0 | - |
360
+ | 1.8287 | 8200 | 0.0 | - |
361
+ | 1.8399 | 8250 | 0.0 | - |
362
+ | 1.8510 | 8300 | 0.0 | - |
363
+ | 1.8622 | 8350 | 0.0 | - |
364
+ | 1.8733 | 8400 | 0.0 | - |
365
+ | 1.8845 | 8450 | 0.0 | - |
366
+ | 1.8956 | 8500 | 0.0 | - |
367
+ | 1.9068 | 8550 | 0.0 | - |
368
+ | 1.9179 | 8600 | 0.0 | - |
369
+ | 1.9291 | 8650 | 0.0 | - |
370
+ | 1.9402 | 8700 | 0.0 | - |
371
+ | 1.9514 | 8750 | 0.0 | - |
372
+ | 1.9625 | 8800 | 0.0 | - |
373
+ | 1.9737 | 8850 | 0.0 | - |
374
+ | 1.9848 | 8900 | 0.0 | - |
375
+ | 1.9960 | 8950 | 0.0 | - |
376
+
377
+ ### Framework Versions
378
+ - Python: 3.11.12
379
+ - SetFit: 1.1.3
380
+ - Sentence Transformers: 4.1.0
381
+ - Transformers: 4.51.3
382
+ - PyTorch: 2.7.0
383
+ - Datasets: 2.19.2
384
+ - Tokenizers: 0.21.1
385
+
386
+ ## Citation
387
+
388
+ ### BibTeX
389
+ ```bibtex
390
+ @article{https://doi.org/10.48550/arxiv.2209.11055,
391
+ doi = {10.48550/ARXIV.2209.11055},
392
+ url = {https://arxiv.org/abs/2209.11055},
393
+ author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
394
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
395
+ title = {Efficient Few-Shot Learning Without Prompts},
396
+ publisher = {arXiv},
397
+ year = {2022},
398
+ copyright = {Creative Commons Attribution 4.0 International}
399
+ }
400
+ ```
401
+
402
+ <!--
403
+ ## Glossary
404
+
405
+ *Clearly define terms in order to be accessible across audiences.*
406
+ -->
407
+
408
+ <!--
409
+ ## Model Card Authors
410
+
411
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
412
+ -->
413
+
414
+ <!--
415
+ ## Model Card Contact
416
+
417
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
418
+ -->
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 3072,
12
+ "layer_norm_eps": 1e-12,
13
+ "max_position_embeddings": 512,
14
+ "model_type": "bert",
15
+ "num_attention_heads": 12,
16
+ "num_hidden_layers": 12,
17
+ "pad_token_id": 0,
18
+ "position_embedding_type": "absolute",
19
+ "torch_dtype": "float32",
20
+ "transformers_version": "4.51.3",
21
+ "type_vocab_size": 2,
22
+ "use_cache": true,
23
+ "vocab_size": 30522
24
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "4.1.0",
4
+ "transformers": "4.51.3",
5
+ "pytorch": "2.7.0"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
+ }
config_setfit.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "normalize_embeddings": false,
3
+ "labels": null
4
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d7272d77af10a7583d669cc13e72e56c5d58fbf64a10dcfaee90b9fcb988f0c
3
+ size 437951328
model_head.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9db26e8a9e044a0144e2a68eca28de7a4eb6b524040c67408a75f62db8e9fbc6
3
+ size 7007
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "max_length": 512,
51
+ "model_max_length": 512,
52
+ "never_split": null,
53
+ "pad_to_multiple_of": null,
54
+ "pad_token": "[PAD]",
55
+ "pad_token_type_id": 0,
56
+ "padding_side": "right",
57
+ "sep_token": "[SEP]",
58
+ "stride": 0,
59
+ "strip_accents": null,
60
+ "tokenize_chinese_chars": true,
61
+ "tokenizer_class": "BertTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
+ "unk_token": "[UNK]"
65
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff