Push model using huggingface_hub.
Browse files- 1_Pooling/config.json +10 -0
- README.md +418 -0
- config.json +24 -0
- config_sentence_transformers.json +10 -0
- config_setfit.json +4 -0
- model.safetensors +3 -0
- model_head.pkl +3 -0
- modules.json +14 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +65 -0
- vocab.txt +0 -0
1_Pooling/config.json
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"word_embedding_dimension": 768,
|
| 3 |
+
"pooling_mode_cls_token": false,
|
| 4 |
+
"pooling_mode_mean_tokens": true,
|
| 5 |
+
"pooling_mode_max_tokens": false,
|
| 6 |
+
"pooling_mode_mean_sqrt_len_tokens": false,
|
| 7 |
+
"pooling_mode_weightedmean_tokens": false,
|
| 8 |
+
"pooling_mode_lasttoken": false,
|
| 9 |
+
"include_prompt": true
|
| 10 |
+
}
|
README.md
ADDED
|
@@ -0,0 +1,418 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- setfit
|
| 4 |
+
- sentence-transformers
|
| 5 |
+
- text-classification
|
| 6 |
+
- generated_from_setfit_trainer
|
| 7 |
+
widget:
|
| 8 |
+
- text: Studiosus N. Krog, Skoleholder i Ribe, haver Leilighed til at tage et par
|
| 9 |
+
unge Mennesker i Kost og Logemente, som skulde behøve Underviisning ved ham, mod
|
| 10 |
+
billig Betaling, da de med Omgangen og Læremaaden skal blive tiente og for nøiede.
|
| 11 |
+
- text: Min Karl Mads Hansen, demitteret fra Landmilicesessionen for Frederichsorg
|
| 12 |
+
Amt den 17 Marti d. A., af 25 lægd No. 77, Hagerup Sogn, 36 Aar, er mod min Vidende
|
| 13 |
+
og Villie undvigt sin Tieneste, var iklædt hvid Vadmels Kiole, rund Hat og Skoe
|
| 14 |
+
paa Fødderne. Da jeg haver grundet Aarsag at forfølge ham, saa advares alle og
|
| 15 |
+
enhver ikke at modtage, huse eller hæle ham, mindre betroe ham noget enten i mit
|
| 16 |
+
eller andres Navn, men tvertimod enten at angive hans Opholdssted, eller mod en
|
| 17 |
+
passende Belønning foranstalte ham tilbagebragt i hans ulovlig forladte Tieneste,
|
| 18 |
+
paa det de ved Deeltagelse ikke skal paadrage sig den Straf, som Lov og Anordninger
|
| 19 |
+
i slig Tilfælde bestemmer. J. F. Menz, Bager, boende ved Amagerbroe.
|
| 20 |
+
- text: En ganske nye Vand-Filtrum af Holms Fabrik i Kjøbenhavn, destillerende 50
|
| 21 |
+
Potter Vand om Dagen er tilkjøbs i Stokkemarke Præstegaard.
|
| 22 |
+
- text: 'Rusland. Den russiske Regjerings heftige Forbittrelse mod Engelland - hedder
|
| 23 |
+
det i en London, ner Avis: bryder nu les i dens Journaler. Den Moskauer Avis paastaaer
|
| 24 |
+
f. Ex., at den næste Fred slutning mellem Rusland og Storbrittanien torde blive
|
| 25 |
+
undertegnet i Calcutta. Denne Trudsel, tilføier den engelske Avis: er intet Pralerie,
|
| 26 |
+
men et Project, hvis Udførelse i flere Aar har beskjæftiget det Petersborgske
|
| 27 |
+
Kabinet. Under det Paaskud at knytte Handelsforbindelser med Lande i det indre
|
| 28 |
+
Asien, have Russerne udvidet deres militaire Rekognosceringer indtil Grændserne
|
| 29 |
+
af det engelske Jndien, og paa en Maade iforveien gjort Udkastet til en Militairvei
|
| 30 |
+
der hen. Flene Eventyrere have, skjulte under allehaandForklædninger, i dette
|
| 31 |
+
Øiemed vovet sig lige til Punab, paa Hindostans Grændse. Vi ville ikke give os
|
| 32 |
+
af med at spaae, men det troe vi, at de næste 10 Aar ville hidføre store og uventede
|
| 33 |
+
Begivenheder i Asien. - Paa Londons Børs var det en heel almindelig Efterretning,
|
| 34 |
+
at den russiske Regjering skal have opfordret endeel tydske Stater til at forhøie
|
| 35 |
+
Jndførselstolden paa engelske Vare saaledes, at den kommer meget nær et Jnførselsforbud.
|
| 36 |
+
Den brittiske Handelstand er naturligviis bleven meget forbittret herover. Ligeledes
|
| 37 |
+
har den Efterretning gjort Opsigt i Wien, at 6 wallakiske Regimenter ere paa Keiser
|
| 38 |
+
Nicolais Befaling blevne indlemmede i den russiske Armee. J St. Petersborg hed
|
| 39 |
+
det, at Don Carlos af Spaniens ældste Søn skal forloves med en nordisk Prindsesse.'
|
| 40 |
+
- text: Hos Undertegnede erholdes i Commission heftede Exemplarer a 2 Rbdlr. i Sedler
|
| 41 |
+
af den nys i Kiøbenhavn udkomne Jule= og Nytaarsgave Lyra ved Jørgen Henrich,
|
| 42 |
+
Berner Rottbøll Sadolin, Boghandler og Bogbinder.
|
| 43 |
+
metrics:
|
| 44 |
+
- accuracy
|
| 45 |
+
- f1
|
| 46 |
+
- precision
|
| 47 |
+
- recall
|
| 48 |
+
pipeline_tag: text-classification
|
| 49 |
+
library_name: setfit
|
| 50 |
+
inference: true
|
| 51 |
+
base_model: JohanHeinsen/Old_News_Segmentation_SBERT_V0.1
|
| 52 |
+
model-index:
|
| 53 |
+
- name: SetFit with JohanHeinsen/Old_News_Segmentation_SBERT_V0.1
|
| 54 |
+
results:
|
| 55 |
+
- task:
|
| 56 |
+
type: text-classification
|
| 57 |
+
name: Text Classification
|
| 58 |
+
dataset:
|
| 59 |
+
name: Unknown
|
| 60 |
+
type: unknown
|
| 61 |
+
split: test
|
| 62 |
+
metrics:
|
| 63 |
+
- type: accuracy
|
| 64 |
+
value: 0.998960498960499
|
| 65 |
+
name: Accuracy
|
| 66 |
+
- type: f1
|
| 67 |
+
value: 0.9915966386554622
|
| 68 |
+
name: F1
|
| 69 |
+
- type: precision
|
| 70 |
+
value: 0.9833333333333333
|
| 71 |
+
name: Precision
|
| 72 |
+
- type: recall
|
| 73 |
+
value: 1.0
|
| 74 |
+
name: Recall
|
| 75 |
+
---
|
| 76 |
+
|
| 77 |
+
# SetFit with JohanHeinsen/Old_News_Segmentation_SBERT_V0.1
|
| 78 |
+
|
| 79 |
+
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [JohanHeinsen/Old_News_Segmentation_SBERT_V0.1](https://huggingface.co/JohanHeinsen/Old_News_Segmentation_SBERT_V0.1) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
|
| 80 |
+
|
| 81 |
+
The model has been trained using an efficient few-shot learning technique that involves:
|
| 82 |
+
|
| 83 |
+
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
|
| 84 |
+
2. Training a classification head with features from the fine-tuned Sentence Transformer.
|
| 85 |
+
|
| 86 |
+
## Model Details
|
| 87 |
+
|
| 88 |
+
### Model Description
|
| 89 |
+
- **Model Type:** SetFit
|
| 90 |
+
- **Sentence Transformer body:** [JohanHeinsen/Old_News_Segmentation_SBERT_V0.1](https://huggingface.co/JohanHeinsen/Old_News_Segmentation_SBERT_V0.1)
|
| 91 |
+
- **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
|
| 92 |
+
- **Maximum Sequence Length:** 512 tokens
|
| 93 |
+
- **Number of Classes:** 2 classes
|
| 94 |
+
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
|
| 95 |
+
<!-- - **Language:** Unknown -->
|
| 96 |
+
<!-- - **License:** Unknown -->
|
| 97 |
+
|
| 98 |
+
### Model Sources
|
| 99 |
+
|
| 100 |
+
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
|
| 101 |
+
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
|
| 102 |
+
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
|
| 103 |
+
|
| 104 |
+
### Model Labels
|
| 105 |
+
| Label | Examples |
|
| 106 |
+
|:------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
| 107 |
+
| 0 | <ul><li>'En meget brav gammel adelig Dame i Augsburg, har legeret 600,000 Gylden til et Pigeinstitues Oprettelse.'</li><li>'Efter indkommen Anmeldelse fra vedkommende Strandtoldbetjent er der løst af Havet inddrevet: paa Østeragger Strand 1 Oxhoved Viin mkt. J Feene paa Tolbøl Strand et Ditto Dito med samme Mærke, paa Hvidberg v. A. Strand 1 Ditto Dito mkt. DL. R. paa Ørum Strand 1 Ditto Dito mkt. 1 Pupaa Steenberg Strand 1 Ditto Dito mkt. NeEieren eller Eierne til fornævnte Oxhoveder Vine indkaldes herved sub poena præclusi et perpetui silentii med Aar og Dags Varsel at indfinde sig ved Amtet for at legitimere Eiendomsretten, hvorefter det indkommende Auctionsbeløb, med Fradrag af alle lovlige Udgifter, skal vorde Vedkommende udbetalt. Thisted Amthuus, den 24de August 1833. Faye.'</li><li>'Ved Tallotteriets 1212te Trækning i Altona den 12te April udkom følgende Nummere:'</li></ul> |
|
| 108 |
+
| 1 | <ul><li>'En Pige 15 Aar gammel, liden af Vext, navnlig Anne Marie, er den 25 May 1761. fra sine Forældre undvigt, og da hende en Arv er tilfalden, saa ombedes hun, eller hvo hende skulde forekomme, at formode hende at indfinde sig hos mig, boende i Nyeboder i Kiøbenhavn paa Elsdyrs-Længden i No. 18, som er hendes Fader, Christen Matros ved 4de Divisions 8de Compagnie.'</li><li>'At fra Kronborg Fæstnings Arbeide den 2 Oct. Sidst er undvigt uærlige Slave Hans Hansen, fød i Roeskilde, 42 Aar gl., liden af Vext, maadelig af Lemmer, blaae af Øine og bruun af Haar, det bekiendtgiøres herved til alle og enhvers Efterretning ligesom man og tillige vil have enhver anmodet at anholde denne for den offentlige Sikkerhed farlige Person, hvor som helst han skulde antræffes, og derefter henbringe ham til nærmeste Arresthuus til Bevaring, hvorfra han, naar saadant Commandant-skabet paa Kronborg tilmeldes, strax skal vorde afhentet, og de paa hans Anholdelse, Arrest og Forplegning anvendte Bekostninger, samt de sædvanlige Opbringerpenge bliver betalt, og tiener tillige til Underretning, at fornævnte Slave ved sin Undvigelse ei havde andet end bare Skiorte paa Kroppen, men Slave Buxer, Strømper og Skoe paa Benene, og en rund Hat paa Hovedet, og har desuden et stort Ar paa det ene Been fra en langvarig Beenskade.'</li><li>'Af Kongens Regiment har Mousqueteer Carl Sverling absenteret sig, samme var klæd i en graa Frakke, rød Manchesters Vest og Buxer, koparret af Ansigt, 23 Aar gl. 65, Tom. Høy; den som tager ham op, levere ham til Casernene imod Douceur efter Forordningen.'</li></ul> |
|
| 109 |
+
|
| 110 |
+
## Evaluation
|
| 111 |
+
|
| 112 |
+
### Metrics
|
| 113 |
+
| Label | Accuracy | F1 | Precision | Recall |
|
| 114 |
+
|:--------|:---------|:-------|:----------|:-------|
|
| 115 |
+
| **all** | 0.9990 | 0.9916 | 0.9833 | 1.0 |
|
| 116 |
+
|
| 117 |
+
## Uses
|
| 118 |
+
|
| 119 |
+
### Direct Use for Inference
|
| 120 |
+
|
| 121 |
+
First install the SetFit library:
|
| 122 |
+
|
| 123 |
+
```bash
|
| 124 |
+
pip install setfit
|
| 125 |
+
```
|
| 126 |
+
|
| 127 |
+
Then you can load this model and run inference.
|
| 128 |
+
|
| 129 |
+
```python
|
| 130 |
+
from setfit import SetFitModel
|
| 131 |
+
|
| 132 |
+
# Download from the 🤗 Hub
|
| 133 |
+
model = SetFitModel.from_pretrained("setfit_model_id")
|
| 134 |
+
# Run inference
|
| 135 |
+
preds = model("En ganske nye Vand-Filtrum af Holms Fabrik i Kjøbenhavn, destillerende 50 Potter Vand om Dagen er tilkjøbs i Stokkemarke Præstegaard.")
|
| 136 |
+
```
|
| 137 |
+
|
| 138 |
+
<!--
|
| 139 |
+
### Downstream Use
|
| 140 |
+
|
| 141 |
+
*List how someone could finetune this model on their own dataset.*
|
| 142 |
+
-->
|
| 143 |
+
|
| 144 |
+
<!--
|
| 145 |
+
### Out-of-Scope Use
|
| 146 |
+
|
| 147 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 148 |
+
-->
|
| 149 |
+
|
| 150 |
+
<!--
|
| 151 |
+
## Bias, Risks and Limitations
|
| 152 |
+
|
| 153 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
| 154 |
+
-->
|
| 155 |
+
|
| 156 |
+
<!--
|
| 157 |
+
### Recommendations
|
| 158 |
+
|
| 159 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
| 160 |
+
-->
|
| 161 |
+
|
| 162 |
+
## Training Details
|
| 163 |
+
|
| 164 |
+
### Training Set Metrics
|
| 165 |
+
| Training set | Min | Median | Max |
|
| 166 |
+
|:-------------|:----|:--------|:-----|
|
| 167 |
+
| Word count | 5 | 88.9318 | 1999 |
|
| 168 |
+
|
| 169 |
+
| Label | Training Sample Count |
|
| 170 |
+
|:------|:----------------------|
|
| 171 |
+
| 0 | 2093 |
|
| 172 |
+
| 1 | 149 |
|
| 173 |
+
|
| 174 |
+
### Training Hyperparameters
|
| 175 |
+
- batch_size: (12, 12)
|
| 176 |
+
- num_epochs: (2, 2)
|
| 177 |
+
- max_steps: -1
|
| 178 |
+
- sampling_strategy: oversampling
|
| 179 |
+
- num_iterations: 12
|
| 180 |
+
- body_learning_rate: (2e-05, 2e-05)
|
| 181 |
+
- head_learning_rate: 2e-05
|
| 182 |
+
- loss: CosineSimilarityLoss
|
| 183 |
+
- distance_metric: cosine_distance
|
| 184 |
+
- margin: 0.25
|
| 185 |
+
- end_to_end: False
|
| 186 |
+
- use_amp: False
|
| 187 |
+
- warmup_proportion: 0.1
|
| 188 |
+
- l2_weight: 0.01
|
| 189 |
+
- seed: 42
|
| 190 |
+
- eval_max_steps: -1
|
| 191 |
+
- load_best_model_at_end: False
|
| 192 |
+
|
| 193 |
+
### Training Results
|
| 194 |
+
| Epoch | Step | Training Loss | Validation Loss |
|
| 195 |
+
|:------:|:----:|:-------------:|:---------------:|
|
| 196 |
+
| 0.0002 | 1 | 0.5665 | - |
|
| 197 |
+
| 0.0112 | 50 | 0.4302 | - |
|
| 198 |
+
| 0.0223 | 100 | 0.3677 | - |
|
| 199 |
+
| 0.0335 | 150 | 0.1981 | - |
|
| 200 |
+
| 0.0446 | 200 | 0.0642 | - |
|
| 201 |
+
| 0.0558 | 250 | 0.0272 | - |
|
| 202 |
+
| 0.0669 | 300 | 0.0083 | - |
|
| 203 |
+
| 0.0781 | 350 | 0.0114 | - |
|
| 204 |
+
| 0.0892 | 400 | 0.0038 | - |
|
| 205 |
+
| 0.1004 | 450 | 0.0036 | - |
|
| 206 |
+
| 0.1115 | 500 | 0.0023 | - |
|
| 207 |
+
| 0.1227 | 550 | 0.005 | - |
|
| 208 |
+
| 0.1338 | 600 | 0.0031 | - |
|
| 209 |
+
| 0.1450 | 650 | 0.0011 | - |
|
| 210 |
+
| 0.1561 | 700 | 0.0038 | - |
|
| 211 |
+
| 0.1673 | 750 | 0.0001 | - |
|
| 212 |
+
| 0.1784 | 800 | 0.0005 | - |
|
| 213 |
+
| 0.1896 | 850 | 0.0019 | - |
|
| 214 |
+
| 0.2007 | 900 | 0.0016 | - |
|
| 215 |
+
| 0.2119 | 950 | 0.0001 | - |
|
| 216 |
+
| 0.2230 | 1000 | 0.0014 | - |
|
| 217 |
+
| 0.2342 | 1050 | 0.0022 | - |
|
| 218 |
+
| 0.2453 | 1100 | 0.0021 | - |
|
| 219 |
+
| 0.2565 | 1150 | 0.0018 | - |
|
| 220 |
+
| 0.2676 | 1200 | 0.0002 | - |
|
| 221 |
+
| 0.2788 | 1250 | 0.0 | - |
|
| 222 |
+
| 0.2899 | 1300 | 0.0019 | - |
|
| 223 |
+
| 0.3011 | 1350 | 0.0 | - |
|
| 224 |
+
| 0.3122 | 1400 | 0.0 | - |
|
| 225 |
+
| 0.3234 | 1450 | 0.0036 | - |
|
| 226 |
+
| 0.3345 | 1500 | 0.0 | - |
|
| 227 |
+
| 0.3457 | 1550 | 0.0 | - |
|
| 228 |
+
| 0.3568 | 1600 | 0.0 | - |
|
| 229 |
+
| 0.3680 | 1650 | 0.0 | - |
|
| 230 |
+
| 0.3791 | 1700 | 0.0 | - |
|
| 231 |
+
| 0.3903 | 1750 | 0.0018 | - |
|
| 232 |
+
| 0.4014 | 1800 | 0.0001 | - |
|
| 233 |
+
| 0.4126 | 1850 | 0.0017 | - |
|
| 234 |
+
| 0.4237 | 1900 | 0.0 | - |
|
| 235 |
+
| 0.4349 | 1950 | 0.0 | - |
|
| 236 |
+
| 0.4460 | 2000 | 0.0 | - |
|
| 237 |
+
| 0.4572 | 2050 | 0.0035 | - |
|
| 238 |
+
| 0.4683 | 2100 | 0.0034 | - |
|
| 239 |
+
| 0.4795 | 2150 | 0.0036 | - |
|
| 240 |
+
| 0.4906 | 2200 | 0.0017 | - |
|
| 241 |
+
| 0.5018 | 2250 | 0.0056 | - |
|
| 242 |
+
| 0.5129 | 2300 | 0.0006 | - |
|
| 243 |
+
| 0.5241 | 2350 | 0.0 | - |
|
| 244 |
+
| 0.5352 | 2400 | 0.0 | - |
|
| 245 |
+
| 0.5464 | 2450 | 0.0 | - |
|
| 246 |
+
| 0.5575 | 2500 | 0.0016 | - |
|
| 247 |
+
| 0.5687 | 2550 | 0.0014 | - |
|
| 248 |
+
| 0.5798 | 2600 | 0.0 | - |
|
| 249 |
+
| 0.5910 | 2650 | 0.0012 | - |
|
| 250 |
+
| 0.6021 | 2700 | 0.0001 | - |
|
| 251 |
+
| 0.6133 | 2750 | 0.0 | - |
|
| 252 |
+
| 0.6244 | 2800 | 0.0 | - |
|
| 253 |
+
| 0.6356 | 2850 | 0.0 | - |
|
| 254 |
+
| 0.6467 | 2900 | 0.0 | - |
|
| 255 |
+
| 0.6579 | 2950 | 0.0 | - |
|
| 256 |
+
| 0.6690 | 3000 | 0.0016 | - |
|
| 257 |
+
| 0.6802 | 3050 | 0.0 | - |
|
| 258 |
+
| 0.6913 | 3100 | 0.0 | - |
|
| 259 |
+
| 0.7025 | 3150 | 0.0 | - |
|
| 260 |
+
| 0.7136 | 3200 | 0.0017 | - |
|
| 261 |
+
| 0.7248 | 3250 | 0.0012 | - |
|
| 262 |
+
| 0.7360 | 3300 | 0.0002 | - |
|
| 263 |
+
| 0.7471 | 3350 | 0.0 | - |
|
| 264 |
+
| 0.7583 | 3400 | 0.0 | - |
|
| 265 |
+
| 0.7694 | 3450 | 0.0 | - |
|
| 266 |
+
| 0.7806 | 3500 | 0.0 | - |
|
| 267 |
+
| 0.7917 | 3550 | 0.0 | - |
|
| 268 |
+
| 0.8029 | 3600 | 0.0 | - |
|
| 269 |
+
| 0.8140 | 3650 | 0.0 | - |
|
| 270 |
+
| 0.8252 | 3700 | 0.0 | - |
|
| 271 |
+
| 0.8363 | 3750 | 0.0 | - |
|
| 272 |
+
| 0.8475 | 3800 | 0.0 | - |
|
| 273 |
+
| 0.8586 | 3850 | 0.0 | - |
|
| 274 |
+
| 0.8698 | 3900 | 0.0 | - |
|
| 275 |
+
| 0.8809 | 3950 | 0.0 | - |
|
| 276 |
+
| 0.8921 | 4000 | 0.0 | - |
|
| 277 |
+
| 0.9032 | 4050 | 0.0 | - |
|
| 278 |
+
| 0.9144 | 4100 | 0.0 | - |
|
| 279 |
+
| 0.9255 | 4150 | 0.0 | - |
|
| 280 |
+
| 0.9367 | 4200 | 0.0 | - |
|
| 281 |
+
| 0.9478 | 4250 | 0.0 | - |
|
| 282 |
+
| 0.9590 | 4300 | 0.0 | - |
|
| 283 |
+
| 0.9701 | 4350 | 0.0 | - |
|
| 284 |
+
| 0.9813 | 4400 | 0.0 | - |
|
| 285 |
+
| 0.9924 | 4450 | 0.0 | - |
|
| 286 |
+
| 1.0036 | 4500 | 0.0 | - |
|
| 287 |
+
| 1.0147 | 4550 | 0.0 | - |
|
| 288 |
+
| 1.0259 | 4600 | 0.0 | - |
|
| 289 |
+
| 1.0370 | 4650 | 0.0 | - |
|
| 290 |
+
| 1.0482 | 4700 | 0.0 | - |
|
| 291 |
+
| 1.0593 | 4750 | 0.0 | - |
|
| 292 |
+
| 1.0705 | 4800 | 0.0 | - |
|
| 293 |
+
| 1.0816 | 4850 | 0.0 | - |
|
| 294 |
+
| 1.0928 | 4900 | 0.0 | - |
|
| 295 |
+
| 1.1039 | 4950 | 0.0 | - |
|
| 296 |
+
| 1.1151 | 5000 | 0.0 | - |
|
| 297 |
+
| 1.1262 | 5050 | 0.0 | - |
|
| 298 |
+
| 1.1374 | 5100 | 0.0 | - |
|
| 299 |
+
| 1.1485 | 5150 | 0.0 | - |
|
| 300 |
+
| 1.1597 | 5200 | 0.0 | - |
|
| 301 |
+
| 1.1708 | 5250 | 0.0 | - |
|
| 302 |
+
| 1.1820 | 5300 | 0.0 | - |
|
| 303 |
+
| 1.1931 | 5350 | 0.0 | - |
|
| 304 |
+
| 1.2043 | 5400 | 0.0 | - |
|
| 305 |
+
| 1.2154 | 5450 | 0.0 | - |
|
| 306 |
+
| 1.2266 | 5500 | 0.0 | - |
|
| 307 |
+
| 1.2377 | 5550 | 0.0 | - |
|
| 308 |
+
| 1.2489 | 5600 | 0.0 | - |
|
| 309 |
+
| 1.2600 | 5650 | 0.0 | - |
|
| 310 |
+
| 1.2712 | 5700 | 0.0 | - |
|
| 311 |
+
| 1.2823 | 5750 | 0.0 | - |
|
| 312 |
+
| 1.2935 | 5800 | 0.0 | - |
|
| 313 |
+
| 1.3046 | 5850 | 0.0 | - |
|
| 314 |
+
| 1.3158 | 5900 | 0.0 | - |
|
| 315 |
+
| 1.3269 | 5950 | 0.0 | - |
|
| 316 |
+
| 1.3381 | 6000 | 0.0 | - |
|
| 317 |
+
| 1.3492 | 6050 | 0.0 | - |
|
| 318 |
+
| 1.3604 | 6100 | 0.0 | - |
|
| 319 |
+
| 1.3715 | 6150 | 0.0 | - |
|
| 320 |
+
| 1.3827 | 6200 | 0.0 | - |
|
| 321 |
+
| 1.3938 | 6250 | 0.0 | - |
|
| 322 |
+
| 1.4050 | 6300 | 0.0 | - |
|
| 323 |
+
| 1.4161 | 6350 | 0.0 | - |
|
| 324 |
+
| 1.4273 | 6400 | 0.0 | - |
|
| 325 |
+
| 1.4384 | 6450 | 0.0 | - |
|
| 326 |
+
| 1.4496 | 6500 | 0.0 | - |
|
| 327 |
+
| 1.4607 | 6550 | 0.0 | - |
|
| 328 |
+
| 1.4719 | 6600 | 0.0 | - |
|
| 329 |
+
| 1.4831 | 6650 | 0.0 | - |
|
| 330 |
+
| 1.4942 | 6700 | 0.0 | - |
|
| 331 |
+
| 1.5054 | 6750 | 0.0 | - |
|
| 332 |
+
| 1.5165 | 6800 | 0.0 | - |
|
| 333 |
+
| 1.5277 | 6850 | 0.0 | - |
|
| 334 |
+
| 1.5388 | 6900 | 0.0 | - |
|
| 335 |
+
| 1.5500 | 6950 | 0.0 | - |
|
| 336 |
+
| 1.5611 | 7000 | 0.0 | - |
|
| 337 |
+
| 1.5723 | 7050 | 0.0 | - |
|
| 338 |
+
| 1.5834 | 7100 | 0.0 | - |
|
| 339 |
+
| 1.5946 | 7150 | 0.0 | - |
|
| 340 |
+
| 1.6057 | 7200 | 0.0 | - |
|
| 341 |
+
| 1.6169 | 7250 | 0.0 | - |
|
| 342 |
+
| 1.6280 | 7300 | 0.0 | - |
|
| 343 |
+
| 1.6392 | 7350 | 0.0 | - |
|
| 344 |
+
| 1.6503 | 7400 | 0.0 | - |
|
| 345 |
+
| 1.6615 | 7450 | 0.0 | - |
|
| 346 |
+
| 1.6726 | 7500 | 0.0 | - |
|
| 347 |
+
| 1.6838 | 7550 | 0.0 | - |
|
| 348 |
+
| 1.6949 | 7600 | 0.0 | - |
|
| 349 |
+
| 1.7061 | 7650 | 0.0 | - |
|
| 350 |
+
| 1.7172 | 7700 | 0.0 | - |
|
| 351 |
+
| 1.7284 | 7750 | 0.0 | - |
|
| 352 |
+
| 1.7395 | 7800 | 0.0 | - |
|
| 353 |
+
| 1.7507 | 7850 | 0.0 | - |
|
| 354 |
+
| 1.7618 | 7900 | 0.0 | - |
|
| 355 |
+
| 1.7730 | 7950 | 0.0 | - |
|
| 356 |
+
| 1.7841 | 8000 | 0.0 | - |
|
| 357 |
+
| 1.7953 | 8050 | 0.0 | - |
|
| 358 |
+
| 1.8064 | 8100 | 0.0 | - |
|
| 359 |
+
| 1.8176 | 8150 | 0.0 | - |
|
| 360 |
+
| 1.8287 | 8200 | 0.0 | - |
|
| 361 |
+
| 1.8399 | 8250 | 0.0 | - |
|
| 362 |
+
| 1.8510 | 8300 | 0.0 | - |
|
| 363 |
+
| 1.8622 | 8350 | 0.0 | - |
|
| 364 |
+
| 1.8733 | 8400 | 0.0 | - |
|
| 365 |
+
| 1.8845 | 8450 | 0.0 | - |
|
| 366 |
+
| 1.8956 | 8500 | 0.0 | - |
|
| 367 |
+
| 1.9068 | 8550 | 0.0 | - |
|
| 368 |
+
| 1.9179 | 8600 | 0.0 | - |
|
| 369 |
+
| 1.9291 | 8650 | 0.0 | - |
|
| 370 |
+
| 1.9402 | 8700 | 0.0 | - |
|
| 371 |
+
| 1.9514 | 8750 | 0.0 | - |
|
| 372 |
+
| 1.9625 | 8800 | 0.0 | - |
|
| 373 |
+
| 1.9737 | 8850 | 0.0 | - |
|
| 374 |
+
| 1.9848 | 8900 | 0.0 | - |
|
| 375 |
+
| 1.9960 | 8950 | 0.0 | - |
|
| 376 |
+
|
| 377 |
+
### Framework Versions
|
| 378 |
+
- Python: 3.11.12
|
| 379 |
+
- SetFit: 1.1.3
|
| 380 |
+
- Sentence Transformers: 4.1.0
|
| 381 |
+
- Transformers: 4.51.3
|
| 382 |
+
- PyTorch: 2.7.0
|
| 383 |
+
- Datasets: 2.19.2
|
| 384 |
+
- Tokenizers: 0.21.1
|
| 385 |
+
|
| 386 |
+
## Citation
|
| 387 |
+
|
| 388 |
+
### BibTeX
|
| 389 |
+
```bibtex
|
| 390 |
+
@article{https://doi.org/10.48550/arxiv.2209.11055,
|
| 391 |
+
doi = {10.48550/ARXIV.2209.11055},
|
| 392 |
+
url = {https://arxiv.org/abs/2209.11055},
|
| 393 |
+
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
|
| 394 |
+
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
|
| 395 |
+
title = {Efficient Few-Shot Learning Without Prompts},
|
| 396 |
+
publisher = {arXiv},
|
| 397 |
+
year = {2022},
|
| 398 |
+
copyright = {Creative Commons Attribution 4.0 International}
|
| 399 |
+
}
|
| 400 |
+
```
|
| 401 |
+
|
| 402 |
+
<!--
|
| 403 |
+
## Glossary
|
| 404 |
+
|
| 405 |
+
*Clearly define terms in order to be accessible across audiences.*
|
| 406 |
+
-->
|
| 407 |
+
|
| 408 |
+
<!--
|
| 409 |
+
## Model Card Authors
|
| 410 |
+
|
| 411 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
| 412 |
+
-->
|
| 413 |
+
|
| 414 |
+
<!--
|
| 415 |
+
## Model Card Contact
|
| 416 |
+
|
| 417 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
| 418 |
+
-->
|
config.json
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"BertModel"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.1,
|
| 6 |
+
"classifier_dropout": null,
|
| 7 |
+
"hidden_act": "gelu",
|
| 8 |
+
"hidden_dropout_prob": 0.1,
|
| 9 |
+
"hidden_size": 768,
|
| 10 |
+
"initializer_range": 0.02,
|
| 11 |
+
"intermediate_size": 3072,
|
| 12 |
+
"layer_norm_eps": 1e-12,
|
| 13 |
+
"max_position_embeddings": 512,
|
| 14 |
+
"model_type": "bert",
|
| 15 |
+
"num_attention_heads": 12,
|
| 16 |
+
"num_hidden_layers": 12,
|
| 17 |
+
"pad_token_id": 0,
|
| 18 |
+
"position_embedding_type": "absolute",
|
| 19 |
+
"torch_dtype": "float32",
|
| 20 |
+
"transformers_version": "4.51.3",
|
| 21 |
+
"type_vocab_size": 2,
|
| 22 |
+
"use_cache": true,
|
| 23 |
+
"vocab_size": 30522
|
| 24 |
+
}
|
config_sentence_transformers.json
ADDED
|
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"__version__": {
|
| 3 |
+
"sentence_transformers": "4.1.0",
|
| 4 |
+
"transformers": "4.51.3",
|
| 5 |
+
"pytorch": "2.7.0"
|
| 6 |
+
},
|
| 7 |
+
"prompts": {},
|
| 8 |
+
"default_prompt_name": null,
|
| 9 |
+
"similarity_fn_name": "cosine"
|
| 10 |
+
}
|
config_setfit.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"normalize_embeddings": false,
|
| 3 |
+
"labels": null
|
| 4 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d7272d77af10a7583d669cc13e72e56c5d58fbf64a10dcfaee90b9fcb988f0c
|
| 3 |
+
size 437951328
|
model_head.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9db26e8a9e044a0144e2a68eca28de7a4eb6b524040c67408a75f62db8e9fbc6
|
| 3 |
+
size 7007
|
modules.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"idx": 0,
|
| 4 |
+
"name": "0",
|
| 5 |
+
"path": "",
|
| 6 |
+
"type": "sentence_transformers.models.Transformer"
|
| 7 |
+
},
|
| 8 |
+
{
|
| 9 |
+
"idx": 1,
|
| 10 |
+
"name": "1",
|
| 11 |
+
"path": "1_Pooling",
|
| 12 |
+
"type": "sentence_transformers.models.Pooling"
|
| 13 |
+
}
|
| 14 |
+
]
|
sentence_bert_config.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"max_seq_length": 512,
|
| 3 |
+
"do_lower_case": false
|
| 4 |
+
}
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,65 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"1": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"2": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"3": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"4": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": true,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_basic_tokenize": true,
|
| 47 |
+
"do_lower_case": true,
|
| 48 |
+
"extra_special_tokens": {},
|
| 49 |
+
"mask_token": "[MASK]",
|
| 50 |
+
"max_length": 512,
|
| 51 |
+
"model_max_length": 512,
|
| 52 |
+
"never_split": null,
|
| 53 |
+
"pad_to_multiple_of": null,
|
| 54 |
+
"pad_token": "[PAD]",
|
| 55 |
+
"pad_token_type_id": 0,
|
| 56 |
+
"padding_side": "right",
|
| 57 |
+
"sep_token": "[SEP]",
|
| 58 |
+
"stride": 0,
|
| 59 |
+
"strip_accents": null,
|
| 60 |
+
"tokenize_chinese_chars": true,
|
| 61 |
+
"tokenizer_class": "BertTokenizer",
|
| 62 |
+
"truncation_side": "right",
|
| 63 |
+
"truncation_strategy": "longest_first",
|
| 64 |
+
"unk_token": "[UNK]"
|
| 65 |
+
}
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|