loganh274 commited on
Commit
48c2549
·
verified ·
1 Parent(s): 3c33138

Push model using huggingface_hub.

Browse files
1_Pooling/config.json CHANGED
@@ -1,10 +1,10 @@
1
- {
2
- "word_embedding_dimension": 768,
3
- "pooling_mode_cls_token": true,
4
- "pooling_mode_mean_tokens": false,
5
- "pooling_mode_max_tokens": false,
6
- "pooling_mode_mean_sqrt_len_tokens": false,
7
- "pooling_mode_weightedmean_tokens": false,
8
- "pooling_mode_lasttoken": false,
9
- "include_prompt": true
10
  }
 
1
+ {
2
+ "word_embedding_dimension": 384,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
  }
README.md CHANGED
@@ -1,228 +1,130 @@
1
- ---
2
- tags:
3
- - setfit
4
- - sentence-transformers
5
- - text-classification
6
- - generated_from_setfit_trainer
7
- widget:
8
- - text: Great to hear that this basic (on the face of it) functionality is being implemented,
9
- it needs to be done thanks.
10
- - text: Greg, I totally agree with you. Regretting moving from Reckon Desktop. Xero
11
- has created twice as much work for me, without the basic functionality that even
12
- Open Source packages have. We have one customer that may have 100 different site
13
- addresses in one year, and Xero is creating a nightmare.
14
- - text: Agree with Benjamin on 16.08.22 - this is such a basic feature. Time to move
15
- a batch of clients back to Sage - hugely embarrassing and a complete waste of
16
- time.
17
- - text: Game changer for our construction projects. Invoicing for different sites
18
- is a breeze now.
19
- - text: "Personally can't believe this functionality hasn't been created given the\
20
- \ amount of businesses that have multiple shipments locations for the same business.\r\
21
- \n\r\ncustomers want a statement for all stores combined and to manually type\
22
- \ in an address each invoice just show PO's can be matched and paid is silly.\
23
- \ Given Xero generally improves efficiency though automation rather than manual\
24
- \ entry"
25
- metrics:
26
- - accuracy
27
- pipeline_tag: text-classification
28
- library_name: setfit
29
- inference: true
30
- base_model: BAAI/bge-base-en-v1.5
31
- model-index:
32
- - name: SetFit with BAAI/bge-base-en-v1.5
33
- results:
34
- - task:
35
- type: text-classification
36
- name: Text Classification
37
- dataset:
38
- name: Unknown
39
- type: unknown
40
- split: test
41
- metrics:
42
- - type: accuracy
43
- value: 0.87
44
- name: Accuracy
45
- ---
46
-
47
- # SetFit with BAAI/bge-base-en-v1.5
48
-
49
- This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [BAAI/bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
50
-
51
- The model has been trained using an efficient few-shot learning technique that involves:
52
-
53
- 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
54
- 2. Training a classification head with features from the fine-tuned Sentence Transformer.
55
-
56
- ## Model Details
57
-
58
- ### Model Description
59
- - **Model Type:** SetFit
60
- - **Sentence Transformer body:** [BAAI/bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5)
61
- - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
62
- - **Maximum Sequence Length:** 512 tokens
63
- - **Number of Classes:** 5 classes
64
- <!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
65
- <!-- - **Language:** Unknown -->
66
- <!-- - **License:** Unknown -->
67
-
68
- ### Model Sources
69
-
70
- - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
71
- - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
72
- - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
73
-
74
- ### Model Labels
75
- | Label | Examples |
76
- |:------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
77
- | 3 | <ul><li>"Glad we don't have to leave for Quickbooks now, this update should solve the issues we were having."</li><li>'This is a solid improvement. Being able to add a secondary address for delivery without editing the main contact is useful.'</li><li>'Good that future company returns can be done, but surely it isnt that hard to build one for trusts and partnerships too?'</li></ul> |
78
- | 0 | <ul><li>'How many more years do you require to have this issue resolved? you may end up loosing clients to competitors over this issue. This issue become more crucial as time goes on.'</li><li>'I asked for this "Improvement" over 10 years ago in this thread (or a previous one). I don\'t believe anyone at Xero actually takes any notice of this forum. its just here so you can vent and think that something will get done. Xero is far too busy developing "New" features to entice new clients rather than supporting existing clients who they have already "Captured"'</li><li>'Too little too late for me Xero. The last price rise was the last straw for me - 80% increase over 3 years. Encouraged me to look at the opposition and realise that I could get significantly more features for half the price. Look back over this issue - a clear indication of how a significant feature gap receives so little attention from Xero. Meanwhile the big news is that the main menu is going to be fiddled with yet again. Really! Too much resource spent rearranging deck chairs I would suggest. Bye.'</li></ul> |
79
- | 1 | <ul><li>"I'd like very much to trial early payment discounts with my clients, but despite using some pretty sophisticated accounting software - by the way, that's your product Xero - I can't add this simple thing to an invoice. Wow! \r\n\r\nIf this functionality hasn't been implemented in 12 years, there must be a reason."</li><li>"@Scott Osburne you technically CAN [CAPS] tab through the fields in the payment pop-up, but to change the date, you have to press space or enter when the date field is highlighted, then tab several times to get to the grid of dates (with the mouse cursor NOT [CAPS] over the grid because if it's over the grid then tab exits the grid), use the arrow keys to find the one you want, then press enter or space (if you press tab it goes to the next field instead of changing the date) when you have the right one highlighted, AND [CAPS] THEN [CAPS] YOU [CAPS] HAVE [CAPS] TO SHIFT [CAPS]-TAB [CAPS] BACK [CAPS] FOUR [CAPS] TIMES [CAPS] TO THE [CAPS] ACCOUNT [CAPS] FIELD [CAPS] because the focus has dissipated and both tab and shift-tab move the focus to the Add button instead of the focus moving in a logical order (WCAG [CAPS] success criterion 2.4.3 failure).\r\n\r\nIf you could enter the date instead of having to tab and arrow etc it would be significantly fewer keypresses, even without the focus order failure."</li><li>'Xero you are bleeding customers due to increased fees. That you aren’t prioritising upgrades that are standard features for your competitors is shocking. Provide a timetable for beta and full deployment schedule for each idea so that we can actually see it working its way through the apps dev pipeline!! [EMPHASIS [CAPS]]'</li></ul> |
80
- | 4 | <ul><li>'Great to hear that this basic (on the face of it) functionality is being implemented, it needs to be done thanks.'</li><li>'Thank you Xero! This worked today!! [EMPHASIS [CAPS]] What a relief to have this feature back. Now that item codes are pulled in for billable expenses, the new invoicing actually represents an improvement for my purposes. Thank you for addressing this request.'</li><li>'Hey Kelly, thank you for the updates. We appreciate you!\r\n\r\nIs there a quarter on your roadmap for when this feature has been allocated to?\r\n\r\nAlso, just checking if this also means that the address selected for an invoice will persist with that invoice, even after a contact adds a new address in the future?'</li></ul> |
81
- | 2 | <ul><li>"Another way Xero could design this is to have multiple sub contacts under a main contact. For example McDonald's is the main contact but then you have McDonald's - Brisbane CBD [CAPS] as a sub contact. Just a thought.. [ELONGATED [CAPS]]"</li><li>'Agreed with all. Please add an option to opt out of email notification. Thank you!'</li><li>"We are a school using Xero we invoice annually and parents can pay monthly so it is really necessary for the online payment method to be linked to statements as well. Doesn't make sense that this is still unavailable."</li></ul> |
82
-
83
- ## Evaluation
84
-
85
- ### Metrics
86
- | Label | Accuracy |
87
- |:--------|:---------|
88
- | **all** | 0.87 |
89
-
90
- ## Uses
91
-
92
- ### Direct Use for Inference
93
-
94
- First install the SetFit library:
95
-
96
- ```bash
97
- pip install setfit
98
- ```
99
-
100
- Then you can load this model and run inference.
101
-
102
- ```python
103
- from setfit import SetFitModel
104
-
105
- # Download from the 🤗 Hub
106
- model = SetFitModel.from_pretrained("setfit_model_id")
107
- # Run inference
108
- preds = model("Game changer for our construction projects. Invoicing for different sites is a breeze now.")
109
- ```
110
-
111
- <!--
112
- ### Downstream Use
113
-
114
- *List how someone could finetune this model on their own dataset.*
115
- -->
116
-
117
- <!--
118
- ### Out-of-Scope Use
119
-
120
- *List how the model may foreseeably be misused and address what users ought not to do with the model.*
121
- -->
122
-
123
- <!--
124
- ## Bias, Risks and Limitations
125
-
126
- *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
127
- -->
128
-
129
- <!--
130
- ### Recommendations
131
-
132
- *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
133
- -->
134
-
135
- ## Training Details
136
-
137
- ### Training Set Metrics
138
- | Training set | Min | Median | Max |
139
- |:-------------|:----|:-------|:----|
140
- | Word count | 1 | 50.106 | 701 |
141
-
142
- | Label | Training Sample Count |
143
- |:------|:----------------------|
144
- | 0 | 100 |
145
- | 1 | 100 |
146
- | 2 | 100 |
147
- | 3 | 100 |
148
- | 4 | 100 |
149
-
150
- ### Training Hyperparameters
151
- - batch_size: (16, 16)
152
- - num_epochs: (1, 16)
153
- - max_steps: -1
154
- - sampling_strategy: oversampling
155
- - num_iterations: 10
156
- - body_learning_rate: (2e-05, 1e-05)
157
- - head_learning_rate: 0.01
158
- - loss: CosineSimilarityLoss
159
- - distance_metric: cosine_distance
160
- - margin: 0.25
161
- - end_to_end: False
162
- - use_amp: False
163
- - warmup_proportion: 0.1
164
- - l2_weight: 0.01
165
- - seed: 42
166
- - eval_max_steps: -1
167
- - load_best_model_at_end: True
168
-
169
- ### Training Results
170
- | Epoch | Step | Training Loss | Validation Loss |
171
- |:------:|:----:|:-------------:|:---------------:|
172
- | 0.0016 | 1 | 0.2036 | - |
173
- | 0.08 | 50 | 0.2392 | - |
174
- | 0.16 | 100 | 0.2148 | - |
175
- | 0.24 | 150 | 0.1638 | - |
176
- | 0.32 | 200 | 0.1284 | - |
177
- | 0.4 | 250 | 0.1046 | - |
178
- | 0.48 | 300 | 0.0822 | - |
179
- | 0.56 | 350 | 0.07 | - |
180
- | 0.64 | 400 | 0.0457 | - |
181
- | 0.72 | 450 | 0.0328 | - |
182
- | 0.8 | 500 | 0.0213 | - |
183
- | 0.88 | 550 | 0.0137 | - |
184
- | 0.96 | 600 | 0.0139 | - |
185
- | 1.0 | 625 | - | 0.1038 |
186
-
187
- ### Framework Versions
188
- - Python: 3.11.9
189
- - SetFit: 1.1.3
190
- - Sentence Transformers: 5.2.0
191
- - Transformers: 4.57.3
192
- - PyTorch: 2.7.1+cu118
193
- - Datasets: 4.4.2
194
- - Tokenizers: 0.22.2
195
-
196
- ## Citation
197
-
198
- ### BibTeX
199
- ```bibtex
200
- @article{https://doi.org/10.48550/arxiv.2209.11055,
201
- doi = {10.48550/ARXIV.2209.11055},
202
- url = {https://arxiv.org/abs/2209.11055},
203
- author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
204
- keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
205
- title = {Efficient Few-Shot Learning Without Prompts},
206
- publisher = {arXiv},
207
- year = {2022},
208
- copyright = {Creative Commons Attribution 4.0 International}
209
- }
210
- ```
211
-
212
- <!--
213
- ## Glossary
214
-
215
- *Clearly define terms in order to be accessible across audiences.*
216
- -->
217
-
218
- <!--
219
- ## Model Card Authors
220
-
221
- *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
222
- -->
223
-
224
- <!--
225
- ## Model Card Contact
226
-
227
- *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
228
  -->
 
1
+ ---
2
+ tags:
3
+ - setfit
4
+ - sentence-transformers
5
+ - text-classification
6
+ - generated_from_setfit_trainer
7
+ widget: []
8
+ metrics:
9
+ - accuracy
10
+ pipeline_tag: text-classification
11
+ library_name: setfit
12
+ inference: true
13
+ ---
14
+
15
+ # SetFit
16
+
17
+ This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
18
+
19
+ The model has been trained using an efficient few-shot learning technique that involves:
20
+
21
+ 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
22
+ 2. Training a classification head with features from the fine-tuned Sentence Transformer.
23
+
24
+ ## Model Details
25
+
26
+ ### Model Description
27
+ - **Model Type:** SetFit
28
+ <!-- - **Sentence Transformer:** [Unknown](https://huggingface.co/unknown) -->
29
+ - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
30
+ - **Maximum Sequence Length:** 256 tokens
31
+ - **Number of Classes:** 5 classes
32
+ <!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
33
+ <!-- - **Language:** Unknown -->
34
+ <!-- - **License:** Unknown -->
35
+
36
+ ### Model Sources
37
+
38
+ - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
39
+ - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
40
+ - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
41
+
42
+ ## Uses
43
+
44
+ ### Direct Use for Inference
45
+
46
+ First install the SetFit library:
47
+
48
+ ```bash
49
+ pip install setfit
50
+ ```
51
+
52
+ Then you can load this model and run inference.
53
+
54
+ ```python
55
+ from setfit import SetFitModel
56
+
57
+ # Download from the 🤗 Hub
58
+ model = SetFitModel.from_pretrained("setfit_model_id")
59
+ # Run inference
60
+ preds = model("I loved the spiderman movie!")
61
+ ```
62
+
63
+ <!--
64
+ ### Downstream Use
65
+
66
+ *List how someone could finetune this model on their own dataset.*
67
+ -->
68
+
69
+ <!--
70
+ ### Out-of-Scope Use
71
+
72
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
73
+ -->
74
+
75
+ <!--
76
+ ## Bias, Risks and Limitations
77
+
78
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
79
+ -->
80
+
81
+ <!--
82
+ ### Recommendations
83
+
84
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
85
+ -->
86
+
87
+ ## Training Details
88
+
89
+ ### Framework Versions
90
+ - Python: 3.11.14
91
+ - SetFit: 1.1.3
92
+ - Sentence Transformers: 5.2.0
93
+ - Transformers: 4.57.5
94
+ - PyTorch: 2.9.1
95
+ - Datasets: 4.4.2
96
+ - Tokenizers: 0.22.2
97
+
98
+ ## Citation
99
+
100
+ ### BibTeX
101
+ ```bibtex
102
+ @article{https://doi.org/10.48550/arxiv.2209.11055,
103
+ doi = {10.48550/ARXIV.2209.11055},
104
+ url = {https://arxiv.org/abs/2209.11055},
105
+ author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
106
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
107
+ title = {Efficient Few-Shot Learning Without Prompts},
108
+ publisher = {arXiv},
109
+ year = {2022},
110
+ copyright = {Creative Commons Attribution 4.0 International}
111
+ }
112
+ ```
113
+
114
+ <!--
115
+ ## Glossary
116
+
117
+ *Clearly define terms in order to be accessible across audiences.*
118
+ -->
119
+
120
+ <!--
121
+ ## Model Card Authors
122
+
123
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
124
+ -->
125
+
126
+ <!--
127
+ ## Model Card Contact
128
+
129
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
130
  -->
config.json CHANGED
@@ -1,31 +1,25 @@
1
- {
2
- "architectures": [
3
- "BertModel"
4
- ],
5
- "attention_probs_dropout_prob": 0.1,
6
- "classifier_dropout": null,
7
- "dtype": "float32",
8
- "gradient_checkpointing": false,
9
- "hidden_act": "gelu",
10
- "hidden_dropout_prob": 0.1,
11
- "hidden_size": 768,
12
- "id2label": {
13
- "0": "LABEL_0"
14
- },
15
- "initializer_range": 0.02,
16
- "intermediate_size": 3072,
17
- "label2id": {
18
- "LABEL_0": 0
19
- },
20
- "layer_norm_eps": 1e-12,
21
- "max_position_embeddings": 512,
22
- "model_type": "bert",
23
- "num_attention_heads": 12,
24
- "num_hidden_layers": 12,
25
- "pad_token_id": 0,
26
- "position_embedding_type": "absolute",
27
- "transformers_version": "4.57.3",
28
- "type_vocab_size": 2,
29
- "use_cache": true,
30
- "vocab_size": 30522
31
- }
 
1
+ {
2
+ "architectures": [
3
+ "BertModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 1536,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.57.5",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 30522
25
+ }
 
 
 
 
 
 
config_sentence_transformers.json CHANGED
@@ -1,14 +1,14 @@
1
- {
2
- "__version__": {
3
- "sentence_transformers": "5.2.0",
4
- "transformers": "4.57.3",
5
- "pytorch": "2.7.1+cu118"
6
- },
7
- "model_type": "SentenceTransformer",
8
- "prompts": {
9
- "query": "",
10
- "document": ""
11
- },
12
- "default_prompt_name": null,
13
- "similarity_fn_name": "cosine"
14
  }
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "5.2.0",
4
+ "transformers": "4.57.5",
5
+ "pytorch": "2.9.1"
6
+ },
7
+ "model_type": "SentenceTransformer",
8
+ "prompts": {
9
+ "query": "",
10
+ "document": ""
11
+ },
12
+ "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
14
  }
config_setfit.json CHANGED
@@ -1,10 +1,10 @@
1
- {
2
- "normalize_embeddings": false,
3
- "labels": [
4
- 0,
5
- 1,
6
- 2,
7
- 3,
8
- 4
9
- ]
10
  }
 
1
+ {
2
+ "labels": [
3
+ 0,
4
+ 1,
5
+ 2,
6
+ 3,
7
+ 4
8
+ ],
9
+ "normalize_embeddings": false
10
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b85322091c0722e761ff13c31f84507915fee2f9a19c8a6280aa1c6691601c6
3
- size 437951328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18a663e92fec7aa0b54431fa5c10a1481be64df6b7f98defb81887c939f13c28
3
+ size 90864192
model_head.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8359dda036c661b2ec116a3d4c3dbed0e415d49b60a97c3e83ba1549aa6438bc
3
- size 31647
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3e83f7cb8466031df74b40f5ca1f8cc3e3d217bf29f4ff1b64b512bc01c5e0a
3
+ size 16287
modules.json CHANGED
@@ -1,20 +1,20 @@
1
- [
2
- {
3
- "idx": 0,
4
- "name": "0",
5
- "path": "",
6
- "type": "sentence_transformers.models.Transformer"
7
- },
8
- {
9
- "idx": 1,
10
- "name": "1",
11
- "path": "1_Pooling",
12
- "type": "sentence_transformers.models.Pooling"
13
- },
14
- {
15
- "idx": 2,
16
- "name": "2",
17
- "path": "2_Normalize",
18
- "type": "sentence_transformers.models.Normalize"
19
- }
20
  ]
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
  ]
sentence_bert_config.json CHANGED
@@ -1,4 +1,4 @@
1
- {
2
- "max_seq_length": 512,
3
- "do_lower_case": true
4
  }
 
1
+ {
2
+ "max_seq_length": 256,
3
+ "do_lower_case": false
4
  }
special_tokens_map.json CHANGED
@@ -1,37 +1,37 @@
1
- {
2
- "cls_token": {
3
- "content": "[CLS]",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "mask_token": {
10
- "content": "[MASK]",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "[PAD]",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "sep_token": {
24
- "content": "[SEP]",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
- "unk_token": {
31
- "content": "[UNK]",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false
36
- }
37
- }
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -1,58 +1,65 @@
1
- {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "100": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "101": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "102": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "103": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
- "clean_up_tokenization_spaces": true,
45
- "cls_token": "[CLS]",
46
- "do_basic_tokenize": true,
47
- "do_lower_case": true,
48
- "extra_special_tokens": {},
49
- "mask_token": "[MASK]",
50
- "model_max_length": 512,
51
- "never_split": null,
52
- "pad_token": "[PAD]",
53
- "sep_token": "[SEP]",
54
- "strip_accents": null,
55
- "tokenize_chinese_chars": true,
56
- "tokenizer_class": "BertTokenizer",
57
- "unk_token": "[UNK]"
58
- }
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "max_length": 128,
51
+ "model_max_length": 256,
52
+ "never_split": null,
53
+ "pad_to_multiple_of": null,
54
+ "pad_token": "[PAD]",
55
+ "pad_token_type_id": 0,
56
+ "padding_side": "right",
57
+ "sep_token": "[SEP]",
58
+ "stride": 0,
59
+ "strip_accents": null,
60
+ "tokenize_chinese_chars": true,
61
+ "tokenizer_class": "BertTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
+ "unk_token": "[UNK]"
65
+ }