mini1013 commited on
Commit
92d8628
·
verified ·
1 Parent(s): 142e653

Push model using huggingface_hub.

Browse files
README.md CHANGED
The diff for this file is too large to render. See raw diff
 
config_setfit.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "labels": null,
3
- "normalize_embeddings": false
4
  }
 
1
  {
2
+ "normalize_embeddings": false,
3
+ "labels": null
4
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71c1d86840d029290a8b85e2b3b6b9f9c7564037134b8f95a1ab89b2e783c0af
3
  size 1112197096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1befd7460067ebd4b7232176d4a98a6ba8072e21554458f0eaf6e46413c53513
3
  size 1112197096
model_head.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ecd9f19b413cfa636dca48670587378caa1bde12333667602ee4e8c57d40cf9
3
  size 1855007
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9df98af145fad89e91291590317935d7560a75953b58a51f8f3095e9efab1885
3
  size 1855007
tokenizer_config.json CHANGED
@@ -47,9 +47,16 @@
47
  "eos_token": "</s>",
48
  "extra_special_tokens": {},
49
  "mask_token": "<mask>",
 
50
  "model_max_length": 512,
 
51
  "pad_token": "<pad>",
 
 
52
  "sep_token": "</s>",
53
- "tokenizer_class": "XLMRobertaTokenizer",
 
 
 
54
  "unk_token": "<unk>"
55
  }
 
47
  "eos_token": "</s>",
48
  "extra_special_tokens": {},
49
  "mask_token": "<mask>",
50
+ "max_length": 512,
51
  "model_max_length": 512,
52
+ "pad_to_multiple_of": null,
53
  "pad_token": "<pad>",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
  "sep_token": "</s>",
57
+ "stride": 0,
58
+ "tokenizer_class": "XLMRobertaTokenizerFast",
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
  "unk_token": "<unk>"
62
  }