mini1013 commited on
Commit
1032569
·
verified ·
1 Parent(s): 5aebce0

Push model using huggingface_hub.

Browse files
README.md CHANGED
The diff for this file is too large to render. See raw diff
 
config_setfit.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "labels": null,
3
- "normalize_embeddings": false
4
  }
 
1
  {
2
+ "normalize_embeddings": false,
3
+ "labels": null
4
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:419cb9d79dded54e974fa646a23d90fb5aff390012bcbabf9bc18866a2878a54
3
  size 1112197096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c35ea800689fd3b1c68f19b06cdfa643eaa154f623fab5a15fb0f289ca13cbe9
3
  size 1112197096
model_head.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:552abd35a4ee26a72ea963ea3d94a4d98c0479a7254a3ee77f2aeb770d2df6b1
3
  size 764679
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e2ec9c660bbd2272105509523162bbf5e06bc80d6fa7a58acec35ceed125903
3
  size 764679
tokenizer_config.json CHANGED
@@ -47,9 +47,16 @@
47
  "eos_token": "</s>",
48
  "extra_special_tokens": {},
49
  "mask_token": "<mask>",
 
50
  "model_max_length": 512,
 
51
  "pad_token": "<pad>",
 
 
52
  "sep_token": "</s>",
53
- "tokenizer_class": "XLMRobertaTokenizer",
 
 
 
54
  "unk_token": "<unk>"
55
  }
 
47
  "eos_token": "</s>",
48
  "extra_special_tokens": {},
49
  "mask_token": "<mask>",
50
+ "max_length": 512,
51
  "model_max_length": 512,
52
+ "pad_to_multiple_of": null,
53
  "pad_token": "<pad>",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
  "sep_token": "</s>",
57
+ "stride": 0,
58
+ "tokenizer_class": "XLMRobertaTokenizerFast",
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
  "unk_token": "<unk>"
62
  }