Gnider commited on
Commit
17bff1b
·
verified ·
1 Parent(s): 1f39436

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "blinoff/roberta-base-russian-v0",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
@@ -30,9 +30,7 @@
30
  "16": "LABEL_16",
31
  "17": "LABEL_17",
32
  "18": "LABEL_18",
33
- "19": "LABEL_19",
34
- "20": "LABEL_20",
35
- "21": "LABEL_21"
36
  },
37
  "initializer_range": 0.02,
38
  "intermediate_size": 3072,
@@ -50,8 +48,6 @@
50
  "LABEL_18": 18,
51
  "LABEL_19": 19,
52
  "LABEL_2": 2,
53
- "LABEL_20": 20,
54
- "LABEL_21": 21,
55
  "LABEL_3": 3,
56
  "LABEL_4": 4,
57
  "LABEL_5": 5,
@@ -60,17 +56,16 @@
60
  "LABEL_8": 8,
61
  "LABEL_9": 9
62
  },
63
- "layer_norm_eps": 1e-12,
64
  "max_position_embeddings": 514,
65
  "model_type": "roberta",
66
  "num_attention_heads": 12,
67
  "num_hidden_layers": 12,
68
  "pad_token_id": 1,
69
  "position_embedding_type": "absolute",
70
- "problem_type": "single_label_classification",
71
  "torch_dtype": "float32",
72
  "transformers_version": "4.41.2",
73
  "type_vocab_size": 1,
74
  "use_cache": true,
75
- "vocab_size": 50021
76
  }
 
1
  {
2
+ "_name_or_path": "roberta-base",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
 
30
  "16": "LABEL_16",
31
  "17": "LABEL_17",
32
  "18": "LABEL_18",
33
+ "19": "LABEL_19"
 
 
34
  },
35
  "initializer_range": 0.02,
36
  "intermediate_size": 3072,
 
48
  "LABEL_18": 18,
49
  "LABEL_19": 19,
50
  "LABEL_2": 2,
 
 
51
  "LABEL_3": 3,
52
  "LABEL_4": 4,
53
  "LABEL_5": 5,
 
56
  "LABEL_8": 8,
57
  "LABEL_9": 9
58
  },
59
+ "layer_norm_eps": 1e-05,
60
  "max_position_embeddings": 514,
61
  "model_type": "roberta",
62
  "num_attention_heads": 12,
63
  "num_hidden_layers": 12,
64
  "pad_token_id": 1,
65
  "position_embedding_type": "absolute",
 
66
  "torch_dtype": "float32",
67
  "transformers_version": "4.41.2",
68
  "type_vocab_size": 1,
69
  "use_cache": true,
70
+ "vocab_size": 50265
71
  }
label_encoder.joblib CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5fa40ddda8e6096d543a261ff2769f0860b599927f2d9e0b41d12239605f8a86
3
- size 1208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a409954c3cfbacdbd5a19099a851a93f386b0f467c894374e6e016095cbdd790
3
+ size 1159
merges.txt CHANGED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54b9bcfaf8c65161ee58537c6181e6a924d33ba239ceb96521e83105b10410c6
3
- size 497924776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37bcbfa3df091fc7184900bcea1d379ed1e68b83a7c5ced857dcfd7269aaa387
3
+ size 498668192
special_tokens_map.json CHANGED
@@ -1,25 +1,7 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "cls_token": {
10
- "content": "<s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "eos_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "mask_token": {
24
  "content": "<mask>",
25
  "lstrip": true,
@@ -27,25 +9,7 @@
27
  "rstrip": false,
28
  "single_word": false
29
  },
30
- "pad_token": {
31
- "content": "<pad>",
32
- "lstrip": false,
33
- "normalized": true,
34
- "rstrip": false,
35
- "single_word": false
36
- },
37
- "sep_token": {
38
- "content": "</s>",
39
- "lstrip": false,
40
- "normalized": true,
41
- "rstrip": false,
42
- "single_word": false
43
- },
44
- "unk_token": {
45
- "content": "<unk>",
46
- "lstrip": false,
47
- "normalized": true,
48
- "rstrip": false,
49
- "single_word": false
50
- }
51
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "mask_token": {
6
  "content": "<mask>",
7
  "lstrip": true,
 
9
  "rstrip": false,
10
  "single_word": false
11
  },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -33,7 +33,7 @@
33
  "single_word": false,
34
  "special": true
35
  },
36
- "4": {
37
  "content": "<mask>",
38
  "lstrip": true,
39
  "normalized": false,
@@ -48,7 +48,6 @@
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
- "max_len": 512,
52
  "model_max_length": 512,
53
  "pad_token": "<pad>",
54
  "sep_token": "</s>",
 
33
  "single_word": false,
34
  "special": true
35
  },
36
+ "50264": {
37
  "content": "<mask>",
38
  "lstrip": true,
39
  "normalized": false,
 
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
 
51
  "model_max_length": 512,
52
  "pad_token": "<pad>",
53
  "sep_token": "</s>",
vocab.json CHANGED
The diff for this file is too large to render. See raw diff