bnunticha commited on
Commit
aca9d41
·
1 Parent(s): 78641f0

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -31,7 +31,7 @@
31
  "position_embedding_type": "absolute",
32
  "problem_type": "single_label_classification",
33
  "torch_dtype": "float32",
34
- "transformers_version": "4.34.0",
35
  "type_vocab_size": 1,
36
  "use_cache": true,
37
  "vocab_size": 25005
 
31
  "position_embedding_type": "absolute",
32
  "problem_type": "single_label_classification",
33
  "torch_dtype": "float32",
34
+ "transformers_version": "4.35.2",
35
  "type_vocab_size": 1,
36
  "use_cache": true,
37
  "vocab_size": 25005
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9de035e6fe518d729ca44e06caf3ce7a6e85dfdf3e672fe510085d61e7c64d4b
3
+ size 421007928
runs/Nov20_03-54-17_bb4c76e02cfc/events.out.tfevents.1700452467.bb4c76e02cfc.2506.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9e62d4eb09f6092d9ad4290bf7859a4efb951c726118a3c64911b2afe9818c8
3
+ size 8998
special_tokens_map.json CHANGED
@@ -7,7 +7,13 @@
7
  "bos_token": "<s>",
8
  "cls_token": "<s>",
9
  "eos_token": "</s>",
10
- "mask_token": "<mask>",
 
 
 
 
 
 
11
  "pad_token": "<pad>",
12
  "sep_token": "</s>",
13
  "unk_token": "<unk>"
 
7
  "bos_token": "<s>",
8
  "cls_token": "<s>",
9
  "eos_token": "</s>",
10
+ "mask_token": {
11
+ "content": "<mask>",
12
+ "lstrip": true,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
  "pad_token": "<pad>",
18
  "sep_token": "</s>",
19
  "unk_token": "<unk>"
tokenizer.json CHANGED
@@ -12,8 +12,8 @@
12
  "id": 0,
13
  "content": "<s>NOTUSED",
14
  "single_word": false,
15
- "lstrip": true,
16
- "rstrip": true,
17
  "normalized": false,
18
  "special": true
19
  },
@@ -30,8 +30,8 @@
30
  "id": 2,
31
  "content": "</s>NOTUSED",
32
  "single_word": false,
33
- "lstrip": true,
34
- "rstrip": true,
35
  "normalized": false,
36
  "special": true
37
  },
@@ -66,8 +66,8 @@
66
  "id": 8,
67
  "content": "<_>",
68
  "single_word": false,
69
- "lstrip": true,
70
- "rstrip": true,
71
  "normalized": false,
72
  "special": true
73
  },
@@ -77,7 +77,7 @@
77
  "single_word": false,
78
  "lstrip": true,
79
  "rstrip": false,
80
- "normalized": true,
81
  "special": true
82
  }
83
  ],
@@ -100,7 +100,8 @@
100
  "pre_tokenizer": {
101
  "type": "Metaspace",
102
  "replacement": "▁",
103
- "add_prefix_space": true
 
104
  },
105
  "post_processor": {
106
  "type": "TemplateProcessing",
@@ -186,7 +187,8 @@
186
  "decoder": {
187
  "type": "Metaspace",
188
  "replacement": "▁",
189
- "add_prefix_space": true
 
190
  },
191
  "model": {
192
  "type": "Unigram",
 
12
  "id": 0,
13
  "content": "<s>NOTUSED",
14
  "single_word": false,
15
+ "lstrip": false,
16
+ "rstrip": false,
17
  "normalized": false,
18
  "special": true
19
  },
 
30
  "id": 2,
31
  "content": "</s>NOTUSED",
32
  "single_word": false,
33
+ "lstrip": false,
34
+ "rstrip": false,
35
  "normalized": false,
36
  "special": true
37
  },
 
66
  "id": 8,
67
  "content": "<_>",
68
  "single_word": false,
69
+ "lstrip": false,
70
+ "rstrip": false,
71
  "normalized": false,
72
  "special": true
73
  },
 
77
  "single_word": false,
78
  "lstrip": true,
79
  "rstrip": false,
80
+ "normalized": false,
81
  "special": true
82
  }
83
  ],
 
100
  "pre_tokenizer": {
101
  "type": "Metaspace",
102
  "replacement": "▁",
103
+ "add_prefix_space": true,
104
+ "prepend_scheme": "always"
105
  },
106
  "post_processor": {
107
  "type": "TemplateProcessing",
 
187
  "decoder": {
188
  "type": "Metaspace",
189
  "replacement": "▁",
190
+ "add_prefix_space": true,
191
+ "prepend_scheme": "always"
192
  },
193
  "model": {
194
  "type": "Unigram",
tokenizer_config.json CHANGED
@@ -2,9 +2,9 @@
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<s>NOTUSED",
5
- "lstrip": true,
6
  "normalized": false,
7
- "rstrip": true,
8
  "single_word": false,
9
  "special": true
10
  },
@@ -18,9 +18,9 @@
18
  },
19
  "2": {
20
  "content": "</s>NOTUSED",
21
- "lstrip": true,
22
  "normalized": false,
23
- "rstrip": true,
24
  "single_word": false,
25
  "special": true
26
  },
@@ -50,16 +50,16 @@
50
  },
51
  "8": {
52
  "content": "<_>",
53
- "lstrip": true,
54
  "normalized": false,
55
- "rstrip": true,
56
  "single_word": false,
57
  "special": true
58
  },
59
  "25004": {
60
  "content": "<mask>",
61
  "lstrip": true,
62
- "normalized": true,
63
  "rstrip": false,
64
  "single_word": false,
65
  "special": true
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<s>NOTUSED",
5
+ "lstrip": false,
6
  "normalized": false,
7
+ "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
 
18
  },
19
  "2": {
20
  "content": "</s>NOTUSED",
21
+ "lstrip": false,
22
  "normalized": false,
23
+ "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
 
50
  },
51
  "8": {
52
  "content": "<_>",
53
+ "lstrip": false,
54
  "normalized": false,
55
+ "rstrip": false,
56
  "single_word": false,
57
  "special": true
58
  },
59
  "25004": {
60
  "content": "<mask>",
61
  "lstrip": true,
62
+ "normalized": false,
63
  "rstrip": false,
64
  "single_word": false,
65
  "special": true
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:065ac3630dbc97acf96a031139d1b85cdececcfc25cf9903f0d5d9756b4e395c
3
- size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a5abd811617cac248f8c52d9bb8037e1fe82883d264c3677d0da6cc6f99cf31
3
+ size 4600