Tanakrit commited on
Commit
928af9b
·
1 Parent(s): f8ec1a2

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,12 +1,13 @@
1
  {
2
- "_name_or_path": "xlm-roberta-base",
3
  "architectures": [
4
- "XLMRobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
 
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
@@ -24,12 +25,12 @@
24
  "LABEL_2": 2,
25
  "LABEL_3": 3
26
  },
27
- "layer_norm_eps": 1e-05,
28
- "max_position_embeddings": 514,
29
- "model_type": "xlm-roberta",
 
30
  "num_attention_heads": 12,
31
  "num_hidden_layers": 12,
32
- "output_past": true,
33
  "pad_token_id": 1,
34
  "position_embedding_type": "absolute",
35
  "problem_type": "single_label_classification",
@@ -37,5 +38,5 @@
37
  "transformers_version": "4.25.1",
38
  "type_vocab_size": 1,
39
  "use_cache": true,
40
- "vocab_size": 250002
41
  }
 
1
  {
2
+ "_name_or_path": "airesearch/wangchanberta-base-att-spm-uncased",
3
  "architectures": [
4
+ "CamembertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 768,
 
25
  "LABEL_2": 2,
26
  "LABEL_3": 3
27
  },
28
+ "layer_norm_eps": 1e-12,
29
+ "max_position_embeddings": 512,
30
+ "model_type": "camembert",
31
+ "num_attention_head": 12,
32
  "num_attention_heads": 12,
33
  "num_hidden_layers": 12,
 
34
  "pad_token_id": 1,
35
  "position_embedding_type": "absolute",
36
  "problem_type": "single_label_classification",
 
38
  "transformers_version": "4.25.1",
39
  "type_vocab_size": 1,
40
  "use_cache": true,
41
+ "vocab_size": 25005
42
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac76ec9e39bb116391f667a8eefeb79e45bb0df5fdf72af8cd6aedf2b7839bd0
3
- size 1112257461
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3267161ae54a89b44f0fa79206215719ad91397a2c2cafdd25955d10119d5e6
3
+ size 421060533
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49c4ba4e495ddf31eb2fdba7fc6aef3c233091d25d35bc9d24694ccf48ae114c
3
+ size 904693
special_tokens_map.json CHANGED
@@ -1,11 +1,16 @@
1
  {
 
 
 
 
 
2
  "bos_token": "<s>",
3
  "cls_token": "<s>",
4
  "eos_token": "</s>",
5
  "mask_token": {
6
  "content": "<mask>",
7
  "lstrip": true,
8
- "normalized": false,
9
  "rstrip": false,
10
  "single_word": false
11
  },
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<s>NOTUSED",
4
+ "</s>NOTUSED",
5
+ "<_>"
6
+ ],
7
  "bos_token": "<s>",
8
  "cls_token": "<s>",
9
  "eos_token": "</s>",
10
  "mask_token": {
11
  "content": "<mask>",
12
  "lstrip": true,
13
+ "normalized": true,
14
  "rstrip": false,
15
  "single_word": false
16
  },
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:62c24cdc13d4c9952d63718d6c9fa4c287974249e16b7ade6d5a85e7bbb75626
3
- size 17082660
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98a9a1aab83c245fe4a62898eaa94e4782ccafa043d53be646a62b1887a02597
3
+ size 2179217
tokenizer_config.json CHANGED
@@ -1,4 +1,9 @@
1
  {
 
 
 
 
 
2
  "bos_token": "<s>",
3
  "cls_token": "<s>",
4
  "eos_token": "</s>",
@@ -10,11 +15,12 @@
10
  "rstrip": false,
11
  "single_word": false
12
  },
13
- "model_max_length": 512,
14
- "name_or_path": "xlm-roberta-base",
15
  "pad_token": "<pad>",
16
  "sep_token": "</s>",
 
17
  "special_tokens_map_file": null,
18
- "tokenizer_class": "XLMRobertaTokenizer",
19
  "unk_token": "<unk>"
20
  }
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<s>NOTUSED",
4
+ "</s>NOTUSED",
5
+ "<_>"
6
+ ],
7
  "bos_token": "<s>",
8
  "cls_token": "<s>",
9
  "eos_token": "</s>",
 
15
  "rstrip": false,
16
  "single_word": false
17
  },
18
+ "model_max_length": 1000000000000000019884624838656,
19
+ "name_or_path": "airesearch/wangchanberta-base-att-spm-uncased",
20
  "pad_token": "<pad>",
21
  "sep_token": "</s>",
22
+ "sp_model_kwargs": {},
23
  "special_tokens_map_file": null,
24
+ "tokenizer_class": "CamembertTokenizer",
25
  "unk_token": "<unk>"
26
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:091c60b06d39055cad3b62266cbdbd12ca8b5a9947178e0510554aa205b20d2a
3
  size 3451
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a5fd828deca4f7e226481be1581298215c8badbeadf5e5aa902fbd98f945ea0
3
  size 3451