Harini2506 commited on
Commit
34b047e
·
1 Parent(s): ed49a98

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,19 +1,15 @@
1
  {
2
- "_name_or_path": "xlnet-base-cased",
3
  "architectures": [
4
- "XLNetForTokenClassification"
5
  ],
6
- "attn_type": "bi",
7
- "bi_data": false,
8
- "bos_token_id": 1,
9
- "clamp_len": -1,
10
- "d_head": 64,
11
- "d_inner": 3072,
12
- "d_model": 768,
13
- "dropout": 0.1,
14
- "end_n_top": 5,
15
  "eos_token_id": 2,
16
- "ff_activation": "gelu",
 
 
17
  "id2label": {
18
  "0": "O",
19
  "1": "B-DatasetName",
@@ -32,6 +28,7 @@
32
  "14": "I-TaskName"
33
  },
34
  "initializer_range": 0.02,
 
35
  "label2id": {
36
  "B-DatasetName": 1,
37
  "B-HyperparameterName": 3,
@@ -49,29 +46,16 @@
49
  "I-TaskName": 14,
50
  "O": 0
51
  },
52
- "layer_norm_eps": 1e-12,
53
- "mem_len": null,
54
- "model_type": "xlnet",
55
- "n_head": 12,
56
- "n_layer": 12,
57
- "pad_token_id": 5,
58
- "reuse_len": null,
59
- "same_length": false,
60
- "start_n_top": 5,
61
- "summary_activation": "tanh",
62
- "summary_last_dropout": 0.1,
63
- "summary_type": "last",
64
- "summary_use_proj": true,
65
- "task_specific_params": {
66
- "text-generation": {
67
- "do_sample": true,
68
- "max_length": 250
69
- }
70
- },
71
  "torch_dtype": "float32",
72
  "transformers_version": "4.34.1",
73
- "untie_r": true,
74
- "use_mems_eval": true,
75
- "use_mems_train": false,
76
- "vocab_size": 32000
77
  }
 
1
  {
2
+ "_name_or_path": "roberta-base",
3
  "architectures": [
4
+ "RobertaForTokenClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
 
 
 
 
 
 
9
  "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
  "id2label": {
14
  "0": "O",
15
  "1": "B-DatasetName",
 
28
  "14": "I-TaskName"
29
  },
30
  "initializer_range": 0.02,
31
+ "intermediate_size": 3072,
32
  "label2id": {
33
  "B-DatasetName": 1,
34
  "B-HyperparameterName": 3,
 
46
  "I-TaskName": 14,
47
  "O": 0
48
  },
49
+ "layer_norm_eps": 1e-05,
50
+ "max_position_embeddings": 514,
51
+ "model_type": "roberta",
52
+ "num_attention_heads": 12,
53
+ "num_hidden_layers": 12,
54
+ "pad_token_id": 1,
55
+ "position_embedding_type": "absolute",
 
 
 
 
 
 
 
 
 
 
 
 
56
  "torch_dtype": "float32",
57
  "transformers_version": "4.34.1",
58
+ "type_vocab_size": 1,
59
+ "use_cache": true,
60
+ "vocab_size": 50265
 
61
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92fa1d909186ae24f05d41550bd2998891299415a48f976b2ddf407df54d2230
3
- size 466989130
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfdd61eb42b8d452e4cf01ee3bac5f2dca1402ecac4e11c3055010d8d0a470e7
3
+ size 496335078
special_tokens_map.json CHANGED
@@ -1,10 +1,6 @@
1
  {
2
- "additional_special_tokens": [
3
- "<eop>",
4
- "<eod>"
5
- ],
6
  "bos_token": "<s>",
7
- "cls_token": "<cls>",
8
  "eos_token": "</s>",
9
  "mask_token": {
10
  "content": "<mask>",
@@ -14,6 +10,6 @@
14
  "single_word": false
15
  },
16
  "pad_token": "<pad>",
17
- "sep_token": "<sep>",
18
  "unk_token": "<unk>"
19
  }
 
1
  {
 
 
 
 
2
  "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
  "eos_token": "</s>",
5
  "mask_token": {
6
  "content": "<mask>",
 
10
  "single_word": false
11
  },
12
  "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
  "unk_token": "<unk>"
15
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,17 +1,18 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "<unk>",
5
  "lstrip": false,
6
- "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
  "1": {
12
- "content": "<s>",
13
  "lstrip": false,
14
- "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
@@ -19,75 +20,38 @@
19
  "2": {
20
  "content": "</s>",
21
  "lstrip": false,
22
- "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
  "3": {
28
- "content": "<cls>",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "4": {
36
- "content": "<sep>",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- },
43
- "5": {
44
- "content": "<pad>",
45
  "lstrip": false,
46
- "normalized": false,
47
  "rstrip": false,
48
  "single_word": false,
49
  "special": true
50
  },
51
- "6": {
52
  "content": "<mask>",
53
  "lstrip": true,
54
  "normalized": false,
55
  "rstrip": false,
56
  "single_word": false,
57
  "special": true
58
- },
59
- "7": {
60
- "content": "<eod>",
61
- "lstrip": false,
62
- "normalized": false,
63
- "rstrip": false,
64
- "single_word": false,
65
- "special": true
66
- },
67
- "8": {
68
- "content": "<eop>",
69
- "lstrip": false,
70
- "normalized": false,
71
- "rstrip": false,
72
- "single_word": false,
73
- "special": true
74
  }
75
  },
76
- "additional_special_tokens": [
77
- "<eop>",
78
- "<eod>"
79
- ],
80
  "bos_token": "<s>",
81
  "clean_up_tokenization_spaces": true,
82
- "cls_token": "<cls>",
83
- "do_lower_case": false,
84
  "eos_token": "</s>",
85
- "keep_accents": false,
86
  "mask_token": "<mask>",
87
- "model_max_length": 1000000000000000019884624838656,
88
  "pad_token": "<pad>",
89
- "remove_space": true,
90
- "sep_token": "<sep>",
91
- "tokenizer_class": "XLNetTokenizer",
92
  "unk_token": "<unk>"
93
  }
 
1
  {
2
+ "add_prefix_space": true,
3
  "added_tokens_decoder": {
4
  "0": {
5
+ "content": "<s>",
6
  "lstrip": false,
7
+ "normalized": true,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
  "1": {
13
+ "content": "<pad>",
14
  "lstrip": false,
15
+ "normalized": true,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
 
20
  "2": {
21
  "content": "</s>",
22
  "lstrip": false,
23
+ "normalized": true,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
  },
28
  "3": {
29
+ "content": "<unk>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  "lstrip": false,
31
+ "normalized": true,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
35
  },
36
+ "50264": {
37
  "content": "<mask>",
38
  "lstrip": true,
39
  "normalized": false,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
  }
44
  },
 
 
 
 
45
  "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
+ "cls_token": "<s>",
 
48
  "eos_token": "</s>",
49
+ "errors": "replace",
50
  "mask_token": "<mask>",
51
+ "model_max_length": 512,
52
  "pad_token": "<pad>",
53
+ "sep_token": "</s>",
54
+ "tokenizer_class": "RobertaTokenizer",
55
+ "trim_offsets": true,
56
  "unk_token": "<unk>"
57
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f6ae9fc25fceece0da7c457194e4b0f540dbe6fcee22e5fd084ce9c3e121d06
3
  size 4472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d05cbdb3ac5b98eff888b7729607f859802eb617fa67d972ddf85b27c1bd1982
3
  size 4472
vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff