Aurelie123 commited on
Commit
650f819
·
verified ·
1 Parent(s): 959850e

Training in progress, step 500

Browse files
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
config.json CHANGED
@@ -1,30 +1,41 @@
1
  {
2
  "architectures": [
3
- "BertForQuestionAnswering"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
- "cell": {},
7
- "classifier_dropout": null,
8
- "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 3072,
14
  "language": "english",
15
- "layer_norm_eps": 1e-12,
 
16
  "max_position_embeddings": 512,
17
- "model_type": "bert",
18
- "name": "Bert",
 
 
19
  "num_attention_heads": 12,
20
- "num_hidden_layers": 6,
21
  "pad_token_id": 0,
22
- "position_embedding_type": "absolute",
23
- "pre_trained": "",
24
- "structure": [],
 
 
 
 
 
 
 
 
 
 
 
 
25
  "torch_dtype": "float32",
26
  "transformers_version": "4.53.2",
27
- "type_vocab_size": 2,
28
- "use_cache": true,
29
- "vocab_size": 30522
30
  }
 
1
  {
2
  "architectures": [
3
+ "DebertaV2ForQuestionAnswering"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
 
 
 
6
  "hidden_act": "gelu",
7
  "hidden_dropout_prob": 0.1,
8
  "hidden_size": 768,
9
  "initializer_range": 0.02,
10
  "intermediate_size": 3072,
11
  "language": "english",
12
+ "layer_norm_eps": 1e-07,
13
+ "legacy": true,
14
  "max_position_embeddings": 512,
15
+ "max_relative_positions": -1,
16
+ "model_type": "deberta-v2",
17
+ "name": "DebertaV2",
18
+ "norm_rel_ebd": "layer_norm",
19
  "num_attention_heads": 12,
20
+ "num_hidden_layers": 12,
21
  "pad_token_id": 0,
22
+ "pooler_dropout": 0,
23
+ "pooler_hidden_act": "gelu",
24
+ "pooler_hidden_size": 768,
25
+ "pos_att_type": [
26
+ "p2c",
27
+ "c2p"
28
+ ],
29
+ "position_biased_input": false,
30
+ "position_buckets": 256,
31
+ "relative_attention": true,
32
+ "share_att_key": true,
33
+ "summary_activation": "tanh",
34
+ "summary_last_dropout": 0,
35
+ "summary_type": "first",
36
+ "summary_use_proj": false,
37
  "torch_dtype": "float32",
38
  "transformers_version": "4.53.2",
39
+ "type_vocab_size": 0,
40
+ "vocab_size": 128100
 
41
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0bb0d6d062a7c49d93cf87a5f525ae23c4579b57316f622c391a477d7d360aa
3
- size 265475832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:321521dbae8739126272b9b2f3fb9cb7e686f1027b23c24472b3e1b91e6c1c2b
3
+ size 735356752
runs/Jul22_14-09-26_2c6cb13cc16e/events.out.tfevents.1753193367.2c6cb13cc16e.714.1 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:890e55213c827c2ad4a30d6738897d62ad18f0c284b277ff5b31794e976d4da3
3
- size 6077
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9ccb988d57a40823b925715bb7bf49e4f237283addff4ceaec28029e4075ef9
3
+ size 6702
runs/Jul22_14-37-53_2c6cb13cc16e/events.out.tfevents.1753195075.2c6cb13cc16e.714.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a40cbd0462d1fd9699704f14e99b04217d1401afef01ba686598f6f37591448
3
+ size 6358
special_tokens_map.json CHANGED
@@ -1,4 +1,11 @@
1
  {
 
 
 
 
 
 
 
2
  "cls_token": {
3
  "content": "[CLS]",
4
  "lstrip": false,
@@ -6,6 +13,13 @@
6
  "rstrip": false,
7
  "single_word": false
8
  },
 
 
 
 
 
 
 
9
  "mask_token": {
10
  "content": "[MASK]",
11
  "lstrip": false,
 
1
  {
2
+ "bos_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
  "cls_token": {
10
  "content": "[CLS]",
11
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "eos_token": {
17
+ "content": "[SEP]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "mask_token": {
24
  "content": "[MASK]",
25
  "lstrip": false,
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88288f3bbc81d03d5cc31434d30f8bd3c80771c9875a75fa763d167a8e34889e
3
- size 711661
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d4c2cda7fd3cb0a4a0460686f9a1f9ef4ca3ce478e30993b1162b64493f55e
3
+ size 8649130
tokenizer_config.json CHANGED
@@ -8,31 +8,31 @@
8
  "single_word": false,
9
  "special": true
10
  },
11
- "100": {
12
- "content": "[UNK]",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
- "101": {
20
- "content": "[CLS]",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
- "102": {
28
- "content": "[SEP]",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
- "103": {
36
  "content": "[MASK]",
37
  "lstrip": false,
38
  "normalized": false,
@@ -41,18 +41,19 @@
41
  "special": true
42
  }
43
  },
 
44
  "clean_up_tokenization_spaces": false,
45
  "cls_token": "[CLS]",
46
- "do_basic_tokenize": true,
47
- "do_lower_case": true,
48
  "extra_special_tokens": {},
49
  "mask_token": "[MASK]",
50
- "model_max_length": 1000000000000000019884624838656,
51
- "never_split": null,
52
  "pad_token": "[PAD]",
53
  "sep_token": "[SEP]",
54
- "strip_accents": null,
55
- "tokenize_chinese_chars": true,
56
- "tokenizer_class": "BertTokenizer",
57
- "unk_token": "[UNK]"
 
58
  }
 
8
  "single_word": false,
9
  "special": true
10
  },
11
+ "1": {
12
+ "content": "[CLS]",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
+ "2": {
20
+ "content": "[SEP]",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "3": {
28
+ "content": "[UNK]",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
+ "128000": {
36
  "content": "[MASK]",
37
  "lstrip": false,
38
  "normalized": false,
 
41
  "special": true
42
  }
43
  },
44
+ "bos_token": "[CLS]",
45
  "clean_up_tokenization_spaces": false,
46
  "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
  "extra_special_tokens": {},
50
  "mask_token": "[MASK]",
51
+ "model_max_length": 512,
 
52
  "pad_token": "[PAD]",
53
  "sep_token": "[SEP]",
54
+ "sp_model_kwargs": {},
55
+ "split_by_punct": false,
56
+ "tokenizer_class": "DebertaV2Tokenizer",
57
+ "unk_token": "[UNK]",
58
+ "vocab_type": "spm"
59
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:670df5286d4561e247870dae44502640f0876c3a672ea41f8485db46a5736acd
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07ae19866cfe2d0d5c037c7e41148938b0f200bfd5edcaeaddf9f96d417f5641
3
  size 5304