Koubra-Gaby commited on
Commit
7809ebb
·
verified ·
1 Parent(s): 7e9592e

Training in progress, step 500

Browse files
config.json CHANGED
@@ -8,7 +8,7 @@
8
  "MarianMTModel"
9
  ],
10
  "attention_dropout": 0.0,
11
- "bos_token_id": 0,
12
  "classif_dropout": 0.0,
13
  "classifier_dropout": 0.0,
14
  "d_model": 512,
@@ -19,12 +19,12 @@
19
  "decoder_start_token_id": 59513,
20
  "decoder_vocab_size": 59514,
21
  "dropout": 0.1,
 
22
  "encoder_attention_heads": 8,
23
  "encoder_ffn_dim": 2048,
24
  "encoder_layerdrop": 0.0,
25
  "encoder_layers": 6,
26
  "eos_token_id": 0,
27
- "forced_eos_token_id": 0,
28
  "gradient_checkpointing": false,
29
  "id2label": {
30
  "0": "LABEL_0",
@@ -32,25 +32,24 @@
32
  "2": "LABEL_2"
33
  },
34
  "init_std": 0.02,
 
35
  "is_encoder_decoder": true,
36
  "label2id": {
37
  "LABEL_0": 0,
38
  "LABEL_1": 1,
39
  "LABEL_2": 2
40
  },
41
- "max_length": null,
42
  "max_position_embeddings": 512,
43
  "model_type": "marian",
44
  "normalize_before": false,
45
  "normalize_embedding": false,
46
- "num_beams": null,
47
  "num_hidden_layers": 6,
48
  "pad_token_id": 59513,
49
  "scale_embedding": true,
50
  "share_encoder_decoder_embeddings": true,
51
  "static_position_embeddings": true,
52
- "torch_dtype": "float32",
53
- "transformers_version": "4.50.3",
54
- "use_cache": true,
55
  "vocab_size": 59514
56
  }
 
8
  "MarianMTModel"
9
  ],
10
  "attention_dropout": 0.0,
11
+ "bos_token_id": null,
12
  "classif_dropout": 0.0,
13
  "classifier_dropout": 0.0,
14
  "d_model": 512,
 
19
  "decoder_start_token_id": 59513,
20
  "decoder_vocab_size": 59514,
21
  "dropout": 0.1,
22
+ "dtype": "float32",
23
  "encoder_attention_heads": 8,
24
  "encoder_ffn_dim": 2048,
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 6,
27
  "eos_token_id": 0,
 
28
  "gradient_checkpointing": false,
29
  "id2label": {
30
  "0": "LABEL_0",
 
32
  "2": "LABEL_2"
33
  },
34
  "init_std": 0.02,
35
+ "is_decoder": false,
36
  "is_encoder_decoder": true,
37
  "label2id": {
38
  "LABEL_0": 0,
39
  "LABEL_1": 1,
40
  "LABEL_2": 2
41
  },
 
42
  "max_position_embeddings": 512,
43
  "model_type": "marian",
44
  "normalize_before": false,
45
  "normalize_embedding": false,
 
46
  "num_hidden_layers": 6,
47
  "pad_token_id": 59513,
48
  "scale_embedding": true,
49
  "share_encoder_decoder_embeddings": true,
50
  "static_position_embeddings": true,
51
+ "tie_word_embeddings": true,
52
+ "transformers_version": "5.2.0",
53
+ "use_cache": false,
54
  "vocab_size": 59514
55
  }
generation_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "assistant_confidence_threshold": 0.4,
3
+ "assistant_lookbehind": 10,
4
+ "bad_words_ids": [
5
+ [
6
+ 59513
7
+ ]
8
+ ],
9
+ "decoder_start_token_id": 59513,
10
+ "diversity_penalty": 0.0,
11
+ "do_sample": false,
12
+ "early_stopping": false,
13
+ "encoder_no_repeat_ngram_size": 0,
14
+ "encoder_repetition_penalty": 1.0,
15
+ "eos_token_id": [
16
+ 0
17
+ ],
18
+ "epsilon_cutoff": 0.0,
19
+ "eta_cutoff": 0.0,
20
+ "forced_eos_token_id": 0,
21
+ "length_penalty": 1.0,
22
+ "max_length": 512,
23
+ "min_length": 0,
24
+ "no_repeat_ngram_size": 0,
25
+ "num_assistant_tokens": 20,
26
+ "num_assistant_tokens_schedule": "constant",
27
+ "num_beam_groups": 1,
28
+ "num_beams": 4,
29
+ "num_return_sequences": 1,
30
+ "output_scores": false,
31
+ "pad_token_id": 59513,
32
+ "remove_invalid_values": false,
33
+ "renormalize_logits": true,
34
+ "repetition_penalty": 1.0,
35
+ "return_dict_in_generate": false,
36
+ "target_lookbehind": 10,
37
+ "temperature": 1.0,
38
+ "top_k": 50,
39
+ "top_p": 1.0,
40
+ "transformers_version": "5.2.0",
41
+ "typical_p": 1.0,
42
+ "use_cache": true
43
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c3b3cf6ed91b0e0eecc40b0cad42cd06552ef43086cc92090f70f9d173fe99c
3
- size 298705768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0265134777bdcecfa031a89d3c89784f92d2669f98cf4eafb44d73f6b50b26a
3
+ size 542475664
tokenizer_config.json CHANGED
@@ -25,9 +25,9 @@
25
  "special": true
26
  }
27
  },
28
- "clean_up_tokenization_spaces": false,
29
  "eos_token": "</s>",
30
- "extra_special_tokens": {},
31
  "model_max_length": 512,
32
  "pad_token": "<pad>",
33
  "separate_vocabs": false,
 
25
  "special": true
26
  }
27
  },
28
+ "backend": "custom",
29
  "eos_token": "</s>",
30
+ "is_local": false,
31
  "model_max_length": 512,
32
  "pad_token": "<pad>",
33
  "separate_vocabs": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a21d18ecfedf33e95e840b94fd7c6ae790e1be1d189a747bd49bea59eafa0db
3
- size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a070c3b91298cfcfd5e260b1232d23f02c86834e4db661ffaf1e5d91bd49714
3
+ size 5329