Gaoussin commited on
Commit
ab7cb4e
·
verified ·
1 Parent(s): 9a1b3ae

Model save

Browse files
README.md CHANGED
@@ -1,9 +1,9 @@
1
  ---
2
  library_name: peft
3
  license: apache-2.0
4
- base_model: google/madlad400-3b-mt
5
  tags:
6
- - base_model:adapter:google/madlad400-3b-mt
7
  - lora
8
  - transformers
9
  model-index:
@@ -14,10 +14,10 @@ model-index:
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/bamalingua-bamalingua/Bamalingua-madlat/runs/l1ah6hmb)
18
  # madlad-bm-fr
19
 
20
- This model is a fine-tuned version of [google/madlad400-3b-mt](https://huggingface.co/google/madlad400-3b-mt) on an unknown dataset.
21
 
22
  ## Model description
23
 
@@ -44,8 +44,8 @@ The following hyperparameters were used during training:
44
  - total_train_batch_size: 16
45
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
- - lr_scheduler_warmup_steps: 100
48
- - training_steps: 1000
49
 
50
  ### Framework versions
51
 
 
1
  ---
2
  library_name: peft
3
  license: apache-2.0
4
+ base_model: jbochi/madlad400-3b-mt
5
  tags:
6
+ - base_model:adapter:jbochi/madlad400-3b-mt
7
  - lora
8
  - transformers
9
  model-index:
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/bamalingua-bamalingua/Bamalingua-madlat/runs/97dtmqsk)
18
  # madlad-bm-fr
19
 
20
+ This model is a fine-tuned version of [jbochi/madlad400-3b-mt](https://huggingface.co/jbochi/madlad400-3b-mt) on an unknown dataset.
21
 
22
  ## Model description
23
 
 
44
  - total_train_batch_size: 16
45
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
+ - lr_scheduler_warmup_steps: 10
48
+ - training_steps: 120
49
 
50
  ### Framework versions
51
 
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "google/madlad400-3b-mt",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -25,13 +25,13 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "q",
29
  "wi_1",
30
- "wi_0",
31
  "k",
 
32
  "v",
33
  "o",
34
- "wo"
 
35
  ],
36
  "target_parameters": null,
37
  "task_type": "SEQ_2_SEQ_LM",
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "jbochi/madlad400-3b-mt",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
28
  "wi_1",
 
29
  "k",
30
+ "wo",
31
  "v",
32
  "o",
33
+ "wi_0",
34
+ "q"
35
  ],
36
  "target_parameters": null,
37
  "task_type": "SEQ_2_SEQ_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c37f03e30916c742c86ffb33b7b40e2a23ee6e8be08bb15418eca556c6fcb989
3
  size 377655120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef0b981bf4cb7b49cd9e57a9f8393fc806aab652c933d344f7d649b19183731d
3
  size 377655120
tokenizer_config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "add_prefix_space": null,
3
  "added_tokens_decoder": {
4
  "0": {
5
  "content": "<unk>",
 
1
  {
2
+ "add_prefix_space": true,
3
  "added_tokens_decoder": {
4
  "0": {
5
  "content": "<unk>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:590112cdb69410d9dc24dab5981a0de779595745d7617f6a79f70c361a2d4e91
3
  size 5969
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96d92d485f95500cc7b523016e3add629a32cc75691dab30405ffccd0ac4cdf5
3
  size 5969