IoakeimE commited on
Commit
7b3148a
·
verified ·
1 Parent(s): 0e6670b

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -4,9 +4,9 @@ library_name: transformers
4
  model_name: sft_normal_simplification_mini
5
  tags:
6
  - generated_from_trainer
7
- - sft
8
- - trl
9
  - unsloth
 
 
10
  licence: license
11
  ---
12
 
@@ -28,7 +28,7 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
-
32
 
33
 
34
  This model was trained with SFT.
@@ -36,7 +36,7 @@ This model was trained with SFT.
36
  ### Framework versions
37
 
38
  - TRL: 0.24.0
39
- - Transformers: 4.57.2
40
  - Pytorch: 2.9.0
41
  - Datasets: 4.3.0
42
  - Tokenizers: 0.22.1
 
4
  model_name: sft_normal_simplification_mini
5
  tags:
6
  - generated_from_trainer
 
 
7
  - unsloth
8
+ - trl
9
+ - sft
10
  licence: license
11
  ---
12
 
 
28
 
29
  ## Training procedure
30
 
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ioakeime-aristotle-university-of-thessaloniki/sft_normal_simplification_mini/runs/6nfctecu)
32
 
33
 
34
  This model was trained with SFT.
 
36
  ### Framework versions
37
 
38
  - TRL: 0.24.0
39
+ - Transformers: 4.57.3
40
  - Pytorch: 2.9.0
41
  - Datasets: 4.3.0
42
  - Tokenizers: 0.22.1
adapter_config.json CHANGED
@@ -33,12 +33,12 @@
33
  "rank_pattern": {},
34
  "revision": null,
35
  "target_modules": [
36
- "gate_proj",
37
  "v_proj",
38
- "o_proj",
39
  "q_proj",
40
  "up_proj",
 
41
  "k_proj",
 
42
  "down_proj"
43
  ],
44
  "target_parameters": null,
 
33
  "rank_pattern": {},
34
  "revision": null,
35
  "target_modules": [
 
36
  "v_proj",
 
37
  "q_proj",
38
  "up_proj",
39
+ "o_proj",
40
  "k_proj",
41
+ "gate_proj",
42
  "down_proj"
43
  ],
44
  "target_parameters": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e277fb005297d0c330731b9924f3ddbb9c99c2a576f578b26f09557a1d7634db
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98849a45d2183ede0484087fabe18552c4b7baef46b044b7b135efe9f6d1b54c
3
  size 167832240
tokenizer_config.json CHANGED
@@ -6179,7 +6179,7 @@
6179
  "legacy": false,
6180
  "model_max_length": 32768,
6181
  "pad_token": "[control_768]",
6182
- "padding_side": "left",
6183
  "sp_model_kwargs": {},
6184
  "spaces_between_special_tokens": false,
6185
  "tokenizer_class": "LlamaTokenizer",
 
6179
  "legacy": false,
6180
  "model_max_length": 32768,
6181
  "pad_token": "[control_768]",
6182
+ "padding_side": "right",
6183
  "sp_model_kwargs": {},
6184
  "spaces_between_special_tokens": false,
6185
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5477b9bf93fb7fdf2fd2c2a74a6b958ac506e68b7f0ab34f1294ae58c89968b
3
  size 6353
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd9f180dd3b7806d68f7fb492ae4ad60f94a9ae8179eefb0c3aab85d229d445d
3
  size 6353