mtzig commited on
Commit
68e63a7
·
verified ·
1 Parent(s): 454c903

Model save

Browse files
README.md CHANGED
@@ -3,10 +3,7 @@ library_name: peft
3
  license: apache-2.0
4
  base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
5
  tags:
6
- - base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v1.0
7
- - lora
8
- - transformers
9
- pipeline_tag: text-generation
10
  model-index:
11
  - name: tinyllama-1.1b-sft-test3
12
  results: []
@@ -19,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.2319
23
 
24
  ## Model description
25
 
@@ -42,7 +39,7 @@ The following hyperparameters were used during training:
42
  - train_batch_size: 16
43
  - eval_batch_size: 10
44
  - seed: 4234
45
- - optimizer: Use paged_adamw_32bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: cosine
47
  - lr_scheduler_warmup_ratio: 0.03
48
  - num_epochs: 1
@@ -51,24 +48,24 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:------:|:----:|:---------------:|
54
- | 0.3307 | 0.0778 | 20 | 0.3030 |
55
- | 0.3519 | 0.1556 | 40 | 0.2559 |
56
- | 0.2853 | 0.2335 | 60 | 0.2719 |
57
- | 0.2972 | 0.3113 | 80 | 0.2569 |
58
- | 0.3116 | 0.3891 | 100 | 0.2501 |
59
- | 0.3691 | 0.4669 | 120 | 0.2374 |
60
- | 0.349 | 0.5447 | 140 | 0.2387 |
61
- | 0.2605 | 0.6226 | 160 | 0.2395 |
62
- | 0.3281 | 0.7004 | 180 | 0.2297 |
63
- | 0.278 | 0.7782 | 200 | 0.2314 |
64
- | 0.2592 | 0.8560 | 220 | 0.2334 |
65
- | 0.2562 | 0.9339 | 240 | 0.2319 |
66
 
67
 
68
  ### Framework versions
69
 
70
- - PEFT 0.17.1
71
- - Transformers 4.56.2
72
- - Pytorch 2.8.0+cu128
73
- - Datasets 4.1.1
74
- - Tokenizers 0.22.1
 
3
  license: apache-2.0
4
  base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
5
  tags:
6
+ - generated_from_trainer
 
 
 
7
  model-index:
8
  - name: tinyllama-1.1b-sft-test3
9
  results: []
 
16
 
17
  This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.2259
20
 
21
  ## Model description
22
 
 
39
  - train_batch_size: 16
40
  - eval_batch_size: 10
41
  - seed: 4234
42
+ - optimizer: Use OptimizerNames.PAGED_ADAMW with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.03
45
  - num_epochs: 1
 
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:------:|:----:|:---------------:|
51
+ | 0.3312 | 0.0778 | 20 | 0.2974 |
52
+ | 0.3523 | 0.1556 | 40 | 0.2538 |
53
+ | 0.2853 | 0.2335 | 60 | 0.2694 |
54
+ | 0.2953 | 0.3113 | 80 | 0.2519 |
55
+ | 0.3104 | 0.3891 | 100 | 0.2379 |
56
+ | 0.3664 | 0.4669 | 120 | 0.2325 |
57
+ | 0.3484 | 0.5447 | 140 | 0.2302 |
58
+ | 0.2604 | 0.6226 | 160 | 0.2322 |
59
+ | 0.3282 | 0.7004 | 180 | 0.2232 |
60
+ | 0.2788 | 0.7782 | 200 | 0.2250 |
61
+ | 0.2589 | 0.8560 | 220 | 0.2260 |
62
+ | 0.2551 | 0.9339 | 240 | 0.2259 |
63
 
64
 
65
  ### Framework versions
66
 
67
+ - PEFT 0.15.2
68
+ - Transformers 4.51.3
69
+ - Pytorch 2.7.0+cu128
70
+ - Datasets 3.5.0
71
+ - Tokenizers 0.21.1
final/README.md CHANGED
@@ -1,11 +1,6 @@
1
  ---
2
  base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
3
  library_name: peft
4
- pipeline_tag: text-generation
5
- tags:
6
- - base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v1.0
7
- - lora
8
- - transformers
9
  ---
10
 
11
  # Model Card for Model ID
@@ -204,4 +199,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
204
  [More Information Needed]
205
  ### Framework versions
206
 
207
- - PEFT 0.17.1
 
1
  ---
2
  base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
3
  library_name: peft
 
 
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
 
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
+ - PEFT 0.15.2
final/adapter_config.json CHANGED
@@ -20,20 +20,17 @@
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
- "qalora_group_size": 16,
24
  "r": 8,
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "v_proj",
29
- "o_proj",
30
  "q_proj",
31
- "k_proj"
 
 
32
  ],
33
- "target_parameters": null,
34
  "task_type": "CAUSAL_LM",
35
  "trainable_token_indices": null,
36
  "use_dora": false,
37
- "use_qalora": false,
38
  "use_rslora": false
39
  }
 
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
 
23
  "r": 8,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
 
27
  "q_proj",
28
+ "k_proj",
29
+ "v_proj",
30
+ "o_proj"
31
  ],
 
32
  "task_type": "CAUSAL_LM",
33
  "trainable_token_indices": null,
34
  "use_dora": false,
 
35
  "use_rslora": false
36
  }
final/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61f68aa2c3a723712896c17effd9c0378c3898d193c09461c45b1334b903cf0a
3
  size 9034304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:965759863b0b55d9d3434a43b226797a448e2d8d47a2000276e763796ec61bba
3
  size 9034304
final/tokenizer_config.json CHANGED
@@ -29,6 +29,7 @@
29
  }
30
  },
31
  "bos_token": "<s>",
 
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
  "extra_special_tokens": {},
 
29
  }
30
  },
31
  "bos_token": "<s>",
32
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "extra_special_tokens": {},
final/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3844ec01a7e5428375f5fa422945a281479583f95ab1b3f54a9ed7709b63560
3
- size 5777
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c5efdf6a4d7499ebc0905143bebac35a1ff8dbc498e16a33ecdd81a101ec6f6
3
+ size 5713