jhj1769 commited on
Commit
9197454
·
verified ·
1 Parent(s): 63df7e6

End of training

Browse files
Files changed (1) hide show
  1. README.md +13 -18
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: peft
3
  license: llama3.2
4
- base_model: meta-llama/Llama-3.2-1B
5
  tags:
6
  - generated_from_trainer
7
  metrics:
@@ -17,11 +17,11 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # mrpc-lora
19
 
20
- This model is a fine-tuned version of [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.8326
23
- - Accuracy: 0.8529
24
- - F1: 0.8921
25
 
26
  ## Model description
27
 
@@ -41,28 +41,23 @@ More information needed
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 2e-05
44
- - train_batch_size: 8
45
- - eval_batch_size: 8
46
  - seed: 42
47
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
- - num_epochs: 10
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
55
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
56
- | 0.6131 | 1.0 | 459 | 0.4045 | 0.8186 | 0.8746 |
57
- | 0.3038 | 2.0 | 918 | 0.3880 | 0.8578 | 0.8993 |
58
- | 0.2163 | 3.0 | 1377 | 0.4428 | 0.8505 | 0.8946 |
59
- | 0.2105 | 4.0 | 1836 | 0.5322 | 0.8529 | 0.8958 |
60
- | 0.4452 | 5.0 | 2295 | 0.6446 | 0.8407 | 0.8799 |
61
- | 0.3287 | 6.0 | 2754 | 0.6270 | 0.8554 | 0.8974 |
62
- | 0.2136 | 7.0 | 3213 | 0.7087 | 0.8456 | 0.8897 |
63
- | 0.1533 | 8.0 | 3672 | 0.7688 | 0.8407 | 0.8825 |
64
- | 0.1739 | 9.0 | 4131 | 0.8213 | 0.8480 | 0.8920 |
65
- | 0.2349 | 10.0 | 4590 | 0.8326 | 0.8529 | 0.8921 |
66
 
67
 
68
  ### Framework versions
 
1
  ---
2
  library_name: peft
3
  license: llama3.2
4
+ base_model: meta-llama/Llama-3.2-3B
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
17
 
18
  # mrpc-lora
19
 
20
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.4694
23
+ - Accuracy: 0.7917
24
+ - F1: 0.8537
25
 
26
  ## Model description
27
 
 
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 2e-05
44
+ - train_batch_size: 16
45
+ - eval_batch_size: 16
46
  - seed: 42
47
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
+ - num_epochs: 5
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
55
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
56
+ | 0.6611 | 1.0 | 230 | 0.6236 | 0.6642 | 0.7720 |
57
+ | 0.5343 | 2.0 | 460 | 0.5438 | 0.7255 | 0.8069 |
58
+ | 0.4298 | 3.0 | 690 | 0.5006 | 0.7745 | 0.8414 |
59
+ | 0.4538 | 4.0 | 920 | 0.4770 | 0.7917 | 0.8557 |
60
+ | 0.4486 | 5.0 | 1150 | 0.4694 | 0.7917 | 0.8537 |
 
 
 
 
 
61
 
62
 
63
  ### Framework versions