EdBerg commited on
Commit
f5ec921
·
verified ·
1 Parent(s): ca9c4b4

End of training

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: meta-llama/Llama-3.2-1B-Instruct
3
  library_name: peft
4
  license: llama3.2
5
  tags:
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # Baha_2A
18
 
19
- This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on an unknown dataset.
20
 
21
  ## Model description
22
 
@@ -44,7 +44,7 @@ The following hyperparameters were used during training:
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.03
47
- - training_steps: 180
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
 
1
  ---
2
+ base_model: meta-llama/Llama-3.2-3B-Instruct
3
  library_name: peft
4
  license: llama3.2
5
  tags:
 
16
 
17
  # Baha_2A
18
 
19
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on an unknown dataset.
20
 
21
  ## Model description
22
 
 
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.03
47
+ - training_steps: 500
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results