Xiangchi commited on
Commit
9afa4fa
·
verified ·
1 Parent(s): ef01b43

Model save

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  tags:
3
  - trl
4
  - sft
@@ -12,10 +13,10 @@ library_name: peft
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/xiangchiyuan/huggingface/runs/8wa5pub5)
16
  # math_without_reason_13bf
17
 
18
- This model was trained from scratch on the None dataset.
19
 
20
  ## Model description
21
 
@@ -57,7 +58,7 @@ The following hyperparameters were used during training:
57
  - total_train_batch_size: 1024
58
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
59
  - lr_scheduler_type: linear
60
- - num_epochs: 25.0
61
 
62
  ### Training results
63
 
 
1
  ---
2
+ base_model: NousResearch/Llama-2-13b-chat-hf
3
  tags:
4
  - trl
5
  - sft
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/xiangchiyuan/huggingface/runs/rw6629e7)
17
  # math_without_reason_13bf
18
 
19
+ This model is a fine-tuned version of [NousResearch/Llama-2-13b-chat-hf](https://huggingface.co/NousResearch/Llama-2-13b-chat-hf) on the None dataset.
20
 
21
  ## Model description
22
 
 
58
  - total_train_batch_size: 1024
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
+ - num_epochs: 10.0
62
 
63
  ### Training results
64