Tuch commited on
Commit
c09a85d
·
verified ·
1 Parent(s): 1e992bd

Model save

Browse files
Files changed (2) hide show
  1. README.md +9 -9
  2. adapter_model.safetensors +1 -1
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: meta-llama/Meta-Llama-3-8B-Instruct
3
  library_name: peft
4
  license: llama3
5
  tags:
@@ -16,13 +16,13 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # results_1
18
 
19
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - eval_loss: 1.3737
22
- - eval_runtime: 77.6899
23
- - eval_samples_per_second: 5.779
24
- - eval_steps_per_second: 0.734
25
- - epoch: 2.4053
26
  - step: 270
27
 
28
  ## Model description
@@ -43,11 +43,11 @@ More information needed
43
 
44
  The following hyperparameters were used during training:
45
  - learning_rate: 1e-05
46
- - train_batch_size: 4
47
  - eval_batch_size: 8
48
  - seed: 42
49
  - gradient_accumulation_steps: 4
50
- - total_train_batch_size: 16
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
  - num_epochs: 12
 
1
  ---
2
+ base_model: scb10x/llama-3-typhoon-v1.5-8b-instruct
3
  library_name: peft
4
  license: llama3
5
  tags:
 
16
 
17
  # results_1
18
 
19
+ This model is a fine-tuned version of [scb10x/llama-3-typhoon-v1.5-8b-instruct](https://huggingface.co/scb10x/llama-3-typhoon-v1.5-8b-instruct) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - eval_loss: 1.1364
22
+ - eval_runtime: 81.6834
23
+ - eval_samples_per_second: 5.497
24
+ - eval_steps_per_second: 0.698
25
+ - epoch: 1.8060
26
  - step: 270
27
 
28
  ## Model description
 
43
 
44
  The following hyperparameters were used during training:
45
  - learning_rate: 1e-05
46
+ - train_batch_size: 3
47
  - eval_batch_size: 8
48
  - seed: 42
49
  - gradient_accumulation_steps: 4
50
+ - total_train_batch_size: 12
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
  - num_epochs: 12
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4322968182d4fcf7806309944a97da21f1b0a613b655fffd2b15f65a1d86a3f
3
  size 125889008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5914ef2ca54f02f064780a95c9270ed42395d0a738046f4e0aedc10ebdb56a1
3
  size 125889008