iproskurina commited on
Commit
3162dde
·
verified ·
1 Parent(s): d443714

Checkpoint after epoch 1

Browse files
README.md CHANGED
@@ -4,6 +4,8 @@ license: llama3.2
4
  base_model: meta-llama/Llama-3.2-1B
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: test
9
  results: []
@@ -15,6 +17,9 @@ should probably proofread and complete it, then remove this comment. -->
15
  # test
16
 
17
  This model is a fine-tuned version of [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) on an unknown dataset.
 
 
 
18
 
19
  ## Model description
20
 
 
4
  base_model: meta-llama/Llama-3.2-1B
5
  tags:
6
  - generated_from_trainer
7
+ metrics:
8
+ - accuracy
9
  model-index:
10
  - name: test
11
  results: []
 
17
  # test
18
 
19
  This model is a fine-tuned version of [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 2.6568
22
+ - Accuracy: 0.5231
23
 
24
  ## Model description
25
 
all_results.json CHANGED
@@ -1,16 +1,14 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_accuracy": 0.5230333635625984,
4
- "eval_loss": 2.656813859939575,
5
- "eval_runtime": 59.9412,
6
- "eval_samples": 5125,
7
- "eval_samples_per_second": 85.5,
8
- "eval_steps_per_second": 10.694,
9
- "perplexity": 14.250811606566634,
10
  "total_flos": 1695049253388288.0,
11
- "train_loss": 2.6765631017550615,
12
- "train_runtime": 106.623,
13
- "train_samples": 567,
14
- "train_samples_per_second": 5.318,
15
- "train_steps_per_second": 1.332
16
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_accuracy": 0.5230967495584936,
4
+ "eval_loss": 2.656780958175659,
5
+ "eval_runtime": 59.9656,
6
+ "eval_samples_per_second": 85.466,
7
+ "eval_steps_per_second": 10.689,
8
+ "perplexity": 14.250342737440896,
 
9
  "total_flos": 1695049253388288.0,
10
+ "train_loss": 2.676516035912742,
11
+ "train_runtime": 105.7579,
12
+ "train_samples_per_second": 5.361,
13
+ "train_steps_per_second": 1.343
 
14
  }
eval_epoch_1_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_accuracy": 0.5230333635625984,
4
- "eval_loss": 2.656813859939575,
5
- "eval_runtime": 59.8639,
6
- "eval_samples_per_second": 85.611,
7
- "eval_steps_per_second": 10.708,
8
- "perplexity": 14.250811606566634
9
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_accuracy": 0.5230967495584936,
4
+ "eval_loss": 2.656780958175659,
5
+ "eval_runtime": 59.9656,
6
+ "eval_samples_per_second": 85.466,
7
+ "eval_steps_per_second": 10.689,
8
+ "perplexity": 14.250342737440896
9
  }
special_tokens_map.json CHANGED
@@ -12,5 +12,6 @@
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
- }
 
16
  }
 
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
+ },
16
+ "pad_token": "<|end_of_text|>"
17
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
- size 17209920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8455cd05329f9cb9895e200605934b714e7fc55873b9969c77d1c3c01ccc60e0
3
+ size 17210188
tokenizer_config.json CHANGED
@@ -2058,5 +2058,6 @@
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
 
2061
  "tokenizer_class": "PreTrainedTokenizerFast"
2062
  }
 
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
2061
+ "pad_token": "<|end_of_text|>",
2062
  "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }
train_epoch_1_results.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
  "total_flos": 1695049253388288.0,
4
- "train_loss": 2.6765631017550615,
5
- "train_runtime": 106.623,
6
- "train_samples": 567,
7
- "train_samples_per_second": 5.318,
8
- "train_steps_per_second": 1.332
9
  }
 
1
  {
2
  "epoch": 1.0,
3
  "total_flos": 1695049253388288.0,
4
+ "train_loss": 2.676516035912742,
5
+ "train_runtime": 105.7579,
6
+ "train_samples_per_second": 5.361,
7
+ "train_steps_per_second": 1.343
 
8
  }
trainer_state.json CHANGED
@@ -11,30 +11,21 @@
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
- "eval_accuracy": 0.5230333635625984,
15
- "eval_loss": 2.656813859939575,
16
- "eval_runtime": 59.8708,
17
- "eval_samples_per_second": 85.601,
18
- "eval_steps_per_second": 10.706,
19
  "step": 142
20
  },
21
  {
22
  "epoch": 1.0,
23
  "step": 142,
24
  "total_flos": 1695049253388288.0,
25
- "train_loss": 2.6765631017550615,
26
- "train_runtime": 106.623,
27
- "train_samples_per_second": 5.318,
28
- "train_steps_per_second": 1.332
29
- },
30
- {
31
- "epoch": 1.0,
32
- "eval_accuracy": 0.5230333635625984,
33
- "eval_loss": 2.656813859939575,
34
- "eval_runtime": 59.8639,
35
- "eval_samples_per_second": 85.611,
36
- "eval_steps_per_second": 10.708,
37
- "step": 142
38
  }
39
  ],
40
  "logging_steps": 500,
 
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
+ "eval_accuracy": 0.5230967495584936,
15
+ "eval_loss": 2.656780958175659,
16
+ "eval_runtime": 59.6737,
17
+ "eval_samples_per_second": 85.884,
18
+ "eval_steps_per_second": 10.742,
19
  "step": 142
20
  },
21
  {
22
  "epoch": 1.0,
23
  "step": 142,
24
  "total_flos": 1695049253388288.0,
25
+ "train_loss": 2.676516035912742,
26
+ "train_runtime": 105.7579,
27
+ "train_samples_per_second": 5.361,
28
+ "train_steps_per_second": 1.343
 
 
 
 
 
 
 
 
 
29
  }
30
  ],
31
  "logging_steps": 500,