MHGanainy commited on
Commit
a759bda
·
verified ·
1 Parent(s): 95c503c

MHGanainy/16-clusters-imbalanced-5

Browse files
README.md CHANGED
@@ -15,6 +15,8 @@ should probably proofread and complete it, then remove this comment. -->
15
  # 16-clusters-imbalanced-5
16
 
17
  This model is a fine-tuned version of [openai-community/gpt2-xl](https://huggingface.co/openai-community/gpt2-xl) on an unknown dataset.
 
 
18
 
19
  ## Model description
20
 
 
15
  # 16-clusters-imbalanced-5
16
 
17
  This model is a fine-tuned version of [openai-community/gpt2-xl](https://huggingface.co/openai-community/gpt2-xl) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 2.1312
20
 
21
  ## Model description
22
 
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 2.1312105655670166,
4
- "eval_runtime": 2.6958,
5
- "eval_samples_per_second": 13.725,
6
- "eval_steps_per_second": 1.855,
7
- "perplexity": 8.425059727805095,
8
  "total_flos": 4161668422041600.0,
9
- "train_loss": 2.3458979244315468,
10
- "train_runtime": 72.3208,
11
- "train_samples_per_second": 6.333,
12
- "train_steps_per_second": 3.166
13
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 2.1312263011932373,
4
+ "eval_runtime": 2.6837,
5
+ "eval_samples_per_second": 13.787,
6
+ "eval_steps_per_second": 1.863,
7
+ "perplexity": 8.425192302438928,
8
  "total_flos": 4161668422041600.0,
9
+ "train_loss": 2.3460041029484513,
10
+ "train_runtime": 71.1285,
11
+ "train_samples_per_second": 6.439,
12
+ "train_steps_per_second": 3.22
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 2.1312105655670166,
4
- "eval_runtime": 2.6958,
5
- "eval_samples_per_second": 13.725,
6
- "eval_steps_per_second": 1.855,
7
- "perplexity": 8.425059727805095
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 2.1312263011932373,
4
+ "eval_runtime": 2.6837,
5
+ "eval_samples_per_second": 13.787,
6
+ "eval_steps_per_second": 1.863,
7
+ "perplexity": 8.425192302438928
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
  "total_flos": 4161668422041600.0,
4
- "train_loss": 2.3458979244315468,
5
- "train_runtime": 72.3208,
6
- "train_samples_per_second": 6.333,
7
- "train_steps_per_second": 3.166
8
  }
 
1
  {
2
  "epoch": 1.0,
3
  "total_flos": 4161668422041600.0,
4
+ "train_loss": 2.3460041029484513,
5
+ "train_runtime": 71.1285,
6
+ "train_samples_per_second": 6.439,
7
+ "train_steps_per_second": 3.22
8
  }
trainer_state.json CHANGED
@@ -10,26 +10,26 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.4366812227074236,
13
- "grad_norm": 0.10677418857812881,
14
  "learning_rate": 6.666666666666667e-06,
15
  "loss": 2.3802,
16
  "step": 100
17
  },
18
  {
19
  "epoch": 0.8733624454148472,
20
- "grad_norm": 0.2271566540002823,
21
  "learning_rate": 1.3333333333333333e-05,
22
- "loss": 2.3234,
23
  "step": 200
24
  },
25
  {
26
  "epoch": 1.0,
27
  "step": 229,
28
  "total_flos": 4161668422041600.0,
29
- "train_loss": 2.3458979244315468,
30
- "train_runtime": 72.3208,
31
- "train_samples_per_second": 6.333,
32
- "train_steps_per_second": 3.166
33
  }
34
  ],
35
  "logging_steps": 100,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.4366812227074236,
13
+ "grad_norm": 0.10466119647026062,
14
  "learning_rate": 6.666666666666667e-06,
15
  "loss": 2.3802,
16
  "step": 100
17
  },
18
  {
19
  "epoch": 0.8733624454148472,
20
+ "grad_norm": 0.2237740308046341,
21
  "learning_rate": 1.3333333333333333e-05,
22
+ "loss": 2.3236,
23
  "step": 200
24
  },
25
  {
26
  "epoch": 1.0,
27
  "step": 229,
28
  "total_flos": 4161668422041600.0,
29
+ "train_loss": 2.3460041029484513,
30
+ "train_runtime": 71.1285,
31
+ "train_samples_per_second": 6.439,
32
+ "train_steps_per_second": 3.22
33
  }
34
  ],
35
  "logging_steps": 100,