cabbage972 commited on
Commit
d86b2ef
·
1 Parent(s): 048c5c1

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -3
README.md CHANGED
@@ -2,6 +2,8 @@
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: expert-uspto
7
  results: []
@@ -13,6 +15,9 @@ should probably proofread and complete it, then remove this comment. -->
13
  # expert-uspto
14
 
15
  This model is a fine-tuned version of [EleutherAI/pythia-1b-deduped](https://huggingface.co/EleutherAI/pythia-1b-deduped) on the None dataset.
 
 
 
16
 
17
  ## Model description
18
 
@@ -33,19 +38,26 @@ More information needed
33
  The following hyperparameters were used during training:
34
  - learning_rate: 0.0001
35
  - train_batch_size: 1
36
- - eval_batch_size: 1
37
  - seed: 42
38
  - distributed_type: multi-GPU
39
  - num_devices: 8
40
  - gradient_accumulation_steps: 8
41
  - total_train_batch_size: 64
42
- - total_eval_batch_size: 8
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
- - training_steps: 1
46
 
47
  ### Training results
48
 
 
 
 
 
 
 
 
49
 
50
 
51
  ### Framework versions
 
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
5
+ metrics:
6
+ - accuracy
7
  model-index:
8
  - name: expert-uspto
9
  results: []
 
15
  # expert-uspto
16
 
17
  This model is a fine-tuned version of [EleutherAI/pythia-1b-deduped](https://huggingface.co/EleutherAI/pythia-1b-deduped) on the None dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 2.2220
20
+ - Accuracy: 0.5362
21
 
22
  ## Model description
23
 
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0001
40
  - train_batch_size: 1
41
+ - eval_batch_size: 8
42
  - seed: 42
43
  - distributed_type: multi-GPU
44
  - num_devices: 8
45
  - gradient_accumulation_steps: 8
46
  - total_train_batch_size: 64
47
+ - total_eval_batch_size: 64
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
+ - training_steps: 1000
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
55
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
56
+ | 2.2735 | 0.01 | 200 | 2.2464 | 0.5325 |
57
+ | 2.2557 | 0.01 | 400 | 2.2417 | 0.5331 |
58
+ | 2.2342 | 0.02 | 600 | 2.2342 | 0.5344 |
59
+ | 2.2241 | 0.03 | 800 | 2.2267 | 0.5355 |
60
+ | 2.229 | 0.03 | 1000 | 2.2220 | 0.5362 |
61
 
62
 
63
  ### Framework versions