allyyc commited on
Commit
401db01
·
1 Parent(s): 3df0d0b
Files changed (7) hide show
  1. README.md +0 -52
  2. all_results.json +0 -8
  3. optimizer.pt +3 -0
  4. rng_state.pth +3 -0
  5. scheduler.pt +3 -0
  6. train_results.json +0 -8
  7. trainer_state.json +0 -9
README.md DELETED
@@ -1,52 +0,0 @@
1
- ---
2
- tags:
3
- - generated_from_trainer
4
- datasets:
5
- - allyc/My-Dataset
6
- model-index:
7
- - name: uspto-gpt2-medium-1
8
- results: []
9
- ---
10
-
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
- # uspto-gpt2-medium-1
15
-
16
- This model is a fine-tuned version of [](https://huggingface.co/) on the allyc/My-Dataset uspto dataset.
17
-
18
- ## Model description
19
-
20
- More information needed
21
-
22
- ## Intended uses & limitations
23
-
24
- More information needed
25
-
26
- ## Training and evaluation data
27
-
28
- More information needed
29
-
30
- ## Training procedure
31
-
32
- ### Training hyperparameters
33
-
34
- The following hyperparameters were used during training:
35
- - learning_rate: 5e-05
36
- - train_batch_size: 32
37
- - eval_batch_size: 32
38
- - seed: 42
39
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
- - lr_scheduler_type: linear
41
- - training_steps: 68000
42
-
43
- ### Training results
44
-
45
-
46
-
47
- ### Framework versions
48
-
49
- - Transformers 4.35.0
50
- - Pytorch 2.0.1+cu117
51
- - Datasets 2.14.6
52
- - Tokenizers 0.14.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
all_results.json DELETED
@@ -1,8 +0,0 @@
1
- {
2
- "epoch": 1.0,
3
- "train_loss": 2.546952809053309,
4
- "train_runtime": 116511.6778,
5
- "train_samples": 68000,
6
- "train_samples_per_second": 18.676,
7
- "train_steps_per_second": 0.584
8
- }
 
 
 
 
 
 
 
 
 
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94751abdb5fe49a9fce5e2d36e4de4eb9a19f85fcc531e2a734a624de7ee1f7f
3
+ size 2838828805
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfdaa80090ee4ec33412e2f2792c86b0232add3b7828c892df95f17a76ea9f72
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d12e4b164e85df734dd297f0ea64d0ab36256e9460357384a2eb5681ab8f560
3
+ size 627
train_results.json DELETED
@@ -1,8 +0,0 @@
1
- {
2
- "epoch": 1.0,
3
- "train_loss": 2.546952809053309,
4
- "train_runtime": 116511.6778,
5
- "train_samples": 68000,
6
- "train_samples_per_second": 18.676,
7
- "train_steps_per_second": 0.584
8
- }
 
 
 
 
 
 
 
 
 
trainer_state.json CHANGED
@@ -823,15 +823,6 @@
823
  "learning_rate": 0.0,
824
  "loss": 2.482,
825
  "step": 68000
826
- },
827
- {
828
- "epoch": 1.0,
829
- "step": 68000,
830
- "total_flos": 4.041705422585856e+18,
831
- "train_loss": 2.546952809053309,
832
- "train_runtime": 116511.6778,
833
- "train_samples_per_second": 18.676,
834
- "train_steps_per_second": 0.584
835
  }
836
  ],
837
  "logging_steps": 500,
 
823
  "learning_rate": 0.0,
824
  "loss": 2.482,
825
  "step": 68000
 
 
 
 
 
 
 
 
 
826
  }
827
  ],
828
  "logging_steps": 500,