davidquarel commited on
Commit
2c8fecb
·
verified ·
1 Parent(s): a829ffe

Upload folder using huggingface_hub

Browse files
Files changed (5) hide show
  1. debug.log +3 -0
  2. eval.log +83 -0
  3. model.json +1 -0
  4. model.safetensors +3 -0
  5. train.log +0 -0
debug.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ name shakespeare_64x4 | device cuda | compile True | data_dir data/shakespeare | should_randomize True | log_interval 10 | eval_interval 250 | eval_steps 100 | batch_size 128 | gradient_accumulation_steps 1 | learning_rate 0.001 | warmup_steps 1000 | max_steps 10000 | decay_lr True | min_lr 1e-05 | weight_decay 0.1 | grad_clip 1.0 | gpt_config {'name': 'ascii_64x4', 'device': device(type='cuda'), 'compile': True, 'block_size': 128, 'vocab_size': 128, 'n_layer': 4, 'n_head': 4, 'n_embd': 64}
2
+ name shakespeare_64x4 | device cuda | compile True | data_dir data/shakespeare | should_randomize True | log_interval 10 | eval_interval 250 | eval_steps 100 | batch_size 128 | gradient_accumulation_steps 1 | learning_rate 0.001 | warmup_steps 1000 | max_steps 10000 | decay_lr True | min_lr 1e-05 | weight_decay 0.1 | grad_clip 1.0 | gpt_config {'name': 'ascii_64x4', 'device': device(type='cuda'), 'compile': True, 'block_size': 128, 'vocab_size': 128, 'n_layer': 4, 'n_head': 4, 'n_embd': 64}
3
+ name shakespeare_64x4 | device cuda | compile True | data_dir data/shakespeare | should_randomize True | log_interval 10 | eval_interval 250 | eval_steps 100 | batch_size 128 | gradient_accumulation_steps 1 | learning_rate 0.001 | warmup_steps 1000 | max_steps 10000 | decay_lr True | min_lr 1e-05 | weight_decay 0.1 | grad_clip 1.0 | gpt_config {'name': 'ascii_64x4', 'device': device(type='cuda'), 'compile': True, 'block_size': 128, 'vocab_size': 128, 'n_layer': 4, 'n_head': 4, 'n_embd': 64}
eval.log ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ type eval | step 0 | loss 4.8501 | checkpoint False
2
+ type eval | step 0 | loss 4.8489 | checkpoint False
3
+ type eval | step 250 | loss 3.1139 | checkpoint True
4
+ type eval | step 500 | loss 2.5083 | checkpoint True
5
+ type eval | step 750 | loss 2.3826 | checkpoint True
6
+ type eval | step 1000 | loss 2.2400 | checkpoint True
7
+ type eval | step 1250 | loss 2.0680 | checkpoint True
8
+ type eval | step 1500 | loss 1.9792 | checkpoint True
9
+ type eval | step 1750 | loss 1.9354 | checkpoint True
10
+ type eval | step 2000 | loss 1.8793 | checkpoint True
11
+ type eval | step 2250 | loss 1.8414 | checkpoint True
12
+ type eval | step 2500 | loss 1.8133 | checkpoint True
13
+ type eval | step 2750 | loss 1.7787 | checkpoint True
14
+ type eval | step 3000 | loss 1.7515 | checkpoint True
15
+ type eval | step 3250 | loss 1.7513 | checkpoint True
16
+ type eval | step 3500 | loss 1.7423 | checkpoint True
17
+ type eval | step 3750 | loss 1.7245 | checkpoint True
18
+ type eval | step 4000 | loss 1.7110 | checkpoint True
19
+ type eval | step 4250 | loss 1.7084 | checkpoint True
20
+ type eval | step 4500 | loss 1.6869 | checkpoint True
21
+ type eval | step 4750 | loss 1.6772 | checkpoint True
22
+ type eval | step 5000 | loss 1.6680 | checkpoint True
23
+ type eval | step 5250 | loss 1.6416 | checkpoint True
24
+ type eval | step 5500 | loss 1.6447 | checkpoint False
25
+ type eval | step 5750 | loss 1.6367 | checkpoint True
26
+ type eval | step 6000 | loss 1.6424 | checkpoint False
27
+ type eval | step 6250 | loss 1.6370 | checkpoint False
28
+ type eval | step 6500 | loss 1.6343 | checkpoint True
29
+ type eval | step 6750 | loss 1.6303 | checkpoint True
30
+ type eval | step 7000 | loss 1.6229 | checkpoint True
31
+ type eval | step 7250 | loss 1.6121 | checkpoint True
32
+ type eval | step 7500 | loss 1.6108 | checkpoint True
33
+ type eval | step 7750 | loss 1.6004 | checkpoint True
34
+ type eval | step 8000 | loss 1.5971 | checkpoint True
35
+ type eval | step 8250 | loss 1.5931 | checkpoint True
36
+ type eval | step 8500 | loss 1.5917 | checkpoint True
37
+ type eval | step 8750 | loss 1.5924 | checkpoint False
38
+ type eval | step 9000 | loss 1.5881 | checkpoint True
39
+ type eval | step 9250 | loss 1.5856 | checkpoint True
40
+ type eval | step 9500 | loss 1.5845 | checkpoint True
41
+ type eval | step 9750 | loss 1.5839 | checkpoint True
42
+ type eval | step 10000 | loss 1.5833 | checkpoint True
43
+ type eval | step 0 | loss 4.8544 | checkpoint False
44
+ type eval | step 250 | loss 3.0956 | checkpoint True
45
+ type eval | step 500 | loss 2.4943 | checkpoint True
46
+ type eval | step 750 | loss 2.3488 | checkpoint True
47
+ type eval | step 1000 | loss 2.1962 | checkpoint True
48
+ type eval | step 1250 | loss 2.0515 | checkpoint True
49
+ type eval | step 1500 | loss 1.9533 | checkpoint True
50
+ type eval | step 1750 | loss 1.9070 | checkpoint True
51
+ type eval | step 2000 | loss 1.8539 | checkpoint True
52
+ type eval | step 2250 | loss 1.8142 | checkpoint True
53
+ type eval | step 2500 | loss 1.7810 | checkpoint True
54
+ type eval | step 2750 | loss 1.7395 | checkpoint True
55
+ type eval | step 3000 | loss 1.7162 | checkpoint True
56
+ type eval | step 3250 | loss 1.7088 | checkpoint True
57
+ type eval | step 3500 | loss 1.6995 | checkpoint True
58
+ type eval | step 3750 | loss 1.6880 | checkpoint True
59
+ type eval | step 4000 | loss 1.6766 | checkpoint True
60
+ type eval | step 4250 | loss 1.6753 | checkpoint True
61
+ type eval | step 4500 | loss 1.6584 | checkpoint True
62
+ type eval | step 4750 | loss 1.6419 | checkpoint True
63
+ type eval | step 5000 | loss 1.6300 | checkpoint True
64
+ type eval | step 5250 | loss 1.6117 | checkpoint True
65
+ type eval | step 5500 | loss 1.6086 | checkpoint True
66
+ type eval | step 5750 | loss 1.6152 | checkpoint False
67
+ type eval | step 6000 | loss 1.6096 | checkpoint False
68
+ type eval | step 6250 | loss 1.6190 | checkpoint False
69
+ type eval | step 6500 | loss 1.6118 | checkpoint False
70
+ type eval | step 6750 | loss 1.6055 | checkpoint True
71
+ type eval | step 7000 | loss 1.6066 | checkpoint False
72
+ type eval | step 7250 | loss 1.6080 | checkpoint False
73
+ type eval | step 7500 | loss 1.6004 | checkpoint True
74
+ type eval | step 7750 | loss 1.5821 | checkpoint True
75
+ type eval | step 8000 | loss 1.5792 | checkpoint True
76
+ type eval | step 8250 | loss 1.5776 | checkpoint True
77
+ type eval | step 8500 | loss 1.5804 | checkpoint False
78
+ type eval | step 8750 | loss 1.5805 | checkpoint False
79
+ type eval | step 9000 | loss 1.5763 | checkpoint True
80
+ type eval | step 9250 | loss 1.5718 | checkpoint True
81
+ type eval | step 9500 | loss 1.5695 | checkpoint True
82
+ type eval | step 9750 | loss 1.5686 | checkpoint True
83
+ type eval | step 10000 | loss 1.5683 | checkpoint True
model.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"block_size": 128, "vocab_size": 128, "n_layer": 4, "n_head": 4, "n_embd": 64}
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b7a990722c440fdbd1e79a78b08a6f65f74fe96acfa49e5ef4cd5c8936a3c85
3
+ size 870720
train.log ADDED
The diff for this file is too large to render. See raw diff