andyjzhao commited on
Commit
0355e3e
·
verified ·
1 Parent(s): a912476

Upload folder using huggingface_hub

Browse files
checkpoint-10/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8ea87b5853946db38bf2851926789b58bcb47715a2ca9212c2d535b156cbae2
3
  size 558664971
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c17840d339873628ec2a78a677ab2cbfbb5c03e76c940568a60226f2ea00041e
3
  size 558664971
checkpoint-10/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed89753effbec333f472b018394cab0286fdba0e80049740efaa89bdbed93fba
3
  size 279336283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:228064a54d4acf279d2af574895d2df9fe663db84d793900c7543edeb6586b58
3
  size 279336283
checkpoint-10/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fe6b873f1da64bb980afe982fc1c6b6bc461ad8c33073d856804323f64b26d9
3
+ size 15429
checkpoint-10/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13db4dd10b22b5464f5ef0b539152b46390eae6a827b355108cb54b1af2eb830
3
+ size 15429
checkpoint-10/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06e293f13fb6e873e8081636b6b07115410b397c55f1ed08b8772b059c8eb74e
3
+ size 15429
checkpoint-10/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c595c077664f94b7f91088561951be20e39483e01865774a83bc46f3971a3152
3
+ size 15429
checkpoint-10/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_global_step": 10,
3
- "best_metric": 7.294625624845928e+17,
4
  "best_model_checkpoint": "/network/scratch/j/jianan.zhao/DNAFM/output/gencode_human_12.8k_12800/CKPT_DEBUG/checkpoint-10",
5
- "epoch": 0.00035457220863028757,
6
  "eval_steps": 10,
7
  "global_step": 10,
8
  "is_hyper_param_search": false,
@@ -12,60 +12,60 @@
12
  {
13
  "comp/rl_weight": 0.03,
14
  "comp/strictness": 0.0,
15
- "epoch": 0.00035457220863028757,
16
- "grad_norm": 1130.100830078125,
17
- "loss": 146.4512,
18
- "loss_ce": 54.19949722290039,
19
- "loss_region": 0.10721047222614288,
20
- "loss_total": 54.30670928955078,
21
  "lr": 2.20454076850486e-05,
22
- "router/selected_tokens_s0": 24.15625,
23
  "router/selected_tokens_s1": 1.0,
24
  "step": 10,
25
- "tokens_trained": 0.00409568
26
  },
27
  {
28
- "epoch": 0.00035457220863028757,
29
- "eval_ppl": 7.294625624845928e+17,
30
- "eval_runtime": 6.6402,
31
  "step": 10,
32
- "tokens_trained": 0.00409568
33
  },
34
  {
35
- "epoch": 0.00035457220863028757,
36
  "eval_F": 7.813110399249941e-05,
37
- "eval_F_cds": 4.744345662481978e-05,
38
  "eval_F_dig": 7.813110399249941e-05,
39
- "eval_F_exon": 6.977782739756614e-05,
40
- "eval_F_intron": 8.580132276805985e-05,
41
- "eval_F_nig": 8.365496566708533e-05,
42
- "eval_F_promoter": 4.770454407466195e-05,
43
- "eval_F_utr": 7.9026626119178e-05,
44
- "eval_G": 0.00016736338285193968,
45
- "eval_G_cds": 0.00016425546555209688,
46
- "eval_G_dig": 0.000140951360132302,
47
- "eval_G_exon": 0.0001619063651334152,
48
- "eval_G_intron": 0.00016938910394235382,
49
- "eval_G_nig": 0.0001631118140770298,
50
- "eval_G_promoter": 0.00016429247622091106,
51
- "eval_G_utr": 0.00019291894904831453,
52
  "eval_avg_bp_per_token": 12799.0,
53
- "eval_bp_per_token/cds": 21077.722222222223,
54
  "eval_bp_per_token/dig": 12799.0,
55
- "eval_bp_per_token/exon": 14331.2,
56
- "eval_bp_per_token/intron": 11654.831973898858,
57
- "eval_bp_per_token/nig": 11953.86301369863,
58
- "eval_bp_per_token/promoter": 20962.363636363636,
59
- "eval_bp_per_token/utr": 12653.962962962964,
60
- "eval_ppl_cds": 2.311600502337928e+20,
61
- "eval_ppl_dig": 9.652397143812314e+16,
62
- "eval_ppl_exon": 2.2747493338461386e+18,
63
- "eval_ppl_intron": 2.793470782953012e+17,
64
- "eval_ppl_nig": 1.1026701972335622e+17,
65
- "eval_ppl_promoter": 1.5383074335283397e+20,
66
- "eval_ppl_utr": 4.222450573520073e+17,
67
  "step": 10,
68
- "tokens_trained": 0.00409568
69
  }
70
  ],
71
  "logging_steps": 10,
 
1
  {
2
  "best_global_step": 10,
3
+ "best_metric": 1.698420053510286e+17,
4
  "best_model_checkpoint": "/network/scratch/j/jianan.zhao/DNAFM/output/gencode_human_12.8k_12800/CKPT_DEBUG/checkpoint-10",
5
+ "epoch": 0.0014184397163120568,
6
  "eval_steps": 10,
7
  "global_step": 10,
8
  "is_hyper_param_search": false,
 
12
  {
13
  "comp/rl_weight": 0.03,
14
  "comp/strictness": 0.0,
15
+ "epoch": 0.0014184397163120568,
16
+ "grad_norm": 904.2174682617188,
17
+ "loss": 144.6514,
18
+ "loss_ce": 54.04480743408203,
19
+ "loss_region": 0.1044125035405159,
20
+ "loss_total": 54.14921951293945,
21
  "lr": 2.20454076850486e-05,
22
+ "router/selected_tokens_s0": 22.09375,
23
  "router/selected_tokens_s1": 1.0,
24
  "step": 10,
25
+ "tokens_trained": 0.01638272
26
  },
27
  {
28
+ "epoch": 0.0014184397163120568,
29
+ "eval_ppl": 1.698420053510286e+17,
30
+ "eval_runtime": 1.7385,
31
  "step": 10,
32
+ "tokens_trained": 0.01638272
33
  },
34
  {
35
+ "epoch": 0.0014184397163120568,
36
  "eval_F": 7.813110399249941e-05,
37
+ "eval_F_cds": 2.439649178448139e-05,
38
  "eval_F_dig": 7.813110399249941e-05,
39
+ "eval_F_exon": 6.169412055031155e-05,
40
+ "eval_F_intron": 8.631567698207511e-05,
41
+ "eval_F_nig": 8.867784676802712e-05,
42
+ "eval_F_promoter": 2.898993785611503e-05,
43
+ "eval_F_utr": 0.00013418766144453016,
44
+ "eval_G": 0.00021740636410105978,
45
+ "eval_G_cds": 0.00017406325095451275,
46
+ "eval_G_dig": 0.00015412581061020393,
47
+ "eval_G_exon": 0.00024593300789684745,
48
+ "eval_G_intron": 0.000209744205032615,
49
+ "eval_G_nig": 0.0001933390126892979,
50
+ "eval_G_promoter": 0.000290331756895652,
51
+ "eval_G_utr": 0.0002710944576302459,
52
  "eval_avg_bp_per_token": 12799.0,
53
+ "eval_bp_per_token/cds": 40989.5,
54
  "eval_bp_per_token/dig": 12799.0,
55
+ "eval_bp_per_token/exon": 16209.0,
56
+ "eval_bp_per_token/intron": 11585.380952380952,
57
+ "eval_bp_per_token/nig": 11276.77358490566,
58
+ "eval_bp_per_token/promoter": 34494.72727272727,
59
+ "eval_bp_per_token/utr": 7452.25,
60
+ "eval_ppl_cds": 2.889123597693874e+19,
61
+ "eval_ppl_dig": 3.430802770811005e+17,
62
+ "eval_ppl_exon": 4.6340490808362584e+16,
63
+ "eval_ppl_intron": 9.889931203838952e+16,
64
+ "eval_ppl_nig": 3.3020604346559384e+16,
65
+ "eval_ppl_promoter": 7.55131215186589e+19,
66
+ "eval_ppl_utr": 1.4233813097089611e+17,
67
  "step": 10,
68
+ "tokens_trained": 0.01638272
69
  }
70
  ],
71
  "logging_steps": 10,
checkpoint-10/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4dbdc20d0327174126de42adf35e1c0844ec1a2a3097ba550495f97e879180f
3
  size 5969
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:912c85f83733bc75ddebd721c1bb21205af4115ec87e6bbd7790ca686f00482a
3
  size 5969