mokcho commited on
Commit
49cb2ac
·
verified ·
1 Parent(s): cf02085

Upload folder using huggingface_hub

Browse files
Aligned-Lambda_0.01/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc5bd77463f6960a25f5e61acd5747ba018bd70789d97a3072991ec5011c2707
3
  size 816784440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2e45c93f2c9a1bc3bbd3debe593d29c4892ad1e22042ba6d421dd6607453334
3
  size 816784440
Aligned-Lambda_0.01/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:355bab60eae9d0d1b496e7f9899ea416cee7589c5dcd28d12b61fecbf5469e27
3
- size 1633695418
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5929e1de42a6585c3fff66c4a8793456272f5f9c4005370a94e2e66fba0838ee
3
+ size 1621691322
Aligned-Lambda_0.01/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:99004c1fd84e57572d9f83eba4950606de6bd58e5072453bb997c850d270379d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8827c23778eb132d4af0295d85486efedc52cc8e7d8260084819c1f5b168cedd
3
  size 14244
Aligned-Lambda_0.01/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5d476c12da381846e3dd23747ce593bb7d9068ba7200583b801a8ee542adb9c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24d1adeb5a474175a1cb9c0c1efabaa9fd87343414b2349dd1c68b0a5ef6eb54
3
  size 1064
Aligned-Lambda_0.01/trainer_state.json CHANGED
@@ -3,49 +3,63 @@
3
  "best_model_checkpoint": null,
4
  "epoch": 5.0,
5
  "eval_steps": 500,
6
- "global_step": 980,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
- "eval_loss": 1.4304096334510379,
14
- "las": 71.23526844172851,
15
- "step": 196,
16
- "uas": 80.8380619816674
 
 
 
 
 
 
 
17
  },
18
  {
19
  "epoch": 2.0,
20
- "eval_loss": 1.6784026655885909,
21
- "las": 71.97730248799651,
22
- "step": 392,
23
- "uas": 81.09995635093846
24
  },
25
  {
26
  "epoch": 2.5510204081632653,
27
- "grad_norm": 3.334646701812744,
28
  "learning_rate": 9.79591836734694e-05,
29
- "loss": 0.6566,
30
- "step": 500
31
  },
32
  {
33
  "epoch": 3.0,
34
- "eval_loss": 1.840693520175086,
35
- "las": 73.98515931907464,
36
- "step": 588,
37
- "uas": 82.19118288956787
 
 
 
 
 
 
 
38
  },
39
  {
40
  "epoch": 4.0,
41
- "eval_loss": 2.0934243467119007,
42
- "las": 75.03273679615889,
43
- "step": 784,
44
- "uas": 83.23876036665212
45
  }
46
  ],
47
  "logging_steps": 500,
48
- "max_steps": 980,
49
  "num_input_tokens_seen": 0,
50
  "num_train_epochs": 5,
51
  "save_steps": 500,
@@ -62,7 +76,7 @@
62
  }
63
  },
64
  "total_flos": 2.157644595068928e+16,
65
- "train_batch_size": 128,
66
  "trial_name": null,
67
  "trial_params": null
68
  }
 
3
  "best_model_checkpoint": null,
4
  "epoch": 5.0,
5
  "eval_steps": 500,
6
+ "global_step": 1960,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
+ "eval_loss": 1.3433195443616972,
14
+ "las": 72.15189873417721,
15
+ "step": 392,
16
+ "uas": 81.66739415102575
17
+ },
18
+ {
19
+ "epoch": 1.2755102040816326,
20
+ "grad_norm": 3.317138433456421,
21
+ "learning_rate": 0.00014897959183673472,
22
+ "loss": 0.8116,
23
+ "step": 500
24
  },
25
  {
26
  "epoch": 2.0,
27
+ "eval_loss": 1.6489886807070837,
28
+ "las": 73.24312527280664,
29
+ "step": 784,
30
+ "uas": 81.92928852029682
31
  },
32
  {
33
  "epoch": 2.5510204081632653,
34
+ "grad_norm": 2.6405985355377197,
35
  "learning_rate": 9.79591836734694e-05,
36
+ "loss": 0.2626,
37
+ "step": 1000
38
  },
39
  {
40
  "epoch": 3.0,
41
+ "eval_loss": 1.7524571418762207,
42
+ "las": 73.15582714971627,
43
+ "step": 1176,
44
+ "uas": 82.36577913574858
45
+ },
46
+ {
47
+ "epoch": 3.826530612244898,
48
+ "grad_norm": 2.158872365951538,
49
+ "learning_rate": 4.6938775510204086e-05,
50
+ "loss": 0.1222,
51
+ "step": 1500
52
  },
53
  {
54
  "epoch": 4.0,
55
+ "eval_loss": 2.07165447043048,
56
+ "las": 75.46922741161065,
57
+ "step": 1568,
58
+ "uas": 84.37363596682671
59
  }
60
  ],
61
  "logging_steps": 500,
62
+ "max_steps": 1960,
63
  "num_input_tokens_seen": 0,
64
  "num_train_epochs": 5,
65
  "save_steps": 500,
 
76
  }
77
  },
78
  "total_flos": 2.157644595068928e+16,
79
+ "train_batch_size": 64,
80
  "trial_name": null,
81
  "trial_params": null
82
  }
Aligned-Lambda_0.01/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af0b7394ee2616ee6e9aafc0bc0af72d9263830be872095653d6f2828f4da29e
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7326d9c14a114d21305aad1da3809b18ba888b85d0a6856daaf707da4ee5a7a0
3
  size 5432