mgh6 commited on
Commit
efe6f13
·
1 Parent(s): 70b670e

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "facebook/esm2_t33_650M_UR50D",
3
  "architectures": [
4
  "EsmForMaskedLM"
5
  ],
@@ -9,16 +9,16 @@
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.0,
12
- "hidden_size": 1280,
13
  "initializer_range": 0.02,
14
- "intermediate_size": 5120,
15
  "is_folding_model": false,
16
  "layer_norm_eps": 1e-05,
17
  "mask_token_id": 32,
18
  "max_position_embeddings": 1026,
19
  "model_type": "esm",
20
  "num_attention_heads": 20,
21
- "num_hidden_layers": 33,
22
  "pad_token_id": 1,
23
  "position_embedding_type": "rotary",
24
  "token_dropout": true,
 
1
  {
2
+ "_name_or_path": "facebook/esm2_t30_150M_UR50D",
3
  "architectures": [
4
  "EsmForMaskedLM"
5
  ],
 
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.0,
12
+ "hidden_size": 640,
13
  "initializer_range": 0.02,
14
+ "intermediate_size": 2560,
15
  "is_folding_model": false,
16
  "layer_norm_eps": 1e-05,
17
  "mask_token_id": 32,
18
  "max_position_embeddings": 1026,
19
  "model_type": "esm",
20
  "num_attention_heads": 20,
21
+ "num_hidden_layers": 30,
22
  "pad_token_id": 1,
23
  "position_embedding_type": "rotary",
24
  "token_dropout": true,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e34b0dad04f31cc5e40e6bc109ba43f342d1a0246b6bfd8d5d31ef60790e44c2
3
- size 2609633793
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da42b87e61ab46a51bc6f77c6e9e07d4f69303ffc2a3e16538dd3ee46f53f94a
3
+ size 595375093
runs/Nov06_15-50-42_bc13u7n1/events.out.tfevents.1699307495.bc13u7n1.55769.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e87be4d7ef39888a6feda96268f90bc699346e42571925890b6ec04c2431da58
3
+ size 4120
runs/Nov06_15-52-35_bc13u7n1/events.out.tfevents.1699307557.bc13u7n1.56372.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ec25b6e1da150ae2578d5feb7ccf504c27a34b592776b9fcfe92c40da1371d4
3
+ size 4122
runs/Nov06_15-54-14_bc13u7n1/events.out.tfevents.1699307656.bc13u7n1.56716.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acef3e2b94e80baa93f0e329a322033acfc60bf11e504cb6bb98d7533e5fea5c
3
+ size 4120
runs/Nov06_15-56-00_bc13u7n1/events.out.tfevents.1699307762.bc13u7n1.57053.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:136df6491dd187532589554c89c6f49ecc96302a38d0dd60e9ebf7f9ca821474
3
+ size 4902
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7087d574ab9d5db5ae1fffb15d24ec5d9060bb9c0284e269f1b57bbeee9cfd9
3
  size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbb86aa6ca77a5d92ca93fcbffe8d7248173551c99f1d31aaf6d6dcc822fe763
3
  size 3963