mgh6 commited on
Commit
47dabed
·
1 Parent(s): 452f604

Training in progress, step 12000

Browse files
config.json CHANGED
@@ -3,12 +3,12 @@
3
  "architectures": [
4
  "EsmForMaskedLM"
5
  ],
6
- "attention_probs_dropout_prob": 0.0,
7
  "classifier_dropout": null,
8
  "emb_layer_norm_before": false,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.0,
12
  "hidden_size": 480,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 1920,
 
3
  "architectures": [
4
  "EsmForMaskedLM"
5
  ],
6
+ "attention_probs_dropout_prob": 0.01,
7
  "classifier_dropout": null,
8
  "emb_layer_norm_before": false,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.01,
12
  "hidden_size": 480,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 1920,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ff650ff2d5d023cccc9a07e3b14226be74bafcc4b8f2f0fa06145eb5e5dd063
3
  size 136000488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2073cb1275b373d00d8a8c56ebe1b5db47ef2ea86942205741292e3a15aff312
3
  size 136000488
runs/Jan07_01-07-27_bc11u19n2/events.out.tfevents.1704611247.bc11u19n2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63867f581538f5b12c38a271c8c0a26a06fe2c743e5f3688c83059427cb76589
3
- size 16859
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc4198bc296cf60785d94503be81ba90af62989327ab465ad841c75e815caedc
3
+ size 17130
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef01ff1b7d7ad1828810ac097acd71e769242be725a347ed694092e764f46278
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5902a7e388c7e70f0556c5bd656e7e6da8e777ef4e0a2adaa4abb451a261743
3
  size 4283