mgh6 commited on
Commit
7dd3c43
·
verified ·
1 Parent(s): 3998f28

Training in progress, step 10

Browse files
config.json CHANGED
@@ -4,11 +4,11 @@
4
  "EsmForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.01,
7
- "classifier_dropout": 0.2,
8
  "emb_layer_norm_before": false,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.2,
12
  "hidden_size": 480,
13
  "id2label": {
14
  "0": "LABEL_0"
 
4
  "EsmForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.01,
7
+ "classifier_dropout": 0.15,
8
  "emb_layer_norm_before": false,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.15,
12
  "hidden_size": 480,
13
  "id2label": {
14
  "0": "LABEL_0"
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ff2e1acdf825d923ba39c77ec992105736efb153b424a3a2a8e49ef2f6a98ae
3
  size 135998360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92b9f02454017a3dd1ced10d0f13a66c2417e009d6cdf60e1ef1e077ae8a77b8
3
  size 135998360
runs/Oct15_06-18-13_torch-flash-large-0-2/events.out.tfevents.1728973391.torch-flash-large-0-2.10244.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:894beef03f4a78e17ad082f51f75c2cbc52b274b4cdbd5a16355676e12c5d73b
3
+ size 306
runs/Oct15_06-25-58_torch-flash-large-0-2/events.out.tfevents.1728973563.torch-flash-large-0-2.10394.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f869dd4940a6b463d6b08397321211974e34d094369ab69607a5df9604fc803
3
+ size 5566
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6b75c699ad6ee47eea326fc5d100132083eedf4cdca50442d4edf9dac61dc25
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c198bc00a032b2738abc96e46549b7b570bb5c4820dc6422c8aae98ca45fe43
3
  size 5240