nazbijari commited on
Commit
1706925
·
1 Parent(s): 883261e

Upload 10 files

Browse files
config.json CHANGED
@@ -3,12 +3,12 @@
3
  "architectures": [
4
  "EsmForSequenceClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.0,
7
  "classifier_dropout": null,
8
  "emb_layer_norm_before": false,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.0,
12
  "hidden_size": 320,
13
  "id2label": {
14
  "0": "LABEL_0",
@@ -45,8 +45,7 @@
45
  "31": "LABEL_31",
46
  "32": "LABEL_32",
47
  "33": "LABEL_33",
48
- "34": "LABEL_34",
49
- "35": "LABEL_35"
50
  },
51
  "initializer_range": 0.02,
52
  "intermediate_size": 1280,
@@ -81,7 +80,6 @@
81
  "LABEL_32": 32,
82
  "LABEL_33": 33,
83
  "LABEL_34": 34,
84
- "LABEL_35": 35,
85
  "LABEL_4": 4,
86
  "LABEL_5": 5,
87
  "LABEL_6": 6,
@@ -100,7 +98,7 @@
100
  "problem_type": "single_label_classification",
101
  "token_dropout": true,
102
  "torch_dtype": "float32",
103
- "transformers_version": "4.27.1",
104
  "use_cache": true,
105
  "vocab_list": null,
106
  "vocab_size": 33
 
3
  "architectures": [
4
  "EsmForSequenceClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
  "emb_layer_norm_before": false,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
  "hidden_size": 320,
13
  "id2label": {
14
  "0": "LABEL_0",
 
45
  "31": "LABEL_31",
46
  "32": "LABEL_32",
47
  "33": "LABEL_33",
48
+ "34": "LABEL_34"
 
49
  },
50
  "initializer_range": 0.02,
51
  "intermediate_size": 1280,
 
80
  "LABEL_32": 32,
81
  "LABEL_33": 33,
82
  "LABEL_34": 34,
 
83
  "LABEL_4": 4,
84
  "LABEL_5": 5,
85
  "LABEL_6": 6,
 
98
  "problem_type": "single_label_classification",
99
  "token_dropout": true,
100
  "torch_dtype": "float32",
101
+ "transformers_version": "4.29.1",
102
  "use_cache": true,
103
  "vocab_list": null,
104
  "vocab_size": 33
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c90e7532b0272bdf5c55db6e70014a52965094981f44e362fa067ebf560f63d3
3
- size 60245413
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0967da3e7ca38ae3fea458f7bd481901eb656987639bb938759380292bc5761
3
+ size 60243653
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2697ec14d37fe31d23927f74d679b2c7bc414a5f4a504b12d1b6d9d4afa6a5e
3
- size 31451615
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb30a6ff634c55a717c8b893c3ea8f504cecb400b9d6764af1f4db7cd2ac8337
3
+ size 31451537
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff0b7d688f72cecc3d6db8a5558dfafab2669c75076ea7f5e8d0318cba901963
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3559bb4af107acbd4e612a25c108e91c43da284afbaf118f0b641f7406c47930
3
+ size 14575
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27b432d74b4d0b03006ee21d4ba448080b3156a1f1e2474e15085f79317224c6
3
- size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9233e9f4636c467a1db89ac591b0bf5917f4bab0044b9a83499b29a2b11c13e
3
+ size 627
tokenizer_config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
 
2
  "model_max_length": 1024,
3
- "special_tokens_map_file": "/home/ubuntu/.cache/huggingface/hub/models--facebook--esm2_t6_8M_UR50D/snapshots/97bbd7b3dc8acb222027cde0d8e911f3de78d5b8/special_tokens_map.json",
4
  "tokenizer_class": "EsmTokenizer"
5
  }
 
1
  {
2
+ "clean_up_tokenization_spaces": true,
3
  "model_max_length": 1024,
 
4
  "tokenizer_class": "EsmTokenizer"
5
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2fa012567c32902a900f6a53cfb9f0d49f5247dc7e97a5e8f4df93cf3909aa60
3
- size 3567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01d4310c8c5489b695a58fd3a5f38aaede700774e709f6ae6e6d5fc9c1efecac
3
+ size 3899