prasadsachin commited on
Commit
0634f0f
·
verified ·
1 Parent(s): ebbd19b

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: keras-hub
3
+ ---
4
+ This is a [`ESM` model](https://keras.io/api/keras_hub/models/esm) uploaded using the KerasHub library and can be used with JAX, TensorFlow, and PyTorch backends.
5
+ This model is related to a `MaskedPLM` task.
6
+
7
+ Model config:
8
+ * **name:** esm_backbone
9
+ * **trainable:** True
10
+ * **vocabulary_size:** 33
11
+ * **num_layers:** 33
12
+ * **num_heads:** 20
13
+ * **hidden_dim:** 1280
14
+ * **intermediate_dim:** 5120
15
+ * **dropout:** 0.0
16
+ * **max_wavelength:** 10000
17
+ * **use_bias:** True
18
+ * **activation:** gelu
19
+ * **layer_norm_eps:** 1e-05
20
+ * **use_pre_layer_norm:** False
21
+ * **position_embedding_type:** rotary
22
+ * **max_sequence_length:** 1026
23
+ * **pad_token_id:** 1
24
+
25
+ This model card has been generated automatically and should be completed by the model author. See [Model Cards documentation](https://huggingface.co/docs/hub/model-cards) for more information.
assets/tokenizer/vocabulary.txt ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <cls>
2
+ <pad>
3
+ <eos>
4
+ <unk>
5
+ L
6
+ A
7
+ G
8
+ V
9
+ S
10
+ E
11
+ R
12
+ T
13
+ I
14
+ D
15
+ P
16
+ K
17
+ Q
18
+ N
19
+ F
20
+ Y
21
+ M
22
+ H
23
+ W
24
+ C
25
+ X
26
+ B
27
+ U
28
+ Z
29
+ O
30
+ .
31
+ -
32
+ <null_1>
33
+ <mask>
config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "module": "keras_hub.src.models.esm.esm_backbone",
3
+ "class_name": "ESMBackbone",
4
+ "config": {
5
+ "name": "esm_backbone",
6
+ "trainable": true,
7
+ "vocabulary_size": 33,
8
+ "num_layers": 33,
9
+ "num_heads": 20,
10
+ "hidden_dim": 1280,
11
+ "intermediate_dim": 5120,
12
+ "dropout": 0.0,
13
+ "max_wavelength": 10000,
14
+ "use_bias": true,
15
+ "activation": "gelu",
16
+ "layer_norm_eps": 1e-05,
17
+ "use_pre_layer_norm": false,
18
+ "position_embedding_type": "rotary",
19
+ "max_sequence_length": 1026,
20
+ "pad_token_id": 1
21
+ },
22
+ "registered_name": "keras_hub>ESMBackbone"
23
+ }
metadata.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "keras_version": "3.10.0",
3
+ "keras_hub_version": "0.22.0.dev0",
4
+ "parameter_count": 649400320,
5
+ "date_saved": "2025-08-11@22:55:00",
6
+ "tasks": [
7
+ "MaskedLM",
8
+ "TextClassifier"
9
+ ]
10
+ }
model.weights.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1f68d3ecebee597087f52effa063742068d7e484be9cb4d06e091359c2779de
3
+ size 2598592024
preprocessor.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "module": "keras_hub.src.models.esm.esm_masked_plm_preprocessor",
3
+ "class_name": "ESMMaskedPLMPreprocessor",
4
+ "config": {
5
+ "name": "esm_masked_plm_preprocessor_1",
6
+ "trainable": true,
7
+ "dtype": {
8
+ "module": "keras",
9
+ "class_name": "DTypePolicy",
10
+ "config": {
11
+ "name": "float32"
12
+ },
13
+ "registered_name": null
14
+ },
15
+ "tokenizer": {
16
+ "module": "keras_hub.src.models.esm.esm_tokenizer",
17
+ "class_name": "ESMTokenizer",
18
+ "config": {
19
+ "name": "esm_tokenizer",
20
+ "trainable": true,
21
+ "dtype": {
22
+ "module": "keras",
23
+ "class_name": "DTypePolicy",
24
+ "config": {
25
+ "name": "int32"
26
+ },
27
+ "registered_name": null
28
+ },
29
+ "config_file": "tokenizer.json",
30
+ "vocabulary": null,
31
+ "sequence_length": null,
32
+ "lowercase": false,
33
+ "strip_accents": false,
34
+ "split": true,
35
+ "suffix_indicator": "##",
36
+ "oov_token": "<unk>",
37
+ "special_tokens": null,
38
+ "special_tokens_in_strings": false
39
+ },
40
+ "registered_name": "keras_hub>ESMTokenizer"
41
+ },
42
+ "config_file": "preprocessor.json",
43
+ "sequence_length": 512,
44
+ "truncate": "round_robin",
45
+ "mask_selection_rate": 0.15,
46
+ "mask_selection_length": 96,
47
+ "mask_token_rate": 0.8,
48
+ "random_token_rate": 0.1
49
+ },
50
+ "registered_name": "keras_hub>ESMMaskedPLMPreprocessor"
51
+ }
task.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "module": "keras_hub.src.models.esm.esm_masked_plm",
3
+ "class_name": "ESMMaskedPLM",
4
+ "config": {
5
+ "backbone": {
6
+ "module": "keras_hub.src.models.esm.esm_backbone",
7
+ "class_name": "ESMBackbone",
8
+ "config": {
9
+ "name": "esm_backbone",
10
+ "trainable": true,
11
+ "vocabulary_size": 33,
12
+ "num_layers": 33,
13
+ "num_heads": 20,
14
+ "hidden_dim": 1280,
15
+ "intermediate_dim": 5120,
16
+ "dropout": 0.0,
17
+ "max_wavelength": 10000,
18
+ "use_bias": true,
19
+ "activation": "gelu",
20
+ "layer_norm_eps": 1e-05,
21
+ "use_pre_layer_norm": false,
22
+ "position_embedding_type": "rotary",
23
+ "max_sequence_length": 1026,
24
+ "pad_token_id": 1
25
+ },
26
+ "registered_name": "keras_hub>ESMBackbone"
27
+ },
28
+ "preprocessor": {
29
+ "module": "keras_hub.src.models.esm.esm_masked_plm_preprocessor",
30
+ "class_name": "ESMMaskedPLMPreprocessor",
31
+ "config": {
32
+ "name": "esm_masked_plm_preprocessor_1",
33
+ "trainable": true,
34
+ "dtype": {
35
+ "module": "keras",
36
+ "class_name": "DTypePolicy",
37
+ "config": {
38
+ "name": "float32"
39
+ },
40
+ "registered_name": null
41
+ },
42
+ "tokenizer": {
43
+ "module": "keras_hub.src.models.esm.esm_tokenizer",
44
+ "class_name": "ESMTokenizer",
45
+ "config": {
46
+ "name": "esm_tokenizer",
47
+ "trainable": true,
48
+ "dtype": {
49
+ "module": "keras",
50
+ "class_name": "DTypePolicy",
51
+ "config": {
52
+ "name": "int32"
53
+ },
54
+ "registered_name": null
55
+ },
56
+ "config_file": "tokenizer.json",
57
+ "vocabulary": null,
58
+ "sequence_length": null,
59
+ "lowercase": false,
60
+ "strip_accents": false,
61
+ "split": true,
62
+ "suffix_indicator": "##",
63
+ "oov_token": "<unk>",
64
+ "special_tokens": null,
65
+ "special_tokens_in_strings": false
66
+ },
67
+ "registered_name": "keras_hub>ESMTokenizer"
68
+ },
69
+ "config_file": "preprocessor.json",
70
+ "sequence_length": 512,
71
+ "truncate": "round_robin",
72
+ "mask_selection_rate": 0.15,
73
+ "mask_selection_length": 96,
74
+ "mask_token_rate": 0.8,
75
+ "random_token_rate": 0.1
76
+ },
77
+ "registered_name": "keras_hub>ESMMaskedPLMPreprocessor"
78
+ },
79
+ "name": "esm_masked_plm"
80
+ },
81
+ "registered_name": "keras_hub>ESMMaskedPLM"
82
+ }
task.weights.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:858d7dea9c135ec9ac15b4c97032ecbca3a21bd575e07d29accbd86ad48c0f7c
3
+ size 2605349592
tokenizer.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "module": "keras_hub.src.models.esm.esm_tokenizer",
3
+ "class_name": "ESMTokenizer",
4
+ "config": {
5
+ "name": "esm_tokenizer",
6
+ "trainable": true,
7
+ "dtype": {
8
+ "module": "keras",
9
+ "class_name": "DTypePolicy",
10
+ "config": {
11
+ "name": "int32"
12
+ },
13
+ "registered_name": null
14
+ },
15
+ "config_file": "tokenizer.json",
16
+ "vocabulary": null,
17
+ "sequence_length": null,
18
+ "lowercase": false,
19
+ "strip_accents": false,
20
+ "split": true,
21
+ "suffix_indicator": "##",
22
+ "oov_token": "<unk>",
23
+ "special_tokens": null,
24
+ "special_tokens_in_strings": false
25
+ },
26
+ "registered_name": "keras_hub>ESMTokenizer"
27
+ }