mohsennp commited on
Commit
b345647
·
verified ·
1 Parent(s): 97bd5d5

Upload DeCodon

Browse files
Files changed (2) hide show
  1. config.json +4 -4
  2. model.safetensors +2 -2
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/large_experiments/goodarzilab/mohsen/biofm/saved_models/CodonGPT/CodonGPT_L2048_l12_a16_b20_r5e-05_wd0.001_g1-w801oh4r/checkpoint-1080000",
3
  "architectures": [
4
  "DeCodon"
5
  ],
@@ -10,9 +10,9 @@
10
  "gamma_init": 1.782709687623856,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
- "hidden_size": 2048,
14
  "initializer_range": 0.02,
15
- "intermediate_size": 8192,
16
  "is_decoder": true,
17
  "layer_norm_eps": 1e-12,
18
  "lm_type": "distilled",
@@ -29,5 +29,5 @@
29
  "use_cache": true,
30
  "use_flash_attn": true,
31
  "use_rotary_emb": true,
32
- "vocab_size": 69
33
  }
 
1
  {
2
+ "_name_or_path": "/large_experiments/goodarzilab/mohsen/biofm/saved_models/CodonGPT/CodonGPT_L2048_l12_a16_b40_r5e-05_wd0.01_g2_taxid-z5jw1bpn/checkpoint-460000",
3
  "architectures": [
4
  "DeCodon"
5
  ],
 
10
  "gamma_init": 1.782709687623856,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 1024,
14
  "initializer_range": 0.02,
15
+ "intermediate_size": 4096,
16
  "is_decoder": true,
17
  "layer_norm_eps": 1e-12,
18
  "lm_type": "distilled",
 
29
  "use_cache": true,
30
  "use_flash_attn": true,
31
  "use_rotary_emb": true,
32
+ "vocab_size": 1736
33
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86cdc82e426c19bc748c5decf2a6bc245b69a92fa21091dc115caaad522de19c
3
- size 2452973200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54744307bb9215d44a912f5a571ba8b36954c0685849331575d69e6387d7ebd2
3
+ size 631984656