jsunn-y commited on
Commit
f6398db
·
1 Parent(s): 73ae5fb

adding new GB1 data and APEXGo checkpoints

Browse files
Files changed (26) hide show
  1. APEXGo/saved_models/CreiLOV/exalted-pine-5_model_state_epoch_241-train-zs.csv +3 -0
  2. APEXGo/saved_models/CreiLOV/exalted-pine-5_model_state_epoch_241.pkl +3 -0
  3. APEXGo/saved_models/GB1/absurd-cosmos-6_model_state_epoch_391-train-zs.csv +3 -0
  4. APEXGo/saved_models/GB1/absurd-cosmos-6_model_state_epoch_391.pkl +3 -0
  5. APEXGo/saved_models/TrpB/dainty-dream-9_model_state_epoch_421-train-zs.csv +3 -0
  6. APEXGo/saved_models/TrpB/dainty-dream-9_model_state_epoch_421.pkl +3 -0
  7. checkpoints/causalLM_finetune/CreiLOV/best/config.json +1 -1
  8. checkpoints/causalLM_finetune/CreiLOV/best/generation_config.json +1 -1
  9. checkpoints/causalLM_finetune/CreiLOV/best/model.safetensors +1 -1
  10. checkpoints/causalLM_finetune/GB1/best/config.json +41 -0
  11. checkpoints/causalLM_finetune/GB1/best/generation_config.json +6 -0
  12. checkpoints/causalLM_finetune/GB1/best/model.safetensors +3 -0
  13. checkpoints/causalLM_finetune/TrpB/best/config.json +1 -1
  14. checkpoints/causalLM_finetune/TrpB/best/generation_config.json +1 -1
  15. checkpoints/causalLM_finetune/TrpB/best/model.safetensors +1 -1
  16. checkpoints/continuous/GB1/best_model.ckpt +3 -0
  17. checkpoints/d3pm_finetune/GB1/best_model.ckpt +3 -0
  18. checkpoints/mdlm/GB1/best_model.ckpt +3 -0
  19. data/hmmer/jackhmmer/GB1.domtbl +3 -0
  20. data/hmmer/jackhmmer/GB1_alignment.a2m +3 -0
  21. data/hmmer/jackhmmer/GB1_alignment.fasta +3 -0
  22. data/hmmer/jackhmmer/GB1_alignment.sto +3 -0
  23. data/hmmer/jackhmmer/GB1_seqs.fasta +3 -0
  24. data/raw_data/GB1_fitness/1-s2.0-S0960982214012688-mmc2.xlsx +3 -0
  25. data/raw_data/GB1_fitness/doubles.csv +3 -0
  26. data/raw_data/GB1_fitness/singles.csv +3 -0
APEXGo/saved_models/CreiLOV/exalted-pine-5_model_state_epoch_241-train-zs.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a2291f9d464ca496a901fcaeb46db57f515b48ea40b09f42389abc5b40d786c
3
+ size 297502
APEXGo/saved_models/CreiLOV/exalted-pine-5_model_state_epoch_241.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a5704d25a941f0cf4c10ba8533607eaf9e530546a4de005db87bb27e87471c0
3
+ size 22073641
APEXGo/saved_models/GB1/absurd-cosmos-6_model_state_epoch_391-train-zs.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57d9d043e270258b64aa5e7fdbbb147b64878b16f6fb3ffdfb01ad936e2b37a1
3
+ size 280468
APEXGo/saved_models/GB1/absurd-cosmos-6_model_state_epoch_391.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c15d93c3f37b55911529d609934ac9a4d252518d6d760d2f0a21a0218d6afa8f
3
+ size 22071782
APEXGo/saved_models/TrpB/dainty-dream-9_model_state_epoch_421-train-zs.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58df7e32e2611422b56b34b75ecf6926424fda128237d2da6fda3269705a869a
3
+ size 285353
APEXGo/saved_models/TrpB/dainty-dream-9_model_state_epoch_421.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54246c8a459d7a190ddb2d31c52a1d2f60db338f99d6b90e74b2c26f979ae58a
3
+ size 22073641
checkpoints/causalLM_finetune/CreiLOV/best/config.json CHANGED
@@ -35,7 +35,7 @@
35
  },
36
  "tokenizer_class": "GPT2Tokenizer",
37
  "torch_dtype": "float32",
38
- "transformers_version": "4.47.0",
39
  "use_cache": true,
40
  "vocab_size": 32
41
  }
 
35
  },
36
  "tokenizer_class": "GPT2Tokenizer",
37
  "torch_dtype": "float32",
38
+ "transformers_version": "4.48.1",
39
  "use_cache": true,
40
  "vocab_size": 32
41
  }
checkpoints/causalLM_finetune/CreiLOV/best/generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.47.0"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.48.1"
6
  }
checkpoints/causalLM_finetune/CreiLOV/best/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72d9578bae32e1b50d52e529b1fe5cadf44de5f6957ddde77dc90764563dddbe
3
  size 617190232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:490da1a18a0142fe05be6c68929e0e56b7524c5d8b1af5817375606b5ea39a28
3
  size 617190232
checkpoints/causalLM_finetune/GB1/best/config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "jsunn-y/ProCALM",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "ProGenForCausalLM"
6
+ ],
7
+ "attn_pdrop": 0.0,
8
+ "bos_token_id": 1,
9
+ "embd_pdrop": 0.0,
10
+ "eos_token_id": 2,
11
+ "gradient_checkpointing": false,
12
+ "initializer_range": 0.02,
13
+ "layer_norm_epsilon": 1e-05,
14
+ "model_type": "progen",
15
+ "n_ctx": 2048,
16
+ "n_embd": 1024,
17
+ "n_head": 16,
18
+ "n_inner": null,
19
+ "n_layer": 12,
20
+ "n_positions": 1024,
21
+ "resid_pdrop": 0.0,
22
+ "rotary_dim": 32,
23
+ "scale_attn_weights": true,
24
+ "summary_activation": null,
25
+ "summary_first_dropout": 0.1,
26
+ "summary_proj_to_labels": true,
27
+ "summary_type": "cls_index",
28
+ "summary_use_proj": true,
29
+ "task_specific_params": {
30
+ "text-generation": {
31
+ "do_sample": true,
32
+ "max_length": 50,
33
+ "temperature": 1.0
34
+ }
35
+ },
36
+ "tokenizer_class": "GPT2Tokenizer",
37
+ "torch_dtype": "float32",
38
+ "transformers_version": "4.48.1",
39
+ "use_cache": true,
40
+ "vocab_size": 32
41
+ }
checkpoints/causalLM_finetune/GB1/best/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.48.1"
6
+ }
checkpoints/causalLM_finetune/GB1/best/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4ae50183db2490718b6febec7f26d5d9e21412eeb0e890341ae0ae02fd87597
3
+ size 617190232
checkpoints/causalLM_finetune/TrpB/best/config.json CHANGED
@@ -35,7 +35,7 @@
35
  },
36
  "tokenizer_class": "GPT2Tokenizer",
37
  "torch_dtype": "float32",
38
- "transformers_version": "4.47.0",
39
  "use_cache": true,
40
  "vocab_size": 32
41
  }
 
35
  },
36
  "tokenizer_class": "GPT2Tokenizer",
37
  "torch_dtype": "float32",
38
+ "transformers_version": "4.48.1",
39
  "use_cache": true,
40
  "vocab_size": 32
41
  }
checkpoints/causalLM_finetune/TrpB/best/generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.47.0"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.48.1"
6
  }
checkpoints/causalLM_finetune/TrpB/best/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1695d3701f56e457cbf06759b07a2c5960809d8a415a73b9d458fe54b0306b63
3
  size 617190232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d165e36e29e33db4602cc6ead4cc33747673bb4b73c0b2cb582474ce846f2e8
3
  size 617190232
checkpoints/continuous/GB1/best_model.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f17e0f967faaa12715dc8a0298d1d3361c2c09c5282c2324a83450dccd52b914
3
+ size 335371182
checkpoints/d3pm_finetune/GB1/best_model.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4953ce06d035646883a459aa3e12b861ab3b480aba04c85d0a161565f37f2294
3
+ size 454937272
checkpoints/mdlm/GB1/best_model.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a43e505385647a5466655a0a7c146373d38b07f5d6bdd2ca389d5dd8e81e416
3
+ size 457406314
data/hmmer/jackhmmer/GB1.domtbl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38cbf2a661194e486a119524c7e6f1bb4a847caa6b9fcc65572099842168d8e1
3
+ size 229823
data/hmmer/jackhmmer/GB1_alignment.a2m ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:264e8cfa06d9de865ca83dc552596ee1434b2e9900c4ee4f149e96eb23f19b53
3
+ size 42990
data/hmmer/jackhmmer/GB1_alignment.fasta ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18a72a99af78038ed72d953f59dff4712918dd03e5a025abd28e044eb8f6c14c
3
+ size 20320
data/hmmer/jackhmmer/GB1_alignment.sto ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b52256b92335be948c1aef2b2ee363823244d42dd68e370abcf536d376a0da5d
3
+ size 77625
data/hmmer/jackhmmer/GB1_seqs.fasta ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cd6de4f55a8c18e08214fc28f999c922a852a063b8fc419650f58581370eeb3
3
+ size 221766
data/raw_data/GB1_fitness/1-s2.0-S0960982214012688-mmc2.xlsx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e912c93cf3d1d95a7d8fc5bb8e8d50a523eb9398c51613b2a74693f07a50fe29
3
+ size 28090836
data/raw_data/GB1_fitness/doubles.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e6480a767a62f7efddc7a579ca1854696c5213423934f32deadd3c3dab8ebd5
3
+ size 17970210
data/raw_data/GB1_fitness/singles.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b47386e433fdc0a9a3174ac50eed601c8bd591450facc16a0fb01a454f3c426
3
+ size 20350