JuIm commited on
Commit
f77b7bc
·
verified ·
1 Parent(s): 865d316

End of training

Browse files
README.md CHANGED
@@ -13,29 +13,37 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # ProGemma2
15
 
16
- This is a custom configuration of Google’s Gemma 2 LLM that is being pre-trained on amino acid sequences of 512 AA or less in length. Periodic updates are made to this page as training reaches new checkpoints.
17
 
18
- The purpose of this model was to investigate the differences between ProGemma and ProtGPT (GPT-2 architecture) as it pertains to sequence generation. Training loss is ~2.2. Perplexity scores as well as AlphaFold 3’s ptm, pLDDT, and iptm scores are generally in line with ProtGPT’s scores for sequence lengths < 250, although the testing phase is still very early. I have yet to do testing for sequence lengths > 250. More robust testing is also required for lengths < 250 AA. In my very preliminary testing, HHblit e-values of ~0.1 are achieved with relatively easily.
19
 
20
- Controlled generation is not a capability of this model, and therefore serves as a method to significantly improve generation as, in principal, a sequence that performs a given function or resides in a particular cellular location can be generated.
21
 
22
- In sequence generation, a top_k of 950 appears to work well as it prevents repetition. This is also seen in ProtGPT.
23
 
24
- Below is code using the Transformers library to generate sequences using ProGemma.
25
 
26
- from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
27
 
28
- model = AutoModelForCausalLM.from_pretrained("JuIm/ProGemma2")
29
 
30
- tokenizer = AutoTokenizer.from_pretrained("JuIm/Amino-Acid-Sequence-Tokenizer")
31
 
32
- progemma = pipeline("text-generation", model=model, tokenizer=tokenizer)
33
 
34
- sequence = progemma("\<bos>", top_k=950, max_length=100, num_return_sequences=1, do_sample=True, repetition_penalty=1.2, eos_token_id=21, pad_token_id=22, bos_token_id=20)
 
 
 
 
 
 
 
 
 
 
35
 
36
- s = sequence[0]['generated_text']
37
 
38
- print(s)
39
 
40
  ### Framework versions
41
 
 
13
 
14
  # ProGemma2
15
 
16
+ This model is a fine-tuned version of [JuIm/ProGemma2](https://huggingface.co/JuIm/ProGemma2) on an unknown dataset.
17
 
18
+ ## Model description
19
 
20
+ More information needed
21
 
22
+ ## Intended uses & limitations
23
 
24
+ More information needed
25
 
26
+ ## Training and evaluation data
27
 
28
+ More information needed
29
 
30
+ ## Training procedure
31
 
32
+ ### Training hyperparameters
33
 
34
+ The following hyperparameters were used during training:
35
+ - learning_rate: 0.001
36
+ - train_batch_size: 2
37
+ - eval_batch_size: 8
38
+ - seed: 42
39
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
+ - lr_scheduler_type: linear
41
+ - lr_scheduler_warmup_ratio: 0.4
42
+ - training_steps: 3500
43
+
44
+ ### Training results
45
 
 
46
 
 
47
 
48
  ### Framework versions
49
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14dd23764e53ee951cb502706939ce26397e56cf833f21b05edd184faa386dfc
3
  size 1342562152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16877ece4417db101bc32d219ffef586569590e4eba05611930a6d044c80660d
3
  size 1342562152
runs/Sep10_11-24-26_1f25d89ff08a/events.out.tfevents.1725967471.1f25d89ff08a.1198.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43b9f70fe58bc7ee90830966b77cc390a53844f95c264f814f5ae0915b5e2337
3
+ size 743400
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbcca3722a8032945b4d4a6d0e35ad76fec6b1bded92f90b198203616d99a832
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7b632e88e3d786acc0c1988111f8746bdd130faf510f762b67848fb02ebf864
3
  size 5176