JuIm commited on
Commit
e5b02db
·
verified ·
1 Parent(s): 7323895

End of training

Browse files
README.md CHANGED
@@ -13,29 +13,37 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # ProGemma2
15
 
16
- TThis is a custom configuration of Google’s Gemma 2 LLM that is being pre-trained on amino acid sequences of 512 AA or less in length. Periodic updates are made to this page as training reaches new checkpoints.
17
 
18
- The purpose of this model was to investigate the differences between ProGemma and ProtGPT (GPT-2 architecture) as it pertains to sequence generation. Training loss is ~2.2. Perplexity scores as well as AlphaFold 3’s ptm, pLDDT, and iptm scores are generally in line with ProtGPT’s scores for sequence lengths < 250, although the testing phase is still very early. I have yet to do testing for sequence lengths > 250. More robust testing is also required for lengths < 250 AA. In my very preliminary testing, HHblit e-values of ~0.1 are achieved with relatively easily.
19
 
20
- Controlled generation is not a capability of this model, and therefore serves as a method to significantly improve generation as, in principal, a sequence that performs a given function or resides in a particular cellular location can be generated.
21
 
22
- In sequence generation, a top_k of 950 appears to work well as it prevents repetition. This is also seen in ProtGPT.
23
 
24
- Below is code using the Transformers library to generate sequences using ProGemma.
25
 
26
- from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
27
 
28
- model = AutoModelForCausalLM.from_pretrained("JuIm/ProGemma2")
29
 
30
- tokenizer = AutoTokenizer.from_pretrained("JuIm/Amino-Acid-Sequence-Tokenizer")
31
 
32
- progemma = pipeline("text-generation", model=model, tokenizer=tokenizer)
33
 
34
- sequence = progemma("\<bos>", top_k=950, max_length=100, num_return_sequences=1, do_sample=True, repetition_penalty=1.2, eos_token_id=21, pad_token_id=22, bos_token_id=20)
 
 
 
 
 
 
 
 
 
 
35
 
36
- s = sequence[0]['generated_text']
37
 
38
- print(s)
39
 
40
  ### Framework versions
41
 
 
13
 
14
  # ProGemma2
15
 
16
+ This model is a fine-tuned version of [JuIm/ProGemma2](https://huggingface.co/JuIm/ProGemma2) on an unknown dataset.
17
 
18
+ ## Model description
19
 
20
+ More information needed
21
 
22
+ ## Intended uses & limitations
23
 
24
+ More information needed
25
 
26
+ ## Training and evaluation data
27
 
28
+ More information needed
29
 
30
+ ## Training procedure
31
 
32
+ ### Training hyperparameters
33
 
34
+ The following hyperparameters were used during training:
35
+ - learning_rate: 0.001
36
+ - train_batch_size: 2
37
+ - eval_batch_size: 8
38
+ - seed: 42
39
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
+ - lr_scheduler_type: linear
41
+ - lr_scheduler_warmup_ratio: 0.4
42
+ - training_steps: 3500
43
+
44
+ ### Training results
45
 
 
46
 
 
47
 
48
  ### Framework versions
49
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac70f7bcec01ad6dd79d18caa3b7f715c73aece56830885fd63efb878412b515
3
  size 1342562152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:016bdbc15602583757f3e25e9cc5e310b0c6f1ef85dc2e4900884042b6ee3c4e
3
  size 1342562152
runs/Sep05_16-47-25_218fb5aa0ae9/events.out.tfevents.1725554850.218fb5aa0ae9.1339.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a23e2a1d4360d8b6db38caed06b6669e7c046cc5fda755412cc36adcc02422d
3
+ size 743400
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c59a1f1b90f9d23065df07ee260b16d0391935dcd5232bcdc4dc1afd27931e9
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbaad44a6fc4f05c77832143bb0ab3c2b4bae12665c1a965419cf27ba178409f
3
  size 5176