JuIm commited on
Commit
aceef28
·
verified ·
1 Parent(s): 9937d6a

End of training

Browse files
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  base_model: JuIm/ProGemma2
3
  tags:
4
  - generated_from_trainer
@@ -12,32 +13,40 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # ProGemma2
14
 
15
- This is a custom configuration of Google’s Gemma 2 LLM that is being pre-trained on amino acid sequences of 512 AA or less in length. Periodic updates are made to this page as training reaches new checkpoints.
16
 
17
- The purpose of this model was to investigate the differences between ProGemma and ProtGPT (GPT-2 architecture) as it pertains to sequence generation. Training loss is ~2.2. Perplexity scores as well as AlphaFold 3’s ptm, pLDDT, and iptm scores are generally in line with ProtGPT’s scores for sequence lengths < 250, although the testing phase is still very early. I have yet to do testing for sequence lengths > 250. More robust testing is also required for lengths < 250 AA. In my very preliminary testing, HHblit e-values of ~0.1 are achieved with relatively easily.
18
 
19
- Controlled generation is not a capability of this model, and therefore serves as a method to significantly improve generation as, in principal, a sequence that performs a given function or resides in a particular cellular location can be generated.
20
 
21
- In sequence generation, a top_k of 950 appears to work well as it prevents repetition. This is also seen in ProtGPT.
22
 
23
- Below is code using the Transformers library to generate sequences using ProGemma.
24
 
25
- from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
26
 
27
- model = AutoModelForCausalLM.from_pretrained("JuIm/ProGemma2")
28
 
29
- tokenizer = AutoTokenizer.from_pretrained("JuIm/Amino-Acid-Sequence-Tokenizer")
30
 
31
- progemma = pipeline("text-generation", model=model, tokenizer=tokenizer)
32
 
33
- sequence = progemma("", top_k=950, max_length=100, num_return_sequences=1, do_sample=True, repetition_penalty=1.2, eos_token_id=21, pad_token_id=22, bos_token_id=20)
 
 
 
 
 
 
 
 
 
 
34
 
35
- s = sequence[0]['generated_text']
36
 
37
- print(s)
38
 
39
  ### Framework versions
40
 
41
- - Transformers 4.42.4
42
  - Pytorch 2.4.0+cu121
43
  - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  base_model: JuIm/ProGemma2
4
  tags:
5
  - generated_from_trainer
 
13
 
14
  # ProGemma2
15
 
16
+ This model is a fine-tuned version of [JuIm/ProGemma2](https://huggingface.co/JuIm/ProGemma2) on an unknown dataset.
17
 
18
+ ## Model description
19
 
20
+ More information needed
21
 
22
+ ## Intended uses & limitations
23
 
24
+ More information needed
25
 
26
+ ## Training and evaluation data
27
 
28
+ More information needed
29
 
30
+ ## Training procedure
31
 
32
+ ### Training hyperparameters
33
 
34
+ The following hyperparameters were used during training:
35
+ - learning_rate: 0.001
36
+ - train_batch_size: 2
37
+ - eval_batch_size: 8
38
+ - seed: 42
39
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
+ - lr_scheduler_type: linear
41
+ - lr_scheduler_warmup_ratio: 0.4
42
+ - training_steps: 3500
43
+
44
+ ### Training results
45
 
 
46
 
 
47
 
48
  ### Framework versions
49
 
50
+ - Transformers 4.44.2
51
  - Pytorch 2.4.0+cu121
52
  - Tokenizers 0.19.1
config.json CHANGED
@@ -26,7 +26,7 @@
26
  "rope_theta": 10000.0,
27
  "sliding_window": 256,
28
  "torch_dtype": "float32",
29
- "transformers_version": "4.42.4",
30
  "use_cache": true,
31
  "vocab_size": 24
32
  }
 
26
  "rope_theta": 10000.0,
27
  "sliding_window": 256,
28
  "torch_dtype": "float32",
29
+ "transformers_version": "4.44.2",
30
  "use_cache": true,
31
  "vocab_size": 24
32
  }
generation_config.json CHANGED
@@ -4,5 +4,5 @@
4
  "cache_implementation": "hybrid",
5
  "eos_token_id": 21,
6
  "pad_token_id": 22,
7
- "transformers_version": "4.42.4"
8
  }
 
4
  "cache_implementation": "hybrid",
5
  "eos_token_id": 21,
6
  "pad_token_id": 22,
7
+ "transformers_version": "4.44.2"
8
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7813bea0859e3db76a1e20d4e262c92c404c3208a943e66f89397852b266cdc
3
  size 1342562152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac70f7bcec01ad6dd79d18caa3b7f715c73aece56830885fd63efb878412b515
3
  size 1342562152
runs/Sep05_14-29-10_d40fbdb96b4d/events.out.tfevents.1725546554.d40fbdb96b4d.1521.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:430f9e9e788b8dc5f36ec8b28f315823087dcdceffb1425b5f809ba85c0dba09
3
+ size 743400
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8469524c5dbe74af7a5d1d467bb3413a99b1129652c1806c99a872214751a37c
3
- size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c59a1f1b90f9d23065df07ee260b16d0391935dcd5232bcdc4dc1afd27931e9
3
+ size 5176