JuIm commited on
Commit
5c35c7f
·
verified ·
1 Parent(s): 1b939b2

End of training

Browse files
README.md CHANGED
@@ -12,31 +12,37 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # ProGemma
14
 
15
- A custom configuration of Google's Gemma 2 LLM pre-trained on amino acid sequences of length 0 - 512 in length. This configuration is 275M parameters.
16
- This model is being trained on Google Colab (free version), so regular updates are made upon hitting new checkpoints.
17
- The dataset is ~500k sequences in total, and the model was trained on about 5% of the dataset as of date 7.28.2024. Training loss at this point is ~2.7.
18
- The tokenizer uses bos, eos, and pad special tokens where each sequence is padded to length 512.
19
 
 
20
 
21
- The purpose of this model was simply to build my own version of NVIDIA's ProtGPT. After this model is completed, I will be looking to add control tags to generate sequences based on a given function for a specific organism.
22
- Upon completion of training, the model will be properly evaluated, looking at perplexity, energy of proteins generated, and AlphaFold 3 pLDDT/pTM scores
23
 
 
24
 
25
- To try this model out for yourself, see the code below:
26
 
27
- from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
28
 
29
- model = AutoModelForCausalLM.from_pretrained("JuIm/ProGemma")
30
- tokenizer = AutoTokenizer.from_pretrained("JuIm/Amino-Acid-Sequence-Tokenizer")
31
 
32
- progemma = pipeline("text-generation", model=model, tokenizer=tokenizer)
33
 
34
- sequence = progemma('<bos', max_length=150, do_sample=True, top_k=950, repetition_penalty=1.2, num_return_sequences=1, eos_token_id=21)
 
 
 
 
 
 
 
 
 
 
 
 
35
 
36
- for i in sequence:
37
- print(sequence)
38
 
39
- 'MLSLFSWFENKLDKTLKKISRIELFRKKITEVICDEHIYVMKPPFSEKTTLTREGYECGSRTMPNLARPDTYLLSRFKENCYGLHYTILGCSKNLLAPFGATFTSMLSVMVIFIFLFTKVEDFIKRCEGAGWVITEFGSTSGVPAVGPG'
40
 
41
  ### Framework versions
42
 
 
12
 
13
  # ProGemma
14
 
15
+ This model is a fine-tuned version of [JuIm/ProGemma](https://huggingface.co/JuIm/ProGemma) on an unknown dataset.
 
 
 
16
 
17
+ ## Model description
18
 
19
+ More information needed
 
20
 
21
+ ## Intended uses & limitations
22
 
23
+ More information needed
24
 
25
+ ## Training and evaluation data
26
 
27
+ More information needed
 
28
 
29
+ ## Training procedure
30
 
31
+ ### Training hyperparameters
32
+
33
+ The following hyperparameters were used during training:
34
+ - learning_rate: 0.001
35
+ - train_batch_size: 1
36
+ - eval_batch_size: 8
37
+ - seed: 42
38
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
+ - lr_scheduler_type: linear
40
+ - lr_scheduler_warmup_ratio: 0.4
41
+ - training_steps: 5000
42
+
43
+ ### Training results
44
 
 
 
45
 
 
46
 
47
  ### Framework versions
48
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f14b657852a05165f09b24a547c40a71542b650a2dd8783cb67b9d03b2176ec8
3
  size 1101271208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de411c75a546aef900000ac61d174346ec545938769b91a3ec8348167fef00f3
3
  size 1101271208
runs/Jul28_20-27-38_f488e2221d62/events.out.tfevents.1722198463.f488e2221d62.651.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7133d35ef5f724893c5fd6994f93ca018a4143b6e94ae90fa5c61b868aaf2c85
3
+ size 1059827
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17bf7c68a2fea6eedabb64345fdef590e65b134c2c5559331816346ce73bd10c
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0de069b7900ed2b617b27a9995235f3d9f9f1e0de81051884a5f99ad7076ad6
3
  size 5112