JuIm commited on
Commit
561c8ed
·
verified ·
1 Parent(s): 7b82a37

End of training

Browse files
README.md CHANGED
@@ -5,41 +5,42 @@ tags:
5
  model-index:
6
  - name: ProGemma
7
  results: []
8
- pipeline_tag: text-generation
9
  ---
10
 
 
 
11
 
12
  # ProGemma
13
 
14
- This is a custom configuration of Google's Gemma 2 model that is being pre-trained on amino acid sequences of lengths 0 to 512.
15
- I used the free version of Google Colab to train this model, so updates are made regularly as the model hits new checkpoints.
16
- As of 07.28.2024, the model has been trained on about 5% of the dataset.
17
 
18
- The model generates amino acids on a letter-by-letter basis.
19
 
20
- Current training loss is about 2.7. Preliminary evaluation of generated sequences on AlphaFold 3 shows pTM scores of ~0.4 and
21
- average pLLDT scores ~60. After training is complete, a proper evaluation will be done to see whether sequences result in proteins with
22
- a low free energy. Perplexity scores will also be calculated.
23
 
24
- The purpose of this model was to see whether I could develop an alternative to NVIDIA's ProtGPT2. ProGemma also serves as a stepping stone
25
- to a new model that will also utilize control tags to generate proteins based on function.
26
 
27
- To use this mode for yourself using the pipeline within the Transformers package, please see the code below:
28
 
29
- from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
30
 
31
- model = AutoModelForCausalLM.from_pretrained("JuIm/ProGemma")
32
- tokenizer = AutoTokenizer.from_pretrained("JuIm/Amino-Acid-Sequence-Tokenizer")
33
 
34
- progemma = pipeline("text-generation", model=model, tokenizer=tokenizer)
35
 
36
- sequence = progemma("bosM", top_k=950, max_length=100, num_return_sequences=1, do_sample=True, repetition_penalty=1.2, eos_token_id=21, pad_token_id=22, bos_token_id=20)
37
-
38
-
39
-
40
- print(sequence)
41
 
 
 
 
 
 
 
 
 
 
42
 
 
43
 
44
 
45
 
@@ -47,4 +48,4 @@ print(sequence)
47
 
48
  - Transformers 4.42.4
49
  - Pytorch 2.3.1+cu121
50
- - Tokenizers 0.19.1
 
5
  model-index:
6
  - name: ProGemma
7
  results: []
 
8
  ---
9
 
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # ProGemma
14
 
15
+ This model is a fine-tuned version of [JuIm/ProGemma](https://huggingface.co/JuIm/ProGemma) on an unknown dataset.
 
 
16
 
17
+ ## Model description
18
 
19
+ More information needed
 
 
20
 
21
+ ## Intended uses & limitations
 
22
 
23
+ More information needed
24
 
25
+ ## Training and evaluation data
26
 
27
+ More information needed
 
28
 
29
+ ## Training procedure
30
 
31
+ ### Training hyperparameters
 
 
 
 
32
 
33
+ The following hyperparameters were used during training:
34
+ - learning_rate: 0.001
35
+ - train_batch_size: 1
36
+ - eval_batch_size: 8
37
+ - seed: 42
38
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
+ - lr_scheduler_type: linear
40
+ - lr_scheduler_warmup_ratio: 0.4
41
+ - training_steps: 5000
42
 
43
+ ### Training results
44
 
45
 
46
 
 
48
 
49
  - Transformers 4.42.4
50
  - Pytorch 2.3.1+cu121
51
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de411c75a546aef900000ac61d174346ec545938769b91a3ec8348167fef00f3
3
  size 1101271208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:421e8d2b9d187571d7f33a061b0159ef8506e62e39b2bfd7d6e4d49ffdc0faeb
3
  size 1101271208
runs/Jul29_16-37-14_cc53d3056a16/events.out.tfevents.1722271039.cc53d3056a16.633.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dd47a96dc0f57392c50d7a5ba59a0a614dc5f1ea1be9b1a204bb7034c588be4
3
+ size 1059827
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0de069b7900ed2b617b27a9995235f3d9f9f1e0de81051884a5f99ad7076ad6
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f1cb5b7c407fa5b732f4672cda3d3aa6944cf6855dc291b8aa8908c79e458e9
3
  size 5112