wabu commited on
Commit
428c239
·
verified ·
1 Parent(s): ba9167b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -5
README.md CHANGED
@@ -10,9 +10,6 @@ model-index:
10
  results: []
11
  ---
12
 
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
-
16
  # AmpGPT2
17
 
18
  AmpGPT2 is a language model capable of generating de novo antimicrobial peptides (AMPs). Generated sequences are predicted to be AMPs 95.83% of the time.
@@ -20,13 +17,20 @@ AmpGPT2 is a language model capable of generating de novo antimicrobial peptides
20
  ## Model description
21
 
22
  AmpGPT2 is a fine-tuned version of [nferruz/ProtGPT2](https://huggingface.co/nferruz/ProtGPT2) based on the GPT2 Transformer architecture.
23
- To validate the results the Antimicrobial Peptide Scanner vr.2 (https://www.dveltri.com/ascan/v2/ascan.html) was used. It is a
 
 
 
24
 
25
  ## Training and evaluation data
26
 
27
  AmpGPT2 was trained using 32014 AMP sequences from the Compass (https://compass.mathematik.uni-marburg.de/) database.
28
 
29
  ## How to use AmpGPT2
 
 
 
 
30
  ```
31
  from transformers import pipeline
32
  from transformers import GPT2LMHeadModel, GPT2Tokenizer
@@ -44,6 +48,7 @@ for i, seq in enumerate(amp_sequences):
44
 
45
  print(f">{sequence_identifier}\n{sequence}")
46
  ```
 
47
  ### Training hyperparameters
48
 
49
  The following hyperparameters were used during training:
@@ -55,8 +60,9 @@ The following hyperparameters were used during training:
55
  - lr_scheduler_type: linear
56
  - num_epochs: 50.0
57
 
 
 
58
  ### Training results
59
- these are the training losses after the final epoch
60
 
61
  | Training Loss | Epoch | Validation Loss | Accuracy |
62
  |:-------------:|:-----:|:---------------:|:--------:|
 
10
  results: []
11
  ---
12
 
 
 
 
13
  # AmpGPT2
14
 
15
  AmpGPT2 is a language model capable of generating de novo antimicrobial peptides (AMPs). Generated sequences are predicted to be AMPs 95.83% of the time.
 
17
  ## Model description
18
 
19
  AmpGPT2 is a fine-tuned version of [nferruz/ProtGPT2](https://huggingface.co/nferruz/ProtGPT2) based on the GPT2 Transformer architecture.
20
+
21
+ To validate the results the Antimicrobial Peptide Scanner vr.2 (https://www.dveltri.com/ascan/v2/ascan.html) was used.
22
+ It is a deep learning tool specifically designed for AMP recognition.
23
+
24
 
25
  ## Training and evaluation data
26
 
27
  AmpGPT2 was trained using 32014 AMP sequences from the Compass (https://compass.mathematik.uni-marburg.de/) database.
28
 
29
  ## How to use AmpGPT2
30
+
31
+ The example code below contains the ideal generation settings found while testing.
32
+ The 'num_return_sequences' parameter specifies the amount of sequences generated. When generating more than 100 sequences at the same time, I recommend doing it in batches.
33
+ The results can then be checked with the peptide scanner (https://www.dveltri.com/ascan/v2/ascan.html).
34
  ```
35
  from transformers import pipeline
36
  from transformers import GPT2LMHeadModel, GPT2Tokenizer
 
48
 
49
  print(f">{sequence_identifier}\n{sequence}")
50
  ```
51
+
52
  ### Training hyperparameters
53
 
54
  The following hyperparameters were used during training:
 
60
  - lr_scheduler_type: linear
61
  - num_epochs: 50.0
62
 
63
+ The model was trained on four NVIDIA A100 GPUs.
64
+
65
  ### Training results
 
66
 
67
  | Training Loss | Epoch | Validation Loss | Accuracy |
68
  |:-------------:|:-----:|:---------------:|:--------:|