Playingyoyo commited on
Commit
c5fdb97
·
verified ·
1 Parent(s): ee4a457

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -0
README.md CHANGED
@@ -21,6 +21,8 @@ To leverage GPepT’s pre-trained weights, input molecules must be converted int
21
 
22
  Unlike traditional protein design models, GPepT is trained in a self-supervised manner, using raw sequence data without explicit annotation. This design enables the model to generalize across diverse sequence spaces, producing functional antimicrobial peptidomimetics upon fine-tuning.
23
 
 
 
24
  ---
25
 
26
  ## **Using GPepT for Sequence Generation**
@@ -28,6 +30,7 @@ GPepT is fully compatible with the HuggingFace Transformers Python library. Inst
28
 
29
  The model excels at generating peptidomimetic sequences in a zero-shot fashion, but it can also be fine-tuned on custom datasets to generate sequences tailored to specific requirements.
30
 
 
31
  ### **Example 1: Zero-Shot Sequence Generation**
32
  GPepT generates sequences that extend from a specified input token (e.g., `<|endoftext|>`). If no input is provided, it selects the start token automatically and generates likely sequences. Here’s a Python example:
33
 
 
21
 
22
  Unlike traditional protein design models, GPepT is trained in a self-supervised manner, using raw sequence data without explicit annotation. This design enables the model to generalize across diverse sequence spaces, producing functional antimicrobial peptidomimetics upon fine-tuning.
23
 
24
+ SMILES representation, and selected chemical properties of each token, which corresponds to a non-canonical amino acid or terminal modification.
25
+
26
  ---
27
 
28
  ## **Using GPepT for Sequence Generation**
 
30
 
31
  The model excels at generating peptidomimetic sequences in a zero-shot fashion, but it can also be fine-tuned on custom datasets to generate sequences tailored to specific requirements.
32
 
33
+
34
  ### **Example 1: Zero-Shot Sequence Generation**
35
  GPepT generates sequences that extend from a specified input token (e.g., `<|endoftext|>`). If no input is provided, it selects the start token automatically and generates likely sequences. Here’s a Python example:
36