File size: 848 Bytes
8d80d34
 
 
 
 
 
 
7e64247
511d848
7e64247
 
8d80d34
ffc2258
5daf4e3
8d80d34
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
---
library_name: transformers
tags: []
---

### Model Description

This repository ships the CodonGPT model checkpoint together with its codon-level Tokenizer and the SynonymousLogitProcessor, so you can reproduce the constrained generation workflow straight from
the model card. The model was pretrained on Ensembl CDS sequences with a GPT-2–style decoder, learns synonymous structure and CAI/GC biases, and is optimized for codon-
aware sequence design. After pulling the snapshot, load the tokenizer and processor from the repo files to enable synonym-aware decoding that encourages biologically equivalent alternatives while preserving
sequence-level realism.

- **Developed by:** Nanil Therapeutics Inc.
- **Model type:** Transformer-based generative language model for protein-coding DNA/mRNA sequences
- **License:** Free for research use